metaflow 2.13.10__py2.py3-none-any.whl → 2.14.1__py2.py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- metaflow/cli.py +0 -1
- metaflow/cli_components/run_cmds.py +15 -14
- metaflow/cli_components/step_cmd.py +2 -0
- metaflow/client/core.py +241 -1
- metaflow/cmd/main_cli.py +1 -1
- metaflow/metadata_provider/heartbeat.py +1 -0
- metaflow/metadata_provider/metadata.py +33 -0
- metaflow/metaflow_config.py +1 -5
- metaflow/mflog/save_logs.py +2 -2
- metaflow/plugins/argo/argo_workflows.py +12 -8
- metaflow/plugins/argo/argo_workflows_cli.py +2 -2
- metaflow/plugins/datatools/s3/s3op.py +4 -4
- metaflow/plugins/env_escape/server.py +7 -0
- metaflow/plugins/env_escape/stub.py +21 -4
- metaflow/plugins/kubernetes/kubernetes_cli.py +1 -1
- metaflow/plugins/metadata_providers/local.py +66 -0
- metaflow/plugins/metadata_providers/service.py +82 -18
- metaflow/plugins/pypi/bootstrap.py +4 -4
- metaflow/runner/click_api.py +6 -9
- metaflow/sidecar/sidecar_worker.py +1 -1
- metaflow/task.py +21 -2
- metaflow/tracing/__init__.py +7 -7
- metaflow/tracing/span_exporter.py +31 -38
- metaflow/tracing/tracing_modules.py +39 -44
- metaflow/version.py +1 -1
- {metaflow-2.13.10.dist-info → metaflow-2.14.1.dist-info}/METADATA +2 -2
- {metaflow-2.13.10.dist-info → metaflow-2.14.1.dist-info}/RECORD +31 -31
- {metaflow-2.13.10.dist-info → metaflow-2.14.1.dist-info}/LICENSE +0 -0
- {metaflow-2.13.10.dist-info → metaflow-2.14.1.dist-info}/WHEEL +0 -0
- {metaflow-2.13.10.dist-info → metaflow-2.14.1.dist-info}/entry_points.txt +0 -0
- {metaflow-2.13.10.dist-info → metaflow-2.14.1.dist-info}/top_level.txt +0 -0
@@ -2,10 +2,12 @@ import collections
|
|
2
2
|
import glob
|
3
3
|
import json
|
4
4
|
import os
|
5
|
+
import re
|
5
6
|
import random
|
6
7
|
import tempfile
|
7
8
|
import time
|
8
9
|
from collections import namedtuple
|
10
|
+
from typing import List
|
9
11
|
|
10
12
|
from metaflow.exception import MetaflowInternalError, MetaflowTaggingError
|
11
13
|
from metaflow.metadata_provider.metadata import ObjectOrder
|
@@ -202,6 +204,70 @@ class LocalMetadataProvider(MetadataProvider):
|
|
202
204
|
"Tagging failed due to too many conflicting updates from other processes"
|
203
205
|
)
|
204
206
|
|
207
|
+
@classmethod
|
208
|
+
def filter_tasks_by_metadata(
|
209
|
+
cls,
|
210
|
+
flow_name: str,
|
211
|
+
run_id: str,
|
212
|
+
step_name: str,
|
213
|
+
field_name: str,
|
214
|
+
pattern: str,
|
215
|
+
) -> List[str]:
|
216
|
+
"""
|
217
|
+
Filter tasks by metadata field and pattern, returning task pathspecs that match criteria.
|
218
|
+
|
219
|
+
Parameters
|
220
|
+
----------
|
221
|
+
flow_name : str
|
222
|
+
Identifier for the flow
|
223
|
+
run_id : str
|
224
|
+
Identifier for the run
|
225
|
+
step_name : str
|
226
|
+
Name of the step to query tasks from
|
227
|
+
field_name : str
|
228
|
+
Name of metadata field to query
|
229
|
+
pattern : str
|
230
|
+
Pattern to match in metadata field value
|
231
|
+
|
232
|
+
Returns
|
233
|
+
-------
|
234
|
+
List[str]
|
235
|
+
List of task pathspecs that match the query criteria
|
236
|
+
"""
|
237
|
+
tasks = cls.get_object("step", "task", {}, None, flow_name, run_id, step_name)
|
238
|
+
if not tasks:
|
239
|
+
return []
|
240
|
+
|
241
|
+
regex = re.compile(pattern)
|
242
|
+
matching_task_pathspecs = []
|
243
|
+
|
244
|
+
for task in tasks:
|
245
|
+
task_id = task.get("task_id")
|
246
|
+
if not task_id:
|
247
|
+
continue
|
248
|
+
|
249
|
+
if pattern == ".*":
|
250
|
+
# If the pattern is ".*", we can match all tasks without reading metadata
|
251
|
+
matching_task_pathspecs.append(
|
252
|
+
f"{flow_name}/{run_id}/{step_name}/{task_id}"
|
253
|
+
)
|
254
|
+
continue
|
255
|
+
|
256
|
+
metadata = cls.get_object(
|
257
|
+
"task", "metadata", {}, None, flow_name, run_id, step_name, task_id
|
258
|
+
)
|
259
|
+
|
260
|
+
if any(
|
261
|
+
meta.get("field_name") == field_name
|
262
|
+
and regex.match(meta.get("value", ""))
|
263
|
+
for meta in metadata
|
264
|
+
):
|
265
|
+
matching_task_pathspecs.append(
|
266
|
+
f"{flow_name}/{run_id}/{step_name}/{task_id}"
|
267
|
+
)
|
268
|
+
|
269
|
+
return matching_task_pathspecs
|
270
|
+
|
205
271
|
@classmethod
|
206
272
|
def _get_object_internal(
|
207
273
|
cls, obj_type, obj_order, sub_type, sub_order, filters, attempt, *args
|
@@ -1,23 +1,20 @@
|
|
1
1
|
import os
|
2
2
|
import random
|
3
|
+
import time
|
3
4
|
|
4
5
|
import requests
|
5
|
-
import time
|
6
6
|
|
7
|
+
from typing import List
|
7
8
|
from metaflow.exception import (
|
8
9
|
MetaflowException,
|
9
|
-
MetaflowTaggingError,
|
10
10
|
MetaflowInternalError,
|
11
|
-
|
12
|
-
from metaflow.metaflow_config import (
|
13
|
-
SERVICE_RETRY_COUNT,
|
14
|
-
SERVICE_HEADERS,
|
15
|
-
SERVICE_URL,
|
11
|
+
MetaflowTaggingError,
|
16
12
|
)
|
17
13
|
from metaflow.metadata_provider import MetadataProvider
|
18
14
|
from metaflow.metadata_provider.heartbeat import HB_URL_KEY
|
15
|
+
from metaflow.metaflow_config import SERVICE_HEADERS, SERVICE_RETRY_COUNT, SERVICE_URL
|
19
16
|
from metaflow.sidecar import Message, MessageTypes, Sidecar
|
20
|
-
|
17
|
+
from urllib.parse import urlencode
|
21
18
|
from metaflow.util import version_parse
|
22
19
|
|
23
20
|
|
@@ -39,6 +36,23 @@ class ServiceException(MetaflowException):
|
|
39
36
|
class ServiceMetadataProvider(MetadataProvider):
|
40
37
|
TYPE = "service"
|
41
38
|
|
39
|
+
_session = requests.Session()
|
40
|
+
_session.mount(
|
41
|
+
"http://",
|
42
|
+
requests.adapters.HTTPAdapter(
|
43
|
+
pool_connections=20,
|
44
|
+
pool_maxsize=20,
|
45
|
+
max_retries=0, # Handle retries explicitly
|
46
|
+
pool_block=False,
|
47
|
+
),
|
48
|
+
)
|
49
|
+
_session.mount(
|
50
|
+
"https://",
|
51
|
+
requests.adapters.HTTPAdapter(
|
52
|
+
pool_connections=20, pool_maxsize=20, max_retries=0, pool_block=False
|
53
|
+
),
|
54
|
+
)
|
55
|
+
|
42
56
|
_supports_attempt_gets = None
|
43
57
|
_supports_tag_mutation = None
|
44
58
|
|
@@ -59,7 +73,9 @@ class ServiceMetadataProvider(MetadataProvider):
|
|
59
73
|
def compute_info(cls, val):
|
60
74
|
v = val.rstrip("/")
|
61
75
|
try:
|
62
|
-
resp =
|
76
|
+
resp = cls._session.get(
|
77
|
+
os.path.join(v, "ping"), headers=SERVICE_HEADERS.copy()
|
78
|
+
)
|
63
79
|
resp.raise_for_status()
|
64
80
|
except: # noqa E722
|
65
81
|
raise ValueError("Metaflow service [%s] unreachable." % v)
|
@@ -304,6 +320,55 @@ class ServiceMetadataProvider(MetadataProvider):
|
|
304
320
|
self._register_system_metadata(run_id, step_name, task["task_id"], attempt)
|
305
321
|
return task["task_id"], did_create
|
306
322
|
|
323
|
+
@classmethod
|
324
|
+
def filter_tasks_by_metadata(
|
325
|
+
cls,
|
326
|
+
flow_name: str,
|
327
|
+
run_id: str,
|
328
|
+
step_name: str,
|
329
|
+
field_name: str,
|
330
|
+
pattern: str,
|
331
|
+
) -> List[str]:
|
332
|
+
"""
|
333
|
+
Filter tasks by metadata field and pattern, returning task pathspecs that match criteria.
|
334
|
+
|
335
|
+
Parameters
|
336
|
+
----------
|
337
|
+
flow_name : str
|
338
|
+
Flow name, that the run belongs to.
|
339
|
+
run_id: str
|
340
|
+
Run id, together with flow_id, that identifies the specific Run whose tasks to query
|
341
|
+
step_name: str
|
342
|
+
Step name to query tasks from
|
343
|
+
field_name: str
|
344
|
+
Metadata field name to query
|
345
|
+
pattern: str
|
346
|
+
Pattern to match in metadata field value
|
347
|
+
|
348
|
+
Returns
|
349
|
+
-------
|
350
|
+
List[str]
|
351
|
+
List of task pathspecs that satisfy the query
|
352
|
+
"""
|
353
|
+
query_params = {
|
354
|
+
"metadata_field_name": field_name,
|
355
|
+
"pattern": pattern,
|
356
|
+
"step_name": step_name,
|
357
|
+
}
|
358
|
+
url = ServiceMetadataProvider._obj_path(flow_name, run_id, step_name)
|
359
|
+
url = f"{url}/filtered_tasks?{urlencode(query_params)}"
|
360
|
+
try:
|
361
|
+
resp = cls._request(None, url, "GET")
|
362
|
+
except Exception as e:
|
363
|
+
if e.http_code == 404:
|
364
|
+
# filter_tasks_by_metadata endpoint does not exist in the version of metadata service
|
365
|
+
# deployed currently. Raise a more informative error message.
|
366
|
+
raise MetaflowInternalError(
|
367
|
+
"The version of metadata service deployed currently does not support filtering tasks by metadata. "
|
368
|
+
"Upgrade Metadata service to version 2.15 or greater to use this feature."
|
369
|
+
) from e
|
370
|
+
return resp
|
371
|
+
|
307
372
|
@staticmethod
|
308
373
|
def _obj_path(
|
309
374
|
flow_name,
|
@@ -412,27 +477,27 @@ class ServiceMetadataProvider(MetadataProvider):
|
|
412
477
|
if method == "GET":
|
413
478
|
if monitor:
|
414
479
|
with monitor.measure("metaflow.service_metadata.get"):
|
415
|
-
resp =
|
480
|
+
resp = cls._session.get(url, headers=SERVICE_HEADERS.copy())
|
416
481
|
else:
|
417
|
-
resp =
|
482
|
+
resp = cls._session.get(url, headers=SERVICE_HEADERS.copy())
|
418
483
|
elif method == "POST":
|
419
484
|
if monitor:
|
420
485
|
with monitor.measure("metaflow.service_metadata.post"):
|
421
|
-
resp =
|
486
|
+
resp = cls._session.post(
|
422
487
|
url, headers=SERVICE_HEADERS.copy(), json=data
|
423
488
|
)
|
424
489
|
else:
|
425
|
-
resp =
|
490
|
+
resp = cls._session.post(
|
426
491
|
url, headers=SERVICE_HEADERS.copy(), json=data
|
427
492
|
)
|
428
493
|
elif method == "PATCH":
|
429
494
|
if monitor:
|
430
495
|
with monitor.measure("metaflow.service_metadata.patch"):
|
431
|
-
resp =
|
496
|
+
resp = cls._session.patch(
|
432
497
|
url, headers=SERVICE_HEADERS.copy(), json=data
|
433
498
|
)
|
434
499
|
else:
|
435
|
-
resp =
|
500
|
+
resp = cls._session.patch(
|
436
501
|
url, headers=SERVICE_HEADERS.copy(), json=data
|
437
502
|
)
|
438
503
|
else:
|
@@ -475,7 +540,6 @@ class ServiceMetadataProvider(MetadataProvider):
|
|
475
540
|
resp.text,
|
476
541
|
)
|
477
542
|
time.sleep(2**i)
|
478
|
-
|
479
543
|
if resp:
|
480
544
|
raise ServiceException(
|
481
545
|
"Metadata request (%s) failed (code %s): %s"
|
@@ -499,9 +563,9 @@ class ServiceMetadataProvider(MetadataProvider):
|
|
499
563
|
try:
|
500
564
|
if monitor:
|
501
565
|
with monitor.measure("metaflow.service_metadata.get"):
|
502
|
-
resp =
|
566
|
+
resp = cls._session.get(url, headers=SERVICE_HEADERS.copy())
|
503
567
|
else:
|
504
|
-
resp =
|
568
|
+
resp = cls._session.get(url, headers=SERVICE_HEADERS.copy())
|
505
569
|
except:
|
506
570
|
if monitor:
|
507
571
|
with monitor.count("metaflow.service_metadata.failed_request"):
|
@@ -10,7 +10,7 @@ import tarfile
|
|
10
10
|
import time
|
11
11
|
from urllib.error import URLError
|
12
12
|
from urllib.request import urlopen
|
13
|
-
from metaflow.metaflow_config import DATASTORE_LOCAL_DIR
|
13
|
+
from metaflow.metaflow_config import DATASTORE_LOCAL_DIR
|
14
14
|
from metaflow.plugins import DATASTORES
|
15
15
|
from metaflow.plugins.pypi.utils import MICROMAMBA_MIRROR_URL, MICROMAMBA_URL
|
16
16
|
from metaflow.util import which
|
@@ -329,8 +329,6 @@ if __name__ == "__main__":
|
|
329
329
|
|
330
330
|
@timer
|
331
331
|
def fast_setup_environment(architecture, storage, env, prefix, pkgs_dir):
|
332
|
-
install_fast_initializer(architecture)
|
333
|
-
|
334
332
|
# Get package urls
|
335
333
|
conda_pkgs = env["conda"]
|
336
334
|
pypi_pkgs = env.get("pypi", [])
|
@@ -381,7 +379,9 @@ if __name__ == "__main__":
|
|
381
379
|
with open(os.path.join(manifest_dir, MAGIC_FILE)) as f:
|
382
380
|
env = json.load(f)[id_][architecture]
|
383
381
|
|
384
|
-
if
|
382
|
+
if datastore_type == "s3":
|
383
|
+
# TODO: Remove this once fast-initializer is ready for all datastores
|
384
|
+
install_fast_initializer(architecture)
|
385
385
|
fast_setup_environment(architecture, storage, env, prefix, pkgs_dir)
|
386
386
|
else:
|
387
387
|
setup_environment(
|
metaflow/runner/click_api.py
CHANGED
@@ -97,12 +97,13 @@ def _method_sanity_check(
|
|
97
97
|
check_type(supplied_v, annotations[supplied_k])
|
98
98
|
except TypeCheckError:
|
99
99
|
raise TypeError(
|
100
|
-
"Invalid type for '%s' (%s), expected: '%s', default is '%s'"
|
100
|
+
"Invalid type for '%s' (%s), expected: '%s', default is '%s' but found '%s'"
|
101
101
|
% (
|
102
102
|
supplied_k,
|
103
103
|
type(supplied_k),
|
104
104
|
annotations[supplied_k],
|
105
105
|
defaults[supplied_k],
|
106
|
+
str(supplied_v),
|
106
107
|
)
|
107
108
|
)
|
108
109
|
|
@@ -218,7 +219,7 @@ def get_inspect_param_obj(p: Union[click.Argument, click.Option], kind: str):
|
|
218
219
|
default=inspect.Parameter.empty if is_vararg else p.default,
|
219
220
|
annotation=annotation,
|
220
221
|
),
|
221
|
-
annotation,
|
222
|
+
Optional[TTuple[annotation]] if is_vararg else annotation,
|
222
223
|
)
|
223
224
|
|
224
225
|
|
@@ -392,7 +393,9 @@ class MetaflowAPI(object):
|
|
392
393
|
options = params.pop("options", {})
|
393
394
|
|
394
395
|
for _, v in args.items():
|
395
|
-
if
|
396
|
+
if v is None:
|
397
|
+
continue
|
398
|
+
if isinstance(v, (list, tuple)):
|
396
399
|
for i in v:
|
397
400
|
components.append(i)
|
398
401
|
else:
|
@@ -527,12 +530,6 @@ def extract_all_params(cmd_obj: Union[click.Command, click.Group]):
|
|
527
530
|
)
|
528
531
|
arg_parameters[each_param.name] = each_param
|
529
532
|
elif isinstance(each_param, click.Option):
|
530
|
-
if each_param.hidden:
|
531
|
-
# Skip hidden options because users should not be setting those.
|
532
|
-
# These are typically internal only options (used by the Runner in part
|
533
|
-
# for example to pass state files or configs to pass local-config-file).
|
534
|
-
continue
|
535
|
-
|
536
533
|
opt_params_sigs[each_param.name], annotations[each_param.name] = (
|
537
534
|
get_inspect_param_obj(each_param, inspect.Parameter.KEYWORD_ONLY)
|
538
535
|
)
|
@@ -49,7 +49,7 @@ def process_messages(worker_type, worker):
|
|
49
49
|
|
50
50
|
|
51
51
|
@click.command(help="Initialize workers")
|
52
|
-
@tracing.
|
52
|
+
@tracing.cli("sidecar")
|
53
53
|
@click.argument("worker-type")
|
54
54
|
def main(worker_type):
|
55
55
|
sidecar_type = SIDECARS.get(worker_type)
|
metaflow/task.py
CHANGED
@@ -493,6 +493,25 @@ class MetaflowTask(object):
|
|
493
493
|
)
|
494
494
|
)
|
495
495
|
|
496
|
+
# Add runtime dag information to the metadata of the task
|
497
|
+
foreach_execution_path = ",".join(
|
498
|
+
[
|
499
|
+
"{}:{}".format(foreach_frame.step, foreach_frame.index)
|
500
|
+
for foreach_frame in foreach_stack
|
501
|
+
]
|
502
|
+
)
|
503
|
+
if foreach_execution_path:
|
504
|
+
metadata.extend(
|
505
|
+
[
|
506
|
+
MetaDatum(
|
507
|
+
field="foreach-execution-path",
|
508
|
+
value=foreach_execution_path,
|
509
|
+
type="foreach-execution-path",
|
510
|
+
tags=metadata_tags,
|
511
|
+
),
|
512
|
+
]
|
513
|
+
)
|
514
|
+
|
496
515
|
self.metadata.register_metadata(
|
497
516
|
run_id,
|
498
517
|
step_name,
|
@@ -559,6 +578,7 @@ class MetaflowTask(object):
|
|
559
578
|
self.flow._success = False
|
560
579
|
self.flow._task_ok = None
|
561
580
|
self.flow._exception = None
|
581
|
+
|
562
582
|
# Note: All internal flow attributes (ie: non-user artifacts)
|
563
583
|
# should either be set prior to running the user code or listed in
|
564
584
|
# FlowSpec._EPHEMERAL to allow for proper merging/importing of
|
@@ -616,7 +636,6 @@ class MetaflowTask(object):
|
|
616
636
|
"graph_info": self.flow._graph_info,
|
617
637
|
}
|
618
638
|
)
|
619
|
-
|
620
639
|
for deco in decorators:
|
621
640
|
deco.task_pre_step(
|
622
641
|
step_name,
|
@@ -728,7 +747,7 @@ class MetaflowTask(object):
|
|
728
747
|
value=attempt_ok,
|
729
748
|
type="internal_attempt_status",
|
730
749
|
tags=["attempt_id:{0}".format(retry_count)],
|
731
|
-
)
|
750
|
+
),
|
732
751
|
],
|
733
752
|
)
|
734
753
|
|
metaflow/tracing/__init__.py
CHANGED
@@ -20,15 +20,15 @@ def post_fork():
|
|
20
20
|
yield
|
21
21
|
|
22
22
|
|
23
|
-
def
|
24
|
-
def
|
23
|
+
def cli(name: str):
|
24
|
+
def cli_wrap(func):
|
25
25
|
@wraps(func)
|
26
26
|
def wrapper_func(*args, **kwargs):
|
27
27
|
return func(*args, **kwargs)
|
28
28
|
|
29
29
|
return wrapper_func
|
30
30
|
|
31
|
-
return
|
31
|
+
return cli_wrap
|
32
32
|
|
33
33
|
|
34
34
|
def inject_tracing_vars(env_dict: Dict[str, str]) -> Dict[str, str]:
|
@@ -40,7 +40,9 @@ def get_trace_id() -> str:
|
|
40
40
|
|
41
41
|
|
42
42
|
@contextlib.contextmanager
|
43
|
-
def traced(name, attrs=
|
43
|
+
def traced(name, attrs=None):
|
44
|
+
if attrs is None:
|
45
|
+
attrs = {}
|
44
46
|
yield
|
45
47
|
|
46
48
|
|
@@ -54,17 +56,15 @@ def tracing(func):
|
|
54
56
|
|
55
57
|
if not DISABLE_TRACING and (CONSOLE_TRACE_ENABLED or OTEL_ENDPOINT or ZIPKIN_ENDPOINT):
|
56
58
|
try:
|
57
|
-
# Overrides No-Op implementations if a specific provider is configured.
|
58
59
|
from .tracing_modules import (
|
59
60
|
init_tracing,
|
60
61
|
post_fork,
|
61
|
-
|
62
|
+
cli,
|
62
63
|
inject_tracing_vars,
|
63
64
|
get_trace_id,
|
64
65
|
traced,
|
65
66
|
tracing,
|
66
67
|
)
|
67
|
-
|
68
68
|
except ImportError as e:
|
69
69
|
# We keep the errors silent by default so that having tracing environment variables present
|
70
70
|
# does not affect users with no need for tracing.
|
@@ -3,60 +3,53 @@ from metaflow.metaflow_config import (
|
|
3
3
|
OTEL_ENDPOINT,
|
4
4
|
ZIPKIN_ENDPOINT,
|
5
5
|
CONSOLE_TRACE_ENABLED,
|
6
|
+
SERVICE_AUTH_KEY,
|
7
|
+
SERVICE_HEADERS,
|
6
8
|
)
|
7
9
|
|
8
10
|
|
9
11
|
def get_span_exporter():
|
10
|
-
|
11
|
-
|
12
|
+
exporter_map = {
|
13
|
+
OTEL_ENDPOINT: _create_otel_exporter,
|
14
|
+
ZIPKIN_ENDPOINT: _create_zipkin_exporter,
|
15
|
+
CONSOLE_TRACE_ENABLED: _create_console_exporter,
|
16
|
+
}
|
12
17
|
|
13
|
-
|
14
|
-
|
18
|
+
for config, create_exporter in exporter_map.items():
|
19
|
+
if config:
|
20
|
+
return create_exporter()
|
15
21
|
|
16
|
-
|
17
|
-
|
18
|
-
else:
|
19
|
-
print("WARNING: endpoints not set up for Opentelemetry", file=sys.stderr)
|
20
|
-
return
|
22
|
+
print("WARNING: endpoints not set up for OpenTelemetry", file=sys.stderr)
|
23
|
+
return None
|
21
24
|
|
22
25
|
|
23
|
-
def
|
26
|
+
def _create_otel_exporter():
|
24
27
|
from opentelemetry.exporter.otlp.proto.http.trace_exporter import OTLPSpanExporter
|
25
28
|
|
26
|
-
|
27
|
-
|
28
|
-
|
29
|
-
|
29
|
+
if not any([SERVICE_AUTH_KEY, SERVICE_HEADERS]):
|
30
|
+
print("WARNING: no auth settings for OpenTelemetry", file=sys.stderr)
|
31
|
+
return None
|
32
|
+
|
33
|
+
config = {
|
34
|
+
"endpoint": OTEL_ENDPOINT,
|
35
|
+
"timeout": 1,
|
36
|
+
}
|
30
37
|
|
31
38
|
if SERVICE_AUTH_KEY:
|
32
|
-
|
33
|
-
endpoint=OTEL_ENDPOINT,
|
34
|
-
headers={"x-api-key": SERVICE_AUTH_KEY},
|
35
|
-
timeout=1,
|
36
|
-
)
|
39
|
+
config["headers"] = {"x-api-key": SERVICE_AUTH_KEY}
|
37
40
|
elif SERVICE_HEADERS:
|
38
|
-
|
39
|
-
|
40
|
-
|
41
|
-
|
42
|
-
|
43
|
-
|
44
|
-
print("WARNING: no auth settings for Opentelemetry", file=sys.stderr)
|
45
|
-
return
|
46
|
-
return span_exporter
|
47
|
-
|
48
|
-
|
49
|
-
def set_zipkin_exporter():
|
41
|
+
config["headers"] = SERVICE_HEADERS
|
42
|
+
|
43
|
+
return OTLPSpanExporter(**config)
|
44
|
+
|
45
|
+
|
46
|
+
def _create_zipkin_exporter():
|
50
47
|
from opentelemetry.exporter.zipkin.proto.http import ZipkinExporter
|
51
48
|
|
52
|
-
|
53
|
-
endpoint=ZIPKIN_ENDPOINT,
|
54
|
-
)
|
55
|
-
return span_exporter
|
49
|
+
return ZipkinExporter(endpoint=ZIPKIN_ENDPOINT)
|
56
50
|
|
57
51
|
|
58
|
-
def
|
52
|
+
def _create_console_exporter():
|
59
53
|
from opentelemetry.sdk.trace.export import ConsoleSpanExporter
|
60
54
|
|
61
|
-
|
62
|
-
return span_exporter
|
55
|
+
return ConsoleSpanExporter()
|