metaflow 2.13.10__py2.py3-none-any.whl → 2.14.1__py2.py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2,10 +2,12 @@ import collections
2
2
  import glob
3
3
  import json
4
4
  import os
5
+ import re
5
6
  import random
6
7
  import tempfile
7
8
  import time
8
9
  from collections import namedtuple
10
+ from typing import List
9
11
 
10
12
  from metaflow.exception import MetaflowInternalError, MetaflowTaggingError
11
13
  from metaflow.metadata_provider.metadata import ObjectOrder
@@ -202,6 +204,70 @@ class LocalMetadataProvider(MetadataProvider):
202
204
  "Tagging failed due to too many conflicting updates from other processes"
203
205
  )
204
206
 
207
+ @classmethod
208
+ def filter_tasks_by_metadata(
209
+ cls,
210
+ flow_name: str,
211
+ run_id: str,
212
+ step_name: str,
213
+ field_name: str,
214
+ pattern: str,
215
+ ) -> List[str]:
216
+ """
217
+ Filter tasks by metadata field and pattern, returning task pathspecs that match criteria.
218
+
219
+ Parameters
220
+ ----------
221
+ flow_name : str
222
+ Identifier for the flow
223
+ run_id : str
224
+ Identifier for the run
225
+ step_name : str
226
+ Name of the step to query tasks from
227
+ field_name : str
228
+ Name of metadata field to query
229
+ pattern : str
230
+ Pattern to match in metadata field value
231
+
232
+ Returns
233
+ -------
234
+ List[str]
235
+ List of task pathspecs that match the query criteria
236
+ """
237
+ tasks = cls.get_object("step", "task", {}, None, flow_name, run_id, step_name)
238
+ if not tasks:
239
+ return []
240
+
241
+ regex = re.compile(pattern)
242
+ matching_task_pathspecs = []
243
+
244
+ for task in tasks:
245
+ task_id = task.get("task_id")
246
+ if not task_id:
247
+ continue
248
+
249
+ if pattern == ".*":
250
+ # If the pattern is ".*", we can match all tasks without reading metadata
251
+ matching_task_pathspecs.append(
252
+ f"{flow_name}/{run_id}/{step_name}/{task_id}"
253
+ )
254
+ continue
255
+
256
+ metadata = cls.get_object(
257
+ "task", "metadata", {}, None, flow_name, run_id, step_name, task_id
258
+ )
259
+
260
+ if any(
261
+ meta.get("field_name") == field_name
262
+ and regex.match(meta.get("value", ""))
263
+ for meta in metadata
264
+ ):
265
+ matching_task_pathspecs.append(
266
+ f"{flow_name}/{run_id}/{step_name}/{task_id}"
267
+ )
268
+
269
+ return matching_task_pathspecs
270
+
205
271
  @classmethod
206
272
  def _get_object_internal(
207
273
  cls, obj_type, obj_order, sub_type, sub_order, filters, attempt, *args
@@ -1,23 +1,20 @@
1
1
  import os
2
2
  import random
3
+ import time
3
4
 
4
5
  import requests
5
- import time
6
6
 
7
+ from typing import List
7
8
  from metaflow.exception import (
8
9
  MetaflowException,
9
- MetaflowTaggingError,
10
10
  MetaflowInternalError,
11
- )
12
- from metaflow.metaflow_config import (
13
- SERVICE_RETRY_COUNT,
14
- SERVICE_HEADERS,
15
- SERVICE_URL,
11
+ MetaflowTaggingError,
16
12
  )
17
13
  from metaflow.metadata_provider import MetadataProvider
18
14
  from metaflow.metadata_provider.heartbeat import HB_URL_KEY
15
+ from metaflow.metaflow_config import SERVICE_HEADERS, SERVICE_RETRY_COUNT, SERVICE_URL
19
16
  from metaflow.sidecar import Message, MessageTypes, Sidecar
20
-
17
+ from urllib.parse import urlencode
21
18
  from metaflow.util import version_parse
22
19
 
23
20
 
@@ -39,6 +36,23 @@ class ServiceException(MetaflowException):
39
36
  class ServiceMetadataProvider(MetadataProvider):
40
37
  TYPE = "service"
41
38
 
39
+ _session = requests.Session()
40
+ _session.mount(
41
+ "http://",
42
+ requests.adapters.HTTPAdapter(
43
+ pool_connections=20,
44
+ pool_maxsize=20,
45
+ max_retries=0, # Handle retries explicitly
46
+ pool_block=False,
47
+ ),
48
+ )
49
+ _session.mount(
50
+ "https://",
51
+ requests.adapters.HTTPAdapter(
52
+ pool_connections=20, pool_maxsize=20, max_retries=0, pool_block=False
53
+ ),
54
+ )
55
+
42
56
  _supports_attempt_gets = None
43
57
  _supports_tag_mutation = None
44
58
 
@@ -59,7 +73,9 @@ class ServiceMetadataProvider(MetadataProvider):
59
73
  def compute_info(cls, val):
60
74
  v = val.rstrip("/")
61
75
  try:
62
- resp = requests.get(os.path.join(v, "ping"), headers=SERVICE_HEADERS.copy())
76
+ resp = cls._session.get(
77
+ os.path.join(v, "ping"), headers=SERVICE_HEADERS.copy()
78
+ )
63
79
  resp.raise_for_status()
64
80
  except: # noqa E722
65
81
  raise ValueError("Metaflow service [%s] unreachable." % v)
@@ -304,6 +320,55 @@ class ServiceMetadataProvider(MetadataProvider):
304
320
  self._register_system_metadata(run_id, step_name, task["task_id"], attempt)
305
321
  return task["task_id"], did_create
306
322
 
323
+ @classmethod
324
+ def filter_tasks_by_metadata(
325
+ cls,
326
+ flow_name: str,
327
+ run_id: str,
328
+ step_name: str,
329
+ field_name: str,
330
+ pattern: str,
331
+ ) -> List[str]:
332
+ """
333
+ Filter tasks by metadata field and pattern, returning task pathspecs that match criteria.
334
+
335
+ Parameters
336
+ ----------
337
+ flow_name : str
338
+ Flow name, that the run belongs to.
339
+ run_id: str
340
+ Run id, together with flow_id, that identifies the specific Run whose tasks to query
341
+ step_name: str
342
+ Step name to query tasks from
343
+ field_name: str
344
+ Metadata field name to query
345
+ pattern: str
346
+ Pattern to match in metadata field value
347
+
348
+ Returns
349
+ -------
350
+ List[str]
351
+ List of task pathspecs that satisfy the query
352
+ """
353
+ query_params = {
354
+ "metadata_field_name": field_name,
355
+ "pattern": pattern,
356
+ "step_name": step_name,
357
+ }
358
+ url = ServiceMetadataProvider._obj_path(flow_name, run_id, step_name)
359
+ url = f"{url}/filtered_tasks?{urlencode(query_params)}"
360
+ try:
361
+ resp = cls._request(None, url, "GET")
362
+ except Exception as e:
363
+ if e.http_code == 404:
364
+ # filter_tasks_by_metadata endpoint does not exist in the version of metadata service
365
+ # deployed currently. Raise a more informative error message.
366
+ raise MetaflowInternalError(
367
+ "The version of metadata service deployed currently does not support filtering tasks by metadata. "
368
+ "Upgrade Metadata service to version 2.15 or greater to use this feature."
369
+ ) from e
370
+ return resp
371
+
307
372
  @staticmethod
308
373
  def _obj_path(
309
374
  flow_name,
@@ -412,27 +477,27 @@ class ServiceMetadataProvider(MetadataProvider):
412
477
  if method == "GET":
413
478
  if monitor:
414
479
  with monitor.measure("metaflow.service_metadata.get"):
415
- resp = requests.get(url, headers=SERVICE_HEADERS.copy())
480
+ resp = cls._session.get(url, headers=SERVICE_HEADERS.copy())
416
481
  else:
417
- resp = requests.get(url, headers=SERVICE_HEADERS.copy())
482
+ resp = cls._session.get(url, headers=SERVICE_HEADERS.copy())
418
483
  elif method == "POST":
419
484
  if monitor:
420
485
  with monitor.measure("metaflow.service_metadata.post"):
421
- resp = requests.post(
486
+ resp = cls._session.post(
422
487
  url, headers=SERVICE_HEADERS.copy(), json=data
423
488
  )
424
489
  else:
425
- resp = requests.post(
490
+ resp = cls._session.post(
426
491
  url, headers=SERVICE_HEADERS.copy(), json=data
427
492
  )
428
493
  elif method == "PATCH":
429
494
  if monitor:
430
495
  with monitor.measure("metaflow.service_metadata.patch"):
431
- resp = requests.patch(
496
+ resp = cls._session.patch(
432
497
  url, headers=SERVICE_HEADERS.copy(), json=data
433
498
  )
434
499
  else:
435
- resp = requests.patch(
500
+ resp = cls._session.patch(
436
501
  url, headers=SERVICE_HEADERS.copy(), json=data
437
502
  )
438
503
  else:
@@ -475,7 +540,6 @@ class ServiceMetadataProvider(MetadataProvider):
475
540
  resp.text,
476
541
  )
477
542
  time.sleep(2**i)
478
-
479
543
  if resp:
480
544
  raise ServiceException(
481
545
  "Metadata request (%s) failed (code %s): %s"
@@ -499,9 +563,9 @@ class ServiceMetadataProvider(MetadataProvider):
499
563
  try:
500
564
  if monitor:
501
565
  with monitor.measure("metaflow.service_metadata.get"):
502
- resp = requests.get(url, headers=SERVICE_HEADERS.copy())
566
+ resp = cls._session.get(url, headers=SERVICE_HEADERS.copy())
503
567
  else:
504
- resp = requests.get(url, headers=SERVICE_HEADERS.copy())
568
+ resp = cls._session.get(url, headers=SERVICE_HEADERS.copy())
505
569
  except:
506
570
  if monitor:
507
571
  with monitor.count("metaflow.service_metadata.failed_request"):
@@ -10,7 +10,7 @@ import tarfile
10
10
  import time
11
11
  from urllib.error import URLError
12
12
  from urllib.request import urlopen
13
- from metaflow.metaflow_config import DATASTORE_LOCAL_DIR, CONDA_USE_FAST_INIT
13
+ from metaflow.metaflow_config import DATASTORE_LOCAL_DIR
14
14
  from metaflow.plugins import DATASTORES
15
15
  from metaflow.plugins.pypi.utils import MICROMAMBA_MIRROR_URL, MICROMAMBA_URL
16
16
  from metaflow.util import which
@@ -329,8 +329,6 @@ if __name__ == "__main__":
329
329
 
330
330
  @timer
331
331
  def fast_setup_environment(architecture, storage, env, prefix, pkgs_dir):
332
- install_fast_initializer(architecture)
333
-
334
332
  # Get package urls
335
333
  conda_pkgs = env["conda"]
336
334
  pypi_pkgs = env.get("pypi", [])
@@ -381,7 +379,9 @@ if __name__ == "__main__":
381
379
  with open(os.path.join(manifest_dir, MAGIC_FILE)) as f:
382
380
  env = json.load(f)[id_][architecture]
383
381
 
384
- if CONDA_USE_FAST_INIT:
382
+ if datastore_type == "s3":
383
+ # TODO: Remove this once fast-initializer is ready for all datastores
384
+ install_fast_initializer(architecture)
385
385
  fast_setup_environment(architecture, storage, env, prefix, pkgs_dir)
386
386
  else:
387
387
  setup_environment(
@@ -97,12 +97,13 @@ def _method_sanity_check(
97
97
  check_type(supplied_v, annotations[supplied_k])
98
98
  except TypeCheckError:
99
99
  raise TypeError(
100
- "Invalid type for '%s' (%s), expected: '%s', default is '%s'"
100
+ "Invalid type for '%s' (%s), expected: '%s', default is '%s' but found '%s'"
101
101
  % (
102
102
  supplied_k,
103
103
  type(supplied_k),
104
104
  annotations[supplied_k],
105
105
  defaults[supplied_k],
106
+ str(supplied_v),
106
107
  )
107
108
  )
108
109
 
@@ -218,7 +219,7 @@ def get_inspect_param_obj(p: Union[click.Argument, click.Option], kind: str):
218
219
  default=inspect.Parameter.empty if is_vararg else p.default,
219
220
  annotation=annotation,
220
221
  ),
221
- annotation,
222
+ Optional[TTuple[annotation]] if is_vararg else annotation,
222
223
  )
223
224
 
224
225
 
@@ -392,7 +393,9 @@ class MetaflowAPI(object):
392
393
  options = params.pop("options", {})
393
394
 
394
395
  for _, v in args.items():
395
- if isinstance(v, list):
396
+ if v is None:
397
+ continue
398
+ if isinstance(v, (list, tuple)):
396
399
  for i in v:
397
400
  components.append(i)
398
401
  else:
@@ -527,12 +530,6 @@ def extract_all_params(cmd_obj: Union[click.Command, click.Group]):
527
530
  )
528
531
  arg_parameters[each_param.name] = each_param
529
532
  elif isinstance(each_param, click.Option):
530
- if each_param.hidden:
531
- # Skip hidden options because users should not be setting those.
532
- # These are typically internal only options (used by the Runner in part
533
- # for example to pass state files or configs to pass local-config-file).
534
- continue
535
-
536
533
  opt_params_sigs[each_param.name], annotations[each_param.name] = (
537
534
  get_inspect_param_obj(each_param, inspect.Parameter.KEYWORD_ONLY)
538
535
  )
@@ -49,7 +49,7 @@ def process_messages(worker_type, worker):
49
49
 
50
50
 
51
51
  @click.command(help="Initialize workers")
52
- @tracing.cli_entrypoint("sidecar")
52
+ @tracing.cli("sidecar")
53
53
  @click.argument("worker-type")
54
54
  def main(worker_type):
55
55
  sidecar_type = SIDECARS.get(worker_type)
metaflow/task.py CHANGED
@@ -493,6 +493,25 @@ class MetaflowTask(object):
493
493
  )
494
494
  )
495
495
 
496
+ # Add runtime dag information to the metadata of the task
497
+ foreach_execution_path = ",".join(
498
+ [
499
+ "{}:{}".format(foreach_frame.step, foreach_frame.index)
500
+ for foreach_frame in foreach_stack
501
+ ]
502
+ )
503
+ if foreach_execution_path:
504
+ metadata.extend(
505
+ [
506
+ MetaDatum(
507
+ field="foreach-execution-path",
508
+ value=foreach_execution_path,
509
+ type="foreach-execution-path",
510
+ tags=metadata_tags,
511
+ ),
512
+ ]
513
+ )
514
+
496
515
  self.metadata.register_metadata(
497
516
  run_id,
498
517
  step_name,
@@ -559,6 +578,7 @@ class MetaflowTask(object):
559
578
  self.flow._success = False
560
579
  self.flow._task_ok = None
561
580
  self.flow._exception = None
581
+
562
582
  # Note: All internal flow attributes (ie: non-user artifacts)
563
583
  # should either be set prior to running the user code or listed in
564
584
  # FlowSpec._EPHEMERAL to allow for proper merging/importing of
@@ -616,7 +636,6 @@ class MetaflowTask(object):
616
636
  "graph_info": self.flow._graph_info,
617
637
  }
618
638
  )
619
-
620
639
  for deco in decorators:
621
640
  deco.task_pre_step(
622
641
  step_name,
@@ -728,7 +747,7 @@ class MetaflowTask(object):
728
747
  value=attempt_ok,
729
748
  type="internal_attempt_status",
730
749
  tags=["attempt_id:{0}".format(retry_count)],
731
- )
750
+ ),
732
751
  ],
733
752
  )
734
753
 
@@ -20,15 +20,15 @@ def post_fork():
20
20
  yield
21
21
 
22
22
 
23
- def cli_entrypoint(name: str):
24
- def cli_entrypoint_wrap(func):
23
+ def cli(name: str):
24
+ def cli_wrap(func):
25
25
  @wraps(func)
26
26
  def wrapper_func(*args, **kwargs):
27
27
  return func(*args, **kwargs)
28
28
 
29
29
  return wrapper_func
30
30
 
31
- return cli_entrypoint_wrap
31
+ return cli_wrap
32
32
 
33
33
 
34
34
  def inject_tracing_vars(env_dict: Dict[str, str]) -> Dict[str, str]:
@@ -40,7 +40,9 @@ def get_trace_id() -> str:
40
40
 
41
41
 
42
42
  @contextlib.contextmanager
43
- def traced(name, attrs={}):
43
+ def traced(name, attrs=None):
44
+ if attrs is None:
45
+ attrs = {}
44
46
  yield
45
47
 
46
48
 
@@ -54,17 +56,15 @@ def tracing(func):
54
56
 
55
57
  if not DISABLE_TRACING and (CONSOLE_TRACE_ENABLED or OTEL_ENDPOINT or ZIPKIN_ENDPOINT):
56
58
  try:
57
- # Overrides No-Op implementations if a specific provider is configured.
58
59
  from .tracing_modules import (
59
60
  init_tracing,
60
61
  post_fork,
61
- cli_entrypoint,
62
+ cli,
62
63
  inject_tracing_vars,
63
64
  get_trace_id,
64
65
  traced,
65
66
  tracing,
66
67
  )
67
-
68
68
  except ImportError as e:
69
69
  # We keep the errors silent by default so that having tracing environment variables present
70
70
  # does not affect users with no need for tracing.
@@ -3,60 +3,53 @@ from metaflow.metaflow_config import (
3
3
  OTEL_ENDPOINT,
4
4
  ZIPKIN_ENDPOINT,
5
5
  CONSOLE_TRACE_ENABLED,
6
+ SERVICE_AUTH_KEY,
7
+ SERVICE_HEADERS,
6
8
  )
7
9
 
8
10
 
9
11
  def get_span_exporter():
10
- if OTEL_ENDPOINT:
11
- return set_otel_exporter()
12
+ exporter_map = {
13
+ OTEL_ENDPOINT: _create_otel_exporter,
14
+ ZIPKIN_ENDPOINT: _create_zipkin_exporter,
15
+ CONSOLE_TRACE_ENABLED: _create_console_exporter,
16
+ }
12
17
 
13
- elif ZIPKIN_ENDPOINT:
14
- return set_zipkin_exporter()
18
+ for config, create_exporter in exporter_map.items():
19
+ if config:
20
+ return create_exporter()
15
21
 
16
- elif CONSOLE_TRACE_ENABLED:
17
- return set_console_exporter()
18
- else:
19
- print("WARNING: endpoints not set up for Opentelemetry", file=sys.stderr)
20
- return
22
+ print("WARNING: endpoints not set up for OpenTelemetry", file=sys.stderr)
23
+ return None
21
24
 
22
25
 
23
- def set_otel_exporter():
26
+ def _create_otel_exporter():
24
27
  from opentelemetry.exporter.otlp.proto.http.trace_exporter import OTLPSpanExporter
25
28
 
26
- from metaflow.metaflow_config import (
27
- SERVICE_AUTH_KEY,
28
- SERVICE_HEADERS,
29
- )
29
+ if not any([SERVICE_AUTH_KEY, SERVICE_HEADERS]):
30
+ print("WARNING: no auth settings for OpenTelemetry", file=sys.stderr)
31
+ return None
32
+
33
+ config = {
34
+ "endpoint": OTEL_ENDPOINT,
35
+ "timeout": 1,
36
+ }
30
37
 
31
38
  if SERVICE_AUTH_KEY:
32
- span_exporter = OTLPSpanExporter(
33
- endpoint=OTEL_ENDPOINT,
34
- headers={"x-api-key": SERVICE_AUTH_KEY},
35
- timeout=1,
36
- )
39
+ config["headers"] = {"x-api-key": SERVICE_AUTH_KEY}
37
40
  elif SERVICE_HEADERS:
38
- span_exporter = OTLPSpanExporter(
39
- endpoint=OTEL_ENDPOINT,
40
- headers=SERVICE_HEADERS,
41
- timeout=1,
42
- )
43
- else:
44
- print("WARNING: no auth settings for Opentelemetry", file=sys.stderr)
45
- return
46
- return span_exporter
47
-
48
-
49
- def set_zipkin_exporter():
41
+ config["headers"] = SERVICE_HEADERS
42
+
43
+ return OTLPSpanExporter(**config)
44
+
45
+
46
+ def _create_zipkin_exporter():
50
47
  from opentelemetry.exporter.zipkin.proto.http import ZipkinExporter
51
48
 
52
- span_exporter = ZipkinExporter(
53
- endpoint=ZIPKIN_ENDPOINT,
54
- )
55
- return span_exporter
49
+ return ZipkinExporter(endpoint=ZIPKIN_ENDPOINT)
56
50
 
57
51
 
58
- def set_console_exporter():
52
+ def _create_console_exporter():
59
53
  from opentelemetry.sdk.trace.export import ConsoleSpanExporter
60
54
 
61
- span_exporter = ConsoleSpanExporter()
62
- return span_exporter
55
+ return ConsoleSpanExporter()