ob-metaflow 2.11.14.1__py2.py3-none-any.whl → 2.11.15.2__py2.py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of ob-metaflow might be problematic. Click here for more details.

metaflow/cli.py CHANGED
@@ -287,126 +287,6 @@ def dump(obj, input_path, private=None, max_value_size=None, include=None, file=
287
287
  echo("Artifacts written to *%s*" % file)
288
288
 
289
289
 
290
- @cli.command(
291
- help="Show stdout/stderr produced by a task or all tasks in a step. "
292
- "The format for input-path is either <run_id>/<step_name> or "
293
- "<run_id>/<step_name>/<task_id>."
294
- )
295
- @click.argument("input-path")
296
- @click.option(
297
- "--stdout/--no-stdout",
298
- default=False,
299
- show_default=True,
300
- help="Show stdout of the task.",
301
- )
302
- @click.option(
303
- "--stderr/--no-stderr",
304
- default=False,
305
- show_default=True,
306
- help="Show stderr of the task.",
307
- )
308
- @click.option(
309
- "--both/--no-both",
310
- default=True,
311
- show_default=True,
312
- help="Show both stdout and stderr of the task.",
313
- )
314
- @click.option(
315
- "--timestamps/--no-timestamps",
316
- default=False,
317
- show_default=True,
318
- help="Show timestamps.",
319
- )
320
- @click.pass_obj
321
- def logs(obj, input_path, stdout=None, stderr=None, both=None, timestamps=False):
322
- types = set()
323
- if stdout:
324
- types.add("stdout")
325
- both = False
326
- if stderr:
327
- types.add("stderr")
328
- both = False
329
- if both:
330
- types.update(("stdout", "stderr"))
331
-
332
- streams = list(sorted(types, reverse=True))
333
-
334
- # Pathspec can either be run_id/step_name or run_id/step_name/task_id.
335
- parts = input_path.split("/")
336
- if len(parts) == 2:
337
- run_id, step_name = parts
338
- task_id = None
339
- elif len(parts) == 3:
340
- run_id, step_name, task_id = parts
341
- else:
342
- raise CommandException(
343
- "input_path should either be run_id/step_name "
344
- "or run_id/step_name/task_id"
345
- )
346
-
347
- datastore_set = TaskDataStoreSet(
348
- obj.flow_datastore, run_id, steps=[step_name], allow_not_done=True
349
- )
350
- if task_id:
351
- ds_list = [
352
- TaskDataStore(
353
- obj.flow_datastore,
354
- run_id=run_id,
355
- step_name=step_name,
356
- task_id=task_id,
357
- mode="r",
358
- allow_not_done=True,
359
- )
360
- ]
361
- else:
362
- ds_list = list(datastore_set) # get all tasks
363
-
364
- if ds_list:
365
-
366
- def echo_unicode(line, **kwargs):
367
- click.secho(line.decode("UTF-8", errors="replace"), **kwargs)
368
-
369
- # old style logs are non mflog-style logs
370
- maybe_old_style = True
371
- for ds in ds_list:
372
- echo(
373
- "Dumping logs of run_id=*{run_id}* "
374
- "step=*{step}* task_id=*{task_id}*".format(
375
- run_id=ds.run_id, step=ds.step_name, task_id=ds.task_id
376
- ),
377
- fg="magenta",
378
- )
379
-
380
- for stream in streams:
381
- echo(stream, bold=True)
382
- logs = ds.load_logs(LOG_SOURCES, stream)
383
- if any(data for _, data in logs):
384
- # attempt to read new, mflog-style logs
385
- for line in mflog.merge_logs([blob for _, blob in logs]):
386
- if timestamps:
387
- ts = mflog.utc_to_local(line.utc_tstamp)
388
- tstamp = ts.strftime("%Y-%m-%d %H:%M:%S.%f")[:-3]
389
- click.secho(tstamp + " ", fg=LOGGER_TIMESTAMP, nl=False)
390
- echo_unicode(line.msg)
391
- maybe_old_style = False
392
- elif maybe_old_style:
393
- # if they are not available, we may be looking at
394
- # a legacy run (unless we have seen new-style data already
395
- # for another stream). This return an empty string if
396
- # nothing is found
397
- log = ds.load_log_legacy(stream)
398
- if log and timestamps:
399
- raise CommandException(
400
- "We can't show --timestamps for old runs. Sorry!"
401
- )
402
- echo_unicode(log, nl=False)
403
- else:
404
- raise CommandException(
405
- "No Tasks found at the given path -- "
406
- "either none exist or none have started yet"
407
- )
408
-
409
-
410
290
  # TODO - move step and init under a separate 'internal' subcommand
411
291
 
412
292
 
@@ -22,7 +22,7 @@ class TaskDataStoreSet(object):
22
22
  prefetch_data_artifacts=None,
23
23
  allow_not_done=False,
24
24
  ):
25
- self.task_datastores = flow_datastore.get_latest_task_datastores(
25
+ self.task_datastores = flow_datastore.get_task_datastores(
26
26
  run_id, steps=steps, pathspecs=pathspecs, allow_not_done=allow_not_done
27
27
  )
28
28
 
@@ -67,8 +67,15 @@ class FlowDataStore(object):
67
67
  def datastore_root(self):
68
68
  return self._storage_impl.datastore_root
69
69
 
70
- def get_latest_task_datastores(
71
- self, run_id=None, steps=None, pathspecs=None, allow_not_done=False
70
+ def get_task_datastores(
71
+ self,
72
+ run_id=None,
73
+ steps=None,
74
+ pathspecs=None,
75
+ allow_not_done=False,
76
+ attempt=None,
77
+ include_prior=False,
78
+ mode="r",
72
79
  ):
73
80
  """
74
81
  Return a list of TaskDataStore for a subset of the tasks.
@@ -93,6 +100,12 @@ class FlowDataStore(object):
93
100
  allow_not_done : bool, optional
94
101
  If True, returns the latest attempt of a task even if that attempt
95
102
  wasn't marked as done, by default False
103
+ attempt : int, optional
104
+ Attempt number of the tasks to return. If not provided, returns latest attempt.
105
+ include_prior : boolean, default False
106
+ If True, returns all attempts up to and including attempt.
107
+ mode : str, default "r"
108
+ Mode to initialize the returned TaskDataStores in.
96
109
 
97
110
  Returns
98
111
  -------
@@ -126,8 +139,13 @@ class FlowDataStore(object):
126
139
  if task.is_file is False
127
140
  ]
128
141
  urls = []
142
+ # parse content urls for specific attempt only, or for all attempts in max range
143
+ attempt_range = range(metaflow_config.MAX_ATTEMPTS)
144
+ # we have no reason to check for attempts greater than MAX_ATTEMPTS, as they do not exist.
145
+ if attempt is not None and attempt <= metaflow_config.MAX_ATTEMPTS - 1:
146
+ attempt_range = range(attempt + 1) if include_prior else [attempt]
129
147
  for task_url in task_urls:
130
- for attempt in range(metaflow_config.MAX_ATTEMPTS):
148
+ for attempt in attempt_range:
131
149
  for suffix in [
132
150
  TaskDataStore.METADATA_DATA_SUFFIX,
133
151
  TaskDataStore.METADATA_ATTEMPT_SUFFIX,
@@ -168,11 +186,19 @@ class FlowDataStore(object):
168
186
  for (run, step, task), attempt in latest_started_attempts.items()
169
187
  )
170
188
  if allow_not_done:
171
- latest_to_fetch = latest_started_attempts
189
+ latest_to_fetch = (
190
+ done_attempts.union(latest_started_attempts)
191
+ if include_prior
192
+ else latest_started_attempts
193
+ )
172
194
  else:
173
- latest_to_fetch = latest_started_attempts & done_attempts
195
+ latest_to_fetch = (
196
+ done_attempts
197
+ if include_prior
198
+ else (latest_started_attempts & done_attempts)
199
+ )
174
200
  latest_to_fetch = [
175
- (v[0], v[1], v[2], v[3], data_objs.get(v), "r", allow_not_done)
201
+ (v[0], v[1], v[2], v[3], data_objs.get(v), mode, allow_not_done)
176
202
  for v in latest_to_fetch
177
203
  ]
178
204
  return list(itertools.starmap(self.get_task_datastore, latest_to_fetch))
@@ -173,6 +173,26 @@ class TaskDataStore(object):
173
173
  if data_obj is not None:
174
174
  self._objects = data_obj.get("objects", {})
175
175
  self._info = data_obj.get("info", {})
176
+ elif self._mode == "d":
177
+ self._objects = {}
178
+ self._info = {}
179
+
180
+ if self._attempt is None:
181
+ for i in range(metaflow_config.MAX_ATTEMPTS):
182
+ check_meta = self._metadata_name_for_attempt(
183
+ self.METADATA_ATTEMPT_SUFFIX, i
184
+ )
185
+ if self.has_metadata(check_meta, add_attempt=False):
186
+ self._attempt = i
187
+
188
+ # Do not allow destructive operations on the datastore if attempt is still in flight
189
+ # and we explicitly did not allow operating on running tasks.
190
+ if not allow_not_done and not self.has_metadata(self.METADATA_DONE_SUFFIX):
191
+ raise DataException(
192
+ "No completed attempts of the task was found for task '%s'"
193
+ % self._path
194
+ )
195
+
176
196
  else:
177
197
  raise DataException("Unknown datastore mode: '%s'" % self._mode)
178
198
 
@@ -750,6 +770,36 @@ class TaskDataStore(object):
750
770
  to_store_dict[n] = data
751
771
  self._save_file(to_store_dict)
752
772
 
773
+ @require_mode("d")
774
+ def scrub_logs(self, logsources, stream, attempt_override=None):
775
+ path_logsources = {
776
+ self._metadata_name_for_attempt(
777
+ self._get_log_location(s, stream),
778
+ attempt_override=attempt_override,
779
+ ): s
780
+ for s in logsources
781
+ }
782
+
783
+ # Legacy log paths
784
+ legacy_log = self._metadata_name_for_attempt(
785
+ "%s.log" % stream, attempt_override
786
+ )
787
+ path_logsources[legacy_log] = stream
788
+
789
+ existing_paths = [
790
+ path
791
+ for path in path_logsources.keys()
792
+ if self.has_metadata(path, add_attempt=False)
793
+ ]
794
+
795
+ # Replace log contents with [REDACTED source stream]
796
+ to_store_dict = {
797
+ path: bytes("[REDACTED %s %s]" % (path_logsources[path], stream), "utf-8")
798
+ for path in existing_paths
799
+ }
800
+
801
+ self._save_file(to_store_dict, add_attempt=False, allow_overwrite=True)
802
+
753
803
  @require_mode("r")
754
804
  def load_log_legacy(self, stream, attempt_override=None):
755
805
  """
@@ -154,6 +154,13 @@ AWS_SECRETS_MANAGER_DEFAULT_REGION = from_conf("AWS_SECRETS_MANAGER_DEFAULT_REGI
154
154
  # - "projects/1234567890/secrets/foo-" -> "projects/1234567890/secrets/foo-mysecret"
155
155
  GCP_SECRET_MANAGER_PREFIX = from_conf("GCP_SECRET_MANAGER_PREFIX")
156
156
 
157
+ # Secrets Backend - Azure Key Vault prefix. With this, users don't have to
158
+ # specify the full https:// vault url in the @secret decorator.
159
+ #
160
+ # It does not make a difference if the prefix ends in a / or not. We will handle either
161
+ # case correctly.
162
+ AZURE_KEY_VAULT_PREFIX = from_conf("AZURE_KEY_VAULT_PREFIX")
163
+
157
164
  # The root directory to save artifact pulls in, when using S3 or Azure
158
165
  ARTIFACT_LOCALROOT = from_conf("ARTIFACT_LOCALROOT", os.getcwd())
159
166
 
@@ -471,6 +478,7 @@ def get_pinned_conda_libs(python_version, datastore_type):
471
478
  elif datastore_type == "azure":
472
479
  pins["azure-identity"] = ">=1.10.0"
473
480
  pins["azure-storage-blob"] = ">=12.12.0"
481
+ pins["azure-keyvault-secrets"] = ">=4.8.0"
474
482
  elif datastore_type == "gs":
475
483
  pins["google-cloud-storage"] = ">=2.5.0"
476
484
  pins["google-auth"] = ">=2.11.0"
@@ -124,7 +124,7 @@ class MetaflowEnvironment(object):
124
124
  cmds.append("%s -m pip install awscli boto3" % self._python())
125
125
  elif datastore_type == "azure":
126
126
  cmds.append(
127
- "%s -m pip install azure-identity azure-storage-blob simple-azure-blob-downloader -qqq"
127
+ "%s -m pip install azure-identity azure-storage-blob azure-keyvault-secrets simple-azure-blob-downloader -qqq"
128
128
  % self._python()
129
129
  )
130
130
  elif datastore_type == "gs":
@@ -14,6 +14,7 @@ CLIS_DESC = [
14
14
  ("argo-workflows", ".argo.argo_workflows_cli.cli"),
15
15
  ("card", ".cards.card_cli.cli"),
16
16
  ("tag", ".tag_cli.cli"),
17
+ ("logs", ".logs_cli.cli"),
17
18
  ]
18
19
 
19
20
  from .test_unbounded_foreach_decorator import InternalTestUnboundedForeachInput
@@ -124,6 +125,10 @@ SECRETS_PROVIDERS_DESC = [
124
125
  "gcp-secret-manager",
125
126
  ".gcp.gcp_secret_manager_secrets_provider.GcpSecretManagerSecretsProvider",
126
127
  ),
128
+ (
129
+ "az-key-vault",
130
+ ".azure.azure_secret_manager_secrets_provider.AzureKeyVaultSecretsProvider",
131
+ ),
127
132
  ]
128
133
 
129
134
  AZURE_CLIENT_PROVIDERS_DESC = [
@@ -32,6 +32,7 @@ from metaflow.metaflow_config import (
32
32
  S3_ENDPOINT_URL,
33
33
  SERVICE_HEADERS,
34
34
  SERVICE_INTERNAL_URL,
35
+ AZURE_KEY_VAULT_PREFIX,
35
36
  )
36
37
 
37
38
  from metaflow.metaflow_config_funcs import config_values
@@ -412,6 +413,9 @@ class Airflow(object):
412
413
  if GCP_SECRET_MANAGER_PREFIX:
413
414
  env["METAFLOW_GCP_SECRET_MANAGER_PREFIX"] = GCP_SECRET_MANAGER_PREFIX
414
415
 
416
+ if AZURE_KEY_VAULT_PREFIX:
417
+ env["METAFLOW_AZURE_KEY_VAULT_PREFIX"] = AZURE_KEY_VAULT_PREFIX
418
+
415
419
  env.update(additional_mf_variables)
416
420
 
417
421
  service_account = (
@@ -33,6 +33,7 @@ from metaflow.metaflow_config import (
33
33
  DEFAULT_METADATA,
34
34
  DEFAULT_SECRETS_BACKEND_TYPE,
35
35
  GCP_SECRET_MANAGER_PREFIX,
36
+ AZURE_KEY_VAULT_PREFIX,
36
37
  KUBERNETES_FETCH_EC2_METADATA,
37
38
  KUBERNETES_LABELS,
38
39
  KUBERNETES_NAMESPACE,
@@ -1420,6 +1421,7 @@ class ArgoWorkflows(object):
1420
1421
  "METAFLOW_AWS_SECRETS_MANAGER_DEFAULT_REGION"
1421
1422
  ] = AWS_SECRETS_MANAGER_DEFAULT_REGION
1422
1423
  env["METAFLOW_GCP_SECRET_MANAGER_PREFIX"] = GCP_SECRET_MANAGER_PREFIX
1424
+ env["METAFLOW_AZURE_KEY_VAULT_PREFIX"] = AZURE_KEY_VAULT_PREFIX
1423
1425
 
1424
1426
  # support for Azure
1425
1427
  env[
@@ -10,7 +10,7 @@ from metaflow.exception import CommandException, METAFLOW_EXIT_DISALLOW_RETRY
10
10
  from metaflow.metadata.util import sync_local_metadata_from_datastore
11
11
  from metaflow.metaflow_config import DATASTORE_LOCAL_DIR
12
12
  from metaflow.mflog import TASK_LOG_SOURCE
13
-
13
+ from metaflow.unbounded_foreach import UBF_CONTROL, UBF_TASK
14
14
  from .batch import Batch, BatchKilledException
15
15
 
16
16
 
@@ -150,8 +150,10 @@ def kill(ctx, run_id, user, my_runs):
150
150
  @click.option("--tmpfs-tempdir", is_flag=True, help="tmpfs requirement for AWS Batch.")
151
151
  @click.option("--tmpfs-size", help="tmpfs requirement for AWS Batch.")
152
152
  @click.option("--tmpfs-path", help="tmpfs requirement for AWS Batch.")
153
- # TODO: Maybe remove it altogether since it's not used here
154
- @click.option("--ubf-context", default=None, type=click.Choice([None, "ubf_control"]))
153
+ # NOTE: ubf-context is not explicitly used, but @parallel decorator tries to pass this so keep it for now
154
+ @click.option(
155
+ "--ubf-context", default=None, type=click.Choice(["none", UBF_CONTROL, UBF_TASK])
156
+ )
155
157
  @click.option("--host-volumes", multiple=True)
156
158
  @click.option("--efs-volumes", multiple=True)
157
159
  @click.option(
@@ -344,7 +346,7 @@ def step(
344
346
  log_options=log_options,
345
347
  num_parallel=num_parallel,
346
348
  )
347
- except Exception as e:
349
+ except Exception:
348
350
  traceback.print_exc()
349
351
  _sync_metadata()
350
352
  sys.exit(METAFLOW_EXIT_DISALLOW_RETRY)
@@ -10,4 +10,4 @@ class MetaflowAzureResourceError(MetaflowException):
10
10
 
11
11
 
12
12
  class MetaflowAzurePackageError(MetaflowException):
13
- headline = "Missing required packages 'azure-identity' and 'azure-storage-blob'"
13
+ headline = "Missing required packages 'azure-identity' and 'azure-storage-blob' and 'azure-keyvault-secrets'"
@@ -0,0 +1,236 @@
1
+ from metaflow.plugins.secrets import SecretsProvider
2
+ import re
3
+ import base64
4
+ import codecs
5
+ from urllib.parse import urlparse
6
+ from metaflow.exception import MetaflowException
7
+ import sys
8
+ from metaflow.metaflow_config import AZURE_KEY_VAULT_PREFIX
9
+ from metaflow.plugins.azure.azure_credential import (
10
+ create_cacheable_azure_credential,
11
+ )
12
+
13
+
14
+ class MetaflowAzureKeyVaultBadVault(MetaflowException):
15
+ """Raised when the secretid is fully qualified but does not have the right key vault domain"""
16
+
17
+
18
+ class MetaflowAzureKeyVaultBadSecretType(MetaflowException):
19
+ """Raised when the secret type is anything except secrets"""
20
+
21
+
22
+ class MetaflowAzureKeyVaultBadSecretPath(MetaflowException):
23
+ """Raised when the secret path does not match to expected length"""
24
+
25
+
26
+ class MetaflowAzureKeyVaultBadSecretName(MetaflowException):
27
+ """Raised when the secret name does not match expected pattern"""
28
+
29
+
30
+ class MetaflowAzureKeyVaultBadSecretVersion(MetaflowException):
31
+ """Raised when the secret version does not match expected pattern"""
32
+
33
+
34
+ class MetaflowAzureKeyVaultBadSecret(MetaflowException):
35
+ """Raised when the secret does not match supported patterns in Metaflow"""
36
+
37
+
38
+ class AzureKeyVaultSecretsProvider(SecretsProvider):
39
+ TYPE = "az-key-vault"
40
+ key_vault_domains = [
41
+ ".vault.azure.net",
42
+ ".vault.azure.cn",
43
+ ".vault.usgovcloudapi.net",
44
+ ".vault.microsoftazure.de",
45
+ ]
46
+ supported_vault_object_types = ["secrets"]
47
+
48
+ # https://learn.microsoft.com/en-us/azure/key-vault/general/about-keys-secrets-certificates has details on vault name structure
49
+ # Vault name and Managed HSM pool name must be a 3-24 character string, containing only 0-9, a-z, A-Z, and not consecutive -.
50
+ def _is_valid_vault_name(self, vault_name):
51
+ vault_name_pattern = r"^(?!.*--)[a-zA-Z0-9-]{3,24}$"
52
+ return re.match(vault_name_pattern, vault_name) is not None
53
+
54
+ # The type of the object can be, "keys", "secrets", or "certificates".
55
+ # Currently only secrets will be supported
56
+ def _is_valid_object_type(self, secret_type):
57
+ for type in self.supported_vault_object_types:
58
+ if secret_type == type:
59
+ return True
60
+ return False
61
+
62
+ # The secret name must be a 1-127 character string, starting with a letter and containing only 0-9, a-z, A-Z, and -.
63
+ def _is_valid_secret_name(self, secret_name):
64
+ secret_name_pattern = r"^[a-zA-Z][a-zA-Z0-9-]{0,126}$"
65
+ return re.match(secret_name_pattern, secret_name) is not None
66
+
67
+ # An object-version is a system-generated, 32 character string identifier that is optionally used to address a unique version of an object.
68
+ def _is_valid_object_version(self, secret_version):
69
+ object_version_pattern = r"^[a-zA-Z0-9]{32}$"
70
+ return re.match(object_version_pattern, secret_version) is not None
71
+
72
+ # This function will check if the secret_id is fully qualified url. It will return True iff the secret_id is of the form:
73
+ # https://myvault.vault.azure.net/secrets/mysecret/ec96f02080254f109c51a1f14cdb1931 OR
74
+ # https://myvault.vault.azure.net/secrets/mysecret/
75
+ # validating the above as per recommendations in https://devblogs.microsoft.com/azure-sdk/guidance-for-applications-using-the-key-vault-libraries/
76
+ def _is_secret_id_fully_qualified_url(self, secret_id):
77
+ # if the secret_id is None/empty/does not start with https then return false
78
+ if secret_id is None or secret_id == "" or not secret_id.startswith("https://"):
79
+ return False
80
+ try:
81
+ parsed_vault_url = urlparse(secret_id)
82
+ except ValueError:
83
+ print(f"invalid vault url", file=sys.stderr)
84
+ return False
85
+ hostname = parsed_vault_url.netloc
86
+
87
+ k_v_domain_found = False
88
+ actual_k_v_domain = ""
89
+ for k_v_domain in self.key_vault_domains:
90
+ if k_v_domain in hostname:
91
+ k_v_domain_found = True
92
+ actual_k_v_domain = k_v_domain
93
+ break
94
+ if not k_v_domain_found:
95
+ # the secret_id started with https:// however the key_vault_domains
96
+ # were not present in the secret_id which means
97
+ raise MetaflowAzureKeyVaultBadVault(f"bad key vault domain {secret_id}")
98
+
99
+ # given the secret_id seems to have a valid key vault domain
100
+ # lets verify that the vault name corresponds to its regex.
101
+ vault_name = hostname[: -len(actual_k_v_domain)]
102
+ # verify the vault name pattern
103
+ if not self._is_valid_vault_name(vault_name):
104
+ raise MetaflowAzureKeyVaultBadVault(f"bad key vault name {vault_name}")
105
+
106
+ path_parts = parsed_vault_url.path.strip("/").split("/")
107
+ total_path_parts = len(path_parts)
108
+ if total_path_parts < 2 or total_path_parts > 3:
109
+ raise MetaflowAzureKeyVaultBadSecretPath(
110
+ f"bad secret uri path {path_parts}"
111
+ )
112
+
113
+ object_type = path_parts[0]
114
+ if not self._is_valid_object_type(object_type):
115
+ raise MetaflowAzureKeyVaultBadSecretType(f"bad secret type {object_type}")
116
+
117
+ secret_name = path_parts[1]
118
+ if not self._is_valid_secret_name(secret_name=secret_name):
119
+ raise MetaflowAzureKeyVaultBadSecretName(f"bad secret name {secret_name}")
120
+
121
+ if total_path_parts == 3:
122
+ if not self._is_valid_object_version(path_parts[2]):
123
+ raise MetaflowAzureKeyVaultBadSecretVersion(
124
+ f"bad secret version {path_parts[2]}"
125
+ )
126
+
127
+ return True
128
+
129
+ # This function will validate the correctness of the partial secret id.
130
+ # It will attempt to construct the fully qualified secret URL internally and
131
+ # call the _is_secret_id_fully_qualified_url to check validity
132
+ def _is_partial_secret_valid(self, secret_id):
133
+ secret_parts = secret_id.strip("/").split("/")
134
+ total_secret_parts = len(secret_parts)
135
+ if total_secret_parts < 1 or total_secret_parts > 2:
136
+ return False
137
+
138
+ # since the secret_id is supposedly a partial id, the AZURE_KEY_VAULT_PREFIX
139
+ # must be set.
140
+ if not AZURE_KEY_VAULT_PREFIX:
141
+ raise ValueError(
142
+ f"cannot use simple secret id without setting METAFLOW_AZURE_KEY_VAULT_PREFIX. {AZURE_KEY_VAULT_PREFIX}"
143
+ )
144
+ domain = AZURE_KEY_VAULT_PREFIX.rstrip("/")
145
+ full_secret = f"{domain}/secrets/{secret_id}"
146
+ if not self._is_secret_id_fully_qualified_url(full_secret):
147
+ return False
148
+
149
+ return True
150
+
151
+ def _sanitize_key_as_env_var(self, key):
152
+ """
153
+ Sanitize a key as an environment variable name.
154
+ This is purely a convenience trade-off to cover common cases well, vs. introducing
155
+ ambiguities (e.g. did the final '_' come from '.', or '-' or is original?).
156
+
157
+ 1/27/2023(jackie):
158
+
159
+ We start with few rules and should *sparingly* add more over time.
160
+ Also, it's TBD whether all possible providers will share the same sanitization logic.
161
+ Therefore we will keep this function private for now
162
+ """
163
+ return key.replace("-", "_").replace(".", "_").replace("/", "_")
164
+
165
+ def get_secret_as_dict(self, secret_id, options={}, role=None):
166
+ # https://learn.microsoft.com/en-us/azure/app-service/app-service-key-vault-references?tabs=azure-cli has a lot of details on
167
+ # the patterns used in key vault
168
+ # Vault names and Managed HSM pool names are selected by the user and are globally unique.
169
+ # Vault name and Managed HSM pool name must be a 3-24 character string, containing only 0-9, a-z, A-Z, and not consecutive -.
170
+ # object-type The type of the object. As of 05/08/24 only "secrets", are supported
171
+ # object-name An object-name is a user provided name for and must be unique within a key vault. The name must be a 1-127 character string, starting with a letter and containing only 0-9, a-z, A-Z, and -.
172
+ # object-version An object-version is a system-generated, 32 character string identifier that is optionally used to address a unique version of an object.
173
+
174
+ # We allow these forms of secret_id:
175
+ #
176
+ # 1. Full path like https://<key-vault-name><.vault-domain>/secrets/<secret-name>/<secret-version>. This is what you
177
+ # see in Azure portal and is easy to copy paste.
178
+ #
179
+ # 2. Full path but without the version like https://<key-vault-name><.vault-domain>/secrets/<secret-name>
180
+ #
181
+ # 3. Simple string like mysecret. This corresponds to the SecretName.
182
+ #
183
+ # 4. Simple string with <secret-name>/<secret-version> suffix like mysecret/123
184
+
185
+ # The latter two forms require METAFLOW_AZURE_KEY_VAULT_PREFIX to be set.
186
+
187
+ # if the secret_id is None/empty/does not start with https then return false
188
+ if secret_id is None or secret_id == "":
189
+ raise MetaflowAzureKeyVaultBadSecret(f"empty secret id is not supported")
190
+
191
+ # check if the passed in secret is a short-form ( #3/#4 in the above comment)
192
+ if not secret_id.startswith("https://"):
193
+ # check if the secret_id is of form `secret_name` OR `secret_name/secret_version`
194
+ if not self._is_partial_secret_valid(secret_id=secret_id):
195
+ raise MetaflowAzureKeyVaultBadSecret(
196
+ f"unsupported partial secret {secret_id}"
197
+ )
198
+
199
+ domain = AZURE_KEY_VAULT_PREFIX.rstrip("/")
200
+ full_secret = f"{domain}/secrets/{secret_id}"
201
+
202
+ # if the secret id is passed as a URL - then check if the url is fully qualified
203
+ if secret_id.startswith("https://"):
204
+ if not self._is_secret_id_fully_qualified_url(secret_id=secret_id):
205
+ raise MetaflowException(f"unsupported secret {secret_id}")
206
+ full_secret = secret_id
207
+
208
+ # at this point I know that the secret URL is good so we can start creating the Secret Client
209
+ az_credentials = create_cacheable_azure_credential()
210
+ res = urlparse(full_secret)
211
+ az_vault_url = f"{res.scheme}://{res.netloc}" # https://myvault.vault.azure.net
212
+ secret_data = res.path.strip("/").split("/")[1:]
213
+ secret_name = secret_data[0]
214
+ secret_version = None
215
+ if len(secret_data) > 1:
216
+ secret_version = secret_data[1]
217
+
218
+ from azure.keyvault.secrets import SecretClient
219
+
220
+ client = SecretClient(vault_url=az_vault_url, credential=az_credentials)
221
+
222
+ key_vault_secret_val = client.get_secret(
223
+ name=secret_name, version=secret_version
224
+ )
225
+
226
+ result = {}
227
+
228
+ if options.get("env_var_name") is not None:
229
+ env_var_name = options["env_var_name"]
230
+ sanitized_key = self._sanitize_key_as_env_var(env_var_name)
231
+ else:
232
+ sanitized_key = self._sanitize_key_as_env_var(key_vault_secret_val.name)
233
+
234
+ response_payload = key_vault_secret_val.value
235
+ result[sanitized_key] = response_payload
236
+ return result
@@ -1245,12 +1245,12 @@ class S3(object):
1245
1245
 
1246
1246
  def _store():
1247
1247
  for key_obj in key_objs:
1248
- if isinstance(key_obj, tuple):
1249
- key = key_obj[0]
1250
- obj = key_obj[1]
1251
- else:
1248
+ if isinstance(key_obj, S3PutObject):
1252
1249
  key = key_obj.key
1253
1250
  obj = key_obj.value
1251
+ else:
1252
+ key = key_obj[0]
1253
+ obj = key_obj[1]
1254
1254
  store_info = {
1255
1255
  "key": key,
1256
1256
  "content_type": getattr(key_obj, "content_type", None),
@@ -1319,12 +1319,12 @@ class S3(object):
1319
1319
 
1320
1320
  def _check():
1321
1321
  for key_path in key_paths:
1322
- if isinstance(key_path, tuple):
1323
- key = key_path[0]
1324
- path = key_path[1]
1325
- else:
1322
+ if isinstance(key_path, S3PutObject):
1326
1323
  key = key_path.key
1327
1324
  path = key_path.path
1325
+ else:
1326
+ key = key_path[0]
1327
+ path = key_path[1]
1328
1328
  store_info = {
1329
1329
  "key": key,
1330
1330
  "content_type": getattr(key_path, "content_type", None),
@@ -32,6 +32,7 @@ from metaflow.metaflow_config import (
32
32
  DEFAULT_METADATA,
33
33
  DEFAULT_SECRETS_BACKEND_TYPE,
34
34
  GCP_SECRET_MANAGER_PREFIX,
35
+ AZURE_KEY_VAULT_PREFIX,
35
36
  KUBERNETES_FETCH_EC2_METADATA,
36
37
  KUBERNETES_LABELS,
37
38
  KUBERNETES_SANDBOX_INIT_SCRIPT,
@@ -261,6 +262,9 @@ class Kubernetes(object):
261
262
  .environment_variable(
262
263
  "METAFLOW_GCP_SECRET_MANAGER_PREFIX", GCP_SECRET_MANAGER_PREFIX
263
264
  )
265
+ .environment_variable(
266
+ "METAFLOW_AZURE_KEY_VAULT_PREFIX", AZURE_KEY_VAULT_PREFIX
267
+ )
264
268
  .environment_variable("METAFLOW_S3_ENDPOINT_URL", S3_ENDPOINT_URL)
265
269
  .environment_variable(
266
270
  "METAFLOW_AZURE_STORAGE_BLOB_SERVICE_ENDPOINT",
@@ -0,0 +1,358 @@
1
+ from metaflow._vendor import click
2
+ from metaflow.cli import LOGGER_TIMESTAMP
3
+
4
+ from ..exception import CommandException
5
+ from ..datastore import TaskDataStoreSet, TaskDataStore
6
+
7
+
8
+ from ..mflog import mflog, LOG_SOURCES
9
+
10
+ # main motivation from https://github.com/pallets/click/issues/430
11
+ # in order to support a default command being called for a Click group.
12
+ #
13
+ # NOTE: We need this in order to not introduce breaking changes to existing CLI, as we wanted to
14
+ # nest both existing `logs` and the new `logs scrub` under a shared group, but `logs` already has
15
+ # a well defined behavior of showing the logs.
16
+ class CustomGroup(click.Group):
17
+ def __init__(self, name=None, commands=None, default_cmd=None, **attrs):
18
+ super(CustomGroup, self).__init__(name, commands, **attrs)
19
+ self.default_cmd = default_cmd
20
+
21
+ def get_command(self, ctx, cmd_name):
22
+ if cmd_name not in self.list_commands(ctx):
23
+ # input from the CLI does not match a command, so we pass that
24
+ # as the args to the default command instead.
25
+ ctx.passed_cmd = cmd_name
26
+ cmd_name = self.default_cmd
27
+ return super(CustomGroup, self).get_command(ctx, cmd_name)
28
+
29
+ def parse_args(self, ctx, args):
30
+ # We first try to parse args as is, to determine whether we need to fall back to the default commmand
31
+ # if any options are supplied, the parse will fail, as the group does not support the options.
32
+ # In this case we fallback to the default command, inserting that as the first arg and parsing again.
33
+ # copy args as trying to parse will destroy them.
34
+ original_args = list(args)
35
+ try:
36
+ super().parse_args(ctx, args)
37
+ args_parseable = True
38
+ except Exception:
39
+ args_parseable = False
40
+ if not args or not args_parseable:
41
+ original_args.insert(0, self.default_cmd)
42
+ return super().parse_args(ctx, original_args)
43
+
44
+ def resolve_command(self, ctx, args):
45
+ cmd_name, cmd_obj, args = super(CustomGroup, self).resolve_command(ctx, args)
46
+ passed_cmd = getattr(ctx, "passed_cmd", None)
47
+ if passed_cmd is not None:
48
+ args.insert(0, passed_cmd)
49
+
50
+ return cmd_name, cmd_obj, args
51
+
52
+ def format_commands(self, ctx, formatter):
53
+ formatter = CustomFormatter(self.default_cmd, formatter)
54
+ return super(CustomGroup, self).format_commands(ctx, formatter)
55
+
56
+
57
+ class CustomFormatter:
58
+ def __init__(self, default_cmd, original_formatter) -> None:
59
+ self.default_cmd = default_cmd
60
+ self.formatter = original_formatter
61
+
62
+ def __getattr__(self, name):
63
+ return getattr(self.formatter, name)
64
+
65
+ def write_dl(self, rows):
66
+ def _format(dup):
67
+ cmd, help = dup
68
+ if cmd == self.default_cmd:
69
+ cmd = cmd + " [Default]"
70
+ return (cmd, help)
71
+
72
+ rows = [_format(dup) for dup in rows]
73
+
74
+ return self.formatter.write_dl(rows)
75
+
76
+
77
+ @click.group()
78
+ def cli():
79
+ pass
80
+
81
+
82
+ @cli.group(cls=CustomGroup, help="Commands related to logs", default_cmd="show")
83
+ @click.pass_context
84
+ def logs(ctx):
85
+ # the logger is configured in cli.py
86
+ global echo
87
+ echo = ctx.obj.echo
88
+
89
+
90
+ @logs.command(
91
+ help="Show stdout/stderr produced by a task or all tasks in a step. "
92
+ "The format for input-path is either <run_id>/<step_name> or "
93
+ "<run_id>/<step_name>/<task_id>."
94
+ )
95
+ @click.argument("input-path")
96
+ @click.option(
97
+ "--stdout/--no-stdout",
98
+ default=False,
99
+ show_default=True,
100
+ help="Show stdout of the task.",
101
+ )
102
+ @click.option(
103
+ "--stderr/--no-stderr",
104
+ default=False,
105
+ show_default=True,
106
+ help="Show stderr of the task.",
107
+ )
108
+ @click.option(
109
+ "--both/--no-both",
110
+ default=True,
111
+ show_default=True,
112
+ help="Show both stdout and stderr of the task.",
113
+ )
114
+ @click.option(
115
+ "--timestamps/--no-timestamps",
116
+ default=False,
117
+ show_default=True,
118
+ help="Show timestamps.",
119
+ )
120
+ @click.option(
121
+ "--attempt",
122
+ default=None,
123
+ type=int,
124
+ show_default=False,
125
+ help="Attempt number of a task to show, defaults to the latest attempt.",
126
+ )
127
+ @click.pass_obj
128
+ def show(
129
+ obj, input_path, stdout=None, stderr=None, both=None, timestamps=False, attempt=None
130
+ ):
131
+ types = set()
132
+ if stdout:
133
+ types.add("stdout")
134
+ both = False
135
+ if stderr:
136
+ types.add("stderr")
137
+ both = False
138
+ if both:
139
+ types.update(("stdout", "stderr"))
140
+
141
+ streams = list(sorted(types, reverse=True))
142
+
143
+ # Pathspec can either be run_id/step_name or run_id/step_name/task_id.
144
+ parts = input_path.split("/")
145
+ if len(parts) == 2:
146
+ run_id, step_name = parts
147
+ task_id = None
148
+ elif len(parts) == 3:
149
+ run_id, step_name, task_id = parts
150
+ else:
151
+ raise CommandException(
152
+ "input_path should either be run_id/step_name "
153
+ "or run_id/step_name/task_id"
154
+ )
155
+
156
+ datastore_set = TaskDataStoreSet(
157
+ obj.flow_datastore, run_id, steps=[step_name], allow_not_done=True
158
+ )
159
+ if task_id:
160
+ ds_list = [
161
+ TaskDataStore(
162
+ obj.flow_datastore,
163
+ run_id=run_id,
164
+ step_name=step_name,
165
+ task_id=task_id,
166
+ mode="r",
167
+ allow_not_done=True,
168
+ )
169
+ ]
170
+ else:
171
+ ds_list = list(datastore_set) # get all tasks
172
+
173
+ if ds_list:
174
+
175
+ def echo_unicode(line, **kwargs):
176
+ click.secho(line.decode("UTF-8", errors="replace"), **kwargs)
177
+
178
+ # old style logs are non mflog-style logs
179
+ maybe_old_style = True
180
+ for ds in ds_list:
181
+ echo(
182
+ "Dumping logs of run_id=*{run_id}* "
183
+ "step=*{step}* task_id=*{task_id}*".format(
184
+ run_id=ds.run_id, step=ds.step_name, task_id=ds.task_id
185
+ ),
186
+ fg="magenta",
187
+ )
188
+
189
+ for stream in streams:
190
+ echo(stream, bold=True)
191
+ logs = ds.load_logs(LOG_SOURCES, stream, attempt_override=attempt)
192
+ if any(data for _, data in logs):
193
+ # attempt to read new, mflog-style logs
194
+ for line in mflog.merge_logs([blob for _, blob in logs]):
195
+ if timestamps:
196
+ ts = mflog.utc_to_local(line.utc_tstamp)
197
+ tstamp = ts.strftime("%Y-%m-%d %H:%M:%S.%f")[:-3]
198
+ click.secho(tstamp + " ", fg=LOGGER_TIMESTAMP, nl=False)
199
+ echo_unicode(line.msg)
200
+ maybe_old_style = False
201
+ elif maybe_old_style:
202
+ # if they are not available, we may be looking at
203
+ # a legacy run (unless we have seen new-style data already
204
+ # for another stream). This return an empty string if
205
+ # nothing is found
206
+ log = ds.load_log_legacy(stream, attempt_override=attempt)
207
+ if log and timestamps:
208
+ raise CommandException(
209
+ "We can't show --timestamps for old runs. Sorry!"
210
+ )
211
+ echo_unicode(log, nl=False)
212
+ else:
213
+ raise CommandException(
214
+ "No Tasks found at the given path -- "
215
+ "either none exist or none have started yet"
216
+ )
217
+
218
+
219
+ @logs.command(
220
+ help="Scrub stdout/stderr produced by a task or all tasks in a step. "
221
+ "The format for input-path is either <run_id>/<step_name> or "
222
+ "<run_id>/<step_name>/<task_id>."
223
+ )
224
+ @click.argument("input-path")
225
+ @click.option(
226
+ "--stdout/--no-stdout",
227
+ default=False,
228
+ show_default=True,
229
+ help="Scrub stdout of the step or task.",
230
+ )
231
+ @click.option(
232
+ "--stderr/--no-stderr",
233
+ default=False,
234
+ show_default=True,
235
+ help="Scrub stderr of the step or task.",
236
+ )
237
+ @click.option(
238
+ "--both/--no-both",
239
+ default=True,
240
+ show_default=True,
241
+ help="Scrub both stdout and stderr of the step or task.",
242
+ )
243
+ @click.option(
244
+ "--attempt",
245
+ default=None,
246
+ type=int,
247
+ show_default=False,
248
+ help="Attempt number of a task to scrub, defaults to the latest attempt.",
249
+ )
250
+ @click.option(
251
+ "--latest/--all",
252
+ default=True,
253
+ show_default=False,
254
+ help="Scrub latest/all attempts of a step or task",
255
+ )
256
+ @click.option(
257
+ "--include-not-done",
258
+ default=False,
259
+ show_default=False,
260
+ is_flag=True,
261
+ help="Also scrub steps or tasks that are not done. Use this for tasks that did not finish correctly, and could not otherwise be scrubbed.",
262
+ )
263
+ @click.pass_obj
264
+ def scrub(
265
+ obj,
266
+ input_path,
267
+ stdout=None,
268
+ stderr=None,
269
+ both=None,
270
+ attempt=None,
271
+ latest=None,
272
+ include_not_done=None,
273
+ ):
274
+ types = set()
275
+ if stdout:
276
+ types.add("stdout")
277
+ both = False
278
+ if stderr:
279
+ types.add("stderr")
280
+ both = False
281
+ if both:
282
+ types.update(("stdout", "stderr"))
283
+
284
+ streams = list(sorted(types, reverse=True))
285
+
286
+ # Pathspec can either be run_id/step_name or run_id/step_name/task_id.
287
+ parts = input_path.split("/")
288
+ if len(parts) == 2:
289
+ run_id, step_name = parts
290
+ task_id = None
291
+ elif len(parts) == 3:
292
+ run_id, step_name, task_id = parts
293
+ else:
294
+ raise CommandException(
295
+ "input_path should either be run_id/step_name "
296
+ "or run_id/step_name/task_id"
297
+ )
298
+
299
+ if task_id:
300
+ if latest:
301
+ ds_list = obj.flow_datastore.get_task_datastores(
302
+ pathspecs=[input_path],
303
+ attempt=attempt,
304
+ mode="d",
305
+ allow_not_done=include_not_done,
306
+ )
307
+ else:
308
+ ds_list = obj.flow_datastore.get_task_datastores(
309
+ pathspecs=[input_path],
310
+ attempt=attempt,
311
+ mode="d",
312
+ allow_not_done=include_not_done,
313
+ include_prior=True,
314
+ )
315
+ else:
316
+ if latest:
317
+ ds_list = obj.flow_datastore.get_task_datastores(
318
+ run_id=run_id,
319
+ steps=[step_name],
320
+ attempt=attempt,
321
+ mode="d",
322
+ allow_not_done=include_not_done,
323
+ )
324
+ else:
325
+ ds_list = obj.flow_datastore.get_task_datastores(
326
+ run_id=run_id,
327
+ steps=[step_name],
328
+ attempt=attempt,
329
+ mode="d",
330
+ allow_not_done=include_not_done,
331
+ include_prior=True,
332
+ )
333
+
334
+ if ds_list:
335
+ for ds in ds_list:
336
+ failures = []
337
+ for stream in streams:
338
+ try:
339
+ ds.scrub_logs(LOG_SOURCES, stream)
340
+ except Exception:
341
+ failures.append(stream)
342
+ if failures:
343
+ obj.echo_always(
344
+ "Failed to scrub %s - attempt %s : *%s*"
345
+ % (ds.pathspec, ds.attempt, ",".join(failures))
346
+ )
347
+ else:
348
+ echo(
349
+ "Logs have been scrubbed for %s - attempt %s"
350
+ % (ds.pathspec, ds.attempt)
351
+ )
352
+
353
+ else:
354
+ raise CommandException(
355
+ "No Tasks found at the given path -- "
356
+ "either none exist or they have not finished yet.\n"
357
+ "If you know the task has finished, you can supply --include-not-done to force scrub it."
358
+ )
metaflow/version.py CHANGED
@@ -1 +1 @@
1
- metaflow_version = "2.11.14.1"
1
+ metaflow_version = "2.11.15.2"
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: ob-metaflow
3
- Version: 2.11.14.1
3
+ Version: 2.11.15.2
4
4
  Summary: Metaflow: More Data Science, Less Engineering
5
5
  Author: Netflix, Outerbounds & the Metaflow Community
6
6
  Author-email: help@outerbounds.co
@@ -12,7 +12,7 @@ Requires-Dist: boto3
12
12
  Requires-Dist: pylint
13
13
  Requires-Dist: kubernetes
14
14
  Provides-Extra: stubs
15
- Requires-Dist: ob-metaflow-stubs ==2.11.14.1 ; extra == 'stubs'
15
+ Requires-Dist: ob-metaflow-stubs ==2.11.15.2 ; extra == 'stubs'
16
16
 
17
17
  ![Metaflow_Logo_Horizontal_FullColor_Ribbon_Dark_RGB](https://user-images.githubusercontent.com/763451/89453116-96a57e00-d713-11ea-9fa6-82b29d4d6eff.png)
18
18
 
@@ -1,7 +1,7 @@
1
1
  metaflow/R.py,sha256=bNcXXpGOe5wquzTRGyU0KS9gJMz7HceKjXxammYPUE0,3841
2
2
  metaflow/__init__.py,sha256=3xaDoR_uEUnf4lzQMR6Akj2IctN03nlOvTKBzFHsiAA,6028
3
3
  metaflow/cards.py,sha256=tP1_RrtmqdFh741pqE4t98S7SA0MtGRlGvRICRZF1Mg,426
4
- metaflow/cli.py,sha256=t8BpXpfr1UtjegySI6ydz6w7TaAyYmwZjbk6K2ZB-ns,35313
4
+ metaflow/cli.py,sha256=2igissDxlfN7f3Bfm0urjL7xtUJJE6jasEJir3OJfsw,31392
5
5
  metaflow/cli_args.py,sha256=lcgBGNTvfaiPxiUnejAe60Upt9swG6lRy1_3OqbU6MY,2616
6
6
  metaflow/clone_util.py,sha256=ar4jSZt2aTd4monBpkIQmcLcsOd0relAB42qTUGt2j8,1810
7
7
  metaflow/cmd_with_io.py,sha256=kl53HkAIyv0ecpItv08wZYczv7u3msD1VCcciqigqf0,588
@@ -15,10 +15,10 @@ metaflow/graph.py,sha256=ZPxyG8uwVMk5YYgX4pQEQaPZtZM5Wy-G4NtJK73IEuA,11818
15
15
  metaflow/includefile.py,sha256=yHczcZ_U0SrasxSNhZb3DIBzx8UZnrJCl3FzvpEQLOA,19753
16
16
  metaflow/integrations.py,sha256=LlsaoePRg03DjENnmLxZDYto3NwWc9z_PtU6nJxLldg,1480
17
17
  metaflow/lint.py,sha256=_kYAbAtsP7IG1Rd0FqNbo8I8Zs66_0WXbaZJFARO3dE,10394
18
- metaflow/metaflow_config.py,sha256=GgzzL_CwZdNfdAen3iy_0RmwqNXrl4hxx9t6DBhgRcU,21117
18
+ metaflow/metaflow_config.py,sha256=XEn3iYMP6c7lleymgBjpfCsYLMR9T2nsx-KiHy5iUcI,21477
19
19
  metaflow/metaflow_config_funcs.py,sha256=pCaiQ2ez9wXixJI3ehmf3QiW9lUqFrZnBZx1my_0wIg,4874
20
20
  metaflow/metaflow_current.py,sha256=sCENPBiji3LcPbwgOG0ukGd_yEc5tST8EowES8DzRtA,7430
21
- metaflow/metaflow_environment.py,sha256=XiMmBZiq3_dwaw0Oi3B8588BahYxzgfqWGMePPZqUUc,7359
21
+ metaflow/metaflow_environment.py,sha256=RHxCI-EcTR0yh352oa13itZeDun45OVF-lOAFxB9uXo,7382
22
22
  metaflow/metaflow_profile.py,sha256=jKPEW-hmAQO-htSxb9hXaeloLacAh41A35rMZH6G8pA,418
23
23
  metaflow/metaflow_version.py,sha256=mPQ6g_3XjNdi0NrxDzwlW8ZH0nMyYpwqmJ04P7TIdP0,4774
24
24
  metaflow/monitor.py,sha256=T0NMaBPvXynlJAO_avKtk8OIIRMyEuMAyF8bIp79aZU,5323
@@ -34,7 +34,7 @@ metaflow/task.py,sha256=ecGaULbK8kXPnyWzH1u6wtGclm0qeJm7K95amEL17sQ,25863
34
34
  metaflow/unbounded_foreach.py,sha256=p184WMbrMJ3xKYHwewj27ZhRUsSj_kw1jlye5gA9xJk,387
35
35
  metaflow/util.py,sha256=RrjsvADLKxSqjL76CxKh_J4OJl840B9Ak3V-vXleGas,13429
36
36
  metaflow/vendor.py,sha256=LZgXrh7ZSDmD32D1T5jj3OKKpXIqqxKzdMAOc5V0SD4,5162
37
- metaflow/version.py,sha256=_gxslYktBiLJ68luN6fxpsnBrMI0qbV6Xyp9YznIpgU,31
37
+ metaflow/version.py,sha256=LJTdJuz7bbVrmaIHT0OQUmxzGPb3ovAAoEl5E7LsPTQ,31
38
38
  metaflow/_vendor/__init__.py,sha256=y_CiwUD3l4eAKvTVDZeqgVujMy31cAM1qjAB-HfI-9s,353
39
39
  metaflow/_vendor/click/__init__.py,sha256=FkyGDQ-cbiQxP_lxgUspyFYS48f2S_pTcfKPz-d_RMo,2463
40
40
  metaflow/_vendor/click/_bashcomplete.py,sha256=9J98IHQYmCAr2Jup6TDshUr5FJEen-AoQCZR0K5nKxQ,12309
@@ -98,12 +98,12 @@ metaflow/cmd/develop/stub_generator.py,sha256=fmiWmr4tXBBvIZdWVEhKvZWtG4vjyIsfre
98
98
  metaflow/cmd/develop/stubs.py,sha256=hhf1giRNNlFGB5zSZdNA8tNvnJcmotXSiNN06N3_WyA,11742
99
99
  metaflow/datastore/__init__.py,sha256=VxP6ddJt3rwiCkpiSfAhyVkUCOe1pgZZsytVEJzFmSQ,155
100
100
  metaflow/datastore/content_addressed_store.py,sha256=dCVFAr4PltlmXNVVYt7UaBGJWe6fWuicCgb68XHqLrA,7643
101
- metaflow/datastore/datastore_set.py,sha256=sjwcxO6ZJgZ7ief_L-TTpjoRJYIlJnGGG_WSaDRij28,2368
101
+ metaflow/datastore/datastore_set.py,sha256=R5pwnxg1DD8kBY9vElvd2eMknrvwTyiSwvQs67_z9bc,2361
102
102
  metaflow/datastore/datastore_storage.py,sha256=7V43QuiWDQ_Q4oHw9y7Z7X9lYj3GI-LV1-xB3d2Tt5k,9038
103
103
  metaflow/datastore/exceptions.py,sha256=r7Ab5FvHIzyFh6kwiptA1lO5nLqWg0xRBoeYGefvapA,373
104
- metaflow/datastore/flow_datastore.py,sha256=ZG1fguspPxgFwKKJj4dHLgZl3ZdfZJDb19EWP2VlRJQ,9183
104
+ metaflow/datastore/flow_datastore.py,sha256=kbJcOLYnvPHgJfZ_WWkD9LJSX1PHI1K6f9oVUu08A9U,10235
105
105
  metaflow/datastore/inputs.py,sha256=i43dXr2xvgtsgKMO9allgCR18bk80GeayeQFyUTH36w,449
106
- metaflow/datastore/task_datastore.py,sha256=aqu68A2ezS2RschhxdX4O82ofML3J6UPwYWbgo-01G8,34312
106
+ metaflow/datastore/task_datastore.py,sha256=RWO-2p_vyJfTV9JtW2dIdt7IW7n_OD8ff43YJwetiW0,36169
107
107
  metaflow/extension_support/__init__.py,sha256=GK3P6YbIN4S7r3rbofzh4xaIJ6wsmDsE7iEMIlbXgMM,49334
108
108
  metaflow/extension_support/_empty_file.py,sha256=HENjnM4uAfeNygxMB_feCCWORFoSat9n_QwzSx2oXPw,109
109
109
  metaflow/extension_support/cmd.py,sha256=hk8iBUUINqvKCDxInKgWpum8ThiRZtHSJP7qBASHzl8,5711
@@ -118,12 +118,13 @@ metaflow/mflog/mflog.py,sha256=VebXxqitOtNAs7VJixnNfziO_i_urG7bsJ5JiB5IXgY,4370
118
118
  metaflow/mflog/save_logs.py,sha256=ZBAF4BMukw4FMAC7odpr9OI2BC_2petPtDX0ca6srC4,2352
119
119
  metaflow/mflog/save_logs_periodically.py,sha256=2Uvk9hi-zlCqXxOQoXmmjH1SCugfw6eG6w70WgfI-ho,1256
120
120
  metaflow/mflog/tee.py,sha256=wTER15qeHuiRpCkOqo-bd-r3Gj-EVlf3IvWRCA4beW4,887
121
- metaflow/plugins/__init__.py,sha256=CGzgHIEK8wp1609kiAOs7unl_r9oZR5bSUP0w3AwyY0,6728
121
+ metaflow/plugins/__init__.py,sha256=A_qoolQowlEQFajlBqfYBRF9qP-hzy33749KZ4R2fgc,6881
122
122
  metaflow/plugins/catch_decorator.py,sha256=UOM2taN_OL2RPpuJhwEOA9ZALm0-hHD0XS2Hn2GUev0,4061
123
123
  metaflow/plugins/debug_logger.py,sha256=mcF5HYzJ0NQmqCMjyVUk3iAP-heroHRIiVWQC6Ha2-I,879
124
124
  metaflow/plugins/debug_monitor.py,sha256=Md5X_sDOSssN9pt2D8YcaIjTK5JaQD55UAYTcF6xYF0,1099
125
125
  metaflow/plugins/environment_decorator.py,sha256=6m9j2B77d-Ja_l_9CTJ__0O6aB2a8Qt_lAZu6UjAcUA,587
126
126
  metaflow/plugins/events_decorator.py,sha256=c2GcH6Mspbey3wBkjM5lqxaNByFOzYDQdllLpXzRNv8,18283
127
+ metaflow/plugins/logs_cli.py,sha256=7-LxLRy8X5s4KcoMSc183cbNIJnIcedf6oWPe13-hQs,11409
127
128
  metaflow/plugins/package_cli.py,sha256=-J6D4cupHfWSZ4GEFo2yy9Je9oL3owRWm5pEJwaiqd4,1649
128
129
  metaflow/plugins/parallel_decorator.py,sha256=yrwfqdTmvEyR3YdvJiPChVOK5vjYTibC09kDs7t6_kg,4444
129
130
  metaflow/plugins/project_decorator.py,sha256=eJOe0Ea7CbUCReEhR_XQvRkhV6jyRqDxM72oZI7EMCk,5336
@@ -134,7 +135,7 @@ metaflow/plugins/tag_cli.py,sha256=O_ZI4ILwGX3xKrLewUUF-zdJjCDi3JmsTb4ow87_RuY,1
134
135
  metaflow/plugins/test_unbounded_foreach_decorator.py,sha256=cB_2OWb38eYfmbVck72ZwU0qgzi6hqJXZAxglpHU_qg,5216
135
136
  metaflow/plugins/timeout_decorator.py,sha256=GGlsnmT1F-5FDaN19pDWKhmcHaN1hstgtZRBipPPu3c,3595
136
137
  metaflow/plugins/airflow/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
137
- metaflow/plugins/airflow/airflow.py,sha256=W4qmLY1clGuSlEf9muGM0qgk8ZbAr7P3squa9_m7ni0,32007
138
+ metaflow/plugins/airflow/airflow.py,sha256=U7jg1fSpGcuXhBCZ9xNh492h0MhBx3EV1WpWQh6RIaI,32147
138
139
  metaflow/plugins/airflow/airflow_cli.py,sha256=fUi6IsRMi6mvL6Twrszk7rZq7_4PmdYr9evJnBpXXPc,14440
139
140
  metaflow/plugins/airflow/airflow_decorator.py,sha256=H9-QnRP4x8tSomLmmpGeuVUI48-CxHR7tlvn_ceX1Zs,1772
140
141
  metaflow/plugins/airflow/airflow_utils.py,sha256=qd6lV2X4VpCO2sLsRc35JMOU4DVz_tQacrM_wWNkQug,28865
@@ -149,7 +150,7 @@ metaflow/plugins/airflow/sensors/s3_sensor.py,sha256=JUKoGNoTCtrO9MNEneEC7ldRNwg
149
150
  metaflow/plugins/argo/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
150
151
  metaflow/plugins/argo/argo_client.py,sha256=MKKhMCbWOPzf6z5zQQiyDRHHkAXcO7ipboDZDqAAvOk,15849
151
152
  metaflow/plugins/argo/argo_events.py,sha256=_C1KWztVqgi3zuH57pInaE9OzABc2NnncC-zdwOMZ-w,5909
152
- metaflow/plugins/argo/argo_workflows.py,sha256=vr0DUFLKUmVXS2IZRThNgqsfRyN2HQMPaPlvj9Ps3Hs,129679
153
+ metaflow/plugins/argo/argo_workflows.py,sha256=niBAizKFjDA-PTQh-PqhSHV4KX68uL83RF2kZN-QtT4,129783
153
154
  metaflow/plugins/argo/argo_workflows_cli.py,sha256=sZTpgfmc50eT3e0qIxpVqUgWhTcYlO1HM4gU6Oaya8g,33259
154
155
  metaflow/plugins/argo/argo_workflows_decorator.py,sha256=K5t4uIk2IXPdK7v7DEjj3buSB8ikLjLycKjbZUYeiaw,6781
155
156
  metaflow/plugins/argo/generate_input_paths.py,sha256=loYsI6RFX9LlFsHb7Fe-mzlTTtRdySoOu7sYDy-uXK0,881
@@ -158,7 +159,7 @@ metaflow/plugins/aws/aws_client.py,sha256=mO8UD6pxFaOnxDb3hTP3HB7Gqb_ZxoR-76LT68
158
159
  metaflow/plugins/aws/aws_utils.py,sha256=BELrOYMURrTHksQQiNgwppP8E8oUp-9xjEKNP-NO67g,7699
159
160
  metaflow/plugins/aws/batch/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
160
161
  metaflow/plugins/aws/batch/batch.py,sha256=e9ssahWM18GnipPK2sqYB-ztx9w7Eoo7YtWyEtufYxs,17787
161
- metaflow/plugins/aws/batch/batch_cli.py,sha256=8j5s9RMZu0aJW76GY2lQkJT5tVDzamg9G_iu1AUpW8o,11632
162
+ metaflow/plugins/aws/batch/batch_cli.py,sha256=6PTbyajRgdy0XmjyJLBTdKdiOB84dcovQQ8sFXlJqko,11749
162
163
  metaflow/plugins/aws/batch/batch_client.py,sha256=s9ZHhxQPPoBQijLUgn6_16QOaD4-22U_44uJbp-yLkI,28565
163
164
  metaflow/plugins/aws/batch/batch_decorator.py,sha256=KUOBrJH1Rl04toRMsbh7D_ThntRZqa-_B9JUQbu8ORk,17319
164
165
  metaflow/plugins/aws/secrets_manager/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
@@ -175,7 +176,8 @@ metaflow/plugins/aws/step_functions/step_functions_client.py,sha256=DKpNwAIWElvW
175
176
  metaflow/plugins/aws/step_functions/step_functions_decorator.py,sha256=9hw_MX36RyFp6IowuAYaJzJg9UC5KCe1FNt1PcG7_J0,3791
176
177
  metaflow/plugins/azure/__init__.py,sha256=GuuhTVC-zSdyAf79a1wiERMq0Zts7fwVT7t9fAf234A,100
177
178
  metaflow/plugins/azure/azure_credential.py,sha256=JmdGEbVzgxy8ucqnQDdTTI_atyMX9WSZUw3qYOo7RhE,2174
178
- metaflow/plugins/azure/azure_exceptions.py,sha256=uvxE3E3nsbQq1dxCx1Yl9O54frbnMS5Elk8Z4qQ2Oh4,404
179
+ metaflow/plugins/azure/azure_exceptions.py,sha256=NnbwpUC23bc61HZjJmeXztY0tBNn_Y_VpIpDDuYWIZ0,433
180
+ metaflow/plugins/azure/azure_secret_manager_secrets_provider.py,sha256=O1osBCO47GPM95HzL4gSt_sjRdQyqX-oLxY8xWaPTDI,11022
179
181
  metaflow/plugins/azure/azure_tail.py,sha256=JAqV4mC42bMpR0O7m6X4cpFuh0peV1ufs_jJXrmicTc,3362
180
182
  metaflow/plugins/azure/azure_utils.py,sha256=j3kAxi2oC-fMpw8YegJvqsAwxi_m7jGPxCaeVwoBZJg,7100
181
183
  metaflow/plugins/azure/blob_service_client_factory.py,sha256=MtyPftBxrXdXMxwhKgLepG6mtlb_2BhJLG_fvbO6D14,6527
@@ -214,7 +216,7 @@ metaflow/plugins/datastores/s3_storage.py,sha256=CZdNqaKtxDXQbEg2YHyphph3hWcLIE5
214
216
  metaflow/plugins/datatools/__init__.py,sha256=ge4L16OBQLy2J_MMvoHg3lMfdm-MluQgRWoyZ5GCRnk,1267
215
217
  metaflow/plugins/datatools/local.py,sha256=67hx3O_vInERlL0aJV0Sd-jUTd_2DOw4sJ4-IyEKNKM,4213
216
218
  metaflow/plugins/datatools/s3/__init__.py,sha256=14tr9fPjN3ULW5IOfKHeG7Uhjmgm7LMtQHfz1SFv-h8,248
217
- metaflow/plugins/datatools/s3/s3.py,sha256=7jxoDSn4hOM4xoqdwk-xcIfBkAt3RfM-ABXA5ZW0q10,66101
219
+ metaflow/plugins/datatools/s3/s3.py,sha256=CQynofOk0l_sJMakTDhb8IlVKqlL-Ko1fmY5mKJTbes,66113
218
220
  metaflow/plugins/datatools/s3/s3op.py,sha256=ZQFSxlaQUt-Ko_kIXMbHOKJc8q4FPXogS3xI6xsDR7Y,43390
219
221
  metaflow/plugins/datatools/s3/s3tail.py,sha256=boQjQGQMI-bvTqcMP2y7uSlSYLcvWOy7J3ZUaF78NAA,2597
220
222
  metaflow/plugins/datatools/s3/s3util.py,sha256=FgRgaVmEq7-i2dV7q8XK5w5PfFt-xJjZa8WrK8IJfdI,3769
@@ -248,7 +250,7 @@ metaflow/plugins/gcp/gs_tail.py,sha256=Jl_wvnzU7dub07A-DOAuP5FeccNIrPM-CeL1xKFs1
248
250
  metaflow/plugins/gcp/gs_utils.py,sha256=ZmIGFse1qYyvAVrwga23PQUzF6dXEDLLsZ2F-YRmvow,2030
249
251
  metaflow/plugins/gcp/includefile_support.py,sha256=vIDeR-MiJuUh-2S2pV7Z7FBkhIWwtHXaRrj76MWGRiY,3869
250
252
  metaflow/plugins/kubernetes/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
251
- metaflow/plugins/kubernetes/kubernetes.py,sha256=5wM8_gRuyyyIv9mbX2lCePsLHoiDgJnbzYwnJy_NEoE,19233
253
+ metaflow/plugins/kubernetes/kubernetes.py,sha256=KIRt-brcAMe7SyhecgooTxOurxF9vTL0Db37LC0reMA,19384
252
254
  metaflow/plugins/kubernetes/kubernetes_cli.py,sha256=wSByGQEaoQo1aV9kJoKYmbVVeFQVsMw9RfG4Bw2sMm8,10274
253
255
  metaflow/plugins/kubernetes/kubernetes_client.py,sha256=irATJpAob4jINkJw0zT_Xoa6JHRtYxx2IOeimlbzvPo,2373
254
256
  metaflow/plugins/kubernetes/kubernetes_decorator.py,sha256=315v32txNvgMQC8QHl1fwf9tslsESlZ8M5KY-qhjgjg,25984
@@ -300,9 +302,9 @@ metaflow/tutorials/07-worldview/README.md,sha256=5vQTrFqulJ7rWN6r20dhot9lI2sVj9W
300
302
  metaflow/tutorials/07-worldview/worldview.ipynb,sha256=ztPZPI9BXxvW1QdS2Tfe7LBuVzvFvv0AToDnsDJhLdE,2237
301
303
  metaflow/tutorials/08-autopilot/README.md,sha256=GnePFp_q76jPs991lMUqfIIh5zSorIeWznyiUxzeUVE,1039
302
304
  metaflow/tutorials/08-autopilot/autopilot.ipynb,sha256=DQoJlILV7Mq9vfPBGW-QV_kNhWPjS5n6SJLqePjFYLY,3191
303
- ob_metaflow-2.11.14.1.dist-info/LICENSE,sha256=nl_Lt5v9VvJ-5lWJDT4ddKAG-VZ-2IaLmbzpgYDz2hU,11343
304
- ob_metaflow-2.11.14.1.dist-info/METADATA,sha256=opXWlALXWsRYlogJ6rG9a-wlWGg2FZYQKJI0fAP1osc,5148
305
- ob_metaflow-2.11.14.1.dist-info/WHEEL,sha256=DZajD4pwLWue70CAfc7YaxT1wLUciNBvN_TTcvXpltE,110
306
- ob_metaflow-2.11.14.1.dist-info/entry_points.txt,sha256=IKwTN1T3I5eJL3uo_vnkyxVffcgnRdFbKwlghZfn27k,57
307
- ob_metaflow-2.11.14.1.dist-info/top_level.txt,sha256=v1pDHoWaSaKeuc5fKTRSfsXCKSdW1zvNVmvA-i0if3o,9
308
- ob_metaflow-2.11.14.1.dist-info/RECORD,,
305
+ ob_metaflow-2.11.15.2.dist-info/LICENSE,sha256=nl_Lt5v9VvJ-5lWJDT4ddKAG-VZ-2IaLmbzpgYDz2hU,11343
306
+ ob_metaflow-2.11.15.2.dist-info/METADATA,sha256=3vdCmlKaPQ9V9EfET5djgsF_JxUOPdIQLV1qmgbIy_Y,5148
307
+ ob_metaflow-2.11.15.2.dist-info/WHEEL,sha256=DZajD4pwLWue70CAfc7YaxT1wLUciNBvN_TTcvXpltE,110
308
+ ob_metaflow-2.11.15.2.dist-info/entry_points.txt,sha256=IKwTN1T3I5eJL3uo_vnkyxVffcgnRdFbKwlghZfn27k,57
309
+ ob_metaflow-2.11.15.2.dist-info/top_level.txt,sha256=v1pDHoWaSaKeuc5fKTRSfsXCKSdW1zvNVmvA-i0if3o,9
310
+ ob_metaflow-2.11.15.2.dist-info/RECORD,,