ob-metaflow 2.12.22.1__py2.py3-none-any.whl → 2.12.24.1__py2.py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of ob-metaflow might be problematic. Click here for more details.
- metaflow/cli.py +24 -19
- metaflow/client/core.py +1 -1
- metaflow/cmd/develop/stub_generator.py +17 -0
- metaflow/cmd/develop/stubs.py +3 -3
- metaflow/metaflow_version.py +8 -5
- metaflow/plugins/argo/argo_client.py +2 -0
- metaflow/plugins/argo/argo_workflows.py +90 -44
- metaflow/plugins/argo/argo_workflows_cli.py +1 -1
- metaflow/plugins/kubernetes/kubernetes.py +17 -1
- metaflow/plugins/kubernetes/kubernetes_client.py +4 -1
- metaflow/plugins/kubernetes/kubernetes_decorator.py +5 -1
- metaflow/plugins/kubernetes/kubernetes_job.py +22 -0
- metaflow/plugins/kubernetes/kubernetes_jobsets.py +7 -1
- metaflow/plugins/pypi/bootstrap.py +1 -1
- metaflow/plugins/pypi/micromamba.py +26 -0
- metaflow/runner/deployer.py +4 -49
- metaflow/runner/metaflow_runner.py +22 -25
- metaflow/runner/subprocess_manager.py +33 -17
- metaflow/runner/utils.py +53 -1
- metaflow/task.py +7 -0
- metaflow/version.py +1 -1
- {ob_metaflow-2.12.22.1.dist-info → ob_metaflow-2.12.24.1.dist-info}/METADATA +2 -2
- {ob_metaflow-2.12.22.1.dist-info → ob_metaflow-2.12.24.1.dist-info}/RECORD +27 -27
- {ob_metaflow-2.12.22.1.dist-info → ob_metaflow-2.12.24.1.dist-info}/LICENSE +0 -0
- {ob_metaflow-2.12.22.1.dist-info → ob_metaflow-2.12.24.1.dist-info}/WHEEL +0 -0
- {ob_metaflow-2.12.22.1.dist-info → ob_metaflow-2.12.24.1.dist-info}/entry_points.txt +0 -0
- {ob_metaflow-2.12.22.1.dist-info → ob_metaflow-2.12.24.1.dist-info}/top_level.txt +0 -0
metaflow/cli.py
CHANGED
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import inspect
|
|
2
|
-
import
|
|
2
|
+
import json
|
|
3
3
|
import sys
|
|
4
4
|
import traceback
|
|
5
5
|
from datetime import datetime
|
|
@@ -7,6 +7,7 @@ from functools import wraps
|
|
|
7
7
|
|
|
8
8
|
import metaflow.tracing as tracing
|
|
9
9
|
from metaflow._vendor import click
|
|
10
|
+
from metaflow.client.core import get_metadata
|
|
10
11
|
|
|
11
12
|
from . import decorators, lint, metaflow_version, namespace, parameters, plugins
|
|
12
13
|
from .cli_args import cli_args
|
|
@@ -698,15 +699,17 @@ def resume(
|
|
|
698
699
|
runtime.print_workflow_info()
|
|
699
700
|
|
|
700
701
|
runtime.persist_constants()
|
|
701
|
-
|
|
702
|
-
|
|
703
|
-
"
|
|
704
|
-
|
|
705
|
-
|
|
706
|
-
|
|
707
|
-
|
|
708
|
-
|
|
709
|
-
|
|
702
|
+
|
|
703
|
+
if runner_attribute_file:
|
|
704
|
+
with open(runner_attribute_file, "w") as f:
|
|
705
|
+
json.dump(
|
|
706
|
+
{
|
|
707
|
+
"run_id": runtime.run_id,
|
|
708
|
+
"flow_name": obj.flow.name,
|
|
709
|
+
"metadata": get_metadata(),
|
|
710
|
+
},
|
|
711
|
+
f,
|
|
712
|
+
)
|
|
710
713
|
|
|
711
714
|
# We may skip clone-only resume if this is not a resume leader,
|
|
712
715
|
# and clone is already complete.
|
|
@@ -774,15 +777,17 @@ def run(
|
|
|
774
777
|
obj.flow._set_constants(obj.graph, kwargs)
|
|
775
778
|
runtime.print_workflow_info()
|
|
776
779
|
runtime.persist_constants()
|
|
777
|
-
|
|
778
|
-
|
|
779
|
-
"
|
|
780
|
-
|
|
781
|
-
|
|
782
|
-
|
|
783
|
-
|
|
784
|
-
|
|
785
|
-
|
|
780
|
+
|
|
781
|
+
if runner_attribute_file:
|
|
782
|
+
with open(runner_attribute_file, "w") as f:
|
|
783
|
+
json.dump(
|
|
784
|
+
{
|
|
785
|
+
"run_id": runtime.run_id,
|
|
786
|
+
"flow_name": obj.flow.name,
|
|
787
|
+
"metadata": get_metadata(),
|
|
788
|
+
},
|
|
789
|
+
f,
|
|
790
|
+
)
|
|
786
791
|
runtime.execute()
|
|
787
792
|
|
|
788
793
|
|
metaflow/client/core.py
CHANGED
|
@@ -115,6 +115,11 @@ class StubGenerator:
|
|
|
115
115
|
:type members_from_other_modules: List[str]
|
|
116
116
|
"""
|
|
117
117
|
|
|
118
|
+
# Let metaflow know we are in stubgen mode. This is sometimes useful to skip
|
|
119
|
+
# some processing like loading libraries, etc. It is used in Metaflow extensions
|
|
120
|
+
# so do not remove even if you do not see a use for it directly in the code.
|
|
121
|
+
os.environ["METAFLOW_STUBGEN"] = "1"
|
|
122
|
+
|
|
118
123
|
self._write_generated_for = include_generated_for
|
|
119
124
|
self._pending_modules = ["metaflow"] # type: List[str]
|
|
120
125
|
self._pending_modules.extend(get_aliased_modules())
|
|
@@ -398,6 +403,18 @@ class StubGenerator:
|
|
|
398
403
|
name_with_module = self._get_element_name_with_module(clazz.__class__)
|
|
399
404
|
buff.write("metaclass=" + name_with_module + "):\n")
|
|
400
405
|
|
|
406
|
+
# Add class docstring
|
|
407
|
+
if clazz.__doc__:
|
|
408
|
+
buff.write('%s"""\n' % TAB)
|
|
409
|
+
my_doc = cast(str, deindent_docstring(clazz.__doc__))
|
|
410
|
+
init_blank = True
|
|
411
|
+
for line in my_doc.split("\n"):
|
|
412
|
+
if init_blank and len(line.strip()) == 0:
|
|
413
|
+
continue
|
|
414
|
+
init_blank = False
|
|
415
|
+
buff.write("%s%s\n" % (TAB, line.rstrip()))
|
|
416
|
+
buff.write('%s"""\n' % TAB)
|
|
417
|
+
|
|
401
418
|
# For NamedTuple, we have __annotations__ but no __init__. In that case,
|
|
402
419
|
# we are going to "create" a __init__ function with the annotations
|
|
403
420
|
# to show what the class takes.
|
metaflow/cmd/develop/stubs.py
CHANGED
|
@@ -170,7 +170,7 @@ def install(ctx: Any, force: bool):
|
|
|
170
170
|
"Metaflow stubs are already installed and valid -- use --force to reinstall"
|
|
171
171
|
)
|
|
172
172
|
return
|
|
173
|
-
mf_version, _ = get_mf_version()
|
|
173
|
+
mf_version, _ = get_mf_version(True)
|
|
174
174
|
with tempfile.TemporaryDirectory() as tmp_dir:
|
|
175
175
|
with open(os.path.join(tmp_dir, "setup.py"), "w") as f:
|
|
176
176
|
f.write(
|
|
@@ -261,10 +261,10 @@ def split_version(vers: str) -> Tuple[str, Optional[str]]:
|
|
|
261
261
|
return vers_split[0], vers_split[1]
|
|
262
262
|
|
|
263
263
|
|
|
264
|
-
def get_mf_version() -> Tuple[str, Optional[str]]:
|
|
264
|
+
def get_mf_version(public: bool = False) -> Tuple[str, Optional[str]]:
|
|
265
265
|
from metaflow.metaflow_version import get_version
|
|
266
266
|
|
|
267
|
-
return split_version(get_version())
|
|
267
|
+
return split_version(get_version(public))
|
|
268
268
|
|
|
269
269
|
|
|
270
270
|
def get_stubs_version(stubs_root_path: Optional[str]) -> Tuple[str, Optional[str]]:
|
metaflow/metaflow_version.py
CHANGED
|
@@ -195,11 +195,14 @@ def get_version(public=False):
|
|
|
195
195
|
if ext_version is None:
|
|
196
196
|
ext_version = getattr(extension_module, "__version__", "<unk>")
|
|
197
197
|
# Update the package information about reported version for the extension
|
|
198
|
-
|
|
199
|
-
|
|
200
|
-
|
|
201
|
-
|
|
202
|
-
|
|
198
|
+
# (only for the full info which is called at least once -- if we update more
|
|
199
|
+
# it will error out since we can only update_package_info once)
|
|
200
|
+
if not public:
|
|
201
|
+
update_package_info(
|
|
202
|
+
package_name=pkg_name,
|
|
203
|
+
extension_name=ext_name,
|
|
204
|
+
package_version=ext_version,
|
|
205
|
+
)
|
|
203
206
|
ext_versions.append("%s(%s)" % (ext_name, ext_version))
|
|
204
207
|
|
|
205
208
|
# We now have all the information about extensions so we can form the final string
|
|
@@ -295,6 +295,8 @@ class ArgoClient(object):
|
|
|
295
295
|
"suspend": schedule is None,
|
|
296
296
|
"schedule": schedule,
|
|
297
297
|
"timezone": timezone,
|
|
298
|
+
"failedJobsHistoryLimit": 10000, # default is unfortunately 1
|
|
299
|
+
"successfulJobsHistoryLimit": 10000, # default is unfortunately 3
|
|
298
300
|
"workflowSpec": {"workflowTemplateRef": {"name": name}},
|
|
299
301
|
},
|
|
300
302
|
}
|
|
@@ -51,7 +51,7 @@ from metaflow.metaflow_config import (
|
|
|
51
51
|
UI_URL,
|
|
52
52
|
PAGERDUTY_TEMPLATE_URL,
|
|
53
53
|
)
|
|
54
|
-
from metaflow.metaflow_config_funcs import config_values
|
|
54
|
+
from metaflow.metaflow_config_funcs import config_values, init_config
|
|
55
55
|
from metaflow.mflog import BASH_SAVE_LOGS, bash_capture_logs, export_mflog_env_vars
|
|
56
56
|
from metaflow.parameters import deploy_time_eval
|
|
57
57
|
from metaflow.plugins.kubernetes.kubernetes import (
|
|
@@ -1491,7 +1491,11 @@ class ArgoWorkflows(object):
|
|
|
1491
1491
|
# {{foo.bar['param_name']}}.
|
|
1492
1492
|
# https://argoproj.github.io/argo-events/tutorials/02-parameterization/
|
|
1493
1493
|
# http://masterminds.github.io/sprig/strings.html
|
|
1494
|
-
|
|
1494
|
+
(
|
|
1495
|
+
"--%s='{{workflow.parameters.%s}}'"
|
|
1496
|
+
if parameter["type"] == "JSON"
|
|
1497
|
+
else "--%s={{workflow.parameters.%s}}"
|
|
1498
|
+
)
|
|
1495
1499
|
% (parameter["name"], parameter["name"])
|
|
1496
1500
|
for parameter in self.parameters.values()
|
|
1497
1501
|
]
|
|
@@ -1818,6 +1822,21 @@ class ArgoWorkflows(object):
|
|
|
1818
1822
|
and k not in set(ARGO_WORKFLOWS_ENV_VARS_TO_SKIP.split(","))
|
|
1819
1823
|
}
|
|
1820
1824
|
|
|
1825
|
+
# get initial configs
|
|
1826
|
+
initial_configs = init_config()
|
|
1827
|
+
for entry in ["OBP_PERIMETER", "OBP_INTEGRATIONS_SECRETS_METADATA_URL"]:
|
|
1828
|
+
if entry not in initial_configs:
|
|
1829
|
+
raise ArgoWorkflowsException(
|
|
1830
|
+
f"{entry} was not found in metaflow config. Please make sure to run `outerbounds configure <...>` command which can be found on the Ourebounds UI or reach out to your Outerbounds support team."
|
|
1831
|
+
)
|
|
1832
|
+
|
|
1833
|
+
additional_obp_configs = {
|
|
1834
|
+
"OBP_PERIMETER": initial_configs["OBP_PERIMETER"],
|
|
1835
|
+
"OBP_INTEGRATIONS_SECRETS_METADATA_URL": initial_configs[
|
|
1836
|
+
"OBP_INTEGRATIONS_SECRETS_METADATA_URL"
|
|
1837
|
+
],
|
|
1838
|
+
}
|
|
1839
|
+
|
|
1821
1840
|
# Tmpfs variables
|
|
1822
1841
|
use_tmpfs = resources["use_tmpfs"]
|
|
1823
1842
|
tmpfs_size = resources["tmpfs_size"]
|
|
@@ -2079,6 +2098,13 @@ class ArgoWorkflows(object):
|
|
|
2079
2098
|
"METAFLOW_KUBERNETES_SERVICE_ACCOUNT_NAME": "spec.serviceAccountName",
|
|
2080
2099
|
"METAFLOW_KUBERNETES_NODE_IP": "status.hostIP",
|
|
2081
2100
|
}.items()
|
|
2101
|
+
]
|
|
2102
|
+
+ [
|
|
2103
|
+
kubernetes_sdk.V1EnvVar(
|
|
2104
|
+
name=k,
|
|
2105
|
+
value=v,
|
|
2106
|
+
)
|
|
2107
|
+
for k, v in additional_obp_configs.items()
|
|
2082
2108
|
],
|
|
2083
2109
|
image=resources["image"],
|
|
2084
2110
|
image_pull_policy=resources["image_pull_policy"],
|
|
@@ -2636,50 +2662,57 @@ class ArgoWorkflows(object):
|
|
|
2636
2662
|
)
|
|
2637
2663
|
from kubernetes import client as kubernetes_sdk
|
|
2638
2664
|
|
|
2639
|
-
return
|
|
2640
|
-
|
|
2641
|
-
|
|
2642
|
-
|
|
2643
|
-
|
|
2644
|
-
|
|
2645
|
-
|
|
2646
|
-
|
|
2647
|
-
|
|
2648
|
-
|
|
2649
|
-
|
|
2650
|
-
|
|
2651
|
-
|
|
2652
|
-
|
|
2653
|
-
|
|
2654
|
-
|
|
2665
|
+
return (
|
|
2666
|
+
DaemonTemplate("heartbeat-daemon")
|
|
2667
|
+
# NOTE: Even though a retry strategy does not work for Argo daemon containers,
|
|
2668
|
+
# this has the side-effect of protecting the exit hooks of the workflow from failing in case the daemon container errors out.
|
|
2669
|
+
.retry_strategy(10, 1)
|
|
2670
|
+
.service_account_name(resources["service_account"])
|
|
2671
|
+
.container(
|
|
2672
|
+
to_camelcase(
|
|
2673
|
+
kubernetes_sdk.V1Container(
|
|
2674
|
+
name="main",
|
|
2675
|
+
# TODO: Make the image configurable
|
|
2676
|
+
image=resources["image"],
|
|
2677
|
+
command=cmds,
|
|
2678
|
+
env=[
|
|
2679
|
+
kubernetes_sdk.V1EnvVar(name=k, value=str(v))
|
|
2680
|
+
for k, v in env.items()
|
|
2681
|
+
],
|
|
2682
|
+
env_from=[
|
|
2683
|
+
kubernetes_sdk.V1EnvFromSource(
|
|
2684
|
+
secret_ref=kubernetes_sdk.V1SecretEnvSource(
|
|
2685
|
+
name=str(k),
|
|
2686
|
+
# optional=True
|
|
2687
|
+
)
|
|
2655
2688
|
)
|
|
2656
|
-
|
|
2657
|
-
|
|
2658
|
-
|
|
2659
|
-
|
|
2660
|
-
|
|
2661
|
-
|
|
2662
|
-
|
|
2663
|
-
|
|
2689
|
+
for k in list(
|
|
2690
|
+
[]
|
|
2691
|
+
if not resources.get("secrets")
|
|
2692
|
+
else (
|
|
2693
|
+
[resources.get("secrets")]
|
|
2694
|
+
if isinstance(resources.get("secrets"), str)
|
|
2695
|
+
else resources.get("secrets")
|
|
2696
|
+
)
|
|
2664
2697
|
)
|
|
2665
|
-
|
|
2666
|
-
|
|
2667
|
-
|
|
2668
|
-
|
|
2669
|
-
|
|
2670
|
-
|
|
2671
|
-
|
|
2672
|
-
|
|
2673
|
-
|
|
2674
|
-
|
|
2675
|
-
|
|
2676
|
-
|
|
2677
|
-
|
|
2678
|
-
|
|
2679
|
-
|
|
2680
|
-
|
|
2681
|
-
|
|
2682
|
-
)
|
|
2698
|
+
+ KUBERNETES_SECRETS.split(",")
|
|
2699
|
+
+ ARGO_WORKFLOWS_KUBERNETES_SECRETS.split(",")
|
|
2700
|
+
if k
|
|
2701
|
+
],
|
|
2702
|
+
resources=kubernetes_sdk.V1ResourceRequirements(
|
|
2703
|
+
# NOTE: base resources for this are kept to a minimum to save on running costs.
|
|
2704
|
+
# This has an adverse effect on startup time for the daemon, which can be completely
|
|
2705
|
+
# alleviated by using a base image that has the required dependencies pre-installed
|
|
2706
|
+
requests={
|
|
2707
|
+
"cpu": "200m",
|
|
2708
|
+
"memory": "100Mi",
|
|
2709
|
+
},
|
|
2710
|
+
limits={
|
|
2711
|
+
"cpu": "200m",
|
|
2712
|
+
"memory": "100Mi",
|
|
2713
|
+
},
|
|
2714
|
+
),
|
|
2715
|
+
)
|
|
2683
2716
|
)
|
|
2684
2717
|
)
|
|
2685
2718
|
)
|
|
@@ -3301,6 +3334,19 @@ class DaemonTemplate(object):
|
|
|
3301
3334
|
self.payload["container"] = container
|
|
3302
3335
|
return self
|
|
3303
3336
|
|
|
3337
|
+
def service_account_name(self, service_account_name):
|
|
3338
|
+
self.payload["serviceAccountName"] = service_account_name
|
|
3339
|
+
return self
|
|
3340
|
+
|
|
3341
|
+
def retry_strategy(self, times, minutes_between_retries):
|
|
3342
|
+
if times > 0:
|
|
3343
|
+
self.payload["retryStrategy"] = {
|
|
3344
|
+
"retryPolicy": "Always",
|
|
3345
|
+
"limit": times,
|
|
3346
|
+
"backoff": {"duration": "%sm" % minutes_between_retries},
|
|
3347
|
+
}
|
|
3348
|
+
return self
|
|
3349
|
+
|
|
3304
3350
|
def to_json(self):
|
|
3305
3351
|
return self.payload
|
|
3306
3352
|
|
|
@@ -45,7 +45,7 @@ from metaflow.metaflow_config import (
|
|
|
45
45
|
SERVICE_INTERNAL_URL,
|
|
46
46
|
)
|
|
47
47
|
from metaflow.unbounded_foreach import UBF_CONTROL, UBF_TASK
|
|
48
|
-
from metaflow.metaflow_config_funcs import config_values
|
|
48
|
+
from metaflow.metaflow_config_funcs import config_values, init_config
|
|
49
49
|
from metaflow.mflog import (
|
|
50
50
|
BASH_SAVE_LOGS,
|
|
51
51
|
bash_capture_logs,
|
|
@@ -311,6 +311,22 @@ class Kubernetes(object):
|
|
|
311
311
|
) + KUBERNETES_SECRETS.split(","):
|
|
312
312
|
jobset.secret(k)
|
|
313
313
|
|
|
314
|
+
initial_configs = init_config()
|
|
315
|
+
for entry in ["OBP_PERIMETER", "OBP_INTEGRATIONS_SECRETS_METADATA_URL"]:
|
|
316
|
+
if entry not in initial_configs:
|
|
317
|
+
raise KubernetesException(
|
|
318
|
+
f"{entry} was not found in metaflow config. Please make sure to run `outerbounds configure <...>` command which can be found on the Ourebounds UI or reach out to your Outerbounds support team."
|
|
319
|
+
)
|
|
320
|
+
|
|
321
|
+
additional_obp_configs = {
|
|
322
|
+
"OBP_PERIMETER": initial_configs["OBP_PERIMETER"],
|
|
323
|
+
"OBP_INTEGRATIONS_SECRETS_METADATA_URL": initial_configs[
|
|
324
|
+
"OBP_INTEGRATIONS_SECRETS_METADATA_URL"
|
|
325
|
+
],
|
|
326
|
+
}
|
|
327
|
+
for k, v in additional_obp_configs.items():
|
|
328
|
+
jobset.environment_variable(k, v)
|
|
329
|
+
|
|
314
330
|
jobset.environment_variables_from_selectors(
|
|
315
331
|
{
|
|
316
332
|
"METAFLOW_KUBERNETES_NAMESPACE": "metadata.namespace",
|
|
@@ -155,7 +155,10 @@ class KubernetesClient(object):
|
|
|
155
155
|
echo("failed to kill pod %s - %s" % (pod.metadata.name, str(e)))
|
|
156
156
|
|
|
157
157
|
with ThreadPoolExecutor() as executor:
|
|
158
|
-
executor.map(_kill_pod,
|
|
158
|
+
operated_pods = list(executor.map(_kill_pod, pods))
|
|
159
|
+
|
|
160
|
+
if not operated_pods:
|
|
161
|
+
echo("No active Kubernetes pods found for run *%s*" % run_id)
|
|
159
162
|
|
|
160
163
|
def jobset(self, **kwargs):
|
|
161
164
|
return KubernetesJobSet(self, **kwargs)
|
|
@@ -558,7 +558,11 @@ class KubernetesDecorator(StepDecorator):
|
|
|
558
558
|
|
|
559
559
|
# TODO: Unify this method with the multi-node setup in @batch
|
|
560
560
|
def _setup_multinode_environment():
|
|
561
|
-
# FIXME
|
|
561
|
+
# TODO [FIXME SOON]
|
|
562
|
+
# Even if Kubernetes may deploy control pods before worker pods, there is always a
|
|
563
|
+
# possibility that the worker pods may start before the control. In the case that this happens,
|
|
564
|
+
# the worker pods will not be able to resolve the control pod's IP address and this will cause
|
|
565
|
+
# the worker pods to fail. This function should account for this in the near future.
|
|
562
566
|
import socket
|
|
563
567
|
|
|
564
568
|
try:
|
|
@@ -9,6 +9,7 @@ from metaflow.exception import MetaflowException
|
|
|
9
9
|
from metaflow.metaflow_config import KUBERNETES_SECRETS
|
|
10
10
|
from metaflow.tracing import inject_tracing_vars
|
|
11
11
|
from metaflow.unbounded_foreach import UBF_CONTROL, UBF_TASK
|
|
12
|
+
from metaflow.metaflow_config_funcs import init_config
|
|
12
13
|
|
|
13
14
|
CLIENT_REFRESH_INTERVAL_SECONDS = 300
|
|
14
15
|
from .kubernetes_jobsets import (
|
|
@@ -74,6 +75,20 @@ class KubernetesJob(object):
|
|
|
74
75
|
if self._kwargs["shared_memory"]
|
|
75
76
|
else None
|
|
76
77
|
)
|
|
78
|
+
initial_configs = init_config()
|
|
79
|
+
for entry in ["OBP_PERIMETER", "OBP_INTEGRATIONS_SECRETS_METADATA_URL"]:
|
|
80
|
+
if entry not in initial_configs:
|
|
81
|
+
raise KubernetesJobException(
|
|
82
|
+
f"{entry} was not found in metaflow config. Please make sure to run `outerbounds configure <...>` command which can be found on the Ourebounds UI or reach out to your Outerbounds support team."
|
|
83
|
+
)
|
|
84
|
+
|
|
85
|
+
additional_obp_configs = {
|
|
86
|
+
"OBP_PERIMETER": initial_configs["OBP_PERIMETER"],
|
|
87
|
+
"OBP_INTEGRATIONS_SECRETS_METADATA_URL": initial_configs[
|
|
88
|
+
"OBP_INTEGRATIONS_SECRETS_METADATA_URL"
|
|
89
|
+
],
|
|
90
|
+
}
|
|
91
|
+
|
|
77
92
|
return client.V1JobSpec(
|
|
78
93
|
# Retries are handled by Metaflow when it is responsible for
|
|
79
94
|
# executing the flow. The responsibility is moved to Kubernetes
|
|
@@ -135,6 +150,13 @@ class KubernetesJob(object):
|
|
|
135
150
|
"METAFLOW_KUBERNETES_NODE_IP": "status.hostIP",
|
|
136
151
|
}.items()
|
|
137
152
|
]
|
|
153
|
+
+ [
|
|
154
|
+
client.V1EnvVar(
|
|
155
|
+
name=k,
|
|
156
|
+
value=v,
|
|
157
|
+
)
|
|
158
|
+
for k, v in additional_obp_configs.items()
|
|
159
|
+
]
|
|
138
160
|
+ [
|
|
139
161
|
client.V1EnvVar(name=k, value=str(v))
|
|
140
162
|
for k, v in inject_tracing_vars({}).items()
|
|
@@ -866,7 +866,13 @@ class KubernetesJobSet(object):
|
|
|
866
866
|
spec=dict(
|
|
867
867
|
replicatedJobs=[self.control.dump(), self.worker.dump()],
|
|
868
868
|
suspend=False,
|
|
869
|
-
startupPolicy=
|
|
869
|
+
startupPolicy=dict(
|
|
870
|
+
# We explicitly set an InOrder Startup policy so that
|
|
871
|
+
# we can ensure that the control pod starts before the worker pods.
|
|
872
|
+
# This is required so that when worker pods try to access the control's IP
|
|
873
|
+
# we are able to resolve the control's IP address.
|
|
874
|
+
startupPolicyOrder="InOrder"
|
|
875
|
+
),
|
|
870
876
|
successPolicy=None,
|
|
871
877
|
# The Failure Policy helps setting the number of retries for the jobset.
|
|
872
878
|
# but we don't rely on it and instead rely on either the local scheduler
|
|
@@ -89,7 +89,7 @@ if __name__ == "__main__":
|
|
|
89
89
|
# TODO: micromamba installation can be pawned off to micromamba.py
|
|
90
90
|
f"""set -e;
|
|
91
91
|
if ! command -v micromamba >/dev/null 2>&1; then
|
|
92
|
-
mkdir micromamba;
|
|
92
|
+
mkdir -p micromamba;
|
|
93
93
|
python -c "import requests, bz2, sys; data = requests.get('https://micro.mamba.pm/api/micromamba/{architecture}/1.5.7').content; sys.stdout.buffer.write(bz2.decompress(data))" | tar -xv -C $(pwd)/micromamba bin/micromamba --strip-components 1;
|
|
94
94
|
export PATH=$PATH:$(pwd)/micromamba;
|
|
95
95
|
if ! command -v micromamba >/dev/null 2>&1; then
|
|
@@ -253,7 +253,33 @@ class Micromamba(object):
|
|
|
253
253
|
try:
|
|
254
254
|
output = json.loads(e.output)
|
|
255
255
|
err = []
|
|
256
|
+
v_pkgs = ["__cuda", "__glibc"]
|
|
256
257
|
for error in output.get("solver_problems", []):
|
|
258
|
+
# raise a specific error message for virtual package related errors
|
|
259
|
+
match = next((p for p in v_pkgs if p in error), None)
|
|
260
|
+
if match is not None:
|
|
261
|
+
vpkg_name = match[2:]
|
|
262
|
+
# try to strip version from error msg which are of the format:
|
|
263
|
+
# nothing provides <__vpkg> >=2.17,<3.0.a0 needed by <pkg_name>
|
|
264
|
+
try:
|
|
265
|
+
vpkg_version = error[
|
|
266
|
+
len("nothing provides %s " % match) : error.index(
|
|
267
|
+
" needed by"
|
|
268
|
+
)
|
|
269
|
+
]
|
|
270
|
+
except ValueError:
|
|
271
|
+
vpkg_version = None
|
|
272
|
+
raise MicromambaException(
|
|
273
|
+
"Please set the environment variable CONDA_OVERRIDE_{var} to a specific version{version} of {name}.\n"
|
|
274
|
+
"Here is an example of supplying environment variables through the command line -\n\n"
|
|
275
|
+
"CONDA_OVERRIDE_{var}=<{name}-version> python flow.py <args>".format(
|
|
276
|
+
var=vpkg_name.upper(),
|
|
277
|
+
version=(
|
|
278
|
+
"" if not vpkg_version else (" (%s)" % vpkg_version)
|
|
279
|
+
),
|
|
280
|
+
name=vpkg_name,
|
|
281
|
+
),
|
|
282
|
+
)
|
|
257
283
|
err.append(error)
|
|
258
284
|
raise MicromambaException(
|
|
259
285
|
msg.format(
|
metaflow/runner/deployer.py
CHANGED
|
@@ -6,56 +6,11 @@ import importlib
|
|
|
6
6
|
import functools
|
|
7
7
|
import tempfile
|
|
8
8
|
|
|
9
|
-
from subprocess import CalledProcessError
|
|
10
9
|
from typing import Optional, Dict, ClassVar
|
|
11
10
|
|
|
12
11
|
from metaflow.exception import MetaflowNotFound
|
|
13
|
-
from metaflow.runner.subprocess_manager import
|
|
14
|
-
from metaflow.runner.utils import
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
def handle_timeout(
|
|
18
|
-
tfp_runner_attribute, command_obj: CommandManager, file_read_timeout: int
|
|
19
|
-
):
|
|
20
|
-
"""
|
|
21
|
-
Handle the timeout for a running subprocess command that reads a file
|
|
22
|
-
and raises an error with appropriate logs if a TimeoutError occurs.
|
|
23
|
-
|
|
24
|
-
Parameters
|
|
25
|
-
----------
|
|
26
|
-
tfp_runner_attribute : NamedTemporaryFile
|
|
27
|
-
Temporary file that stores runner attribute data.
|
|
28
|
-
command_obj : CommandManager
|
|
29
|
-
Command manager object that encapsulates the running command details.
|
|
30
|
-
file_read_timeout : int
|
|
31
|
-
Timeout for reading the file.
|
|
32
|
-
|
|
33
|
-
Returns
|
|
34
|
-
-------
|
|
35
|
-
str
|
|
36
|
-
Content read from the temporary file.
|
|
37
|
-
|
|
38
|
-
Raises
|
|
39
|
-
------
|
|
40
|
-
RuntimeError
|
|
41
|
-
If a TimeoutError occurs, it raises a RuntimeError with the command's
|
|
42
|
-
stdout and stderr logs.
|
|
43
|
-
"""
|
|
44
|
-
try:
|
|
45
|
-
content = read_from_file_when_ready(
|
|
46
|
-
tfp_runner_attribute.name, command_obj, timeout=file_read_timeout
|
|
47
|
-
)
|
|
48
|
-
return content
|
|
49
|
-
except (CalledProcessError, TimeoutError) as e:
|
|
50
|
-
stdout_log = open(command_obj.log_files["stdout"]).read()
|
|
51
|
-
stderr_log = open(command_obj.log_files["stderr"]).read()
|
|
52
|
-
command = " ".join(command_obj.command)
|
|
53
|
-
error_message = "Error executing: '%s':\n" % command
|
|
54
|
-
if stdout_log.strip():
|
|
55
|
-
error_message += "\nStdout:\n%s\n" % stdout_log
|
|
56
|
-
if stderr_log.strip():
|
|
57
|
-
error_message += "\nStderr:\n%s\n" % stderr_log
|
|
58
|
-
raise RuntimeError(error_message) from e
|
|
12
|
+
from metaflow.runner.subprocess_manager import SubprocessManager
|
|
13
|
+
from metaflow.runner.utils import handle_timeout
|
|
59
14
|
|
|
60
15
|
|
|
61
16
|
def get_lower_level_group(
|
|
@@ -209,7 +164,7 @@ class TriggeredRun(object):
|
|
|
209
164
|
elif callable(v):
|
|
210
165
|
setattr(self, k, functools.partial(v, self))
|
|
211
166
|
else:
|
|
212
|
-
setattr(self
|
|
167
|
+
setattr(self, k, v)
|
|
213
168
|
|
|
214
169
|
def wait_for_run(self, timeout=None):
|
|
215
170
|
"""
|
|
@@ -287,7 +242,7 @@ class DeployedFlow(object):
|
|
|
287
242
|
elif callable(v):
|
|
288
243
|
setattr(self, k, functools.partial(v, self))
|
|
289
244
|
else:
|
|
290
|
-
setattr(self
|
|
245
|
+
setattr(self, k, v)
|
|
291
246
|
|
|
292
247
|
|
|
293
248
|
class DeployerImpl(object):
|
|
@@ -1,14 +1,14 @@
|
|
|
1
1
|
import importlib
|
|
2
2
|
import os
|
|
3
3
|
import sys
|
|
4
|
+
import json
|
|
4
5
|
import tempfile
|
|
5
6
|
|
|
6
|
-
from subprocess import CalledProcessError
|
|
7
7
|
from typing import Dict, Iterator, Optional, Tuple
|
|
8
8
|
|
|
9
9
|
from metaflow import Run, metadata
|
|
10
10
|
|
|
11
|
-
from .utils import
|
|
11
|
+
from .utils import handle_timeout, clear_and_set_os_environ
|
|
12
12
|
from .subprocess_manager import CommandManager, SubprocessManager
|
|
13
13
|
|
|
14
14
|
|
|
@@ -102,16 +102,19 @@ class ExecutingRun(object):
|
|
|
102
102
|
for executing the run.
|
|
103
103
|
|
|
104
104
|
The return value is one of the following strings:
|
|
105
|
+
- `timeout` indicates that the run timed out.
|
|
105
106
|
- `running` indicates a currently executing run.
|
|
106
107
|
- `failed` indicates a failed run.
|
|
107
|
-
- `successful` a successful run.
|
|
108
|
+
- `successful` indicates a successful run.
|
|
108
109
|
|
|
109
110
|
Returns
|
|
110
111
|
-------
|
|
111
112
|
str
|
|
112
113
|
The current status of the run.
|
|
113
114
|
"""
|
|
114
|
-
if self.command_obj.
|
|
115
|
+
if self.command_obj.timeout:
|
|
116
|
+
return "timeout"
|
|
117
|
+
elif self.command_obj.process.returncode is None:
|
|
115
118
|
return "running"
|
|
116
119
|
elif self.command_obj.process.returncode != 0:
|
|
117
120
|
return "failed"
|
|
@@ -271,28 +274,22 @@ class Runner(object):
|
|
|
271
274
|
|
|
272
275
|
# It is thus necessary to set them to correct values before we return
|
|
273
276
|
# the Run object.
|
|
274
|
-
try:
|
|
275
|
-
# Set the environment variables to what they were before the run executed.
|
|
276
|
-
clear_and_set_os_environ(self.old_env)
|
|
277
277
|
|
|
278
|
-
|
|
279
|
-
|
|
280
|
-
|
|
281
|
-
|
|
282
|
-
|
|
283
|
-
|
|
284
|
-
|
|
285
|
-
|
|
286
|
-
|
|
287
|
-
|
|
288
|
-
|
|
289
|
-
|
|
290
|
-
|
|
291
|
-
|
|
292
|
-
|
|
293
|
-
if stderr_log.strip():
|
|
294
|
-
error_message += "\nStderr:\n%s\n" % stderr_log
|
|
295
|
-
raise RuntimeError(error_message) from e
|
|
278
|
+
content = handle_timeout(
|
|
279
|
+
tfp_runner_attribute, command_obj, self.file_read_timeout
|
|
280
|
+
)
|
|
281
|
+
content = json.loads(content)
|
|
282
|
+
pathspec = "%s/%s" % (content.get("flow_name"), content.get("run_id"))
|
|
283
|
+
|
|
284
|
+
# Set the environment variables to what they were before the run executed.
|
|
285
|
+
clear_and_set_os_environ(self.old_env)
|
|
286
|
+
|
|
287
|
+
# Set the correct metadata from the runner_attribute file corresponding to this run.
|
|
288
|
+
metadata_for_flow = content.get("metadata")
|
|
289
|
+
metadata(metadata_for_flow)
|
|
290
|
+
|
|
291
|
+
run_object = Run(pathspec, _namespace_check=False)
|
|
292
|
+
return ExecutingRun(self, command_obj, run_object)
|
|
296
293
|
|
|
297
294
|
def run(self, **kwargs) -> ExecutingRun:
|
|
298
295
|
"""
|
|
@@ -42,6 +42,19 @@ class SubprocessManager(object):
|
|
|
42
42
|
def __init__(self):
|
|
43
43
|
self.commands: Dict[int, CommandManager] = {}
|
|
44
44
|
|
|
45
|
+
try:
|
|
46
|
+
loop = asyncio.get_running_loop()
|
|
47
|
+
loop.add_signal_handler(
|
|
48
|
+
signal.SIGINT,
|
|
49
|
+
lambda: self._handle_sigint(signum=signal.SIGINT, frame=None),
|
|
50
|
+
)
|
|
51
|
+
except RuntimeError:
|
|
52
|
+
signal.signal(signal.SIGINT, self._handle_sigint)
|
|
53
|
+
|
|
54
|
+
def _handle_sigint(self, signum, frame):
|
|
55
|
+
for each_command in self.commands.values():
|
|
56
|
+
each_command.kill(termination_timeout=2)
|
|
57
|
+
|
|
45
58
|
async def __aenter__(self) -> "SubprocessManager":
|
|
46
59
|
return self
|
|
47
60
|
|
|
@@ -83,6 +96,7 @@ class SubprocessManager(object):
|
|
|
83
96
|
command_obj = CommandManager(command, env, cwd)
|
|
84
97
|
pid = command_obj.run(show_output=show_output)
|
|
85
98
|
self.commands[pid] = command_obj
|
|
99
|
+
command_obj.sync_wait()
|
|
86
100
|
return pid
|
|
87
101
|
|
|
88
102
|
async def async_run_command(
|
|
@@ -169,11 +183,12 @@ class CommandManager(object):
|
|
|
169
183
|
self.cwd = cwd if cwd is not None else os.getcwd()
|
|
170
184
|
|
|
171
185
|
self.process = None
|
|
186
|
+
self.stdout_thread = None
|
|
187
|
+
self.stderr_thread = None
|
|
172
188
|
self.run_called: bool = False
|
|
189
|
+
self.timeout: bool = False
|
|
173
190
|
self.log_files: Dict[str, str] = {}
|
|
174
191
|
|
|
175
|
-
signal.signal(signal.SIGINT, self._handle_sigint)
|
|
176
|
-
|
|
177
192
|
async def __aenter__(self) -> "CommandManager":
|
|
178
193
|
return self
|
|
179
194
|
|
|
@@ -214,13 +229,22 @@ class CommandManager(object):
|
|
|
214
229
|
else:
|
|
215
230
|
await asyncio.wait_for(self.emit_logs(stream), timeout)
|
|
216
231
|
except asyncio.TimeoutError:
|
|
232
|
+
self.timeout = True
|
|
217
233
|
command_string = " ".join(self.command)
|
|
218
|
-
|
|
234
|
+
self.kill(termination_timeout=2)
|
|
219
235
|
print(
|
|
220
236
|
"Timeout: The process (PID %d; command: '%s') did not complete "
|
|
221
237
|
"within %s seconds." % (self.process.pid, command_string, timeout)
|
|
222
238
|
)
|
|
223
239
|
|
|
240
|
+
def sync_wait(self):
|
|
241
|
+
if not self.run_called:
|
|
242
|
+
raise RuntimeError("No command run yet to wait for...")
|
|
243
|
+
|
|
244
|
+
self.process.wait()
|
|
245
|
+
self.stdout_thread.join()
|
|
246
|
+
self.stderr_thread.join()
|
|
247
|
+
|
|
224
248
|
def run(self, show_output: bool = False):
|
|
225
249
|
"""
|
|
226
250
|
Run the subprocess synchronously. This can only be called once.
|
|
@@ -265,22 +289,17 @@ class CommandManager(object):
|
|
|
265
289
|
|
|
266
290
|
self.run_called = True
|
|
267
291
|
|
|
268
|
-
stdout_thread = threading.Thread(
|
|
292
|
+
self.stdout_thread = threading.Thread(
|
|
269
293
|
target=stream_to_stdout_and_file,
|
|
270
294
|
args=(self.process.stdout, stdout_logfile),
|
|
271
295
|
)
|
|
272
|
-
stderr_thread = threading.Thread(
|
|
296
|
+
self.stderr_thread = threading.Thread(
|
|
273
297
|
target=stream_to_stdout_and_file,
|
|
274
298
|
args=(self.process.stderr, stderr_logfile),
|
|
275
299
|
)
|
|
276
300
|
|
|
277
|
-
stdout_thread.start()
|
|
278
|
-
stderr_thread.start()
|
|
279
|
-
|
|
280
|
-
self.process.wait()
|
|
281
|
-
|
|
282
|
-
stdout_thread.join()
|
|
283
|
-
stderr_thread.join()
|
|
301
|
+
self.stdout_thread.start()
|
|
302
|
+
self.stderr_thread.start()
|
|
284
303
|
|
|
285
304
|
return self.process.pid
|
|
286
305
|
except Exception as e:
|
|
@@ -441,13 +460,13 @@ class CommandManager(object):
|
|
|
441
460
|
if self.run_called:
|
|
442
461
|
shutil.rmtree(self.temp_dir, ignore_errors=True)
|
|
443
462
|
|
|
444
|
-
|
|
463
|
+
def kill(self, termination_timeout: float = 2):
|
|
445
464
|
"""
|
|
446
465
|
Kill the subprocess and its descendants.
|
|
447
466
|
|
|
448
467
|
Parameters
|
|
449
468
|
----------
|
|
450
|
-
termination_timeout : float, default
|
|
469
|
+
termination_timeout : float, default 2
|
|
451
470
|
The time to wait after sending a SIGTERM to the process and its descendants
|
|
452
471
|
before sending a SIGKILL.
|
|
453
472
|
"""
|
|
@@ -457,9 +476,6 @@ class CommandManager(object):
|
|
|
457
476
|
else:
|
|
458
477
|
print("No process to kill.")
|
|
459
478
|
|
|
460
|
-
def _handle_sigint(self, signum, frame):
|
|
461
|
-
asyncio.create_task(self.kill())
|
|
462
|
-
|
|
463
479
|
|
|
464
480
|
async def main():
|
|
465
481
|
flow_file = "../try.py"
|
metaflow/runner/utils.py
CHANGED
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
import os
|
|
2
2
|
import ast
|
|
3
3
|
import time
|
|
4
|
+
import asyncio
|
|
4
5
|
|
|
5
6
|
from subprocess import CalledProcessError
|
|
6
7
|
from typing import Dict, TYPE_CHECKING
|
|
@@ -40,6 +41,13 @@ def clear_and_set_os_environ(env: Dict):
|
|
|
40
41
|
os.environ.update(env)
|
|
41
42
|
|
|
42
43
|
|
|
44
|
+
def check_process_status(command_obj: "CommandManager"):
|
|
45
|
+
if isinstance(command_obj.process, asyncio.subprocess.Process):
|
|
46
|
+
return command_obj.process.returncode is not None
|
|
47
|
+
else:
|
|
48
|
+
return command_obj.process.poll() is not None
|
|
49
|
+
|
|
50
|
+
|
|
43
51
|
def read_from_file_when_ready(
|
|
44
52
|
file_path: str, command_obj: "CommandManager", timeout: float = 5
|
|
45
53
|
):
|
|
@@ -47,7 +55,7 @@ def read_from_file_when_ready(
|
|
|
47
55
|
with open(file_path, "r", encoding="utf-8") as file_pointer:
|
|
48
56
|
content = file_pointer.read()
|
|
49
57
|
while not content:
|
|
50
|
-
if command_obj
|
|
58
|
+
if check_process_status(command_obj):
|
|
51
59
|
# Check to make sure the file hasn't been read yet to avoid a race
|
|
52
60
|
# where the file is written between the end of this while loop and the
|
|
53
61
|
# poll call above.
|
|
@@ -64,3 +72,47 @@ def read_from_file_when_ready(
|
|
|
64
72
|
time.sleep(0.1)
|
|
65
73
|
content = file_pointer.read()
|
|
66
74
|
return content
|
|
75
|
+
|
|
76
|
+
|
|
77
|
+
def handle_timeout(
|
|
78
|
+
tfp_runner_attribute, command_obj: "CommandManager", file_read_timeout: int
|
|
79
|
+
):
|
|
80
|
+
"""
|
|
81
|
+
Handle the timeout for a running subprocess command that reads a file
|
|
82
|
+
and raises an error with appropriate logs if a TimeoutError occurs.
|
|
83
|
+
|
|
84
|
+
Parameters
|
|
85
|
+
----------
|
|
86
|
+
tfp_runner_attribute : NamedTemporaryFile
|
|
87
|
+
Temporary file that stores runner attribute data.
|
|
88
|
+
command_obj : CommandManager
|
|
89
|
+
Command manager object that encapsulates the running command details.
|
|
90
|
+
file_read_timeout : int
|
|
91
|
+
Timeout for reading the file.
|
|
92
|
+
|
|
93
|
+
Returns
|
|
94
|
+
-------
|
|
95
|
+
str
|
|
96
|
+
Content read from the temporary file.
|
|
97
|
+
|
|
98
|
+
Raises
|
|
99
|
+
------
|
|
100
|
+
RuntimeError
|
|
101
|
+
If a TimeoutError occurs, it raises a RuntimeError with the command's
|
|
102
|
+
stdout and stderr logs.
|
|
103
|
+
"""
|
|
104
|
+
try:
|
|
105
|
+
content = read_from_file_when_ready(
|
|
106
|
+
tfp_runner_attribute.name, command_obj, timeout=file_read_timeout
|
|
107
|
+
)
|
|
108
|
+
return content
|
|
109
|
+
except (CalledProcessError, TimeoutError) as e:
|
|
110
|
+
stdout_log = open(command_obj.log_files["stdout"]).read()
|
|
111
|
+
stderr_log = open(command_obj.log_files["stderr"]).read()
|
|
112
|
+
command = " ".join(command_obj.command)
|
|
113
|
+
error_message = "Error executing: '%s':\n" % command
|
|
114
|
+
if stdout_log.strip():
|
|
115
|
+
error_message += "\nStdout:\n%s\n" % stdout_log
|
|
116
|
+
if stderr_log.strip():
|
|
117
|
+
error_message += "\nStderr:\n%s\n" % stderr_log
|
|
118
|
+
raise RuntimeError(error_message) from e
|
metaflow/task.py
CHANGED
|
@@ -180,6 +180,10 @@ class MetaflowTask(object):
|
|
|
180
180
|
# 2) join - pop the topmost frame from the stack
|
|
181
181
|
# 3) step following a split - push a new frame in the stack
|
|
182
182
|
|
|
183
|
+
# We have a non-modifying case (case 4)) where we propagate the
|
|
184
|
+
# foreach-stack information to all tasks in the foreach. This is
|
|
185
|
+
# then used later to write the foreach-stack metadata for that task
|
|
186
|
+
|
|
183
187
|
# case 1) - reset the stack
|
|
184
188
|
if step_name == "start":
|
|
185
189
|
self.flow._foreach_stack = []
|
|
@@ -264,6 +268,9 @@ class MetaflowTask(object):
|
|
|
264
268
|
stack = inputs[0]["_foreach_stack"]
|
|
265
269
|
stack.append(frame)
|
|
266
270
|
self.flow._foreach_stack = stack
|
|
271
|
+
# case 4) - propagate in the foreach nest
|
|
272
|
+
elif "_foreach_stack" in inputs[0]:
|
|
273
|
+
self.flow._foreach_stack = inputs[0]["_foreach_stack"]
|
|
267
274
|
|
|
268
275
|
def _clone_flow(self, datastore):
|
|
269
276
|
x = self.flow.__class__(use_cli=False)
|
metaflow/version.py
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
metaflow_version = "2.12.
|
|
1
|
+
metaflow_version = "2.12.24.1"
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.1
|
|
2
2
|
Name: ob-metaflow
|
|
3
|
-
Version: 2.12.
|
|
3
|
+
Version: 2.12.24.1
|
|
4
4
|
Summary: Metaflow: More Data Science, Less Engineering
|
|
5
5
|
Author: Netflix, Outerbounds & the Metaflow Community
|
|
6
6
|
Author-email: help@outerbounds.co
|
|
@@ -12,7 +12,7 @@ Requires-Dist: boto3
|
|
|
12
12
|
Requires-Dist: pylint
|
|
13
13
|
Requires-Dist: kubernetes
|
|
14
14
|
Provides-Extra: stubs
|
|
15
|
-
Requires-Dist: metaflow-stubs==2.12.
|
|
15
|
+
Requires-Dist: metaflow-stubs==2.12.24.1; extra == "stubs"
|
|
16
16
|
|
|
17
17
|

|
|
18
18
|
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
metaflow/R.py,sha256=CqVfIatvmjciuICNnoyyNGrwE7Va9iXfLdFbQa52hwA,3958
|
|
2
2
|
metaflow/__init__.py,sha256=mrvSL4mgwJOGkNWmk0cdSJSFz1aIhF5K4mdLgOfQbME,6010
|
|
3
3
|
metaflow/cards.py,sha256=tP1_RrtmqdFh741pqE4t98S7SA0MtGRlGvRICRZF1Mg,426
|
|
4
|
-
metaflow/cli.py,sha256=
|
|
4
|
+
metaflow/cli.py,sha256=vz8flftUkmRBdjHHwWREsFecfNqlFF0YoAKSzexE30w,34494
|
|
5
5
|
metaflow/cli_args.py,sha256=lcgBGNTvfaiPxiUnejAe60Upt9swG6lRy1_3OqbU6MY,2616
|
|
6
6
|
metaflow/clone_util.py,sha256=XfUX0vssu_hPlyZfhFl1AOnKkLqvt33Qp8xNrmdocGg,2057
|
|
7
7
|
metaflow/cmd_with_io.py,sha256=kl53HkAIyv0ecpItv08wZYczv7u3msD1VCcciqigqf0,588
|
|
@@ -21,7 +21,7 @@ metaflow/metaflow_config_funcs.py,sha256=5GlvoafV6SxykwfL8D12WXSfwjBN_NsyuKE_Q3g
|
|
|
21
21
|
metaflow/metaflow_current.py,sha256=pC-EMnAsnvBLvLd61W6MvfiCKcboryeui9f6r8z_sg8,7161
|
|
22
22
|
metaflow/metaflow_environment.py,sha256=rojFyGdyY56sN1HaEb1-0XX53Q3XPNnl0SaH-8xXZ8w,7987
|
|
23
23
|
metaflow/metaflow_profile.py,sha256=jKPEW-hmAQO-htSxb9hXaeloLacAh41A35rMZH6G8pA,418
|
|
24
|
-
metaflow/metaflow_version.py,sha256=
|
|
24
|
+
metaflow/metaflow_version.py,sha256=duhIzfKZtcxMVMs2uiBqBvUarSHJqyWDwMhaBOQd_g0,7491
|
|
25
25
|
metaflow/monitor.py,sha256=T0NMaBPvXynlJAO_avKtk8OIIRMyEuMAyF8bIp79aZU,5323
|
|
26
26
|
metaflow/multicore_utils.py,sha256=vdTNgczVLODifscUbbveJbuSDOl3Y9pAxhr7sqYiNf4,4760
|
|
27
27
|
metaflow/package.py,sha256=QutDP6WzjwGk1UCKXqBfXa9F10Q--FlRr0J7fwlple0,7399
|
|
@@ -31,12 +31,12 @@ metaflow/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
|
31
31
|
metaflow/pylint_wrapper.py,sha256=zzBY9YaSUZOGH-ypDKAv2B_7XcoyMZj-zCoCrmYqNRc,2865
|
|
32
32
|
metaflow/runtime.py,sha256=fbBObJJciagHWPzR3T7x9e_jez_RBnLZIHsXMvYnW_M,68875
|
|
33
33
|
metaflow/tagging_util.py,sha256=ctyf0Q1gBi0RyZX6J0e9DQGNkNHblV_CITfy66axXB4,2346
|
|
34
|
-
metaflow/task.py,sha256=
|
|
34
|
+
metaflow/task.py,sha256=5-Qy6wAm7dt7DnhnX1KVhVAKl4DWb3IZWIN5YCdRGIg,29043
|
|
35
35
|
metaflow/tuple_util.py,sha256=_G5YIEhuugwJ_f6rrZoelMFak3DqAR2tt_5CapS1XTY,830
|
|
36
36
|
metaflow/unbounded_foreach.py,sha256=p184WMbrMJ3xKYHwewj27ZhRUsSj_kw1jlye5gA9xJk,387
|
|
37
37
|
metaflow/util.py,sha256=olAvJK3y1it_k99MhLulTaAJo7OFVt5rnrD-ulIFLCU,13616
|
|
38
38
|
metaflow/vendor.py,sha256=FchtA9tH22JM-eEtJ2c9FpUdMn8sSb1VHuQS56EcdZk,5139
|
|
39
|
-
metaflow/version.py,sha256=
|
|
39
|
+
metaflow/version.py,sha256=d372RV8qNQ_4WoLsDeosYcVFOzaXBqQDVoqK6m525cM,31
|
|
40
40
|
metaflow/_vendor/__init__.py,sha256=y_CiwUD3l4eAKvTVDZeqgVujMy31cAM1qjAB-HfI-9s,353
|
|
41
41
|
metaflow/_vendor/typing_extensions.py,sha256=0nUs5p1A_UrZigrAVBoOEM6TxU37zzPDUtiij1ZwpNc,110417
|
|
42
42
|
metaflow/_vendor/zipp.py,sha256=ajztOH-9I7KA_4wqDYygtHa6xUBVZgFpmZ8FE74HHHI,8425
|
|
@@ -111,7 +111,7 @@ metaflow/_vendor/v3_6/importlib_metadata/_meta.py,sha256=_F48Hu_jFxkfKWz5wcYS8vO
|
|
|
111
111
|
metaflow/_vendor/v3_6/importlib_metadata/_text.py,sha256=HCsFksZpJLeTP3NEk_ngrAeXVRRtTrtyh9eOABoRP4A,2166
|
|
112
112
|
metaflow/_vendor/v3_6/importlib_metadata/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
113
113
|
metaflow/client/__init__.py,sha256=1GtQB4Y_CBkzaxg32L1syNQSlfj762wmLrfrDxGi1b8,226
|
|
114
|
-
metaflow/client/core.py,sha256=
|
|
114
|
+
metaflow/client/core.py,sha256=vDRmLhoRXOfFiIplY2Xp3Go5lnn-CKeJdPqtOjWIX4Y,74173
|
|
115
115
|
metaflow/client/filecache.py,sha256=Wy0yhhCqC1JZgebqi7z52GCwXYnkAqMZHTtxThvwBgM,15229
|
|
116
116
|
metaflow/cmd/__init__.py,sha256=AbpHGcgLb-kRsJGnwFEktk7uzpZOCcBY74-YBdrKVGs,1
|
|
117
117
|
metaflow/cmd/configure_cmd.py,sha256=o-DKnUf2FBo_HiMVyoyzQaGBSMtpbEPEdFTQZ0hkU-k,33396
|
|
@@ -119,8 +119,8 @@ metaflow/cmd/main_cli.py,sha256=E546zT_jYQKysmjwfpEgzZd5QMsyirs28M2s0OPU93E,2966
|
|
|
119
119
|
metaflow/cmd/tutorials_cmd.py,sha256=8FdlKkicTOhCIDKcBR5b0Oz6giDvS-EMY3o9skIrRqw,5156
|
|
120
120
|
metaflow/cmd/util.py,sha256=jS_0rUjOnGGzPT65fzRLdGjrYAOOLA4jU2S0HJLV0oc,406
|
|
121
121
|
metaflow/cmd/develop/__init__.py,sha256=p1Sy8yU1MEKSrH5ttOWOZvNcI1qYu6J6jghdTHwPgOw,689
|
|
122
|
-
metaflow/cmd/develop/stub_generator.py,sha256
|
|
123
|
-
metaflow/cmd/develop/stubs.py,sha256=
|
|
122
|
+
metaflow/cmd/develop/stub_generator.py,sha256=_P_80CRFxyYjoMFynwg0IhAiexL9Wh2WqsnagiaVYVw,48050
|
|
123
|
+
metaflow/cmd/develop/stubs.py,sha256=JX2qNZDvG0upvPueAcLhoR_zyLtRranZMwY05tLdpRQ,11884
|
|
124
124
|
metaflow/datastore/__init__.py,sha256=VxP6ddJt3rwiCkpiSfAhyVkUCOe1pgZZsytVEJzFmSQ,155
|
|
125
125
|
metaflow/datastore/content_addressed_store.py,sha256=6T7tNqL29kpmecyMLHF35RhoSBOb-OZcExnsB65AvnI,7641
|
|
126
126
|
metaflow/datastore/datastore_set.py,sha256=R5pwnxg1DD8kBY9vElvd2eMknrvwTyiSwvQs67_z9bc,2361
|
|
@@ -173,10 +173,10 @@ metaflow/plugins/airflow/sensors/base_sensor.py,sha256=s-OQBfPWZ_T3wn96Ua59CCEj1
|
|
|
173
173
|
metaflow/plugins/airflow/sensors/external_task_sensor.py,sha256=zhYlrZnXT20KW8-fVk0fCNtTyNiKJB5PMVASacu30r0,6034
|
|
174
174
|
metaflow/plugins/airflow/sensors/s3_sensor.py,sha256=iDReG-7FKnumrtQg-HY6cCUAAqNA90nARrjjjEEk_x4,3275
|
|
175
175
|
metaflow/plugins/argo/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
176
|
-
metaflow/plugins/argo/argo_client.py,sha256=
|
|
176
|
+
metaflow/plugins/argo/argo_client.py,sha256=KTUpP0DmnmNsMp4tbdNyKX_zOdTFRVpUkrf7Vv79d-o,16011
|
|
177
177
|
metaflow/plugins/argo/argo_events.py,sha256=_C1KWztVqgi3zuH57pInaE9OzABc2NnncC-zdwOMZ-w,5909
|
|
178
|
-
metaflow/plugins/argo/argo_workflows.py,sha256=
|
|
179
|
-
metaflow/plugins/argo/argo_workflows_cli.py,sha256=
|
|
178
|
+
metaflow/plugins/argo/argo_workflows.py,sha256=AFP0ZmtsxoAAnReYEgfkGYOIWDdG6Ee-hu37CLMo4gY,175217
|
|
179
|
+
metaflow/plugins/argo/argo_workflows_cli.py,sha256=0qAGo0YlC1Y9-1zqYAzhVCpCcITotfOI421VOIRpseM,37232
|
|
180
180
|
metaflow/plugins/argo/argo_workflows_decorator.py,sha256=yprszMdbE3rBTcEA9VR0IEnPjTprUauZBc4SBb-Q7sA,7878
|
|
181
181
|
metaflow/plugins/argo/argo_workflows_deployer.py,sha256=wSSZtThn_VPvE_Wu6NB1L0Q86LmBJh9g009v_lpvBPM,8125
|
|
182
182
|
metaflow/plugins/argo/capture_error.py,sha256=Ys9dscGrTpW-ZCirLBU0gD9qBM0BjxyxGlUMKcwewQc,1852
|
|
@@ -280,20 +280,20 @@ metaflow/plugins/gcp/gs_utils.py,sha256=ZmIGFse1qYyvAVrwga23PQUzF6dXEDLLsZ2F-YRm
|
|
|
280
280
|
metaflow/plugins/gcp/includefile_support.py,sha256=vIDeR-MiJuUh-2S2pV7Z7FBkhIWwtHXaRrj76MWGRiY,3869
|
|
281
281
|
metaflow/plugins/kubernetes/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
282
282
|
metaflow/plugins/kubernetes/kube_utils.py,sha256=fYDlvqi8jYPsWijDwT6Z2qhQswyFqv7tiwtic_I80Vg,749
|
|
283
|
-
metaflow/plugins/kubernetes/kubernetes.py,sha256=
|
|
283
|
+
metaflow/plugins/kubernetes/kubernetes.py,sha256=HYNIicCfzTL-9q-nEhwURe2z7_dNIe8eOSyzJEBnd4o,32108
|
|
284
284
|
metaflow/plugins/kubernetes/kubernetes_cli.py,sha256=sFZ9Zrjef85vCO0MGpUF-em8Pw3dePFb3hbX3PtAH4I,13463
|
|
285
|
-
metaflow/plugins/kubernetes/kubernetes_client.py,sha256=
|
|
286
|
-
metaflow/plugins/kubernetes/kubernetes_decorator.py,sha256=
|
|
287
|
-
metaflow/plugins/kubernetes/kubernetes_job.py,sha256=
|
|
288
|
-
metaflow/plugins/kubernetes/kubernetes_jobsets.py,sha256=
|
|
285
|
+
metaflow/plugins/kubernetes/kubernetes_client.py,sha256=tuvXP-QKpdeSmzVolB2R_TaacOr5DIb0j642eKcjsiM,6491
|
|
286
|
+
metaflow/plugins/kubernetes/kubernetes_decorator.py,sha256=Pr4nzigafYyHgZK4gPnQaQuyxViHozp_idl-AWY3cgk,26871
|
|
287
|
+
metaflow/plugins/kubernetes/kubernetes_job.py,sha256=E2Aa-IItwYErlwAbPgwdqJI_l6axBQ-IzKY__psW-nE,32865
|
|
288
|
+
metaflow/plugins/kubernetes/kubernetes_jobsets.py,sha256=wb0sK1OxW7pRbKdj6bWB4JsskXDsoKKqjyUWo4N9Y6E,41196
|
|
289
289
|
metaflow/plugins/metadata/__init__.py,sha256=AbpHGcgLb-kRsJGnwFEktk7uzpZOCcBY74-YBdrKVGs,1
|
|
290
290
|
metaflow/plugins/metadata/local.py,sha256=YhLJC5zjVJrvQFIyQ92ZBByiUmhCC762RUX7ITX12O8,22428
|
|
291
291
|
metaflow/plugins/metadata/service.py,sha256=ihq5F7KQZlxvYwzH_-jyP2aWN_I96i2vp92j_d697s8,20204
|
|
292
292
|
metaflow/plugins/pypi/__init__.py,sha256=0YFZpXvX7HCkyBFglatual7XGifdA1RwC3U4kcizyak,1037
|
|
293
|
-
metaflow/plugins/pypi/bootstrap.py,sha256=
|
|
293
|
+
metaflow/plugins/pypi/bootstrap.py,sha256=FI-itExqIz7DUzLnnkGwoB60rFBviygpIFThUtqk_4E,5227
|
|
294
294
|
metaflow/plugins/pypi/conda_decorator.py,sha256=fPeXxvmg51oSFTnlguNlcWUIdXHA9OuMnp9ElaxQPFo,15695
|
|
295
295
|
metaflow/plugins/pypi/conda_environment.py,sha256=NSBcN4oD_RZONOUrnubLh-ur8EfS0h9PM4r7GNZczyc,19685
|
|
296
|
-
metaflow/plugins/pypi/micromamba.py,sha256=
|
|
296
|
+
metaflow/plugins/pypi/micromamba.py,sha256=HQIxsixkLjqs0ukWGTlATNu5DrbisReOr39Qd21_GZo,13737
|
|
297
297
|
metaflow/plugins/pypi/pip.py,sha256=7B06mPOs5MvY33xbzPVYZlBr1iKMYaN-n8uulL9zSVg,13649
|
|
298
298
|
metaflow/plugins/pypi/pypi_decorator.py,sha256=rDMbHl7r81Ye7-TuIlKAVJ_CDnfjl9jV44ZPws-UsTY,7229
|
|
299
299
|
metaflow/plugins/pypi/pypi_environment.py,sha256=FYMg8kF3lXqcLfRYWD83a9zpVjcoo_TARqMGZ763rRk,230
|
|
@@ -303,12 +303,12 @@ metaflow/plugins/secrets/inline_secrets_provider.py,sha256=EChmoBGA1i7qM3jtYwPpL
|
|
|
303
303
|
metaflow/plugins/secrets/secrets_decorator.py,sha256=s-sFzPWOjahhpr5fMj-ZEaHkDYAPTO0isYXGvaUwlG8,11273
|
|
304
304
|
metaflow/runner/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
305
305
|
metaflow/runner/click_api.py,sha256=Qfg4BOz5K2LaXTYBsi1y4zTfNIsGGHBVF3UkorX_-o8,13878
|
|
306
|
-
metaflow/runner/deployer.py,sha256=
|
|
307
|
-
metaflow/runner/metaflow_runner.py,sha256=
|
|
306
|
+
metaflow/runner/deployer.py,sha256=xNqxFIezWz3AcVqR4jeL-JnInGtefwEYMbXttxgB_I8,12276
|
|
307
|
+
metaflow/runner/metaflow_runner.py,sha256=I7Ao0GHHfP55nUCE8g6CpTPGjuWgXedSc9EZX-tIE2c,15001
|
|
308
308
|
metaflow/runner/nbdeploy.py,sha256=fP1s_5MeiDyT_igP82pB5EUqX9rOy2s06Hyc-OUbOvQ,4115
|
|
309
309
|
metaflow/runner/nbrun.py,sha256=lmvhzMCz7iC9LSPGRijifW1wMXxa4RW_jVmpdjQi22E,7261
|
|
310
|
-
metaflow/runner/subprocess_manager.py,sha256=
|
|
311
|
-
metaflow/runner/utils.py,sha256=
|
|
310
|
+
metaflow/runner/subprocess_manager.py,sha256=jC_PIYIeAp_G__lf6WHZF3Lxzpp-WAQleMrRZq9j7nc,20467
|
|
311
|
+
metaflow/runner/utils.py,sha256=aQ6WiNz9b-pqWWE14PdcAqti7_Zh_MIPlEA8zXJ6tXo,3807
|
|
312
312
|
metaflow/sidecar/__init__.py,sha256=1mmNpmQ5puZCpRmmYlCOeieZ4108Su9XQ4_EqF1FGOU,131
|
|
313
313
|
metaflow/sidecar/sidecar.py,sha256=EspKXvPPNiyRToaUZ51PS5TT_PzrBNAurn_wbFnmGr0,1334
|
|
314
314
|
metaflow/sidecar/sidecar_messages.py,sha256=zPsCoYgDIcDkkvdC9MEpJTJ3y6TSGm2JWkRc4vxjbFA,1071
|
|
@@ -345,9 +345,9 @@ metaflow/tutorials/07-worldview/README.md,sha256=5vQTrFqulJ7rWN6r20dhot9lI2sVj9W
|
|
|
345
345
|
metaflow/tutorials/07-worldview/worldview.ipynb,sha256=ztPZPI9BXxvW1QdS2Tfe7LBuVzvFvv0AToDnsDJhLdE,2237
|
|
346
346
|
metaflow/tutorials/08-autopilot/README.md,sha256=GnePFp_q76jPs991lMUqfIIh5zSorIeWznyiUxzeUVE,1039
|
|
347
347
|
metaflow/tutorials/08-autopilot/autopilot.ipynb,sha256=DQoJlILV7Mq9vfPBGW-QV_kNhWPjS5n6SJLqePjFYLY,3191
|
|
348
|
-
ob_metaflow-2.12.
|
|
349
|
-
ob_metaflow-2.12.
|
|
350
|
-
ob_metaflow-2.12.
|
|
351
|
-
ob_metaflow-2.12.
|
|
352
|
-
ob_metaflow-2.12.
|
|
353
|
-
ob_metaflow-2.12.
|
|
348
|
+
ob_metaflow-2.12.24.1.dist-info/LICENSE,sha256=nl_Lt5v9VvJ-5lWJDT4ddKAG-VZ-2IaLmbzpgYDz2hU,11343
|
|
349
|
+
ob_metaflow-2.12.24.1.dist-info/METADATA,sha256=gvYCBrqcVdy7KaQQ1EF7OeXJH0lSKbfZDJc49usMgkk,5143
|
|
350
|
+
ob_metaflow-2.12.24.1.dist-info/WHEEL,sha256=AHX6tWk3qWuce7vKLrj7lnulVHEdWoltgauo8bgCXgU,109
|
|
351
|
+
ob_metaflow-2.12.24.1.dist-info/entry_points.txt,sha256=IKwTN1T3I5eJL3uo_vnkyxVffcgnRdFbKwlghZfn27k,57
|
|
352
|
+
ob_metaflow-2.12.24.1.dist-info/top_level.txt,sha256=v1pDHoWaSaKeuc5fKTRSfsXCKSdW1zvNVmvA-i0if3o,9
|
|
353
|
+
ob_metaflow-2.12.24.1.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|