llamactl 0.3.24__py3-none-any.whl → 0.3.25__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- llama_deploy/cli/commands/dev.py +79 -1
- {llamactl-0.3.24.dist-info → llamactl-0.3.25.dist-info}/METADATA +3 -3
- {llamactl-0.3.24.dist-info → llamactl-0.3.25.dist-info}/RECORD +5 -5
- {llamactl-0.3.24.dist-info → llamactl-0.3.25.dist-info}/WHEEL +0 -0
- {llamactl-0.3.24.dist-info → llamactl-0.3.25.dist-info}/entry_points.txt +0 -0
llama_deploy/cli/commands/dev.py
CHANGED
|
@@ -79,6 +79,69 @@ def validate_command(deployment_file: Path, interactive: bool) -> None:
|
|
|
79
79
|
rprint(f"[green]Validated workflows in {config_dir} successfully.[/green]")
|
|
80
80
|
|
|
81
81
|
|
|
82
|
+
@dev.command(
|
|
83
|
+
"export-json-graph",
|
|
84
|
+
help="Produce a JSON graph representation of registered workflows",
|
|
85
|
+
hidden=True, # perhaps expose if we have a built in visualization (mermaid, etc.)
|
|
86
|
+
)
|
|
87
|
+
@click.argument(
|
|
88
|
+
"deployment_file",
|
|
89
|
+
required=False,
|
|
90
|
+
default=DEFAULT_DEPLOYMENT_FILE_PATH,
|
|
91
|
+
type=_ClickPath(dir_okay=True, resolve_path=True, path_type=Path),
|
|
92
|
+
)
|
|
93
|
+
@click.option(
|
|
94
|
+
"--output",
|
|
95
|
+
help=(
|
|
96
|
+
"File where output JSON graph will be saved. "
|
|
97
|
+
"Defaults to workflows.json in the current directory."
|
|
98
|
+
),
|
|
99
|
+
required=False,
|
|
100
|
+
default=None,
|
|
101
|
+
type=_ClickPath(dir_okay=True, resolve_path=True, path_type=Path),
|
|
102
|
+
)
|
|
103
|
+
@interactive_option
|
|
104
|
+
@global_options
|
|
105
|
+
def export_json_graph_command(
|
|
106
|
+
deployment_file: Path,
|
|
107
|
+
output: Path | None,
|
|
108
|
+
interactive: bool,
|
|
109
|
+
) -> None:
|
|
110
|
+
"""Export the configured workflows to a JSON document that may be used for graph visualization."""
|
|
111
|
+
if not deployment_file.exists():
|
|
112
|
+
rprint(f"[red]Deployment file '{deployment_file}' does not exist[/red]")
|
|
113
|
+
raise click.Abort()
|
|
114
|
+
|
|
115
|
+
_ensure_project_layout(
|
|
116
|
+
deployment_file, command_name="llamactl dev export-json-graph"
|
|
117
|
+
)
|
|
118
|
+
_maybe_inject_llama_cloud_credentials(
|
|
119
|
+
deployment_file, interactive, require_cloud=False
|
|
120
|
+
)
|
|
121
|
+
|
|
122
|
+
prepare_server(
|
|
123
|
+
deployment_file=deployment_file,
|
|
124
|
+
install=True,
|
|
125
|
+
build=False,
|
|
126
|
+
install_ui_deps=False,
|
|
127
|
+
)
|
|
128
|
+
|
|
129
|
+
wd = Path.cwd()
|
|
130
|
+
if output is None:
|
|
131
|
+
output = wd / "workflows.json"
|
|
132
|
+
|
|
133
|
+
try:
|
|
134
|
+
start_export_json_graph_in_target_venv(
|
|
135
|
+
cwd=wd,
|
|
136
|
+
deployment_file=deployment_file,
|
|
137
|
+
output=output,
|
|
138
|
+
)
|
|
139
|
+
except subprocess.CalledProcessError as exc:
|
|
140
|
+
rprint("[red]Workflow JSON graph export failed. See errors above.[/red]")
|
|
141
|
+
raise Exit(exc.returncode)
|
|
142
|
+
rprint(f"[green]Exported workflow JSON graph to {output}[/green]")
|
|
143
|
+
|
|
144
|
+
|
|
82
145
|
@dev.command(
|
|
83
146
|
"run",
|
|
84
147
|
help=(
|
|
@@ -199,6 +262,21 @@ def start_preflight_in_target_venv(*, cwd: Path, deployment_file: Path) -> None:
|
|
|
199
262
|
_start_preflight_in_target_venv(cwd=cwd, deployment_file=deployment_file)
|
|
200
263
|
|
|
201
264
|
|
|
265
|
+
def start_export_json_graph_in_target_venv(
|
|
266
|
+
*, cwd: Path, deployment_file: Path, output: Path
|
|
267
|
+
) -> None:
|
|
268
|
+
"""Thin wrapper so tests can monkeypatch `dev.start_export_json_graph_in_target_venv`."""
|
|
269
|
+
from llama_deploy.appserver.app import (
|
|
270
|
+
start_export_json_graph_in_target_venv as _start_export_json_graph_in_target_venv,
|
|
271
|
+
)
|
|
272
|
+
|
|
273
|
+
_start_export_json_graph_in_target_venv(
|
|
274
|
+
cwd=cwd,
|
|
275
|
+
deployment_file=deployment_file,
|
|
276
|
+
output=output,
|
|
277
|
+
)
|
|
278
|
+
|
|
279
|
+
|
|
202
280
|
def parse_environment_variables(
|
|
203
281
|
config: DeploymentConfig, config_parent: Path
|
|
204
282
|
) -> dict[str, str]:
|
|
@@ -210,4 +288,4 @@ def parse_environment_variables(
|
|
|
210
288
|
return _parse_environment_variables(config, config_parent)
|
|
211
289
|
|
|
212
290
|
|
|
213
|
-
__all__ = ["dev", "validate_command", "run_command"]
|
|
291
|
+
__all__ = ["dev", "validate_command", "run_command", "export_json_graph_command"]
|
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
Metadata-Version: 2.3
|
|
2
2
|
Name: llamactl
|
|
3
|
-
Version: 0.3.
|
|
3
|
+
Version: 0.3.25
|
|
4
4
|
Summary: A command-line interface for managing LlamaDeploy projects and deployments
|
|
5
5
|
Author: Adrian Lyjak
|
|
6
6
|
Author-email: Adrian Lyjak <adrianlyjak@gmail.com>
|
|
7
7
|
License: MIT
|
|
8
|
-
Requires-Dist: llama-deploy-core[client]>=0.3.
|
|
9
|
-
Requires-Dist: llama-deploy-appserver>=0.3.
|
|
8
|
+
Requires-Dist: llama-deploy-core[client]>=0.3.25,<0.4.0
|
|
9
|
+
Requires-Dist: llama-deploy-appserver>=0.3.25,<0.4.0
|
|
10
10
|
Requires-Dist: vibe-llama-core>=0.1.0
|
|
11
11
|
Requires-Dist: rich>=13.0.0
|
|
12
12
|
Requires-Dist: questionary>=2.0.0
|
|
@@ -5,7 +5,7 @@ llama_deploy/cli/client.py,sha256=61c04ff808374913bf2fc1fc5838c498b4f8c779d4e056
|
|
|
5
5
|
llama_deploy/cli/commands/aliased_group.py,sha256=364d846d9ceec465e6f2f47051ad06e1ccfbea1d7526654c1ffbd7b7ab7e6af0,1302
|
|
6
6
|
llama_deploy/cli/commands/auth.py,sha256=c8b94de8c0647e241b0083782b8e241225c0c68ee2d32f85d54c29ae0d7dcb1b,26891
|
|
7
7
|
llama_deploy/cli/commands/deployment.py,sha256=dc5d039224409708446b91db482c20da648eba720c1527cfdb2952a1bb07ad3e,15567
|
|
8
|
-
llama_deploy/cli/commands/dev.py,sha256=
|
|
8
|
+
llama_deploy/cli/commands/dev.py,sha256=10f394bc91ea71c3d1f23d280919482794ddf918e5676e7a3305f49d2a71f646,9436
|
|
9
9
|
llama_deploy/cli/commands/env.py,sha256=ae8f94eb2651a10615bac37afa16447ad1d78cb78c83ad8b8ae75e878733d323,7478
|
|
10
10
|
llama_deploy/cli/commands/init.py,sha256=afdb65b5e70cfaf3bdbc923d13db1a31ad23c14605e5bcd55ddaab8fff6e69a4,17514
|
|
11
11
|
llama_deploy/cli/commands/pkg.py,sha256=f91a87220c1d527e02a183dac5ca52c58608128e29bedf664362af3d2d31c461,4084
|
|
@@ -41,7 +41,7 @@ llama_deploy/cli/utils/env_inject.py,sha256=01911758bcc3cf22aad0db0d1ade56aece48
|
|
|
41
41
|
llama_deploy/cli/utils/redact.py,sha256=1e768d76b4a6708230c34f7ce8a5a82ab52795bb3d6ab0387071ab4e8d7e7934,863
|
|
42
42
|
llama_deploy/cli/utils/retry.py,sha256=62ca6f286cb4de38cc5efcef3f376b12c2e6eb9b3e5ebe555d2a60aeb0957c19,1526
|
|
43
43
|
llama_deploy/cli/utils/version.py,sha256=bf01a6dda948b868cc08c93701ed44cd36b487402404af8451d4c0996a2edb31,364
|
|
44
|
-
llamactl-0.3.
|
|
45
|
-
llamactl-0.3.
|
|
46
|
-
llamactl-0.3.
|
|
47
|
-
llamactl-0.3.
|
|
44
|
+
llamactl-0.3.25.dist-info/WHEEL,sha256=66530aef82d5020ef5af27ae0123c71abb9261377c5bc519376c671346b12918,79
|
|
45
|
+
llamactl-0.3.25.dist-info/entry_points.txt,sha256=b67e1eb64305058751a651a80f2d2268b5f7046732268421e796f64d4697f83c,52
|
|
46
|
+
llamactl-0.3.25.dist-info/METADATA,sha256=d9a13b3e596b86e5eb582b5756230c2994ea9a4aec6d12be7cab534e542b1022,3329
|
|
47
|
+
llamactl-0.3.25.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|