llama-deploy-core 0.3.0a16__tar.gz → 0.3.0a18__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (21) hide show
  1. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/PKG-INFO +2 -2
  2. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/pyproject.toml +2 -2
  3. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/deployment_config.py +19 -0
  4. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/README.md +0 -0
  5. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/__init__.py +0 -0
  6. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/client/manage_client.py +0 -0
  7. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/config.py +0 -0
  8. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/git/git_util.py +0 -0
  9. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/path_util.py +0 -0
  10. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/py.typed +0 -0
  11. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/schema/__init__.py +0 -0
  12. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/schema/base.py +0 -0
  13. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/schema/deployments.py +0 -0
  14. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/schema/git_validation.py +0 -0
  15. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/schema/projects.py +0 -0
  16. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/schema/public.py +0 -0
  17. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/server/manage_api/__init__.py +0 -0
  18. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/server/manage_api/_abstract_deployments_service.py +0 -0
  19. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/server/manage_api/_create_deployments_router.py +0 -0
  20. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/server/manage_api/_exceptions.py +0 -0
  21. {llama_deploy_core-0.3.0a16 → llama_deploy_core-0.3.0a18}/src/llama_deploy/core/ui_build.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: llama-deploy-core
3
- Version: 0.3.0a16
3
+ Version: 0.3.0a18
4
4
  Summary: Core models and schemas for LlamaDeploy
5
5
  License: MIT
6
6
  Requires-Dist: fastapi>=0.115.0
@@ -8,7 +8,7 @@ Requires-Dist: overrides>=7.7.0
8
8
  Requires-Dist: pydantic>=2.0.0
9
9
  Requires-Dist: pyyaml>=6.0.2
10
10
  Requires-Dist: types-pyyaml>=6.0.12.20250822
11
- Requires-Dist: httpx>=0.27.0 ; extra == 'client'
11
+ Requires-Dist: httpx>=0.24.0,<1.0.0 ; extra == 'client'
12
12
  Requires-Dist: fastapi>=0.115.0 ; extra == 'server'
13
13
  Requires-Python: >=3.11, <4
14
14
  Provides-Extra: client
@@ -1,6 +1,6 @@
1
1
  [project]
2
2
  name = "llama-deploy-core"
3
- version = "0.3.0a16"
3
+ version = "0.3.0a18"
4
4
  description = "Core models and schemas for LlamaDeploy"
5
5
  readme = "README.md"
6
6
  license = { text = "MIT" }
@@ -18,7 +18,7 @@ server = [
18
18
  "fastapi>=0.115.0",
19
19
  ]
20
20
  client = [
21
- "httpx>=0.27.0",
21
+ "httpx>=0.24.0,<1.0.0",
22
22
  ]
23
23
 
24
24
  [build-system]
@@ -6,12 +6,26 @@ from pathlib import Path
6
6
  from typing import Any, TypeVar
7
7
 
8
8
  import yaml
9
+ from llama_deploy.core.git.git_util import get_git_root, is_git_repo
9
10
  from llama_deploy.core.path_util import validate_path_traversal
10
11
  from pydantic import BaseModel, ConfigDict, Field, ValidationError, model_validator
11
12
 
12
13
  DEFAULT_DEPLOYMENT_NAME = "default"
13
14
 
14
15
 
16
+ def read_deployment_config_from_git_root_or_cwd(
17
+ cwd: Path, config_path: Path
18
+ ) -> "DeploymentConfig":
19
+ """
20
+ Read the deployment config from the git root or cwd.
21
+ """
22
+ if is_git_repo():
23
+ git_root = get_git_root()
24
+ relative_cwd_path = cwd.relative_to(git_root)
25
+ return read_deployment_config(git_root, relative_cwd_path / config_path)
26
+ return read_deployment_config(cwd, config_path)
27
+
28
+
15
29
  def read_deployment_config(source_root: Path, config_path: Path) -> "DeploymentConfig":
16
30
  """
17
31
  Read the deployment config from the config directory.
@@ -160,6 +174,10 @@ class DeploymentConfig(BaseModel):
160
174
  default=DEFAULT_DEPLOYMENT_NAME,
161
175
  description="The url safe path name of the deployment.",
162
176
  )
177
+ llama_cloud: bool = Field(
178
+ default=False,
179
+ description="If true, serving locally expects Llama Cloud access and will inject credentials when possible.",
180
+ )
163
181
  app: str | None = Field(
164
182
  None,
165
183
  description="A full bundle of all workflows as an 'app'. \"path.to_import:app_name\"",
@@ -186,6 +204,7 @@ class DeploymentConfig(BaseModel):
186
204
 
187
205
  return DeploymentConfig(
188
206
  name=_pick_non_default(self.name, config.name, "default"),
207
+ llama_cloud=self.llama_cloud or config.llama_cloud,
189
208
  app=self.app or config.app,
190
209
  workflows={**self.workflows, **config.workflows},
191
210
  env_files=list(set(self.env_files + config.env_files)),