llama-deploy-core 0.3.10__tar.gz → 0.3.12__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/PKG-INFO +2 -1
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/pyproject.toml +2 -1
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/client/manage_client.py +4 -1
- llama_deploy_core-0.3.12/src/llama_deploy/core/client/ssl_util.py +32 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/deployments.py +12 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/README.md +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/__init__.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/config.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/deployment_config.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/git/git_util.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/iter_utils.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/path_util.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/py.typed +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/__init__.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/base.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/git_validation.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/projects.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/public.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/server/manage_api/__init__.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/server/manage_api/_abstract_deployments_service.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/server/manage_api/_create_deployments_router.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/server/manage_api/_exceptions.py +0 -0
- {llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/ui_build.py +0 -0
|
@@ -1,12 +1,13 @@
|
|
|
1
1
|
Metadata-Version: 2.3
|
|
2
2
|
Name: llama-deploy-core
|
|
3
|
-
Version: 0.3.
|
|
3
|
+
Version: 0.3.12
|
|
4
4
|
Summary: Core models and schemas for LlamaDeploy
|
|
5
5
|
License: MIT
|
|
6
6
|
Requires-Dist: fastapi>=0.115.0
|
|
7
7
|
Requires-Dist: overrides>=7.7.0
|
|
8
8
|
Requires-Dist: pydantic>=2.0.0
|
|
9
9
|
Requires-Dist: pyyaml>=6.0.2
|
|
10
|
+
Requires-Dist: truststore>=0.10.4
|
|
10
11
|
Requires-Dist: types-pyyaml>=6.0.12.20250822
|
|
11
12
|
Requires-Dist: httpx>=0.24.0,<1.0.0 ; extra == 'client'
|
|
12
13
|
Requires-Dist: fastapi>=0.115.0 ; extra == 'server'
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
[project]
|
|
2
2
|
name = "llama-deploy-core"
|
|
3
|
-
version = "0.3.
|
|
3
|
+
version = "0.3.12"
|
|
4
4
|
description = "Core models and schemas for LlamaDeploy"
|
|
5
5
|
readme = "README.md"
|
|
6
6
|
license = { text = "MIT" }
|
|
@@ -10,6 +10,7 @@ dependencies = [
|
|
|
10
10
|
"overrides>=7.7.0",
|
|
11
11
|
"pydantic>=2.0.0",
|
|
12
12
|
"pyyaml>=6.0.2",
|
|
13
|
+
"truststore>=0.10.4",
|
|
13
14
|
"types-pyyaml>=6.0.12.20250822",
|
|
14
15
|
]
|
|
15
16
|
|
{llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/client/manage_client.py
RENAMED
|
@@ -4,6 +4,7 @@ from contextlib import asynccontextmanager
|
|
|
4
4
|
from typing import AsyncIterator, Callable, List
|
|
5
5
|
|
|
6
6
|
import httpx
|
|
7
|
+
from llama_deploy.core.client.ssl_util import get_httpx_verify_param
|
|
7
8
|
from llama_deploy.core.schema import LogEvent
|
|
8
9
|
from llama_deploy.core.schema.deployments import (
|
|
9
10
|
DeploymentCreate,
|
|
@@ -30,13 +31,15 @@ class BaseClient:
|
|
|
30
31
|
if api_key:
|
|
31
32
|
headers["Authorization"] = f"Bearer {api_key}"
|
|
32
33
|
|
|
34
|
+
verify = get_httpx_verify_param()
|
|
33
35
|
self.client = httpx.AsyncClient(
|
|
34
36
|
base_url=self.base_url,
|
|
35
37
|
headers=headers,
|
|
36
38
|
auth=auth,
|
|
39
|
+
verify=verify,
|
|
37
40
|
)
|
|
38
41
|
self.hookless_client = httpx.AsyncClient(
|
|
39
|
-
base_url=self.base_url, headers=headers, auth=auth
|
|
42
|
+
base_url=self.base_url, headers=headers, auth=auth, verify=verify
|
|
40
43
|
)
|
|
41
44
|
|
|
42
45
|
async def aclose(self) -> None:
|
|
@@ -0,0 +1,32 @@
|
|
|
1
|
+
"""Utility functions for SSL/TLS configuration with optional truststore support."""
|
|
2
|
+
|
|
3
|
+
from __future__ import annotations
|
|
4
|
+
|
|
5
|
+
import os
|
|
6
|
+
import ssl
|
|
7
|
+
from typing import Any
|
|
8
|
+
|
|
9
|
+
import truststore
|
|
10
|
+
|
|
11
|
+
|
|
12
|
+
def get_ssl_context() -> ssl.SSLContext | bool:
|
|
13
|
+
"""Get SSL context for httpx clients.
|
|
14
|
+
|
|
15
|
+
Returns an SSL context using truststore if LLAMA_DEPLOY_USE_TRUSTSTORE is set,
|
|
16
|
+
otherwise returns True (default SSL verification).
|
|
17
|
+
|
|
18
|
+
Truststore allows Python to use the system certificate store, which is useful
|
|
19
|
+
for corporate environments with MITM proxies.
|
|
20
|
+
"""
|
|
21
|
+
if os.getenv("LLAMA_DEPLOY_USE_TRUSTSTORE", "").lower() in ("1", "true", "yes"):
|
|
22
|
+
return truststore.SSLContext(ssl.PROTOCOL_TLS_CLIENT)
|
|
23
|
+
return True
|
|
24
|
+
|
|
25
|
+
|
|
26
|
+
def get_httpx_verify_param() -> Any:
|
|
27
|
+
"""Get the verify parameter for httpx clients.
|
|
28
|
+
|
|
29
|
+
Returns an SSL context using truststore if configured, otherwise returns True.
|
|
30
|
+
This can be passed directly to httpx.Client/AsyncClient's verify parameter.
|
|
31
|
+
"""
|
|
32
|
+
return get_ssl_context()
|
{llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/deployments.py
RENAMED
|
@@ -41,6 +41,8 @@ class DeploymentResponse(Base):
|
|
|
41
41
|
status: LlamaDeploymentPhase
|
|
42
42
|
warning: str | None = None
|
|
43
43
|
events: list[DeploymentEvent] | None = None
|
|
44
|
+
# version selector for appserver image
|
|
45
|
+
llama_deploy_version: str | None = None
|
|
44
46
|
|
|
45
47
|
|
|
46
48
|
class DeploymentsListResponse(Base):
|
|
@@ -54,6 +56,8 @@ class DeploymentCreate(Base):
|
|
|
54
56
|
git_ref: str | None = None
|
|
55
57
|
personal_access_token: str | None = None
|
|
56
58
|
secrets: dict[str, str] | None = None
|
|
59
|
+
# optional version selector; if omitted, server may set based on client version
|
|
60
|
+
llama_deploy_version: str | None = None
|
|
57
61
|
|
|
58
62
|
|
|
59
63
|
class LlamaDeploymentMetadata(Base):
|
|
@@ -83,6 +87,8 @@ class LlamaDeploymentSpec(Base):
|
|
|
83
87
|
secretName: str | None = None
|
|
84
88
|
# when true, the deployment will prebuild the UI assets and serve them from a static file server
|
|
85
89
|
staticAssetsPath: str | None = None
|
|
90
|
+
# explicit imageTag (operator will use this if provided)
|
|
91
|
+
imageTag: str | None = None
|
|
86
92
|
|
|
87
93
|
|
|
88
94
|
class LlamaDeploymentStatus(Base):
|
|
@@ -122,6 +128,8 @@ class DeploymentUpdate(Base):
|
|
|
122
128
|
personal_access_token: str | None = None
|
|
123
129
|
secrets: dict[str, str | None] | None = None
|
|
124
130
|
static_assets_path: Path | None = None
|
|
131
|
+
# allow updating version selector
|
|
132
|
+
llama_deploy_version: str | None = None
|
|
125
133
|
|
|
126
134
|
|
|
127
135
|
class DeploymentUpdateResult(Base):
|
|
@@ -200,6 +208,10 @@ def apply_deployment_update(
|
|
|
200
208
|
# String value means add/update this secret
|
|
201
209
|
secret_adds[key] = value
|
|
202
210
|
|
|
211
|
+
# Handle version selector
|
|
212
|
+
if update.llama_deploy_version is not None:
|
|
213
|
+
updated_spec.imageTag = f"appserver-{update.llama_deploy_version}"
|
|
214
|
+
|
|
203
215
|
return DeploymentUpdateResult(
|
|
204
216
|
updated_spec=updated_spec,
|
|
205
217
|
secret_adds=secret_adds,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/deployment_config.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/__init__.py
RENAMED
|
File without changes
|
|
File without changes
|
{llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/git_validation.py
RENAMED
|
File without changes
|
{llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/projects.py
RENAMED
|
File without changes
|
{llama_deploy_core-0.3.10 → llama_deploy_core-0.3.12}/src/llama_deploy/core/schema/public.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|