llama-deploy-core 0.3.26__tar.gz → 0.4.0__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/PKG-INFO +1 -1
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/pyproject.toml +1 -1
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/_compat.py +3 -3
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/iter_utils.py +5 -3
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/README.md +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/__init__.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/client/manage_client.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/client/ssl_util.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/config.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/deployment_config.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/git/git_util.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/path_util.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/py.typed +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/__init__.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/base.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/deployments.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/git_validation.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/projects.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/public.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/server/manage_api/__init__.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/server/manage_api/_abstract_deployments_service.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/server/manage_api/_create_deployments_router.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/server/manage_api/_exceptions.py +0 -0
- {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/ui_build.py +0 -0
|
@@ -7,10 +7,10 @@ from typing import Any, BinaryIO
|
|
|
7
7
|
|
|
8
8
|
if sys.version_info >= (3, 11):
|
|
9
9
|
# Stdlib TOML parser (Python 3.11+)
|
|
10
|
-
import tomllib as _toml_backend
|
|
10
|
+
import tomllib as _toml_backend
|
|
11
11
|
else:
|
|
12
12
|
# Lightweight TOML backport for Python 3.10
|
|
13
|
-
import tomli as _toml_backend # type: ignore
|
|
13
|
+
import tomli as _toml_backend # type: ignore
|
|
14
14
|
|
|
15
15
|
|
|
16
16
|
def get_logging_level_mapping() -> Mapping[str, int]:
|
|
@@ -21,7 +21,7 @@ def get_logging_level_mapping() -> Mapping[str, int]:
|
|
|
21
21
|
|
|
22
22
|
return {
|
|
23
23
|
name: level
|
|
24
|
-
for name, level in logging._nameToLevel.items()
|
|
24
|
+
for name, level in logging._nameToLevel.items()
|
|
25
25
|
if isinstance(level, int)
|
|
26
26
|
}
|
|
27
27
|
|
|
@@ -4,7 +4,7 @@ from __future__ import annotations
|
|
|
4
4
|
|
|
5
5
|
import asyncio
|
|
6
6
|
import time
|
|
7
|
-
from typing import Any, AsyncGenerator, Callable, TypeVar
|
|
7
|
+
from typing import Any, AsyncGenerator, Callable, TypeVar, cast
|
|
8
8
|
|
|
9
9
|
from typing_extensions import Literal
|
|
10
10
|
|
|
@@ -41,11 +41,13 @@ async def debounced_sorted_prefix(
|
|
|
41
41
|
yield buffered_item
|
|
42
42
|
buffer = []
|
|
43
43
|
else:
|
|
44
|
+
# item is T after checking != "__COMPLETE__"
|
|
45
|
+
actual_item = cast(T, item)
|
|
44
46
|
if debouncer.is_complete:
|
|
45
|
-
yield
|
|
47
|
+
yield actual_item
|
|
46
48
|
else:
|
|
47
49
|
debouncer.extend_window()
|
|
48
|
-
buffer.append(
|
|
50
|
+
buffer.append(actual_item)
|
|
49
51
|
|
|
50
52
|
|
|
51
53
|
COMPLETE = Literal["__COMPLETE__"]
|
|
File without changes
|
|
File without changes
|
{llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/client/manage_client.py
RENAMED
|
File without changes
|
{llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/client/ssl_util.py
RENAMED
|
File without changes
|
|
File without changes
|
{llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/deployment_config.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/__init__.py
RENAMED
|
File without changes
|
|
File without changes
|
{llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/deployments.py
RENAMED
|
File without changes
|
{llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/git_validation.py
RENAMED
|
File without changes
|
{llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/projects.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|