llama-deploy-core 0.3.26__tar.gz → 0.4.0__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (24) hide show
  1. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/PKG-INFO +1 -1
  2. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/pyproject.toml +1 -1
  3. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/_compat.py +3 -3
  4. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/iter_utils.py +5 -3
  5. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/README.md +0 -0
  6. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/__init__.py +0 -0
  7. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/client/manage_client.py +0 -0
  8. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/client/ssl_util.py +0 -0
  9. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/config.py +0 -0
  10. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/deployment_config.py +0 -0
  11. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/git/git_util.py +0 -0
  12. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/path_util.py +0 -0
  13. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/py.typed +0 -0
  14. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/__init__.py +0 -0
  15. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/base.py +0 -0
  16. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/deployments.py +0 -0
  17. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/git_validation.py +0 -0
  18. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/projects.py +0 -0
  19. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/schema/public.py +0 -0
  20. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/server/manage_api/__init__.py +0 -0
  21. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/server/manage_api/_abstract_deployments_service.py +0 -0
  22. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/server/manage_api/_create_deployments_router.py +0 -0
  23. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/server/manage_api/_exceptions.py +0 -0
  24. {llama_deploy_core-0.3.26 → llama_deploy_core-0.4.0}/src/llama_deploy/core/ui_build.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: llama-deploy-core
3
- Version: 0.3.26
3
+ Version: 0.4.0
4
4
  Summary: Core models and schemas for LlamaDeploy
5
5
  License: MIT
6
6
  Requires-Dist: fastapi>=0.115.0
@@ -13,7 +13,7 @@ dev = [
13
13
 
14
14
  [project]
15
15
  name = "llama-deploy-core"
16
- version = "0.3.26"
16
+ version = "0.4.0"
17
17
  description = "Core models and schemas for LlamaDeploy"
18
18
  readme = "README.md"
19
19
  license = {text = "MIT"}
@@ -7,10 +7,10 @@ from typing import Any, BinaryIO
7
7
 
8
8
  if sys.version_info >= (3, 11):
9
9
  # Stdlib TOML parser (Python 3.11+)
10
- import tomllib as _toml_backend # type: ignore[import-not-found]
10
+ import tomllib as _toml_backend
11
11
  else:
12
12
  # Lightweight TOML backport for Python 3.10
13
- import tomli as _toml_backend # type: ignore[import-not-found]
13
+ import tomli as _toml_backend # type: ignore
14
14
 
15
15
 
16
16
  def get_logging_level_mapping() -> Mapping[str, int]:
@@ -21,7 +21,7 @@ def get_logging_level_mapping() -> Mapping[str, int]:
21
21
 
22
22
  return {
23
23
  name: level
24
- for name, level in logging._nameToLevel.items() # type: ignore[attr-defined]
24
+ for name, level in logging._nameToLevel.items()
25
25
  if isinstance(level, int)
26
26
  }
27
27
 
@@ -4,7 +4,7 @@ from __future__ import annotations
4
4
 
5
5
  import asyncio
6
6
  import time
7
- from typing import Any, AsyncGenerator, Callable, TypeVar
7
+ from typing import Any, AsyncGenerator, Callable, TypeVar, cast
8
8
 
9
9
  from typing_extensions import Literal
10
10
 
@@ -41,11 +41,13 @@ async def debounced_sorted_prefix(
41
41
  yield buffered_item
42
42
  buffer = []
43
43
  else:
44
+ # item is T after checking != "__COMPLETE__"
45
+ actual_item = cast(T, item)
44
46
  if debouncer.is_complete:
45
- yield item # type: ignore[misc] # item is T after checking != "__COMPLETE__"
47
+ yield actual_item
46
48
  else:
47
49
  debouncer.extend_window()
48
- buffer.append(item) # type: ignore[arg-type] # item is T after checking != "__COMPLETE__"
50
+ buffer.append(actual_item)
49
51
 
50
52
 
51
53
  COMPLETE = Literal["__COMPLETE__"]