llamactl 0.3.0a6__py3-none-any.whl → 0.3.0a8__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -6,7 +6,10 @@ import webbrowser
6
6
  from textwrap import dedent
7
7
  from typing import Any, Dict, cast
8
8
 
9
- from aiohttp import web
9
+ from aiohttp.web_app import Application
10
+ from aiohttp.web_request import Request
11
+ from aiohttp.web_response import Response
12
+ from aiohttp.web_runner import AppRunner, TCPSite
10
13
 
11
14
  logger = logging.getLogger(__name__)
12
15
 
@@ -18,9 +21,9 @@ class GitHubCallbackServer:
18
21
  self.port = port
19
22
  self.callback_data: Dict[str, Any] = {}
20
23
  self.callback_received = asyncio.Event()
21
- self.app: web.Application | None = None
22
- self.runner: web.AppRunner | None = None
23
- self.site: web.TCPSite | None = None
24
+ self.app: Application | None = None
25
+ self.runner: AppRunner | None = None
26
+ self.site: TCPSite | None = None
24
27
 
25
28
  async def start_and_wait(self, timeout: float = 300) -> Dict[str, Any]:
26
29
  """Start the server and wait for a callback with timeout"""
@@ -38,19 +41,19 @@ class GitHubCallbackServer:
38
41
 
39
42
  async def _start_server(self) -> None:
40
43
  """Start the aiohttp server"""
41
- self.app = web.Application()
44
+ self.app = Application()
42
45
  self.app.router.add_get("/", self._handle_callback)
43
46
  self.app.router.add_get("/{path:.*}", self._handle_callback)
44
47
 
45
- self.runner = web.AppRunner(self.app, logger=None) # Suppress server logs
48
+ self.runner = AppRunner(self.app, logger=None) # Suppress server logs
46
49
  await self.runner.setup()
47
50
 
48
- self.site = web.TCPSite(self.runner, "localhost", self.port)
51
+ self.site = TCPSite(self.runner, "localhost", self.port)
49
52
  await self.site.start()
50
53
 
51
54
  logger.debug(f"GitHub callback server started on port {self.port}")
52
55
 
53
- async def _handle_callback(self, request: web.Request) -> web.Response:
56
+ async def _handle_callback(self, request: Request) -> Response:
54
57
  """Handle the GitHub callback"""
55
58
  # Capture query parameters
56
59
  query_params: dict[str, str] = dict(cast(Any, request.query))
@@ -62,7 +65,7 @@ class GitHubCallbackServer:
62
65
 
63
66
  # Return success page
64
67
  html_response = self._get_success_html()
65
- return web.Response(text=html_response, content_type="text/html")
68
+ return Response(text=html_response, content_type="text/html")
66
69
 
67
70
  async def stop(self) -> None:
68
71
  """Stop the server and cleanup"""
@@ -2,7 +2,6 @@
2
2
 
3
3
  from dataclasses import dataclass
4
4
  from pathlib import Path
5
- from typing import Optional
6
5
 
7
6
  from textual.app import App, ComposeResult
8
7
  from textual.containers import (
@@ -114,7 +114,7 @@ class SecretsWidget(Widget):
114
114
 
115
115
  # Short help text for textarea format
116
116
  yield Static(
117
- "Format: one per line, KEY=VALUE (e.g., API_KEY=abc).",
117
+ "Format: one per line, KEY=VALUE",
118
118
  classes="secret-label",
119
119
  )
120
120
 
@@ -1,19 +1,19 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: llamactl
3
- Version: 0.3.0a6
3
+ Version: 0.3.0a8
4
4
  Summary: A command-line interface for managing LlamaDeploy projects and deployments
5
5
  Author: Adrian Lyjak
6
6
  Author-email: Adrian Lyjak <adrianlyjak@gmail.com>
7
7
  License: MIT
8
- Requires-Dist: llama-deploy-core>=0.3.0a6,<0.4.0
9
- Requires-Dist: llama-deploy-appserver>=0.3.0a6,<0.4.0
8
+ Requires-Dist: llama-deploy-core>=0.3.0a8,<0.4.0
9
+ Requires-Dist: llama-deploy-appserver>=0.3.0a8,<0.4.0
10
10
  Requires-Dist: httpx>=0.24.0
11
11
  Requires-Dist: rich>=13.0.0
12
12
  Requires-Dist: questionary>=2.0.0
13
13
  Requires-Dist: click>=8.2.1
14
14
  Requires-Dist: python-dotenv>=1.0.0
15
15
  Requires-Dist: tenacity>=9.1.2
16
- Requires-Dist: textual>=4.0.0
16
+ Requires-Dist: textual>=5.3.0
17
17
  Requires-Dist: aiohttp>=3.12.14
18
18
  Requires-Dist: copier>=9.9.0
19
19
  Requires-Python: >=3.12, <4
@@ -0,0 +1,26 @@
1
+ llama_deploy/cli/__init__.py,sha256=274c45e48048bf60668ab564ae8e7c5e6daf1d7779005f87d07ce9fa7d04936c,422
2
+ llama_deploy/cli/app.py,sha256=5200b4ac01b0ad0c405ce841fc01a12ed32f7b6474472f00a7d6c75fe274ea45,2324
3
+ llama_deploy/cli/client.py,sha256=03d73191a2b89e33573b7df81b9d7586227e871451891f737d9f601d3aa47ffe,10032
4
+ llama_deploy/cli/commands/aliased_group.py,sha256=6e2457cdea51de83bb7f02b37abb77cb9b5bff0a61bdddd66c43240b66b13f13,986
5
+ llama_deploy/cli/commands/deployment.py,sha256=7874f4a499ce1bfd6ae14833410cc75c4c954463d96064cfd045421358479d4c,8810
6
+ llama_deploy/cli/commands/init.py,sha256=da6aecb5ebc8e4cf15421227fcafd98f573f601c52f8849c00d243b572c9f56a,6285
7
+ llama_deploy/cli/commands/profile.py,sha256=933d7a434c2684c7b47bfbd7340a09e4b34d56d20624886e15fdb4e0af97ce0b,6765
8
+ llama_deploy/cli/commands/serve.py,sha256=22227f383bb5a9d43de7c788139c685c7370e24f495c9b1929faae80b87d4ded,2232
9
+ llama_deploy/cli/config.py,sha256=ebec8cf9e2112378ee6ecd626166711f3fba8cfa27cd1c931fe899c0b2a047b3,6241
10
+ llama_deploy/cli/debug.py,sha256=e85a72d473bbe1645eb31772f7349bde703d45704166f767385895c440afc762,496
11
+ llama_deploy/cli/env.py,sha256=6ebc24579815b3787829c81fd5bb9f31698a06e62c0128a788559f962b33a7af,1016
12
+ llama_deploy/cli/interactive_prompts/utils.py,sha256=db78eba78bf347738feb89ac3eeb77a1d11f4003980f81cf3c13842f8d41afeb,2463
13
+ llama_deploy/cli/options.py,sha256=38bb4a231ad0436d8b910c98ff659c0736f619efdf56c402d60bb3f755df38e0,598
14
+ llama_deploy/cli/textual/deployment_form.py,sha256=5566e2545ef9548b14b7d3d2b0c1bda1dcd99aca814d34823da6b9da1903b8df,20890
15
+ llama_deploy/cli/textual/deployment_help.py,sha256=d43e9ff29db71a842cf8b491545763d581ede3132b8af518c73af85a40950046,2464
16
+ llama_deploy/cli/textual/deployment_monitor.py,sha256=1e1ea3381575d19e655a2a9eda8253c7e7fe9a02a2d637fd1fdce94500dde168,15044
17
+ llama_deploy/cli/textual/git_validation.py,sha256=44e359d16aa879f4566a0077d025fdd799f500862a8462b5ed3586e528f7a273,13300
18
+ llama_deploy/cli/textual/github_callback_server.py,sha256=dc74c510f8a98ef6ffaab0f6d11c7ea86ee77ca5adbc7725a2a29112bae24191,7556
19
+ llama_deploy/cli/textual/llama_loader.py,sha256=468213a504057f21838b01f48d51f52e60aa622d6f0fe5bb800d76ced846cea9,1245
20
+ llama_deploy/cli/textual/profile_form.py,sha256=4410678edbd59b014f937ce760bafa51ae86f6dd58bec88f048a9eda273446aa,5956
21
+ llama_deploy/cli/textual/secrets_form.py,sha256=a43fbd81aad034d0d60906bfd917c107f9ace414648b0f63ac0b29eeba4050db,7061
22
+ llama_deploy/cli/textual/styles.tcss,sha256=536cec7627d2a16dd03bf25bb9b6e4d53f1e0d18272b07ec0dc3bf76b0a7c2e0,3056
23
+ llamactl-0.3.0a8.dist-info/WHEEL,sha256=66530aef82d5020ef5af27ae0123c71abb9261377c5bc519376c671346b12918,79
24
+ llamactl-0.3.0a8.dist-info/entry_points.txt,sha256=b67e1eb64305058751a651a80f2d2268b5f7046732268421e796f64d4697f83c,52
25
+ llamactl-0.3.0a8.dist-info/METADATA,sha256=f4d04dd44460d1065bbc21979f97b7194b05b6bee247ee2748a2704d857d5ea5,3166
26
+ llamactl-0.3.0a8.dist-info/RECORD,,
@@ -1,24 +0,0 @@
1
- llama_deploy/cli/__init__.py,sha256=a7ac1d286680f8b95d046ed3f116c54d6301cf524ba39f0c19b39f9c95978d22,366
2
- llama_deploy/cli/app.py,sha256=5200b4ac01b0ad0c405ce841fc01a12ed32f7b6474472f00a7d6c75fe274ea45,2324
3
- llama_deploy/cli/client.py,sha256=b0a1f002bd3feab188459c6e56f75a995048016d135666a17e53c4588beea9cc,6911
4
- llama_deploy/cli/commands/aliased_group.py,sha256=6e2457cdea51de83bb7f02b37abb77cb9b5bff0a61bdddd66c43240b66b13f13,986
5
- llama_deploy/cli/commands/deployment.py,sha256=d1aacc8c6cbe4d73b3284d04e805fb044af591f9b323cc5c4acb2c07b0ad649c,8463
6
- llama_deploy/cli/commands/profile.py,sha256=933d7a434c2684c7b47bfbd7340a09e4b34d56d20624886e15fdb4e0af97ce0b,6765
7
- llama_deploy/cli/commands/serve.py,sha256=38e29816529af97bf00b8abea3ba6eb39e9a51e6b03616c2d8611b557079ddf2,1945
8
- llama_deploy/cli/config.py,sha256=1759e502ee77e72cce5358a4a85296f3e4182dd1495d21a3d9f308f92690e702,6251
9
- llama_deploy/cli/debug.py,sha256=e85a72d473bbe1645eb31772f7349bde703d45704166f767385895c440afc762,496
10
- llama_deploy/cli/env.py,sha256=6ebc24579815b3787829c81fd5bb9f31698a06e62c0128a788559f962b33a7af,1016
11
- llama_deploy/cli/interactive_prompts/utils.py,sha256=4ecab983c3f4869e7b8fcbb385e3ea91fde3bf86e8db7afe72d2febfd45d551e,2492
12
- llama_deploy/cli/options.py,sha256=38bb4a231ad0436d8b910c98ff659c0736f619efdf56c402d60bb3f755df38e0,598
13
- llama_deploy/cli/textual/deployment_form.py,sha256=9a7a364ab9735585db83d5955f4bb46faf0776b050c3477e2e7e5cfca5b11b2d,19465
14
- llama_deploy/cli/textual/deployment_help.py,sha256=7dae54ffbdaea201362928883cc206b496b1386023fb7f1c78d168132543e592,2339
15
- llama_deploy/cli/textual/git_validation.py,sha256=44e359d16aa879f4566a0077d025fdd799f500862a8462b5ed3586e528f7a273,13300
16
- llama_deploy/cli/textual/github_callback_server.py,sha256=031929592f448c5005e11cea19268fd696d60d92573c385a0397cfabacb06eb7,7444
17
- llama_deploy/cli/textual/llama_loader.py,sha256=468213a504057f21838b01f48d51f52e60aa622d6f0fe5bb800d76ced846cea9,1245
18
- llama_deploy/cli/textual/profile_form.py,sha256=6c2d55b7c0a1712796eebc8f05bb0597e955cf1dd6f96d37a6f0dce7865a2554,5984
19
- llama_deploy/cli/textual/secrets_form.py,sha256=b46b0e5999cd7c92433b91dcfb560bb5f8de9e5c4e96abbad63ee68544c387be,7082
20
- llama_deploy/cli/textual/styles.tcss,sha256=536cec7627d2a16dd03bf25bb9b6e4d53f1e0d18272b07ec0dc3bf76b0a7c2e0,3056
21
- llamactl-0.3.0a6.dist-info/WHEEL,sha256=66530aef82d5020ef5af27ae0123c71abb9261377c5bc519376c671346b12918,79
22
- llamactl-0.3.0a6.dist-info/entry_points.txt,sha256=b67e1eb64305058751a651a80f2d2268b5f7046732268421e796f64d4697f83c,52
23
- llamactl-0.3.0a6.dist-info/METADATA,sha256=6bf285b99000994431ec30eb1d6a97881e6fbd2a85785ca1d36dcb25156a4abc,3166
24
- llamactl-0.3.0a6.dist-info/RECORD,,