llama-deploy-appserver 0.2.7a1__py3-none-any.whl → 0.3.0a1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- llama_deploy/appserver/__main__.py +0 -4
- llama_deploy/appserver/app.py +105 -25
- llama_deploy/appserver/bootstrap.py +76 -24
- llama_deploy/appserver/deployment.py +7 -421
- llama_deploy/appserver/deployment_config_parser.py +35 -59
- llama_deploy/appserver/routers/__init__.py +4 -3
- llama_deploy/appserver/routers/deployments.py +162 -385
- llama_deploy/appserver/routers/status.py +4 -31
- llama_deploy/appserver/routers/ui_proxy.py +213 -0
- llama_deploy/appserver/settings.py +57 -55
- llama_deploy/appserver/types.py +0 -3
- llama_deploy/appserver/workflow_loader.py +383 -0
- {llama_deploy_appserver-0.2.7a1.dist-info → llama_deploy_appserver-0.3.0a1.dist-info}/METADATA +3 -6
- llama_deploy_appserver-0.3.0a1.dist-info/RECORD +17 -0
- {llama_deploy_appserver-0.2.7a1.dist-info → llama_deploy_appserver-0.3.0a1.dist-info}/WHEEL +1 -1
- llama_deploy/appserver/client/__init__.py +0 -3
- llama_deploy/appserver/client/base.py +0 -30
- llama_deploy/appserver/client/client.py +0 -49
- llama_deploy/appserver/client/models/__init__.py +0 -4
- llama_deploy/appserver/client/models/apiserver.py +0 -356
- llama_deploy/appserver/client/models/model.py +0 -82
- llama_deploy/appserver/run_autodeploy.py +0 -141
- llama_deploy/appserver/server.py +0 -60
- llama_deploy/appserver/source_managers/__init__.py +0 -5
- llama_deploy/appserver/source_managers/base.py +0 -33
- llama_deploy/appserver/source_managers/git.py +0 -48
- llama_deploy/appserver/source_managers/local.py +0 -51
- llama_deploy/appserver/tracing.py +0 -237
- llama_deploy_appserver-0.2.7a1.dist-info/RECORD +0 -28
@@ -1,433 +1,210 @@
|
|
1
1
|
import asyncio
|
2
2
|
import json
|
3
|
-
import
|
4
|
-
from typing import Annotated, AsyncGenerator, List, Optional
|
3
|
+
from typing import AsyncGenerator
|
5
4
|
|
6
|
-
import httpx
|
7
|
-
import websockets
|
8
5
|
from fastapi import (
|
9
6
|
APIRouter,
|
10
|
-
Depends,
|
11
|
-
File,
|
12
7
|
HTTPException,
|
13
|
-
Request,
|
14
|
-
UploadFile,
|
15
|
-
WebSocket,
|
16
8
|
)
|
17
|
-
from fastapi.responses import JSONResponse, StreamingResponse
|
9
|
+
from fastapi.responses import JSONResponse, RedirectResponse, StreamingResponse
|
18
10
|
from llama_deploy.appserver.deployment import Deployment
|
19
|
-
from llama_deploy.appserver.deployment_config_parser import DeploymentConfig
|
20
|
-
from llama_deploy.appserver.server import manager
|
21
11
|
from llama_deploy.appserver.types import (
|
22
|
-
DeploymentDefinition,
|
23
12
|
EventDefinition,
|
24
13
|
SessionDefinition,
|
25
14
|
TaskDefinition,
|
26
15
|
TaskResult,
|
27
16
|
generate_id,
|
28
17
|
)
|
29
|
-
from
|
18
|
+
from llama_deploy.appserver.workflow_loader import DEFAULT_SERVICE_ID
|
30
19
|
from workflows import Context
|
31
20
|
from workflows.context import JsonSerializer
|
32
21
|
from workflows.handler import WorkflowHandler
|
33
22
|
|
34
|
-
deployments_router = APIRouter(
|
35
|
-
prefix="/deployments",
|
36
|
-
)
|
37
|
-
logger = logging.getLogger(__name__)
|
38
|
-
|
39
|
-
|
40
|
-
def deployment(deployment_name: str) -> Deployment:
|
41
|
-
"""FastAPI dependency to retrieve a Deployment instance"""
|
42
|
-
deployment = manager.get_deployment(deployment_name)
|
43
|
-
if deployment is None:
|
44
|
-
raise HTTPException(status_code=404, detail="Deployment not found")
|
45
|
-
return deployment
|
46
|
-
|
47
|
-
|
48
|
-
@deployments_router.get("/")
|
49
|
-
async def read_deployments() -> list[DeploymentDefinition]:
|
50
|
-
"""Returns a list of active deployments."""
|
51
|
-
return [DeploymentDefinition(name=k) for k in manager._deployments.keys()]
|
52
|
-
|
53
|
-
|
54
|
-
@deployments_router.get("/{deployment_name}")
|
55
|
-
async def read_deployment(
|
56
|
-
deployment: Annotated[Deployment, Depends(deployment)],
|
57
|
-
) -> DeploymentDefinition:
|
58
|
-
"""Returns the details of a specific deployment."""
|
59
|
-
|
60
|
-
return DeploymentDefinition(name=deployment.name)
|
61
|
-
|
62
23
|
|
63
|
-
|
64
|
-
|
65
|
-
|
66
|
-
|
67
|
-
reload: bool = False,
|
68
|
-
local: bool = False,
|
69
|
-
) -> DeploymentDefinition:
|
70
|
-
"""Creates a new deployment by uploading a configuration file."""
|
71
|
-
config = DeploymentConfig.from_yaml_bytes(await config_file.read())
|
72
|
-
await manager.deploy(config, base_path, reload, local)
|
24
|
+
def create_base_router(name: str) -> APIRouter:
|
25
|
+
base_router = APIRouter(
|
26
|
+
prefix="",
|
27
|
+
)
|
73
28
|
|
74
|
-
|
29
|
+
@base_router.get("/", response_model=None)
|
30
|
+
async def root() -> RedirectResponse:
|
31
|
+
return RedirectResponse(f"/deployments/{name}/ui")
|
75
32
|
|
33
|
+
return base_router
|
76
34
|
|
77
|
-
@deployments_router.post("/{deployment_name}/tasks/run")
|
78
|
-
async def create_deployment_task(
|
79
|
-
deployment: Annotated[Deployment, Depends(deployment)],
|
80
|
-
task_definition: TaskDefinition,
|
81
|
-
session_id: str | None = None,
|
82
|
-
) -> JSONResponse:
|
83
|
-
"""Create a task for the deployment, wait for result and delete associated session."""
|
84
35
|
|
85
|
-
|
86
|
-
|
87
|
-
|
88
|
-
|
89
|
-
|
90
|
-
|
36
|
+
def create_deployments_router(
|
37
|
+
name: str, deployment: Deployment, serve_static: bool = False
|
38
|
+
) -> APIRouter:
|
39
|
+
deployments_router = APIRouter(
|
40
|
+
prefix="/deployments",
|
41
|
+
)
|
91
42
|
|
92
|
-
|
93
|
-
|
94
|
-
|
95
|
-
|
96
|
-
|
43
|
+
@deployments_router.post(f"/{name}/tasks/run")
|
44
|
+
async def create_deployment_task(
|
45
|
+
task_definition: TaskDefinition,
|
46
|
+
session_id: str | None = None,
|
47
|
+
) -> JSONResponse:
|
48
|
+
"""Create a task for the deployment, wait for result and delete associated session."""
|
49
|
+
|
50
|
+
service_id = task_definition.service_id or DEFAULT_SERVICE_ID
|
51
|
+
|
52
|
+
if service_id not in deployment.service_names:
|
53
|
+
raise HTTPException(
|
54
|
+
status_code=404,
|
55
|
+
detail=(
|
56
|
+
"There is no default service for this deployment. service_id is required"
|
57
|
+
if not task_definition.service_id
|
58
|
+
else f"Service '{service_id}' not found in deployment 'deployment_name'"
|
59
|
+
),
|
60
|
+
)
|
97
61
|
|
98
|
-
|
99
|
-
|
100
|
-
|
101
|
-
)
|
102
|
-
return JSONResponse(result)
|
103
|
-
|
104
|
-
|
105
|
-
@deployments_router.post("/{deployment_name}/tasks/create")
|
106
|
-
async def create_deployment_task_nowait(
|
107
|
-
deployment: Annotated[Deployment, Depends(deployment)],
|
108
|
-
task_definition: TaskDefinition,
|
109
|
-
session_id: str | None = None,
|
110
|
-
) -> TaskDefinition:
|
111
|
-
"""Create a task for the deployment but don't wait for result."""
|
112
|
-
service_id = task_definition.service_id or deployment.default_service
|
113
|
-
if service_id is None:
|
114
|
-
raise HTTPException(
|
115
|
-
status_code=400,
|
116
|
-
detail="Service is None and deployment has no default service",
|
62
|
+
run_kwargs = json.loads(task_definition.input) if task_definition.input else {}
|
63
|
+
result = await deployment.run_workflow(
|
64
|
+
service_id=service_id, session_id=session_id, **run_kwargs
|
117
65
|
)
|
66
|
+
return JSONResponse(result)
|
67
|
+
|
68
|
+
@deployments_router.post(f"/{name}/tasks/create")
|
69
|
+
async def create_deployment_task_nowait(
|
70
|
+
task_definition: TaskDefinition,
|
71
|
+
session_id: str | None = None,
|
72
|
+
) -> TaskDefinition:
|
73
|
+
"""Create a task for the deployment but don't wait for result."""
|
74
|
+
service_id = task_definition.service_id or DEFAULT_SERVICE_ID
|
75
|
+
if service_id not in deployment.service_names:
|
76
|
+
raise HTTPException(
|
77
|
+
status_code=404,
|
78
|
+
detail=(
|
79
|
+
"There is no default service for this deployment. service_id is required"
|
80
|
+
if not task_definition.service_id
|
81
|
+
else f"Service '{service_id}' not found in deployment 'deployment_name'"
|
82
|
+
),
|
83
|
+
)
|
118
84
|
|
119
|
-
|
120
|
-
|
121
|
-
|
122
|
-
detail=f"Service '{task_definition.service_id}' not found in deployment 'deployment_name'",
|
85
|
+
run_kwargs = json.loads(task_definition.input) if task_definition.input else {}
|
86
|
+
handler_id, session_id = deployment.run_workflow_no_wait(
|
87
|
+
service_id=service_id, session_id=session_id, **run_kwargs
|
123
88
|
)
|
124
89
|
|
125
|
-
|
126
|
-
|
127
|
-
service_id=service_id, session_id=session_id, **run_kwargs
|
128
|
-
)
|
129
|
-
|
130
|
-
task_definition.session_id = session_id
|
131
|
-
task_definition.task_id = handler_id
|
90
|
+
task_definition.session_id = session_id
|
91
|
+
task_definition.task_id = handler_id
|
132
92
|
|
133
|
-
|
93
|
+
return task_definition
|
134
94
|
|
135
|
-
|
136
|
-
|
137
|
-
|
138
|
-
|
139
|
-
|
140
|
-
|
141
|
-
|
142
|
-
|
143
|
-
"""Send a human response event to a service for a specific task and session."""
|
144
|
-
ctx = deployment._contexts[session_id]
|
145
|
-
serializer = JsonSerializer()
|
146
|
-
event = serializer.deserialize(event_def.event_obj_str)
|
147
|
-
ctx.send_event(event)
|
148
|
-
|
149
|
-
return event_def
|
150
|
-
|
151
|
-
|
152
|
-
@deployments_router.get("/{deployment_name}/tasks/{task_id}/events")
|
153
|
-
async def get_events(
|
154
|
-
deployment: Annotated[Deployment, Depends(deployment)],
|
155
|
-
session_id: str,
|
156
|
-
task_id: str,
|
157
|
-
raw_event: bool = False,
|
158
|
-
) -> StreamingResponse:
|
159
|
-
"""
|
160
|
-
Get the stream of events from a given task and session.
|
161
|
-
|
162
|
-
Args:
|
163
|
-
raw_event (bool, default=False): Whether to return the raw event object
|
164
|
-
or just the event data.
|
165
|
-
"""
|
166
|
-
|
167
|
-
async def event_stream(handler: WorkflowHandler) -> AsyncGenerator[str, None]:
|
95
|
+
@deployments_router.post(f"/{name}/tasks/{{task_id}}/events")
|
96
|
+
async def send_event(
|
97
|
+
task_id: str,
|
98
|
+
session_id: str,
|
99
|
+
event_def: EventDefinition,
|
100
|
+
) -> EventDefinition:
|
101
|
+
"""Send a human response event to a service for a specific task and session."""
|
102
|
+
ctx = deployment._contexts[session_id]
|
168
103
|
serializer = JsonSerializer()
|
169
|
-
|
170
|
-
|
171
|
-
|
172
|
-
|
173
|
-
|
174
|
-
|
175
|
-
|
176
|
-
|
177
|
-
|
178
|
-
|
179
|
-
|
180
|
-
|
181
|
-
|
182
|
-
|
183
|
-
|
184
|
-
|
185
|
-
|
186
|
-
|
187
|
-
|
188
|
-
|
104
|
+
event = serializer.deserialize(event_def.event_obj_str)
|
105
|
+
ctx.send_event(event)
|
106
|
+
|
107
|
+
return event_def
|
108
|
+
|
109
|
+
@deployments_router.get(f"/{name}/tasks/{{task_id}}/events")
|
110
|
+
async def get_events(
|
111
|
+
session_id: str,
|
112
|
+
task_id: str,
|
113
|
+
raw_event: bool = False,
|
114
|
+
) -> StreamingResponse:
|
115
|
+
"""
|
116
|
+
Get the stream of events from a given task and session.
|
117
|
+
|
118
|
+
Args:
|
119
|
+
raw_event (bool, default=False): Whether to return the raw event object
|
120
|
+
or just the event data.
|
121
|
+
"""
|
122
|
+
|
123
|
+
async def event_stream(handler: WorkflowHandler) -> AsyncGenerator[str, None]:
|
124
|
+
serializer = JsonSerializer()
|
125
|
+
# this will hang indefinitely if done and queue is empty. Bail
|
126
|
+
if (
|
127
|
+
handler.is_done()
|
128
|
+
and handler.ctx is not None
|
129
|
+
and handler.ctx.streaming_queue.empty()
|
130
|
+
):
|
131
|
+
return
|
132
|
+
async for event in handler.stream_events():
|
133
|
+
data = json.loads(serializer.serialize(event))
|
134
|
+
if raw_event:
|
135
|
+
yield json.dumps(data) + "\n"
|
136
|
+
else:
|
137
|
+
yield json.dumps(data.get("value")) + "\n"
|
138
|
+
await asyncio.sleep(0.01)
|
139
|
+
await handler
|
189
140
|
|
190
|
-
|
191
|
-
|
192
|
-
|
193
|
-
deployment: Annotated[Deployment, Depends(deployment)],
|
194
|
-
session_id: str,
|
195
|
-
task_id: str,
|
196
|
-
) -> TaskResult | None:
|
197
|
-
"""Get the task result associated with a task and session."""
|
198
|
-
|
199
|
-
handler = deployment._handlers[task_id]
|
200
|
-
return TaskResult(task_id=task_id, history=[], result=await handler)
|
201
|
-
|
202
|
-
|
203
|
-
@deployments_router.get("/{deployment_name}/tasks")
|
204
|
-
async def get_tasks(
|
205
|
-
deployment: Annotated[Deployment, Depends(deployment)],
|
206
|
-
) -> list[TaskDefinition]:
|
207
|
-
"""Get all the tasks from all the sessions in a given deployment."""
|
208
|
-
|
209
|
-
tasks: list[TaskDefinition] = []
|
210
|
-
for task_id, handler in deployment._handlers.items():
|
211
|
-
if handler.is_done():
|
212
|
-
continue
|
213
|
-
tasks.append(
|
214
|
-
TaskDefinition(
|
215
|
-
task_id=task_id,
|
216
|
-
input=deployment._handler_inputs[task_id],
|
217
|
-
)
|
141
|
+
return StreamingResponse(
|
142
|
+
event_stream(deployment._handlers[task_id]),
|
143
|
+
media_type="application/x-ndjson",
|
218
144
|
)
|
219
145
|
|
220
|
-
|
221
|
-
|
222
|
-
|
223
|
-
|
224
|
-
|
225
|
-
|
226
|
-
|
227
|
-
|
228
|
-
|
229
|
-
|
230
|
-
|
231
|
-
|
232
|
-
|
233
|
-
|
234
|
-
|
235
|
-
|
236
|
-
|
237
|
-
|
238
|
-
|
239
|
-
|
240
|
-
|
241
|
-
|
242
|
-
|
243
|
-
deployment: Annotated[Deployment, Depends(deployment)],
|
244
|
-
) -> SessionDefinition:
|
245
|
-
"""Create a new session for a deployment."""
|
246
|
-
|
247
|
-
workflow = deployment._workflow_services[deployment.default_service]
|
248
|
-
session_id = generate_id()
|
249
|
-
deployment._contexts[session_id] = Context(workflow)
|
250
|
-
|
251
|
-
return SessionDefinition(session_id=session_id)
|
252
|
-
|
253
|
-
|
254
|
-
@deployments_router.post("/{deployment_name}/sessions/delete")
|
255
|
-
async def delete_session(
|
256
|
-
deployment: Annotated[Deployment, Depends(deployment)], session_id: str
|
257
|
-
) -> None:
|
258
|
-
"""Get the active sessions in a deployment and service."""
|
259
|
-
|
260
|
-
deployment._contexts.pop(session_id)
|
261
|
-
|
262
|
-
|
263
|
-
async def _ws_proxy(ws: WebSocket, upstream_url: str) -> None:
|
264
|
-
"""Proxy WebSocket connection to upstream server."""
|
265
|
-
await ws.accept()
|
266
|
-
|
267
|
-
# Forward most headers except WebSocket-specific ones
|
268
|
-
header_blacklist = {
|
269
|
-
"host",
|
270
|
-
"connection",
|
271
|
-
"upgrade",
|
272
|
-
"sec-websocket-key",
|
273
|
-
"sec-websocket-version",
|
274
|
-
"sec-websocket-extensions",
|
275
|
-
}
|
276
|
-
hdrs = [(k, v) for k, v in ws.headers.items() if k.lower() not in header_blacklist]
|
277
|
-
|
278
|
-
try:
|
279
|
-
# Parse subprotocols if present
|
280
|
-
subprotocols: Optional[List[websockets.Subprotocol]] = None
|
281
|
-
if "sec-websocket-protocol" in ws.headers:
|
282
|
-
# Parse comma-separated subprotocols
|
283
|
-
subprotocols = [
|
284
|
-
websockets.Subprotocol(p.strip())
|
285
|
-
for p in ws.headers["sec-websocket-protocol"].split(",")
|
286
|
-
]
|
287
|
-
|
288
|
-
# Open upstream WebSocket connection
|
289
|
-
async with websockets.connect(
|
290
|
-
upstream_url,
|
291
|
-
additional_headers=hdrs,
|
292
|
-
subprotocols=subprotocols,
|
293
|
-
open_timeout=None,
|
294
|
-
ping_interval=None,
|
295
|
-
) as upstream:
|
296
|
-
|
297
|
-
async def client_to_upstream() -> None:
|
298
|
-
try:
|
299
|
-
while True:
|
300
|
-
msg = await ws.receive()
|
301
|
-
if msg["type"] == "websocket.receive":
|
302
|
-
if "text" in msg:
|
303
|
-
await upstream.send(msg["text"])
|
304
|
-
elif "bytes" in msg:
|
305
|
-
await upstream.send(msg["bytes"])
|
306
|
-
elif msg["type"] == "websocket.disconnect":
|
307
|
-
break
|
308
|
-
except Exception as e:
|
309
|
-
logger.debug(f"Client to upstream connection ended: {e}")
|
310
|
-
|
311
|
-
async def upstream_to_client() -> None:
|
312
|
-
try:
|
313
|
-
async for message in upstream:
|
314
|
-
if isinstance(message, str):
|
315
|
-
await ws.send_text(message)
|
316
|
-
else:
|
317
|
-
await ws.send_bytes(message)
|
318
|
-
except Exception as e:
|
319
|
-
logger.debug(f"Upstream to client connection ended: {e}")
|
320
|
-
|
321
|
-
# Pump both directions concurrently
|
322
|
-
await asyncio.gather(
|
323
|
-
client_to_upstream(), upstream_to_client(), return_exceptions=True
|
146
|
+
@deployments_router.get(f"/{name}/tasks/{{task_id}}/results")
|
147
|
+
async def get_task_result(
|
148
|
+
session_id: str,
|
149
|
+
task_id: str,
|
150
|
+
) -> TaskResult | None:
|
151
|
+
"""Get the task result associated with a task and session."""
|
152
|
+
|
153
|
+
handler = deployment._handlers[task_id]
|
154
|
+
return TaskResult(task_id=task_id, history=[], result=await handler)
|
155
|
+
|
156
|
+
@deployments_router.get(f"/{name}/tasks")
|
157
|
+
async def get_tasks() -> list[TaskDefinition]:
|
158
|
+
"""Get all the tasks from all the sessions in a given deployment."""
|
159
|
+
|
160
|
+
tasks: list[TaskDefinition] = []
|
161
|
+
for task_id, handler in deployment._handlers.items():
|
162
|
+
if handler.is_done():
|
163
|
+
continue
|
164
|
+
tasks.append(
|
165
|
+
TaskDefinition(
|
166
|
+
task_id=task_id,
|
167
|
+
input=deployment._handler_inputs[task_id],
|
168
|
+
)
|
324
169
|
)
|
325
170
|
|
326
|
-
|
327
|
-
logger.error(f"WebSocket proxy error: {e}")
|
328
|
-
finally:
|
329
|
-
try:
|
330
|
-
await ws.close()
|
331
|
-
except Exception as e:
|
332
|
-
logger.debug(f"Error closing client connection: {e}")
|
333
|
-
|
334
|
-
|
335
|
-
@deployments_router.websocket("/{deployment_name}/ui/{path:path}")
|
336
|
-
@deployments_router.websocket("/{deployment_name}/ui")
|
337
|
-
async def websocket_proxy(
|
338
|
-
websocket: WebSocket,
|
339
|
-
deployment: Annotated[Deployment, Depends(deployment)],
|
340
|
-
path: str | None = None,
|
341
|
-
) -> None:
|
342
|
-
if deployment._config.ui is None:
|
343
|
-
raise HTTPException(status_code=404, detail="Deployment has no ui configured")
|
171
|
+
return tasks
|
344
172
|
|
345
|
-
|
346
|
-
|
347
|
-
|
173
|
+
@deployments_router.get(f"/{name}/sessions")
|
174
|
+
async def get_sessions() -> list[SessionDefinition]:
|
175
|
+
"""Get the active sessions in a deployment and service."""
|
348
176
|
|
349
|
-
|
350
|
-
upstream_url = f"ws://localhost:{deployment._config.ui.port}{upstream_path}"
|
351
|
-
if websocket.url.query:
|
352
|
-
upstream_url += f"?{websocket.url.query}"
|
177
|
+
return [SessionDefinition(session_id=k) for k in deployment._contexts.keys()]
|
353
178
|
|
354
|
-
|
179
|
+
@deployments_router.get(f"/{name}/sessions/{{session_id}}")
|
180
|
+
async def get_session(
|
181
|
+
session_id: str,
|
182
|
+
) -> SessionDefinition:
|
183
|
+
"""Get the definition of a session by ID."""
|
355
184
|
|
356
|
-
|
185
|
+
return SessionDefinition(session_id=session_id)
|
357
186
|
|
187
|
+
@deployments_router.post(f"/{name}/sessions/create")
|
188
|
+
async def create_session() -> SessionDefinition:
|
189
|
+
"""Create a new session for a deployment."""
|
358
190
|
|
359
|
-
|
360
|
-
|
361
|
-
|
362
|
-
|
363
|
-
|
364
|
-
|
365
|
-
|
366
|
-
)
|
367
|
-
async def proxy(
|
368
|
-
request: Request,
|
369
|
-
deployment: Annotated[Deployment, Depends(deployment)],
|
370
|
-
path: str | None = None,
|
371
|
-
) -> StreamingResponse:
|
372
|
-
if deployment._config.ui is None:
|
373
|
-
raise HTTPException(status_code=404, detail="Deployment has no ui configured")
|
374
|
-
|
375
|
-
# Build the upstream URL using FastAPI's extracted path parameter
|
376
|
-
slash_path = f"/{path}" if path else ""
|
377
|
-
upstream_path = f"/deployments/{deployment.name}/ui{slash_path}"
|
378
|
-
|
379
|
-
upstream_url = httpx.URL(
|
380
|
-
f"http://localhost:{deployment._config.ui.port}{upstream_path}"
|
381
|
-
).copy_with(params=request.query_params)
|
382
|
-
|
383
|
-
# Debug logging
|
384
|
-
logger.debug(f"Proxying {request.method} {request.url} -> {upstream_url}")
|
385
|
-
|
386
|
-
# Strip hop-by-hop headers + host
|
387
|
-
hop_by_hop = {
|
388
|
-
"connection",
|
389
|
-
"keep-alive",
|
390
|
-
"proxy-authenticate",
|
391
|
-
"proxy-authorization",
|
392
|
-
"te", # codespell:ignore
|
393
|
-
"trailers",
|
394
|
-
"transfer-encoding",
|
395
|
-
"upgrade",
|
396
|
-
"host",
|
397
|
-
}
|
398
|
-
headers = {k: v for k, v in request.headers.items() if k.lower() not in hop_by_hop}
|
399
|
-
|
400
|
-
try:
|
401
|
-
client = httpx.AsyncClient(timeout=None)
|
402
|
-
|
403
|
-
req = client.build_request(
|
404
|
-
request.method,
|
405
|
-
upstream_url,
|
406
|
-
headers=headers,
|
407
|
-
content=request.stream(), # stream uploads
|
408
|
-
)
|
409
|
-
upstream = await client.send(req, stream=True)
|
191
|
+
workflow = deployment.default_service
|
192
|
+
if workflow is None:
|
193
|
+
raise HTTPException(
|
194
|
+
status_code=400,
|
195
|
+
detail="There is no default service for this deployment",
|
196
|
+
)
|
197
|
+
session_id = generate_id()
|
198
|
+
deployment._contexts[session_id] = Context(workflow)
|
410
199
|
|
411
|
-
|
412
|
-
k: v for k, v in upstream.headers.items() if k.lower() not in hop_by_hop
|
413
|
-
}
|
200
|
+
return SessionDefinition(session_id=session_id)
|
414
201
|
|
415
|
-
|
416
|
-
|
417
|
-
|
418
|
-
|
202
|
+
@deployments_router.post(f"/{name}/sessions/delete")
|
203
|
+
async def delete_session(
|
204
|
+
session_id: str,
|
205
|
+
) -> None:
|
206
|
+
"""Get the active sessions in a deployment and service."""
|
419
207
|
|
420
|
-
|
421
|
-
upstream.aiter_raw(), # stream downloads
|
422
|
-
status_code=upstream.status_code,
|
423
|
-
headers=resp_headers,
|
424
|
-
background=BackgroundTask(cleanup), # tidy up when finished
|
425
|
-
)
|
208
|
+
deployment._contexts.pop(session_id)
|
426
209
|
|
427
|
-
|
428
|
-
raise HTTPException(status_code=502, detail="Upstream server unavailable")
|
429
|
-
except httpx.TimeoutException:
|
430
|
-
raise HTTPException(status_code=504, detail="Upstream server timeout")
|
431
|
-
except Exception as e:
|
432
|
-
logger.error(f"Proxy error: {e}")
|
433
|
-
raise HTTPException(status_code=502, detail="Proxy error")
|
210
|
+
return deployments_router
|
@@ -1,40 +1,13 @@
|
|
1
|
-
import httpx
|
2
1
|
from fastapi import APIRouter
|
3
|
-
from fastapi.exceptions import HTTPException
|
4
|
-
from fastapi.responses import PlainTextResponse
|
5
|
-
from llama_deploy.appserver.server import manager
|
6
|
-
from llama_deploy.appserver.settings import settings
|
7
2
|
from llama_deploy.appserver.types import Status, StatusEnum
|
8
3
|
|
9
|
-
|
10
|
-
prefix="/
|
4
|
+
health_router = APIRouter(
|
5
|
+
prefix="/health",
|
11
6
|
)
|
12
7
|
|
13
8
|
|
14
|
-
@
|
15
|
-
async def
|
9
|
+
@health_router.get("")
|
10
|
+
async def health() -> Status:
|
16
11
|
return Status(
|
17
12
|
status=StatusEnum.HEALTHY,
|
18
|
-
max_deployments=manager._max_deployments,
|
19
|
-
deployments=list(manager._deployments.keys()),
|
20
|
-
status_message="",
|
21
13
|
)
|
22
|
-
|
23
|
-
|
24
|
-
@status_router.get("/metrics")
|
25
|
-
async def metrics() -> PlainTextResponse:
|
26
|
-
"""Proxies the Prometheus metrics endpoint through the API Server.
|
27
|
-
|
28
|
-
This endpoint is mostly used in serverless environments where the LlamaDeploy
|
29
|
-
container cannot expose more than one port (e.g. Knative, Google Cloud Run).
|
30
|
-
If Prometheus is not enabled, this endpoint returns an empty HTTP-204 response.
|
31
|
-
"""
|
32
|
-
if not settings.prometheus_enabled:
|
33
|
-
return PlainTextResponse(status_code=204)
|
34
|
-
|
35
|
-
try:
|
36
|
-
async with httpx.AsyncClient() as client:
|
37
|
-
response = await client.get(f"http://127.0.0.1:{settings.prometheus_port}/")
|
38
|
-
return PlainTextResponse(content=response.text)
|
39
|
-
except httpx.RequestError as exc:
|
40
|
-
raise HTTPException(status_code=500, detail=str(exc))
|