llama-stack 0.4.3__py3-none-any.whl → 0.4.4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- llama_stack/distributions/dell/doc_template.md +209 -0
- llama_stack/distributions/meta-reference-gpu/doc_template.md +119 -0
- llama_stack/distributions/nvidia/doc_template.md +170 -0
- llama_stack/distributions/oci/doc_template.md +140 -0
- llama_stack/models/llama/llama3/dog.jpg +0 -0
- llama_stack/models/llama/llama3/pasta.jpeg +0 -0
- llama_stack/models/llama/resources/dog.jpg +0 -0
- llama_stack/models/llama/resources/pasta.jpeg +0 -0
- llama_stack/models/llama/resources/small_dog.jpg +0 -0
- llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py +136 -11
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl/LocalInference.h +9 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl/LocalInference.swift +189 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl/Parsing.swift +238 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl/PromptTemplate.swift +12 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl/SystemPrompts.swift +89 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.pbxproj +550 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.xcworkspace/contents.xcworkspacedata +7 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.xcworkspace/xcshareddata/IDEWorkspaceChecks.plist +8 -0
- llama_stack/providers/remote/datasetio/nvidia/README.md +74 -0
- llama_stack/providers/remote/eval/nvidia/README.md +134 -0
- llama_stack/providers/remote/files/s3/README.md +266 -0
- llama_stack/providers/remote/inference/nvidia/NVIDIA.md +203 -0
- llama_stack/providers/remote/post_training/nvidia/README.md +151 -0
- llama_stack/providers/remote/safety/nvidia/README.md +78 -0
- llama_stack/providers/utils/responses/responses_store.py +34 -0
- {llama_stack-0.4.3.dist-info → llama_stack-0.4.4.dist-info}/METADATA +2 -2
- {llama_stack-0.4.3.dist-info → llama_stack-0.4.4.dist-info}/RECORD +31 -142
- llama_stack-0.4.4.dist-info/top_level.txt +1 -0
- llama_stack-0.4.3.dist-info/top_level.txt +0 -2
- llama_stack_api/__init__.py +0 -945
- llama_stack_api/admin/__init__.py +0 -45
- llama_stack_api/admin/api.py +0 -72
- llama_stack_api/admin/fastapi_routes.py +0 -117
- llama_stack_api/admin/models.py +0 -113
- llama_stack_api/agents.py +0 -173
- llama_stack_api/batches/__init__.py +0 -40
- llama_stack_api/batches/api.py +0 -53
- llama_stack_api/batches/fastapi_routes.py +0 -113
- llama_stack_api/batches/models.py +0 -78
- llama_stack_api/benchmarks/__init__.py +0 -43
- llama_stack_api/benchmarks/api.py +0 -39
- llama_stack_api/benchmarks/fastapi_routes.py +0 -109
- llama_stack_api/benchmarks/models.py +0 -109
- llama_stack_api/common/__init__.py +0 -5
- llama_stack_api/common/content_types.py +0 -101
- llama_stack_api/common/errors.py +0 -95
- llama_stack_api/common/job_types.py +0 -38
- llama_stack_api/common/responses.py +0 -77
- llama_stack_api/common/training_types.py +0 -47
- llama_stack_api/common/type_system.py +0 -146
- llama_stack_api/connectors.py +0 -146
- llama_stack_api/conversations.py +0 -270
- llama_stack_api/datasetio.py +0 -55
- llama_stack_api/datasets/__init__.py +0 -61
- llama_stack_api/datasets/api.py +0 -35
- llama_stack_api/datasets/fastapi_routes.py +0 -104
- llama_stack_api/datasets/models.py +0 -152
- llama_stack_api/datatypes.py +0 -373
- llama_stack_api/eval.py +0 -137
- llama_stack_api/file_processors/__init__.py +0 -27
- llama_stack_api/file_processors/api.py +0 -64
- llama_stack_api/file_processors/fastapi_routes.py +0 -78
- llama_stack_api/file_processors/models.py +0 -42
- llama_stack_api/files/__init__.py +0 -35
- llama_stack_api/files/api.py +0 -51
- llama_stack_api/files/fastapi_routes.py +0 -124
- llama_stack_api/files/models.py +0 -107
- llama_stack_api/inference.py +0 -1169
- llama_stack_api/inspect_api/__init__.py +0 -37
- llama_stack_api/inspect_api/api.py +0 -25
- llama_stack_api/inspect_api/fastapi_routes.py +0 -76
- llama_stack_api/inspect_api/models.py +0 -28
- llama_stack_api/internal/__init__.py +0 -9
- llama_stack_api/internal/kvstore.py +0 -28
- llama_stack_api/internal/sqlstore.py +0 -81
- llama_stack_api/llama_stack_api/__init__.py +0 -945
- llama_stack_api/llama_stack_api/admin/__init__.py +0 -45
- llama_stack_api/llama_stack_api/admin/api.py +0 -72
- llama_stack_api/llama_stack_api/admin/fastapi_routes.py +0 -117
- llama_stack_api/llama_stack_api/admin/models.py +0 -113
- llama_stack_api/llama_stack_api/agents.py +0 -173
- llama_stack_api/llama_stack_api/batches/__init__.py +0 -40
- llama_stack_api/llama_stack_api/batches/api.py +0 -53
- llama_stack_api/llama_stack_api/batches/fastapi_routes.py +0 -113
- llama_stack_api/llama_stack_api/batches/models.py +0 -78
- llama_stack_api/llama_stack_api/benchmarks/__init__.py +0 -43
- llama_stack_api/llama_stack_api/benchmarks/api.py +0 -39
- llama_stack_api/llama_stack_api/benchmarks/fastapi_routes.py +0 -109
- llama_stack_api/llama_stack_api/benchmarks/models.py +0 -109
- llama_stack_api/llama_stack_api/common/__init__.py +0 -5
- llama_stack_api/llama_stack_api/common/content_types.py +0 -101
- llama_stack_api/llama_stack_api/common/errors.py +0 -95
- llama_stack_api/llama_stack_api/common/job_types.py +0 -38
- llama_stack_api/llama_stack_api/common/responses.py +0 -77
- llama_stack_api/llama_stack_api/common/training_types.py +0 -47
- llama_stack_api/llama_stack_api/common/type_system.py +0 -146
- llama_stack_api/llama_stack_api/connectors.py +0 -146
- llama_stack_api/llama_stack_api/conversations.py +0 -270
- llama_stack_api/llama_stack_api/datasetio.py +0 -55
- llama_stack_api/llama_stack_api/datasets/__init__.py +0 -61
- llama_stack_api/llama_stack_api/datasets/api.py +0 -35
- llama_stack_api/llama_stack_api/datasets/fastapi_routes.py +0 -104
- llama_stack_api/llama_stack_api/datasets/models.py +0 -152
- llama_stack_api/llama_stack_api/datatypes.py +0 -373
- llama_stack_api/llama_stack_api/eval.py +0 -137
- llama_stack_api/llama_stack_api/file_processors/__init__.py +0 -27
- llama_stack_api/llama_stack_api/file_processors/api.py +0 -64
- llama_stack_api/llama_stack_api/file_processors/fastapi_routes.py +0 -78
- llama_stack_api/llama_stack_api/file_processors/models.py +0 -42
- llama_stack_api/llama_stack_api/files/__init__.py +0 -35
- llama_stack_api/llama_stack_api/files/api.py +0 -51
- llama_stack_api/llama_stack_api/files/fastapi_routes.py +0 -124
- llama_stack_api/llama_stack_api/files/models.py +0 -107
- llama_stack_api/llama_stack_api/inference.py +0 -1169
- llama_stack_api/llama_stack_api/inspect_api/__init__.py +0 -37
- llama_stack_api/llama_stack_api/inspect_api/api.py +0 -25
- llama_stack_api/llama_stack_api/inspect_api/fastapi_routes.py +0 -76
- llama_stack_api/llama_stack_api/inspect_api/models.py +0 -28
- llama_stack_api/llama_stack_api/internal/__init__.py +0 -9
- llama_stack_api/llama_stack_api/internal/kvstore.py +0 -28
- llama_stack_api/llama_stack_api/internal/sqlstore.py +0 -81
- llama_stack_api/llama_stack_api/models.py +0 -171
- llama_stack_api/llama_stack_api/openai_responses.py +0 -1468
- llama_stack_api/llama_stack_api/post_training.py +0 -370
- llama_stack_api/llama_stack_api/prompts.py +0 -203
- llama_stack_api/llama_stack_api/providers/__init__.py +0 -33
- llama_stack_api/llama_stack_api/providers/api.py +0 -16
- llama_stack_api/llama_stack_api/providers/fastapi_routes.py +0 -57
- llama_stack_api/llama_stack_api/providers/models.py +0 -24
- llama_stack_api/llama_stack_api/py.typed +0 -0
- llama_stack_api/llama_stack_api/rag_tool.py +0 -168
- llama_stack_api/llama_stack_api/resource.py +0 -37
- llama_stack_api/llama_stack_api/router_utils.py +0 -160
- llama_stack_api/llama_stack_api/safety.py +0 -132
- llama_stack_api/llama_stack_api/schema_utils.py +0 -208
- llama_stack_api/llama_stack_api/scoring.py +0 -93
- llama_stack_api/llama_stack_api/scoring_functions.py +0 -211
- llama_stack_api/llama_stack_api/shields.py +0 -93
- llama_stack_api/llama_stack_api/tools.py +0 -226
- llama_stack_api/llama_stack_api/vector_io.py +0 -941
- llama_stack_api/llama_stack_api/vector_stores.py +0 -53
- llama_stack_api/llama_stack_api/version.py +0 -9
- llama_stack_api/models.py +0 -171
- llama_stack_api/openai_responses.py +0 -1468
- llama_stack_api/post_training.py +0 -370
- llama_stack_api/prompts.py +0 -203
- llama_stack_api/providers/__init__.py +0 -33
- llama_stack_api/providers/api.py +0 -16
- llama_stack_api/providers/fastapi_routes.py +0 -57
- llama_stack_api/providers/models.py +0 -24
- llama_stack_api/py.typed +0 -0
- llama_stack_api/rag_tool.py +0 -168
- llama_stack_api/resource.py +0 -37
- llama_stack_api/router_utils.py +0 -160
- llama_stack_api/safety.py +0 -132
- llama_stack_api/schema_utils.py +0 -208
- llama_stack_api/scoring.py +0 -93
- llama_stack_api/scoring_functions.py +0 -211
- llama_stack_api/shields.py +0 -93
- llama_stack_api/tools.py +0 -226
- llama_stack_api/vector_io.py +0 -941
- llama_stack_api/vector_stores.py +0 -53
- llama_stack_api/version.py +0 -9
- {llama_stack-0.4.3.dist-info → llama_stack-0.4.4.dist-info}/WHEEL +0 -0
- {llama_stack-0.4.3.dist-info → llama_stack-0.4.4.dist-info}/entry_points.txt +0 -0
- {llama_stack-0.4.3.dist-info → llama_stack-0.4.4.dist-info}/licenses/LICENSE +0 -0
|
@@ -0,0 +1,151 @@
|
|
|
1
|
+
# NVIDIA Post-Training Provider for LlamaStack
|
|
2
|
+
|
|
3
|
+
This provider enables fine-tuning of LLMs using NVIDIA's NeMo Customizer service.
|
|
4
|
+
|
|
5
|
+
## Features
|
|
6
|
+
|
|
7
|
+
- Supervised fine-tuning of Llama models
|
|
8
|
+
- LoRA fine-tuning support
|
|
9
|
+
- Job management and status tracking
|
|
10
|
+
|
|
11
|
+
## Getting Started
|
|
12
|
+
|
|
13
|
+
### Prerequisites
|
|
14
|
+
|
|
15
|
+
- LlamaStack with NVIDIA configuration
|
|
16
|
+
- Access to Hosted NVIDIA NeMo Customizer service
|
|
17
|
+
- Dataset registered in the Hosted NVIDIA NeMo Customizer service
|
|
18
|
+
- Base model downloaded and available in the Hosted NVIDIA NeMo Customizer service
|
|
19
|
+
|
|
20
|
+
### Setup
|
|
21
|
+
|
|
22
|
+
Build the NVIDIA environment:
|
|
23
|
+
|
|
24
|
+
```bash
|
|
25
|
+
uv pip install llama-stack-client
|
|
26
|
+
uv run llama stack list-deps nvidia | xargs -L1 uv pip install
|
|
27
|
+
```
|
|
28
|
+
|
|
29
|
+
### Basic Usage using the LlamaStack Python Client
|
|
30
|
+
|
|
31
|
+
### Create Customization Job
|
|
32
|
+
|
|
33
|
+
#### Initialize the client
|
|
34
|
+
|
|
35
|
+
```python
|
|
36
|
+
import os
|
|
37
|
+
|
|
38
|
+
os.environ["NVIDIA_API_KEY"] = "your-api-key"
|
|
39
|
+
os.environ["NVIDIA_CUSTOMIZER_URL"] = "http://nemo.test"
|
|
40
|
+
os.environ["NVIDIA_DATASET_NAMESPACE"] = "default"
|
|
41
|
+
os.environ["NVIDIA_PROJECT_ID"] = "test-project"
|
|
42
|
+
os.environ["NVIDIA_OUTPUT_MODEL_DIR"] = "test-example-model@v1"
|
|
43
|
+
|
|
44
|
+
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
|
45
|
+
|
|
46
|
+
client = LlamaStackAsLibraryClient("nvidia")
|
|
47
|
+
client.initialize()
|
|
48
|
+
```
|
|
49
|
+
|
|
50
|
+
#### Configure fine-tuning parameters
|
|
51
|
+
|
|
52
|
+
```python
|
|
53
|
+
from llama_stack_client.types.post_training_supervised_fine_tune_params import (
|
|
54
|
+
TrainingConfig,
|
|
55
|
+
TrainingConfigDataConfig,
|
|
56
|
+
TrainingConfigOptimizerConfig,
|
|
57
|
+
)
|
|
58
|
+
from llama_stack_client.types.algorithm_config_param import LoraFinetuningConfig
|
|
59
|
+
```
|
|
60
|
+
|
|
61
|
+
#### Set up LoRA configuration
|
|
62
|
+
|
|
63
|
+
```python
|
|
64
|
+
algorithm_config = LoraFinetuningConfig(type="LoRA", adapter_dim=16)
|
|
65
|
+
```
|
|
66
|
+
|
|
67
|
+
#### Configure training data
|
|
68
|
+
|
|
69
|
+
```python
|
|
70
|
+
data_config = TrainingConfigDataConfig(
|
|
71
|
+
dataset_id="your-dataset-id", # Use client.datasets.list() to see available datasets
|
|
72
|
+
batch_size=16,
|
|
73
|
+
)
|
|
74
|
+
```
|
|
75
|
+
|
|
76
|
+
#### Configure optimizer
|
|
77
|
+
|
|
78
|
+
```python
|
|
79
|
+
optimizer_config = TrainingConfigOptimizerConfig(
|
|
80
|
+
lr=0.0001,
|
|
81
|
+
)
|
|
82
|
+
```
|
|
83
|
+
|
|
84
|
+
#### Set up training configuration
|
|
85
|
+
|
|
86
|
+
```python
|
|
87
|
+
training_config = TrainingConfig(
|
|
88
|
+
n_epochs=2,
|
|
89
|
+
data_config=data_config,
|
|
90
|
+
optimizer_config=optimizer_config,
|
|
91
|
+
)
|
|
92
|
+
```
|
|
93
|
+
|
|
94
|
+
#### Start fine-tuning job
|
|
95
|
+
|
|
96
|
+
```python
|
|
97
|
+
training_job = client.post_training.supervised_fine_tune(
|
|
98
|
+
job_uuid="unique-job-id",
|
|
99
|
+
model="meta-llama/Llama-3.1-8B-Instruct",
|
|
100
|
+
checkpoint_dir="",
|
|
101
|
+
algorithm_config=algorithm_config,
|
|
102
|
+
training_config=training_config,
|
|
103
|
+
logger_config={},
|
|
104
|
+
hyperparam_search_config={},
|
|
105
|
+
)
|
|
106
|
+
```
|
|
107
|
+
|
|
108
|
+
### List all jobs
|
|
109
|
+
|
|
110
|
+
```python
|
|
111
|
+
jobs = client.post_training.job.list()
|
|
112
|
+
```
|
|
113
|
+
|
|
114
|
+
### Check job status
|
|
115
|
+
|
|
116
|
+
```python
|
|
117
|
+
job_status = client.post_training.job.status(job_uuid="your-job-id")
|
|
118
|
+
```
|
|
119
|
+
|
|
120
|
+
### Cancel a job
|
|
121
|
+
|
|
122
|
+
```python
|
|
123
|
+
client.post_training.job.cancel(job_uuid="your-job-id")
|
|
124
|
+
```
|
|
125
|
+
|
|
126
|
+
### Inference with the fine-tuned model
|
|
127
|
+
|
|
128
|
+
#### 1. Register the model
|
|
129
|
+
|
|
130
|
+
```python
|
|
131
|
+
from llama_stack_api.models import Model, ModelType
|
|
132
|
+
|
|
133
|
+
client.models.register(
|
|
134
|
+
model_id="test-example-model@v1",
|
|
135
|
+
provider_id="nvidia",
|
|
136
|
+
provider_model_id="test-example-model@v1",
|
|
137
|
+
model_type=ModelType.llm,
|
|
138
|
+
)
|
|
139
|
+
```
|
|
140
|
+
|
|
141
|
+
#### 2. Inference with the fine-tuned model
|
|
142
|
+
|
|
143
|
+
```python
|
|
144
|
+
response = client.completions.create(
|
|
145
|
+
prompt="Complete the sentence using one word: Roses are red, violets are ",
|
|
146
|
+
stream=False,
|
|
147
|
+
model="test-example-model@v1",
|
|
148
|
+
max_tokens=50,
|
|
149
|
+
)
|
|
150
|
+
print(response.choices[0].text)
|
|
151
|
+
```
|
|
@@ -0,0 +1,78 @@
|
|
|
1
|
+
# NVIDIA Safety Provider for LlamaStack
|
|
2
|
+
|
|
3
|
+
This provider enables safety checks and guardrails for LLM interactions using NVIDIA's NeMo Guardrails service.
|
|
4
|
+
|
|
5
|
+
## Features
|
|
6
|
+
|
|
7
|
+
- Run safety checks for messages
|
|
8
|
+
|
|
9
|
+
## Getting Started
|
|
10
|
+
|
|
11
|
+
### Prerequisites
|
|
12
|
+
|
|
13
|
+
- LlamaStack with NVIDIA configuration
|
|
14
|
+
- Access to NVIDIA NeMo Guardrails service
|
|
15
|
+
- NIM for model to use for safety check is deployed
|
|
16
|
+
|
|
17
|
+
### Setup
|
|
18
|
+
|
|
19
|
+
Build the NVIDIA environment:
|
|
20
|
+
|
|
21
|
+
```bash
|
|
22
|
+
uv pip install llama-stack-client
|
|
23
|
+
uv run llama stack list-deps nvidia | xargs -L1 uv pip install
|
|
24
|
+
```
|
|
25
|
+
|
|
26
|
+
### Basic Usage using the LlamaStack Python Client
|
|
27
|
+
|
|
28
|
+
#### Initialize the client
|
|
29
|
+
|
|
30
|
+
```python
|
|
31
|
+
import os
|
|
32
|
+
|
|
33
|
+
os.environ["NVIDIA_API_KEY"] = "your-api-key"
|
|
34
|
+
os.environ["NVIDIA_GUARDRAILS_URL"] = "http://guardrails.test"
|
|
35
|
+
|
|
36
|
+
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
|
37
|
+
|
|
38
|
+
client = LlamaStackAsLibraryClient("nvidia")
|
|
39
|
+
client.initialize()
|
|
40
|
+
```
|
|
41
|
+
|
|
42
|
+
#### Create a safety shield
|
|
43
|
+
|
|
44
|
+
```python
|
|
45
|
+
from llama_stack_api.safety import Shield
|
|
46
|
+
from llama_stack_api.inference import Message
|
|
47
|
+
|
|
48
|
+
# Create a safety shield
|
|
49
|
+
shield = Shield(
|
|
50
|
+
shield_id="your-shield-id",
|
|
51
|
+
provider_resource_id="safety-model-id", # The model to use for safety checks
|
|
52
|
+
description="Safety checks for content moderation",
|
|
53
|
+
)
|
|
54
|
+
|
|
55
|
+
# Register the shield
|
|
56
|
+
await client.safety.register_shield(shield)
|
|
57
|
+
```
|
|
58
|
+
|
|
59
|
+
#### Run safety checks
|
|
60
|
+
|
|
61
|
+
```python
|
|
62
|
+
# Messages to check
|
|
63
|
+
messages = [Message(role="user", content="Your message to check")]
|
|
64
|
+
|
|
65
|
+
# Run safety check
|
|
66
|
+
response = await client.safety.run_shield(
|
|
67
|
+
shield_id="your-shield-id",
|
|
68
|
+
messages=messages,
|
|
69
|
+
)
|
|
70
|
+
|
|
71
|
+
# Check for violations
|
|
72
|
+
if response.violation:
|
|
73
|
+
print(f"Safety violation detected: {response.violation.user_message}")
|
|
74
|
+
print(f"Violation level: {response.violation.violation_level}")
|
|
75
|
+
print(f"Metadata: {response.violation.metadata}")
|
|
76
|
+
else:
|
|
77
|
+
print("No safety violations detected")
|
|
78
|
+
```
|
|
@@ -89,6 +89,40 @@ class ResponsesStore:
|
|
|
89
89
|
) -> None:
|
|
90
90
|
await self._write_response_object(response_object, input, messages)
|
|
91
91
|
|
|
92
|
+
async def upsert_response_object(
|
|
93
|
+
self,
|
|
94
|
+
response_object: OpenAIResponseObject,
|
|
95
|
+
input: list[OpenAIResponseInput],
|
|
96
|
+
messages: list[OpenAIMessageParam],
|
|
97
|
+
) -> None:
|
|
98
|
+
"""Upsert response object using INSERT on first call, UPDATE on subsequent calls.
|
|
99
|
+
|
|
100
|
+
This method enables incremental persistence during streaming, allowing clients
|
|
101
|
+
to poll GET /v1/responses/{response_id} and see in-progress turn state.
|
|
102
|
+
|
|
103
|
+
:param response_object: The response object to store/update.
|
|
104
|
+
:param input: The input items for the response.
|
|
105
|
+
:param messages: The chat completion messages (for conversation continuity).
|
|
106
|
+
"""
|
|
107
|
+
if self.sql_store is None:
|
|
108
|
+
raise ValueError("Responses store is not initialized")
|
|
109
|
+
|
|
110
|
+
data = response_object.model_dump()
|
|
111
|
+
data["input"] = [input_item.model_dump() for input_item in input]
|
|
112
|
+
data["messages"] = [msg.model_dump() for msg in messages]
|
|
113
|
+
|
|
114
|
+
await self.sql_store.upsert(
|
|
115
|
+
table="openai_responses",
|
|
116
|
+
data={
|
|
117
|
+
"id": data["id"],
|
|
118
|
+
"created_at": data["created_at"],
|
|
119
|
+
"model": data["model"],
|
|
120
|
+
"response_object": data,
|
|
121
|
+
},
|
|
122
|
+
conflict_columns=["id"],
|
|
123
|
+
update_columns=["response_object"],
|
|
124
|
+
)
|
|
125
|
+
|
|
92
126
|
async def _write_response_object(
|
|
93
127
|
self,
|
|
94
128
|
response_object: OpenAIResponseObject,
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: llama_stack
|
|
3
|
-
Version: 0.4.
|
|
3
|
+
Version: 0.4.4
|
|
4
4
|
Summary: Llama Stack
|
|
5
5
|
Author-email: Meta Llama <llama-oss@meta.com>
|
|
6
6
|
License: MIT
|
|
@@ -46,7 +46,7 @@ Requires-Dist: psycopg2-binary
|
|
|
46
46
|
Requires-Dist: tornado>=6.5.3
|
|
47
47
|
Requires-Dist: urllib3>=2.6.3
|
|
48
48
|
Provides-Extra: client
|
|
49
|
-
Requires-Dist: llama-stack-client==0.4.
|
|
49
|
+
Requires-Dist: llama-stack-client==0.4.4; extra == "client"
|
|
50
50
|
Dynamic: license-file
|
|
51
51
|
|
|
52
52
|
# Llama Stack
|
|
@@ -104,17 +104,21 @@ llama_stack/distributions/template.py,sha256=2__m-cBZ-F1rBoR08diSy3fBGP5QEJbqzql
|
|
|
104
104
|
llama_stack/distributions/dell/__init__.py,sha256=XCPA_Cx4PWtWZ3-X2FBiFlHoteVPfnuzT1HiQtVAPCY,259
|
|
105
105
|
llama_stack/distributions/dell/config.yaml,sha256=s6KpD2dJK8s4MX9KSERfQ84NXeKgwmyEmXrOo3EH8AA,3262
|
|
106
106
|
llama_stack/distributions/dell/dell.py,sha256=O8LIKjZPS5tU87_B4Md-w2H0smDT_Hak5ZtE5QhMXso,5540
|
|
107
|
+
llama_stack/distributions/dell/doc_template.md,sha256=V6M6Jt5oHXKcsGljm6EggCpKwqmESK3U_njyL7vCHls,6126
|
|
107
108
|
llama_stack/distributions/dell/run-with-safety.yaml,sha256=yU1Uc-j81sTAzUa847aTkpF2q4PJ7T8AiGq40eI1kaI,3485
|
|
108
109
|
llama_stack/distributions/meta-reference-gpu/__init__.py,sha256=GziH9wzF_zUGaAZatKMwy-CvqtdNWB5H8X4qdULPuvs,269
|
|
109
110
|
llama_stack/distributions/meta-reference-gpu/config.yaml,sha256=tsn8oyHDCpuJgf3wTmmg5Vbdm4Psc6VMmrHd2Dv9yr0,3682
|
|
111
|
+
llama_stack/distributions/meta-reference-gpu/doc_template.md,sha256=7IWY_4U1n1FWWfmlZtlpH_SDO6534r6ntBEcvVT5ZI0,3276
|
|
110
112
|
llama_stack/distributions/meta-reference-gpu/meta_reference.py,sha256=sulbjnyT8057lWSawLnbtlyWjAmg7Bg6_iaKJnMieuM,6147
|
|
111
113
|
llama_stack/distributions/meta-reference-gpu/run-with-safety.yaml,sha256=U4uNvd0TejpfDrBkDc_Mt_xctPXK2ylskG6iCNvwktc,4220
|
|
112
114
|
llama_stack/distributions/nvidia/__init__.py,sha256=gE4p5a53QK02FQIZbVgi8kKvA5qOSQcq6DjlQ1YC-5k,261
|
|
113
115
|
llama_stack/distributions/nvidia/config.yaml,sha256=5kvNVeX8MOTiWUZMoNMLGwpaRRqt_VYoU7J2JguQWTg,3106
|
|
116
|
+
llama_stack/distributions/nvidia/doc_template.md,sha256=rMbz6-dD16DWUzLqW5pfZMxnjKSkcDkQAQpSAgheC-E,8083
|
|
114
117
|
llama_stack/distributions/nvidia/nvidia.py,sha256=4CMVA7jHb9EhIeUtKaj4QxECVEDT1bcaqVO703voCnA,5809
|
|
115
118
|
llama_stack/distributions/nvidia/run-with-safety.yaml,sha256=ctvAk-e6-JPWmc6GWlolN8Dq86vjqgFMibuAudmpUh8,3722
|
|
116
119
|
llama_stack/distributions/oci/__init__.py,sha256=Xil18y0rIa7YOBLQQs1vaPmegzpKnV7ua6zdjPCB2Gc,258
|
|
117
120
|
llama_stack/distributions/oci/config.yaml,sha256=JboZ2Ub5OwDIi-aTi25-B43DDmi3QSAxrudXsjq3Yn4,3452
|
|
121
|
+
llama_stack/distributions/oci/doc_template.md,sha256=oBItZ4qgn728U5wZkD2kHV7cJcWDDdaTG6CdXUW_NQw,5207
|
|
118
122
|
llama_stack/distributions/oci/oci.py,sha256=KpXep4QSt2PdkCK1HipMplPDoHyPCFcf56OmtyeGYoE,4231
|
|
119
123
|
llama_stack/distributions/open-benchmark/__init__.py,sha256=dKa2d24FDbturDyW04FKwyRYGGem6FPrFStB-oCFwAQ,269
|
|
120
124
|
llama_stack/distributions/open-benchmark/config.yaml,sha256=FxEblNi0HgKZSAMkiJvpoaq_C2p1JEv1UIzflXz_728,6997
|
|
@@ -144,9 +148,11 @@ llama_stack/models/llama/tokenizer_utils.py,sha256=ERY4A-spxh7tcQ8VzvpRxLJL48lGW
|
|
|
144
148
|
llama_stack/models/llama/llama3/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
145
149
|
llama_stack/models/llama/llama3/args.py,sha256=5bRB8pKIXhLIX0dev1um4UUykZ06rRhcHMZ54NAg0pk,2104
|
|
146
150
|
llama_stack/models/llama/llama3/chat_format.py,sha256=iEgJeXq4S_zq-f0yydZkoZL8Gq1jwX1isQQBGxvsfSc,10376
|
|
151
|
+
llama_stack/models/llama/llama3/dog.jpg,sha256=MOrU3JpVoSqMPipctb9ai3KFTbzd2rVoCSjVzEx0byY,40215
|
|
147
152
|
llama_stack/models/llama/llama3/generation.py,sha256=8hV7PLDXss8udOKJMiZuaIrNWxT6QYXcqC2mSIy-iHA,14366
|
|
148
153
|
llama_stack/models/llama/llama3/interface.py,sha256=NpY5ATMfybONjjM92RdaIPdmgt6wJTCJwnv2QX40l-8,7310
|
|
149
154
|
llama_stack/models/llama/llama3/model.py,sha256=SG5jKNMgMiTNqsPjrxLtDQNA6nfJDqGoBrEyJx6m7zE,11092
|
|
155
|
+
llama_stack/models/llama/llama3/pasta.jpeg,sha256=j7zVfjFge3ugK2Gi75wP6SCQCIeMOG5g2bt3VxfoEqs,448611
|
|
150
156
|
llama_stack/models/llama/llama3/template_data.py,sha256=GvOxOyigTv5jHLX6xVDip40QIuiPiQUr_iGJznKVDag,2960
|
|
151
157
|
llama_stack/models/llama/llama3/tokenizer.model,sha256=gunTGXnpKrkpzVREQPEp2ezXl7aeMn-A8X4cUNVVG1U,2183982
|
|
152
158
|
llama_stack/models/llama/llama3/tokenizer.py,sha256=UCS0QcZfYljrucyzJfU-Tk7uK1-mCv_iKqrQ-uvjhiY,6831
|
|
@@ -192,6 +198,9 @@ llama_stack/models/llama/llama4/quantization/loader.py,sha256=V8x8xTBkfRNIH20_bc
|
|
|
192
198
|
llama_stack/models/llama/llama4/vision/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
193
199
|
llama_stack/models/llama/llama4/vision/embedding.py,sha256=yg8e2ztlXKuKjOjvJksaNs9B9J0s7GM7ZxZJqfd_wfg,7194
|
|
194
200
|
llama_stack/models/llama/llama4/vision/encoder.py,sha256=nHIw0qvNlMY5Wzi0A5VufHwSbY3ZU1BHGS8rPMerTFM,14271
|
|
201
|
+
llama_stack/models/llama/resources/dog.jpg,sha256=MOrU3JpVoSqMPipctb9ai3KFTbzd2rVoCSjVzEx0byY,40215
|
|
202
|
+
llama_stack/models/llama/resources/pasta.jpeg,sha256=j7zVfjFge3ugK2Gi75wP6SCQCIeMOG5g2bt3VxfoEqs,448611
|
|
203
|
+
llama_stack/models/llama/resources/small_dog.jpg,sha256=CMdzEuiKdVxhD0oYJCJ3OTlvmDIKDW4k3ElCOGV8zxs,41823
|
|
195
204
|
llama_stack/providers/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
196
205
|
llama_stack/providers/inline/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
197
206
|
llama_stack/providers/inline/agents/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
@@ -200,7 +209,7 @@ llama_stack/providers/inline/agents/meta_reference/agents.py,sha256=Q1fqMRZ2Yfb4
|
|
|
200
209
|
llama_stack/providers/inline/agents/meta_reference/config.py,sha256=OY82xSuMfh8bMDfV7BZO4P-MMzzjUaMxOayNn4hNPj4,1375
|
|
201
210
|
llama_stack/providers/inline/agents/meta_reference/safety.py,sha256=e5taz2bA9Ks3nlRgLp0CGHsgU_Rv_6ed4bGemQlGEv8,1684
|
|
202
211
|
llama_stack/providers/inline/agents/meta_reference/responses/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
203
|
-
llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py,sha256=
|
|
212
|
+
llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py,sha256=dgWqY1OUqgpIUnoOp0TX55_SREl-TGMj8YOptQ7W_2A,30932
|
|
204
213
|
llama_stack/providers/inline/agents/meta_reference/responses/streaming.py,sha256=PsHByA2JUBU2KESWXpHq4rBfV4wMBwZ4JMSLZR_kxns,70310
|
|
205
214
|
llama_stack/providers/inline/agents/meta_reference/responses/tool_executor.py,sha256=2aEeqhWP8kCEubFwj9zFj_QzfXD1MssHDedhjvuMZLs,23202
|
|
206
215
|
llama_stack/providers/inline/agents/meta_reference/responses/types.py,sha256=VFUUSbaH15cHzy72pYcxDfuEdKbUxT3F3jg5X-42kbE,8721
|
|
@@ -232,6 +241,14 @@ llama_stack/providers/inline/inference/meta_reference/parallel_utils.py,sha256=i
|
|
|
232
241
|
llama_stack/providers/inline/inference/sentence_transformers/__init__.py,sha256=Yr4N2r8xdC8XcA4ASaKlcffQuCtqHsnQDjCa-_xlhR0,631
|
|
233
242
|
llama_stack/providers/inline/inference/sentence_transformers/config.py,sha256=pp-oKUeCJ5WxwLIYbHnduxt_5_NDD2ynPw7qTlxu548,407
|
|
234
243
|
llama_stack/providers/inline/inference/sentence_transformers/sentence_transformers.py,sha256=ON5HaefyMD1coh6nOz6IDz43HU9IutDvhUi2AEZs20E,2347
|
|
244
|
+
llama_stack/providers/inline/ios/inference/LocalInferenceImpl/LocalInference.h,sha256=YogYgEyYYkrfof1dpSzHTvfQdHklZAk0f3syMV_DZsk,394
|
|
245
|
+
llama_stack/providers/inline/ios/inference/LocalInferenceImpl/LocalInference.swift,sha256=KFRhTWgW5WPy9hQqT6XVGA-QV8gwwyw6Vx0dQ1PDqMc,6564
|
|
246
|
+
llama_stack/providers/inline/ios/inference/LocalInferenceImpl/Parsing.swift,sha256=5UMN70TvZK3lmZs4QSCFWUOu_GEIBmJIqpPKQqzKGxE,6216
|
|
247
|
+
llama_stack/providers/inline/ios/inference/LocalInferenceImpl/PromptTemplate.swift,sha256=M91HD2E5KiKQ17hJor-amS-gH1U2Mw5HudqpPHfKZec,268
|
|
248
|
+
llama_stack/providers/inline/ios/inference/LocalInferenceImpl/SystemPrompts.swift,sha256=u1rDMpFsLMaK92isxca_VsArHN7rZ01cl2OOhIrOEmQ,2811
|
|
249
|
+
llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.pbxproj,sha256=kD3yXS1_2eOOMTc7ez6XGTkweOQDGqW0w9tNsASN4sE,21605
|
|
250
|
+
llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.xcworkspace/contents.xcworkspacedata,sha256=fzsAtcP9tFJC17h-HlxOJdH6gSmhbJQpXsxOjqIjXF8,135
|
|
251
|
+
llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.xcworkspace/xcshareddata/IDEWorkspaceChecks.plist,sha256=36D5u4W5XMupxoEa2t-MtAf7XVIIjonbg-vDCOxEhUI,238
|
|
235
252
|
llama_stack/providers/inline/post_training/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
236
253
|
llama_stack/providers/inline/post_training/common/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
237
254
|
llama_stack/providers/inline/post_training/common/utils.py,sha256=3Qq35XOyNiUDfrZ0R4nv3pSPw5hF58P9nqjHZ5FZIZU,1251
|
|
@@ -344,16 +361,19 @@ llama_stack/providers/remote/datasetio/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa
|
|
|
344
361
|
llama_stack/providers/remote/datasetio/huggingface/__init__.py,sha256=X0NAiDAOSxOZFEehJuhu4I-KXt6NJcxzd3w3CYp9zMk,475
|
|
345
362
|
llama_stack/providers/remote/datasetio/huggingface/config.py,sha256=iowl39mwxoi5ESmjITn_psdaaTtcQkNPulBOC5IOzmQ,703
|
|
346
363
|
llama_stack/providers/remote/datasetio/huggingface/huggingface.py,sha256=EoCbD7s-mGnBALF26dWzWsireeSuysOqLWEGil6erE8,3489
|
|
364
|
+
llama_stack/providers/remote/datasetio/nvidia/README.md,sha256=4_DtUqhAyprbkPKNv7eTr0iQ_lcPK7-Bf6X3ha2U580,1723
|
|
347
365
|
llama_stack/providers/remote/datasetio/nvidia/__init__.py,sha256=aIpVjR9WVrc28zdZbkHwbpKdm_RYRx4gGKa_PE3B46U,615
|
|
348
366
|
llama_stack/providers/remote/datasetio/nvidia/config.py,sha256=g8-c-zd5-kveKs_60g23SUFl-o65eiEWxHqwyxw2GoA,2164
|
|
349
367
|
llama_stack/providers/remote/datasetio/nvidia/datasetio.py,sha256=KMx0dJkLk8uhyfLZEoNldiCE60mXKVoRCjSfgqUbGyk,3783
|
|
350
368
|
llama_stack/providers/remote/eval/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
369
|
+
llama_stack/providers/remote/eval/nvidia/README.md,sha256=e2nMDEfBvNx4lYud-FK3FIi1hjB1KTG9_fK6eqmZQgY,2516
|
|
351
370
|
llama_stack/providers/remote/eval/nvidia/__init__.py,sha256=vGqXSYDw8ruoLMlRHa0q_0NFTom_Vkd0GGUCFD60np0,715
|
|
352
371
|
llama_stack/providers/remote/eval/nvidia/config.py,sha256=BHyuojT2Zt_bQcI5y4a52DpCXQoxywn0JIs7viJKhN4,912
|
|
353
372
|
llama_stack/providers/remote/eval/nvidia/eval.py,sha256=QRyWmHmDioaSeEvfobhFMUTQKeu-M2uJdRsBhWLqoy4,5748
|
|
354
373
|
llama_stack/providers/remote/files/openai/__init__.py,sha256=X__X3MJF46wXY356ZCerk6DL4tALJUuEIzrIB_oWEKQ,579
|
|
355
374
|
llama_stack/providers/remote/files/openai/config.py,sha256=pBxTuDv2rMTMH6BsKxwi2kIh7gb5tUkMaWTxyDFmCiQ,953
|
|
356
375
|
llama_stack/providers/remote/files/openai/files.py,sha256=PhDqSw0yZW0rTrJmz5D7Rof-jRuJA93I8DfHtyPOb3o,8439
|
|
376
|
+
llama_stack/providers/remote/files/s3/README.md,sha256=M9jsDONmruw5oGGzscTbQP72ClAqpV-Ekdoedpb4fQs,6367
|
|
357
377
|
llama_stack/providers/remote/files/s3/__init__.py,sha256=Js4E7QuYzJYOzd7-9hwuOyv-pZxhFGqtYG2WkzM5YFY,563
|
|
358
378
|
llama_stack/providers/remote/files/s3/config.py,sha256=As_MCoVRua4s9qjD-ds1yyusfPuuisnJjxyOnZgpGHk,1925
|
|
359
379
|
llama_stack/providers/remote/files/s3/files.py,sha256=5ATTDb2DA3e8tE7AXLhHcVLPjqAcY7BxJP6t959Lepc,12326
|
|
@@ -385,6 +405,7 @@ llama_stack/providers/remote/inference/groq/groq.py,sha256=Bpx-sO2qiZ7qQKhKVp4vy
|
|
|
385
405
|
llama_stack/providers/remote/inference/llama_openai_compat/__init__.py,sha256=zDBa9R0i-va7EjpXD1n7N0bwSlNAvSwUzn5uxm8FTgs,501
|
|
386
406
|
llama_stack/providers/remote/inference/llama_openai_compat/config.py,sha256=YhZjUOzi9L9bqED2jj2BEcWpfSrF3oU7mn0v8quOJ3k,1052
|
|
387
407
|
llama_stack/providers/remote/inference/llama_openai_compat/llama.py,sha256=S7FFCJCQqSBYtjJ8A-wY8hdfphtC5dkcZ6argbKmZOw,1459
|
|
408
|
+
llama_stack/providers/remote/inference/nvidia/NVIDIA.md,sha256=XOgNOZ5xacKXdYqY9SEfXTIQMNTSdosZHtGDBMSHDy4,5469
|
|
388
409
|
llama_stack/providers/remote/inference/nvidia/__init__.py,sha256=nLgtvfGFjJtYEHKvl34gqSzrO_YkGQFtSEb-Je32lG8,753
|
|
389
410
|
llama_stack/providers/remote/inference/nvidia/config.py,sha256=StPKhd-l6TrWXGcG6v-TGZMrW4c3aAHkmE1fxJ4Unw8,2865
|
|
390
411
|
llama_stack/providers/remote/inference/nvidia/nvidia.py,sha256=-3onTye9BcAsd6mw1FmU2FFQ1CLJ7Fcydg3eIZkl0EM,6425
|
|
@@ -424,6 +445,7 @@ llama_stack/providers/remote/inference/watsonx/__init__.py,sha256=05ruRuPiOkniMY
|
|
|
424
445
|
llama_stack/providers/remote/inference/watsonx/config.py,sha256=7BU_x1uYJmFSYvPrF4HSG8eIUNZU1ao1iJbBfVYyrqQ,1425
|
|
425
446
|
llama_stack/providers/remote/inference/watsonx/watsonx.py,sha256=DOQyUMLDK7R-uapn34XRby9BvDLd9TXTO8aTzkHiOkU,11109
|
|
426
447
|
llama_stack/providers/remote/post_training/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
448
|
+
llama_stack/providers/remote/post_training/nvidia/README.md,sha256=_RE7P29RsR1ankLCiCxEVM_JOfUn68VQ08i3mFBaEUE,3303
|
|
427
449
|
llama_stack/providers/remote/post_training/nvidia/__init__.py,sha256=jChqHlLy0TY5fz4eHip0aujp9wH8SUn7JvyinQGfPzM,637
|
|
428
450
|
llama_stack/providers/remote/post_training/nvidia/config.py,sha256=xpIKIqRwNPaUJZxoZcGfJQq9UVh8dppHSdWp9uT_rKc,3447
|
|
429
451
|
llama_stack/providers/remote/post_training/nvidia/models.py,sha256=chuDZwPF9WmasAART5ZAoTra4rSRhWp44L1_YSia5qI,736
|
|
@@ -433,6 +455,7 @@ llama_stack/providers/remote/safety/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8
|
|
|
433
455
|
llama_stack/providers/remote/safety/bedrock/__init__.py,sha256=_u71SoCLPh97YDj8dZobQNJKJAhe-LW7FrMzG5smF2w,470
|
|
434
456
|
llama_stack/providers/remote/safety/bedrock/bedrock.py,sha256=dLCN0Yx5EcTaet0-CdM8OTk7qifqhfo8Y11lRAAiCOg,4136
|
|
435
457
|
llama_stack/providers/remote/safety/bedrock/config.py,sha256=f6XKxmt9mtd-9S8mxQ6EiOgpoG8vf2fHHF9JD0saBnM,395
|
|
458
|
+
llama_stack/providers/remote/safety/nvidia/README.md,sha256=su8iNvHS8kF_1TNm99SUk4TWCKDdOccv-BMxZBMhvU8,1831
|
|
436
459
|
llama_stack/providers/remote/safety/nvidia/__init__.py,sha256=rP0yxZsZ8NGA574mkUDPnPMDtynjoGlscTOY4QPfGfI,465
|
|
437
460
|
llama_stack/providers/remote/safety/nvidia/config.py,sha256=QhKrnOIkiIOQane-Bug2idIqKbo5OgyvUgx0L6aad9I,1523
|
|
438
461
|
llama_stack/providers/remote/safety/nvidia/nvidia.py,sha256=ZVVyBoZgxWYmNbU0ZfFI5UP6fPExShYeUQmXcg2p9iE,6014
|
|
@@ -498,7 +521,7 @@ llama_stack/providers/utils/memory/file_utils.py,sha256=MsjispuPO0cMXmRqAoTJ-dwM
|
|
|
498
521
|
llama_stack/providers/utils/memory/openai_vector_store_mixin.py,sha256=7yVvuq_uc9DxOd7SLqcJOVmu2c0AkBGmPmXZDaM1TgQ,60115
|
|
499
522
|
llama_stack/providers/utils/memory/vector_store.py,sha256=HiNCtZ4OzvPk4RGuRNks7CnbrDoppYyrJdjST-emnZQ,11941
|
|
500
523
|
llama_stack/providers/utils/responses/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
501
|
-
llama_stack/providers/utils/responses/responses_store.py,sha256=
|
|
524
|
+
llama_stack/providers/utils/responses/responses_store.py,sha256=DP_iIM25QPF_S0tptm1QPbcNc-CaRQekPecD2J9wLgg,11736
|
|
502
525
|
llama_stack/providers/utils/scoring/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
503
526
|
llama_stack/providers/utils/scoring/aggregation_utils.py,sha256=S2XR7DLXEVf7FCTOoEtAFjGO4arAP9UtQKpC-l56xK8,2720
|
|
504
527
|
llama_stack/providers/utils/scoring/base_scoring_fn.py,sha256=Upd-Hu8ftoZArfAuHx00uDkvEtmVKWg8U9tgkArro4o,4098
|
|
@@ -513,143 +536,9 @@ llama_stack/telemetry/constants.py,sha256=LtXE61xwNL3cBYZXKcXcbwD_Uh1jazP3V8a0od
|
|
|
513
536
|
llama_stack/telemetry/helpers.py,sha256=7uarMIHL5ngOUXQZxkH96corFxE7Jk5JaizRQ8Z8Ok0,1694
|
|
514
537
|
llama_stack/testing/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
515
538
|
llama_stack/testing/api_recorder.py,sha256=oGGTrzzBYNNvOIcvcFZenNPthr0yziJ7hlGPtckx460,39240
|
|
516
|
-
llama_stack-0.4.
|
|
517
|
-
|
|
518
|
-
|
|
519
|
-
|
|
520
|
-
|
|
521
|
-
|
|
522
|
-
llama_stack_api/datatypes.py,sha256=S7qOix_CBofuCEU6Gmm9qogZnnIO-WlN1kfO3D4Xlnc,12590
|
|
523
|
-
llama_stack_api/eval.py,sha256=PjgrSNk_Q8MmnN5hGKr1mMHTdrouuCItVJko32vxT6M,5095
|
|
524
|
-
llama_stack_api/inference.py,sha256=-Zy6F6R6NWI9Wq7acxz84K5C5RfOD9c2ytn8MrlK76s,41432
|
|
525
|
-
llama_stack_api/models.py,sha256=6RLvp94GDNBcMYya06SefIF6whIqAmm0Igsp1MoqLLA,5206
|
|
526
|
-
llama_stack_api/openai_responses.py,sha256=IslBagXUoebtBCYKATr9w7YR72GBjM7gYLNBPGDST4E,53967
|
|
527
|
-
llama_stack_api/post_training.py,sha256=94C4xbjG7Y9w7TRAcfXPOR1Um11QQ7KopU5y1lwCiX4,12991
|
|
528
|
-
llama_stack_api/prompts.py,sha256=D7wa6wZB4LslUGgIQSUezFtYap16qjQ-d33-6SUzTaw,7063
|
|
529
|
-
llama_stack_api/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
530
|
-
llama_stack_api/rag_tool.py,sha256=EtfHzPaGjxutdbJ3Ymx6QLtzBNHfCM6W6UGZ9TaV7UU,5695
|
|
531
|
-
llama_stack_api/resource.py,sha256=WDLMV9yeHYwSwxJSt-x-bWEMJU3Dgrl0fwzufTZuyWE,1088
|
|
532
|
-
llama_stack_api/router_utils.py,sha256=ylbRZ16gylyFCCHvS-B9cFpl9E1yRsYL8YlsuIFGP8Y,6949
|
|
533
|
-
llama_stack_api/safety.py,sha256=JXz6gwcl0YlKBMgkAVg89Atq0AtugvubRaQomAHmTzM,4319
|
|
534
|
-
llama_stack_api/schema_utils.py,sha256=YThcm7VlaQdkpOxNvIkn51FfGRlvdVt1TiV-KVBKkyA,7661
|
|
535
|
-
llama_stack_api/scoring.py,sha256=ejVkQbmeBBtbBuy8Xgg-b4aHFe6l8zwYnr5R7GV5gn0,2867
|
|
536
|
-
llama_stack_api/scoring_functions.py,sha256=0lP_ZENUh12i12ibg-_XNNPKLHi_TvB8H5LyEtBLhSE,7789
|
|
537
|
-
llama_stack_api/shields.py,sha256=9dNMyTVL0xcR8_BXCHb_zuAJC7Cz8pX8htRwW2-EDSw,2823
|
|
538
|
-
llama_stack_api/tools.py,sha256=eCyZx806VfpBJgsuJF9R3urA8ljF3g0kLapNpx9YRzY,7518
|
|
539
|
-
llama_stack_api/vector_io.py,sha256=3tYy8xLhVvx_rMtfi5Pxv0GwTMm1TfMYwq82tFqRz1U,36517
|
|
540
|
-
llama_stack_api/vector_stores.py,sha256=mILSO3k2X-Hg4G3YEdq54fKAenCuAzRAXqpNg-_D_Ng,1832
|
|
541
|
-
llama_stack_api/version.py,sha256=V3jdW3iFPdfOt4jWzJA-di7v0zHLYsn11hNtRzkY7uQ,297
|
|
542
|
-
llama_stack_api/admin/__init__.py,sha256=VnJn9fbk-dFkRrm1P5UWlAOcZDA2jf6dx9W5nt-WgOY,1049
|
|
543
|
-
llama_stack_api/admin/api.py,sha256=m14f4iBUJf-G0qITj66o-TFKCSUiD9U12XRnZ1Slr_w,1961
|
|
544
|
-
llama_stack_api/admin/fastapi_routes.py,sha256=3CPWhB86UMlYl3pQ0ZkbF5FLIKIzG2I61esXavoGEjY,3739
|
|
545
|
-
llama_stack_api/admin/models.py,sha256=aoDiI1mtM_XemHwqRFFwiVD64LbenXiYU-QK52IJDQU,3932
|
|
546
|
-
llama_stack_api/batches/__init__.py,sha256=vnHvv8mzJnFlCGa3V-lTiC0k2mVPDLOGZTqgUDovwKg,999
|
|
547
|
-
llama_stack_api/batches/api.py,sha256=49aBQJPOB-x6ohKVWwJ7SORmfm9QSsWak7OBE6L0cMM,1416
|
|
548
|
-
llama_stack_api/batches/fastapi_routes.py,sha256=1b0eSB2Wb2K6gvrhpBFTexsOpxuVU1urgfIOnwxx1fc,3864
|
|
549
|
-
llama_stack_api/batches/models.py,sha256=Dv9cHaaCqaLi_g5wIkKoy-Mn282Gqh711G2swb5ufGM,2692
|
|
550
|
-
llama_stack_api/benchmarks/__init__.py,sha256=_1Vs5xcZb179BxRjTv7OUVL1yvUdzWL92Bsx1aXYMSU,1119
|
|
551
|
-
llama_stack_api/benchmarks/api.py,sha256=j0zaRRBCdJS0XSq5dMthK1nUhiRqzPsJaHkQF61KlFc,933
|
|
552
|
-
llama_stack_api/benchmarks/fastapi_routes.py,sha256=_ZQ74esmXQRC8dbCb3vypT4z-7KdNoouDDH5T2LmD_c,3560
|
|
553
|
-
llama_stack_api/benchmarks/models.py,sha256=h5fWO3KUTnoFzLeIB_lYEVtcgw3D53Rx44WPHE0M7is,3644
|
|
554
|
-
llama_stack_api/common/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
555
|
-
llama_stack_api/common/content_types.py,sha256=lwc4VlPKWpRSTBO_U1MHdyItmQUzyNAqoaV9g3wKzF4,2693
|
|
556
|
-
llama_stack_api/common/errors.py,sha256=zrOjWerYj5BweLoyoqAbc3HGVSiaXLt10sw6TIJHnZ8,3725
|
|
557
|
-
llama_stack_api/common/job_types.py,sha256=1ifNdcNPqWPWw64R58zkhAnVWCj7oYg3utImbvf4NIc,1031
|
|
558
|
-
llama_stack_api/common/responses.py,sha256=qhwUdKKYzIhnlPPIah36rN3vVgMXEld3kS14XjtwFC0,2505
|
|
559
|
-
llama_stack_api/common/training_types.py,sha256=47eJdnLGfFEesnzRLYr0wysolfql7jpGz7Uh8X-hEec,1468
|
|
560
|
-
llama_stack_api/common/type_system.py,sha256=hTfEKuCXU16X0dBNWbzydhAMgKpPVm6lMM6L28gc9gw,3374
|
|
561
|
-
llama_stack_api/datasets/__init__.py,sha256=Cy5e0m2kU0rCyRcizrEC60gP1BEdD65-XFBvcCEpRWo,1436
|
|
562
|
-
llama_stack_api/datasets/api.py,sha256=DRJAwf8ZYjwVcYoE0pbHZGDHnHsrQJQiVcljvE9qkLc,1046
|
|
563
|
-
llama_stack_api/datasets/fastapi_routes.py,sha256=_F_-nnXeYwo8c5nFAEw7z3b8WPhSnGN_Uy61Cxv1F9A,3096
|
|
564
|
-
llama_stack_api/datasets/models.py,sha256=-Pkz8nD7se10Z_JzSKuRRwY-vcwAwU6UhWSajwfem_U,4648
|
|
565
|
-
llama_stack_api/file_processors/__init__.py,sha256=s9H1EQdDPm5MAmZiZDQbAgY0XXsdo10Bw3WlDu390B4,766
|
|
566
|
-
llama_stack_api/file_processors/api.py,sha256=MxrxuEDjTaqEdMu5kxMuAwwaGZy3yiAFku7VtORdWjk,2775
|
|
567
|
-
llama_stack_api/file_processors/fastapi_routes.py,sha256=NT1D_goFVmtAXdurOjY2ctgi6aAr4nHtgplz2Nhg5cg,2925
|
|
568
|
-
llama_stack_api/file_processors/models.py,sha256=a6_evBoh3PEZVrxJ1lDkWKUy5bZkjCHbydiyMZB9E50,1366
|
|
569
|
-
llama_stack_api/files/__init__.py,sha256=7ncmkC_-3WKYu3FIseApV5w4ER7PHyG1M2E6pb2mduo,839
|
|
570
|
-
llama_stack_api/files/api.py,sha256=79tc1hRe78AE_QA_BdOfpNpjfYTzLVYg6h4dXNkKu3I,1258
|
|
571
|
-
llama_stack_api/files/fastapi_routes.py,sha256=-FadxkQZKXUlYSJtmfZCXCBExAG9HBHttT-j_i0d8Ig,4177
|
|
572
|
-
llama_stack_api/files/models.py,sha256=Uz-gPoMZSV8P7eVHdKSDGMTE-B3dFUdM3BXU9s0PdGY,4239
|
|
573
|
-
llama_stack_api/inspect_api/__init__.py,sha256=0jRDcUhEmVtXqK3BDX8I2qtcN0S4lFAAcLI-dMpGQ-w,861
|
|
574
|
-
llama_stack_api/inspect_api/api.py,sha256=XkdM7jJ3_UlEIE4woEVi5mO2O1aNn9_FPtb18NTnWSM,726
|
|
575
|
-
llama_stack_api/inspect_api/fastapi_routes.py,sha256=I7R8roy6einYDzrPN8wNjrRokpoSNZi9zrtmLHS1vDw,2575
|
|
576
|
-
llama_stack_api/inspect_api/models.py,sha256=EW69EHkOG8i0GS8KW8Kz6WaPZV74hzwad8dGXWrrKhs,683
|
|
577
|
-
llama_stack_api/internal/__init__.py,sha256=hZiF7mONpu54guvMUTW9XpfkETUO55u6hqYOYkz8Bt0,307
|
|
578
|
-
llama_stack_api/internal/kvstore.py,sha256=mgNJz6r8_ju3I3JT2Pz5fSX_9DLv_OupsS2NnJe3usY,833
|
|
579
|
-
llama_stack_api/internal/sqlstore.py,sha256=FBIQhG7VOVMMSTe24uMigfxEWXnarY0hzx9HjrNXVnI,2262
|
|
580
|
-
llama_stack_api/llama_stack_api/__init__.py,sha256=5XNQGpundjXTutLgnYp6B1t6KITWXH_of626GciNma4,28103
|
|
581
|
-
llama_stack_api/llama_stack_api/agents.py,sha256=u0sg3AoWCip5o8T4DMTM8uqP3BsdbkKbor3PmxKTg0g,7143
|
|
582
|
-
llama_stack_api/llama_stack_api/connectors.py,sha256=PcAwndbVQC6pm5HGSlNprqYFTZzhCM7SYHPyRkSIoaQ,4644
|
|
583
|
-
llama_stack_api/llama_stack_api/conversations.py,sha256=pLQD2ZT6rSWF2IIQUtdSvkq50w9-piCMVr9hgdMmlBw,10290
|
|
584
|
-
llama_stack_api/llama_stack_api/datasetio.py,sha256=n4wQRv01rl8K_Ig2_Ln5hZBfdbmptKtDGsNE1igJ1-E,2075
|
|
585
|
-
llama_stack_api/llama_stack_api/datatypes.py,sha256=S7qOix_CBofuCEU6Gmm9qogZnnIO-WlN1kfO3D4Xlnc,12590
|
|
586
|
-
llama_stack_api/llama_stack_api/eval.py,sha256=PjgrSNk_Q8MmnN5hGKr1mMHTdrouuCItVJko32vxT6M,5095
|
|
587
|
-
llama_stack_api/llama_stack_api/inference.py,sha256=-Zy6F6R6NWI9Wq7acxz84K5C5RfOD9c2ytn8MrlK76s,41432
|
|
588
|
-
llama_stack_api/llama_stack_api/models.py,sha256=6RLvp94GDNBcMYya06SefIF6whIqAmm0Igsp1MoqLLA,5206
|
|
589
|
-
llama_stack_api/llama_stack_api/openai_responses.py,sha256=IslBagXUoebtBCYKATr9w7YR72GBjM7gYLNBPGDST4E,53967
|
|
590
|
-
llama_stack_api/llama_stack_api/post_training.py,sha256=94C4xbjG7Y9w7TRAcfXPOR1Um11QQ7KopU5y1lwCiX4,12991
|
|
591
|
-
llama_stack_api/llama_stack_api/prompts.py,sha256=D7wa6wZB4LslUGgIQSUezFtYap16qjQ-d33-6SUzTaw,7063
|
|
592
|
-
llama_stack_api/llama_stack_api/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
593
|
-
llama_stack_api/llama_stack_api/rag_tool.py,sha256=EtfHzPaGjxutdbJ3Ymx6QLtzBNHfCM6W6UGZ9TaV7UU,5695
|
|
594
|
-
llama_stack_api/llama_stack_api/resource.py,sha256=WDLMV9yeHYwSwxJSt-x-bWEMJU3Dgrl0fwzufTZuyWE,1088
|
|
595
|
-
llama_stack_api/llama_stack_api/router_utils.py,sha256=ylbRZ16gylyFCCHvS-B9cFpl9E1yRsYL8YlsuIFGP8Y,6949
|
|
596
|
-
llama_stack_api/llama_stack_api/safety.py,sha256=JXz6gwcl0YlKBMgkAVg89Atq0AtugvubRaQomAHmTzM,4319
|
|
597
|
-
llama_stack_api/llama_stack_api/schema_utils.py,sha256=YThcm7VlaQdkpOxNvIkn51FfGRlvdVt1TiV-KVBKkyA,7661
|
|
598
|
-
llama_stack_api/llama_stack_api/scoring.py,sha256=ejVkQbmeBBtbBuy8Xgg-b4aHFe6l8zwYnr5R7GV5gn0,2867
|
|
599
|
-
llama_stack_api/llama_stack_api/scoring_functions.py,sha256=0lP_ZENUh12i12ibg-_XNNPKLHi_TvB8H5LyEtBLhSE,7789
|
|
600
|
-
llama_stack_api/llama_stack_api/shields.py,sha256=9dNMyTVL0xcR8_BXCHb_zuAJC7Cz8pX8htRwW2-EDSw,2823
|
|
601
|
-
llama_stack_api/llama_stack_api/tools.py,sha256=eCyZx806VfpBJgsuJF9R3urA8ljF3g0kLapNpx9YRzY,7518
|
|
602
|
-
llama_stack_api/llama_stack_api/vector_io.py,sha256=3tYy8xLhVvx_rMtfi5Pxv0GwTMm1TfMYwq82tFqRz1U,36517
|
|
603
|
-
llama_stack_api/llama_stack_api/vector_stores.py,sha256=mILSO3k2X-Hg4G3YEdq54fKAenCuAzRAXqpNg-_D_Ng,1832
|
|
604
|
-
llama_stack_api/llama_stack_api/version.py,sha256=V3jdW3iFPdfOt4jWzJA-di7v0zHLYsn11hNtRzkY7uQ,297
|
|
605
|
-
llama_stack_api/llama_stack_api/admin/__init__.py,sha256=VnJn9fbk-dFkRrm1P5UWlAOcZDA2jf6dx9W5nt-WgOY,1049
|
|
606
|
-
llama_stack_api/llama_stack_api/admin/api.py,sha256=m14f4iBUJf-G0qITj66o-TFKCSUiD9U12XRnZ1Slr_w,1961
|
|
607
|
-
llama_stack_api/llama_stack_api/admin/fastapi_routes.py,sha256=3CPWhB86UMlYl3pQ0ZkbF5FLIKIzG2I61esXavoGEjY,3739
|
|
608
|
-
llama_stack_api/llama_stack_api/admin/models.py,sha256=aoDiI1mtM_XemHwqRFFwiVD64LbenXiYU-QK52IJDQU,3932
|
|
609
|
-
llama_stack_api/llama_stack_api/batches/__init__.py,sha256=vnHvv8mzJnFlCGa3V-lTiC0k2mVPDLOGZTqgUDovwKg,999
|
|
610
|
-
llama_stack_api/llama_stack_api/batches/api.py,sha256=49aBQJPOB-x6ohKVWwJ7SORmfm9QSsWak7OBE6L0cMM,1416
|
|
611
|
-
llama_stack_api/llama_stack_api/batches/fastapi_routes.py,sha256=1b0eSB2Wb2K6gvrhpBFTexsOpxuVU1urgfIOnwxx1fc,3864
|
|
612
|
-
llama_stack_api/llama_stack_api/batches/models.py,sha256=Dv9cHaaCqaLi_g5wIkKoy-Mn282Gqh711G2swb5ufGM,2692
|
|
613
|
-
llama_stack_api/llama_stack_api/benchmarks/__init__.py,sha256=_1Vs5xcZb179BxRjTv7OUVL1yvUdzWL92Bsx1aXYMSU,1119
|
|
614
|
-
llama_stack_api/llama_stack_api/benchmarks/api.py,sha256=j0zaRRBCdJS0XSq5dMthK1nUhiRqzPsJaHkQF61KlFc,933
|
|
615
|
-
llama_stack_api/llama_stack_api/benchmarks/fastapi_routes.py,sha256=_ZQ74esmXQRC8dbCb3vypT4z-7KdNoouDDH5T2LmD_c,3560
|
|
616
|
-
llama_stack_api/llama_stack_api/benchmarks/models.py,sha256=h5fWO3KUTnoFzLeIB_lYEVtcgw3D53Rx44WPHE0M7is,3644
|
|
617
|
-
llama_stack_api/llama_stack_api/common/__init__.py,sha256=vUvqRS2CXhASaFzYVspRYa5q8usSCzjKUlZhzNLuiKg,200
|
|
618
|
-
llama_stack_api/llama_stack_api/common/content_types.py,sha256=lwc4VlPKWpRSTBO_U1MHdyItmQUzyNAqoaV9g3wKzF4,2693
|
|
619
|
-
llama_stack_api/llama_stack_api/common/errors.py,sha256=zrOjWerYj5BweLoyoqAbc3HGVSiaXLt10sw6TIJHnZ8,3725
|
|
620
|
-
llama_stack_api/llama_stack_api/common/job_types.py,sha256=1ifNdcNPqWPWw64R58zkhAnVWCj7oYg3utImbvf4NIc,1031
|
|
621
|
-
llama_stack_api/llama_stack_api/common/responses.py,sha256=qhwUdKKYzIhnlPPIah36rN3vVgMXEld3kS14XjtwFC0,2505
|
|
622
|
-
llama_stack_api/llama_stack_api/common/training_types.py,sha256=47eJdnLGfFEesnzRLYr0wysolfql7jpGz7Uh8X-hEec,1468
|
|
623
|
-
llama_stack_api/llama_stack_api/common/type_system.py,sha256=hTfEKuCXU16X0dBNWbzydhAMgKpPVm6lMM6L28gc9gw,3374
|
|
624
|
-
llama_stack_api/llama_stack_api/datasets/__init__.py,sha256=Cy5e0m2kU0rCyRcizrEC60gP1BEdD65-XFBvcCEpRWo,1436
|
|
625
|
-
llama_stack_api/llama_stack_api/datasets/api.py,sha256=DRJAwf8ZYjwVcYoE0pbHZGDHnHsrQJQiVcljvE9qkLc,1046
|
|
626
|
-
llama_stack_api/llama_stack_api/datasets/fastapi_routes.py,sha256=_F_-nnXeYwo8c5nFAEw7z3b8WPhSnGN_Uy61Cxv1F9A,3096
|
|
627
|
-
llama_stack_api/llama_stack_api/datasets/models.py,sha256=-Pkz8nD7se10Z_JzSKuRRwY-vcwAwU6UhWSajwfem_U,4648
|
|
628
|
-
llama_stack_api/llama_stack_api/file_processors/__init__.py,sha256=s9H1EQdDPm5MAmZiZDQbAgY0XXsdo10Bw3WlDu390B4,766
|
|
629
|
-
llama_stack_api/llama_stack_api/file_processors/api.py,sha256=MxrxuEDjTaqEdMu5kxMuAwwaGZy3yiAFku7VtORdWjk,2775
|
|
630
|
-
llama_stack_api/llama_stack_api/file_processors/fastapi_routes.py,sha256=NT1D_goFVmtAXdurOjY2ctgi6aAr4nHtgplz2Nhg5cg,2925
|
|
631
|
-
llama_stack_api/llama_stack_api/file_processors/models.py,sha256=a6_evBoh3PEZVrxJ1lDkWKUy5bZkjCHbydiyMZB9E50,1366
|
|
632
|
-
llama_stack_api/llama_stack_api/files/__init__.py,sha256=7ncmkC_-3WKYu3FIseApV5w4ER7PHyG1M2E6pb2mduo,839
|
|
633
|
-
llama_stack_api/llama_stack_api/files/api.py,sha256=79tc1hRe78AE_QA_BdOfpNpjfYTzLVYg6h4dXNkKu3I,1258
|
|
634
|
-
llama_stack_api/llama_stack_api/files/fastapi_routes.py,sha256=-FadxkQZKXUlYSJtmfZCXCBExAG9HBHttT-j_i0d8Ig,4177
|
|
635
|
-
llama_stack_api/llama_stack_api/files/models.py,sha256=Uz-gPoMZSV8P7eVHdKSDGMTE-B3dFUdM3BXU9s0PdGY,4239
|
|
636
|
-
llama_stack_api/llama_stack_api/inspect_api/__init__.py,sha256=0jRDcUhEmVtXqK3BDX8I2qtcN0S4lFAAcLI-dMpGQ-w,861
|
|
637
|
-
llama_stack_api/llama_stack_api/inspect_api/api.py,sha256=XkdM7jJ3_UlEIE4woEVi5mO2O1aNn9_FPtb18NTnWSM,726
|
|
638
|
-
llama_stack_api/llama_stack_api/inspect_api/fastapi_routes.py,sha256=I7R8roy6einYDzrPN8wNjrRokpoSNZi9zrtmLHS1vDw,2575
|
|
639
|
-
llama_stack_api/llama_stack_api/inspect_api/models.py,sha256=EW69EHkOG8i0GS8KW8Kz6WaPZV74hzwad8dGXWrrKhs,683
|
|
640
|
-
llama_stack_api/llama_stack_api/internal/__init__.py,sha256=hZiF7mONpu54guvMUTW9XpfkETUO55u6hqYOYkz8Bt0,307
|
|
641
|
-
llama_stack_api/llama_stack_api/internal/kvstore.py,sha256=mgNJz6r8_ju3I3JT2Pz5fSX_9DLv_OupsS2NnJe3usY,833
|
|
642
|
-
llama_stack_api/llama_stack_api/internal/sqlstore.py,sha256=FBIQhG7VOVMMSTe24uMigfxEWXnarY0hzx9HjrNXVnI,2262
|
|
643
|
-
llama_stack_api/llama_stack_api/providers/__init__.py,sha256=a_187ghsdPNYJ5xLizqKYREJJLBa-lpcIhLp8spgsH8,841
|
|
644
|
-
llama_stack_api/llama_stack_api/providers/api.py,sha256=ytwxri9s6p8j9ClFKgN9mfa1TF0VZh1o8W5cVZR49rc,534
|
|
645
|
-
llama_stack_api/llama_stack_api/providers/fastapi_routes.py,sha256=jb1yrXEk1MdtcgWCToSZtaB-wjKqv5uVKIkvduXoKlM,1962
|
|
646
|
-
llama_stack_api/llama_stack_api/providers/models.py,sha256=nqBzh9je_dou35XFjYGD43hwKgjWy6HIRmGWUrcGqOw,653
|
|
647
|
-
llama_stack_api/providers/__init__.py,sha256=a_187ghsdPNYJ5xLizqKYREJJLBa-lpcIhLp8spgsH8,841
|
|
648
|
-
llama_stack_api/providers/api.py,sha256=ytwxri9s6p8j9ClFKgN9mfa1TF0VZh1o8W5cVZR49rc,534
|
|
649
|
-
llama_stack_api/providers/fastapi_routes.py,sha256=jb1yrXEk1MdtcgWCToSZtaB-wjKqv5uVKIkvduXoKlM,1962
|
|
650
|
-
llama_stack_api/providers/models.py,sha256=nqBzh9je_dou35XFjYGD43hwKgjWy6HIRmGWUrcGqOw,653
|
|
651
|
-
llama_stack-0.4.3.dist-info/METADATA,sha256=i3L_soHo8hFMex4qrg30BHHC4f79XoW3jdR3Zza76Yk,12464
|
|
652
|
-
llama_stack-0.4.3.dist-info/WHEEL,sha256=wUyA8OaulRlbfwMtmQsvNngGrxQHAvkKcvRmdizlJi0,92
|
|
653
|
-
llama_stack-0.4.3.dist-info/entry_points.txt,sha256=E5xoyAM9064aW_y96eSSwZCNT_ANctrvrhLMJnMQlw0,141
|
|
654
|
-
llama_stack-0.4.3.dist-info/top_level.txt,sha256=pyNYneZU5w62BaExic-GC1ph5kk8JI2mJFwzqiZy2cU,28
|
|
655
|
-
llama_stack-0.4.3.dist-info/RECORD,,
|
|
539
|
+
llama_stack-0.4.4.dist-info/licenses/LICENSE,sha256=42g1gBn9gHYdBt5e6e1aFYhnc-JT9trU9qBD84oUAlY,1087
|
|
540
|
+
llama_stack-0.4.4.dist-info/METADATA,sha256=Wz0lOkkHDVebZVPs_uEeU-jFBV5nvmhD_YU32gFSxF0,12464
|
|
541
|
+
llama_stack-0.4.4.dist-info/WHEEL,sha256=wUyA8OaulRlbfwMtmQsvNngGrxQHAvkKcvRmdizlJi0,92
|
|
542
|
+
llama_stack-0.4.4.dist-info/entry_points.txt,sha256=E5xoyAM9064aW_y96eSSwZCNT_ANctrvrhLMJnMQlw0,141
|
|
543
|
+
llama_stack-0.4.4.dist-info/top_level.txt,sha256=2-nbQ1CAn4_w76YD_O6N6ofvjmk4DX5NFaBuApSx5N0,12
|
|
544
|
+
llama_stack-0.4.4.dist-info/RECORD,,
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
llama_stack
|