llama-stack 0.4.2__py3-none-any.whl → 0.4.4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- llama_stack/core/library_client.py +80 -3
- llama_stack/core/routing_tables/common.py +11 -0
- llama_stack/core/routing_tables/vector_stores.py +4 -0
- llama_stack/core/stack.py +16 -1
- llama_stack/core/storage/kvstore/kvstore.py +11 -0
- llama_stack/core/storage/kvstore/mongodb/mongodb.py +5 -0
- llama_stack/core/storage/kvstore/postgres/postgres.py +8 -0
- llama_stack/core/storage/kvstore/redis/redis.py +5 -0
- llama_stack/core/storage/sqlstore/sqlalchemy_sqlstore.py +8 -0
- llama_stack/core/storage/sqlstore/sqlstore.py +8 -0
- llama_stack/distributions/dell/doc_template.md +209 -0
- llama_stack/distributions/meta-reference-gpu/doc_template.md +119 -0
- llama_stack/distributions/nvidia/doc_template.md +170 -0
- llama_stack/distributions/oci/doc_template.md +140 -0
- llama_stack/models/llama/llama3/dog.jpg +0 -0
- llama_stack/models/llama/llama3/pasta.jpeg +0 -0
- llama_stack/models/llama/resources/dog.jpg +0 -0
- llama_stack/models/llama/resources/pasta.jpeg +0 -0
- llama_stack/models/llama/resources/small_dog.jpg +0 -0
- llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py +184 -33
- llama_stack/providers/inline/agents/meta_reference/responses/streaming.py +4 -0
- llama_stack/providers/inline/agents/meta_reference/responses/tool_executor.py +9 -1
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl/LocalInference.h +9 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl/LocalInference.swift +189 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl/Parsing.swift +238 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl/PromptTemplate.swift +12 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl/SystemPrompts.swift +89 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.pbxproj +550 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.xcworkspace/contents.xcworkspacedata +7 -0
- llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.xcworkspace/xcshareddata/IDEWorkspaceChecks.plist +8 -0
- llama_stack/providers/remote/datasetio/nvidia/README.md +74 -0
- llama_stack/providers/remote/eval/nvidia/README.md +134 -0
- llama_stack/providers/remote/files/s3/README.md +266 -0
- llama_stack/providers/remote/inference/nvidia/NVIDIA.md +203 -0
- llama_stack/providers/remote/post_training/nvidia/README.md +151 -0
- llama_stack/providers/remote/safety/nvidia/README.md +78 -0
- llama_stack/providers/remote/vector_io/pgvector/pgvector.py +13 -1
- llama_stack/providers/utils/inference/embedding_mixin.py +20 -16
- llama_stack/providers/utils/memory/openai_vector_store_mixin.py +33 -0
- llama_stack/providers/utils/responses/responses_store.py +34 -0
- llama_stack/providers/utils/tools/mcp.py +258 -16
- {llama_stack-0.4.2.dist-info → llama_stack-0.4.4.dist-info}/METADATA +2 -2
- {llama_stack-0.4.2.dist-info → llama_stack-0.4.4.dist-info}/RECORD +47 -158
- {llama_stack-0.4.2.dist-info → llama_stack-0.4.4.dist-info}/WHEEL +1 -1
- llama_stack-0.4.4.dist-info/top_level.txt +1 -0
- llama_stack-0.4.2.dist-info/top_level.txt +0 -2
- llama_stack_api/__init__.py +0 -945
- llama_stack_api/admin/__init__.py +0 -45
- llama_stack_api/admin/api.py +0 -72
- llama_stack_api/admin/fastapi_routes.py +0 -117
- llama_stack_api/admin/models.py +0 -113
- llama_stack_api/agents.py +0 -173
- llama_stack_api/batches/__init__.py +0 -40
- llama_stack_api/batches/api.py +0 -53
- llama_stack_api/batches/fastapi_routes.py +0 -113
- llama_stack_api/batches/models.py +0 -78
- llama_stack_api/benchmarks/__init__.py +0 -43
- llama_stack_api/benchmarks/api.py +0 -39
- llama_stack_api/benchmarks/fastapi_routes.py +0 -109
- llama_stack_api/benchmarks/models.py +0 -109
- llama_stack_api/common/__init__.py +0 -5
- llama_stack_api/common/content_types.py +0 -101
- llama_stack_api/common/errors.py +0 -95
- llama_stack_api/common/job_types.py +0 -38
- llama_stack_api/common/responses.py +0 -77
- llama_stack_api/common/training_types.py +0 -47
- llama_stack_api/common/type_system.py +0 -146
- llama_stack_api/connectors.py +0 -146
- llama_stack_api/conversations.py +0 -270
- llama_stack_api/datasetio.py +0 -55
- llama_stack_api/datasets/__init__.py +0 -61
- llama_stack_api/datasets/api.py +0 -35
- llama_stack_api/datasets/fastapi_routes.py +0 -104
- llama_stack_api/datasets/models.py +0 -152
- llama_stack_api/datatypes.py +0 -373
- llama_stack_api/eval.py +0 -137
- llama_stack_api/file_processors/__init__.py +0 -27
- llama_stack_api/file_processors/api.py +0 -64
- llama_stack_api/file_processors/fastapi_routes.py +0 -78
- llama_stack_api/file_processors/models.py +0 -42
- llama_stack_api/files/__init__.py +0 -35
- llama_stack_api/files/api.py +0 -51
- llama_stack_api/files/fastapi_routes.py +0 -124
- llama_stack_api/files/models.py +0 -107
- llama_stack_api/inference.py +0 -1169
- llama_stack_api/inspect_api/__init__.py +0 -37
- llama_stack_api/inspect_api/api.py +0 -25
- llama_stack_api/inspect_api/fastapi_routes.py +0 -76
- llama_stack_api/inspect_api/models.py +0 -28
- llama_stack_api/internal/__init__.py +0 -9
- llama_stack_api/internal/kvstore.py +0 -26
- llama_stack_api/internal/sqlstore.py +0 -79
- llama_stack_api/llama_stack_api/__init__.py +0 -945
- llama_stack_api/llama_stack_api/admin/__init__.py +0 -45
- llama_stack_api/llama_stack_api/admin/api.py +0 -72
- llama_stack_api/llama_stack_api/admin/fastapi_routes.py +0 -117
- llama_stack_api/llama_stack_api/admin/models.py +0 -113
- llama_stack_api/llama_stack_api/agents.py +0 -173
- llama_stack_api/llama_stack_api/batches/__init__.py +0 -40
- llama_stack_api/llama_stack_api/batches/api.py +0 -53
- llama_stack_api/llama_stack_api/batches/fastapi_routes.py +0 -113
- llama_stack_api/llama_stack_api/batches/models.py +0 -78
- llama_stack_api/llama_stack_api/benchmarks/__init__.py +0 -43
- llama_stack_api/llama_stack_api/benchmarks/api.py +0 -39
- llama_stack_api/llama_stack_api/benchmarks/fastapi_routes.py +0 -109
- llama_stack_api/llama_stack_api/benchmarks/models.py +0 -109
- llama_stack_api/llama_stack_api/common/__init__.py +0 -5
- llama_stack_api/llama_stack_api/common/content_types.py +0 -101
- llama_stack_api/llama_stack_api/common/errors.py +0 -95
- llama_stack_api/llama_stack_api/common/job_types.py +0 -38
- llama_stack_api/llama_stack_api/common/responses.py +0 -77
- llama_stack_api/llama_stack_api/common/training_types.py +0 -47
- llama_stack_api/llama_stack_api/common/type_system.py +0 -146
- llama_stack_api/llama_stack_api/connectors.py +0 -146
- llama_stack_api/llama_stack_api/conversations.py +0 -270
- llama_stack_api/llama_stack_api/datasetio.py +0 -55
- llama_stack_api/llama_stack_api/datasets/__init__.py +0 -61
- llama_stack_api/llama_stack_api/datasets/api.py +0 -35
- llama_stack_api/llama_stack_api/datasets/fastapi_routes.py +0 -104
- llama_stack_api/llama_stack_api/datasets/models.py +0 -152
- llama_stack_api/llama_stack_api/datatypes.py +0 -373
- llama_stack_api/llama_stack_api/eval.py +0 -137
- llama_stack_api/llama_stack_api/file_processors/__init__.py +0 -27
- llama_stack_api/llama_stack_api/file_processors/api.py +0 -64
- llama_stack_api/llama_stack_api/file_processors/fastapi_routes.py +0 -78
- llama_stack_api/llama_stack_api/file_processors/models.py +0 -42
- llama_stack_api/llama_stack_api/files/__init__.py +0 -35
- llama_stack_api/llama_stack_api/files/api.py +0 -51
- llama_stack_api/llama_stack_api/files/fastapi_routes.py +0 -124
- llama_stack_api/llama_stack_api/files/models.py +0 -107
- llama_stack_api/llama_stack_api/inference.py +0 -1169
- llama_stack_api/llama_stack_api/inspect_api/__init__.py +0 -37
- llama_stack_api/llama_stack_api/inspect_api/api.py +0 -25
- llama_stack_api/llama_stack_api/inspect_api/fastapi_routes.py +0 -76
- llama_stack_api/llama_stack_api/inspect_api/models.py +0 -28
- llama_stack_api/llama_stack_api/internal/__init__.py +0 -9
- llama_stack_api/llama_stack_api/internal/kvstore.py +0 -26
- llama_stack_api/llama_stack_api/internal/sqlstore.py +0 -79
- llama_stack_api/llama_stack_api/models.py +0 -171
- llama_stack_api/llama_stack_api/openai_responses.py +0 -1468
- llama_stack_api/llama_stack_api/post_training.py +0 -370
- llama_stack_api/llama_stack_api/prompts.py +0 -203
- llama_stack_api/llama_stack_api/providers/__init__.py +0 -33
- llama_stack_api/llama_stack_api/providers/api.py +0 -16
- llama_stack_api/llama_stack_api/providers/fastapi_routes.py +0 -57
- llama_stack_api/llama_stack_api/providers/models.py +0 -24
- llama_stack_api/llama_stack_api/py.typed +0 -0
- llama_stack_api/llama_stack_api/rag_tool.py +0 -168
- llama_stack_api/llama_stack_api/resource.py +0 -37
- llama_stack_api/llama_stack_api/router_utils.py +0 -160
- llama_stack_api/llama_stack_api/safety.py +0 -132
- llama_stack_api/llama_stack_api/schema_utils.py +0 -208
- llama_stack_api/llama_stack_api/scoring.py +0 -93
- llama_stack_api/llama_stack_api/scoring_functions.py +0 -211
- llama_stack_api/llama_stack_api/shields.py +0 -93
- llama_stack_api/llama_stack_api/tools.py +0 -226
- llama_stack_api/llama_stack_api/vector_io.py +0 -941
- llama_stack_api/llama_stack_api/vector_stores.py +0 -51
- llama_stack_api/llama_stack_api/version.py +0 -9
- llama_stack_api/models.py +0 -171
- llama_stack_api/openai_responses.py +0 -1468
- llama_stack_api/post_training.py +0 -370
- llama_stack_api/prompts.py +0 -203
- llama_stack_api/providers/__init__.py +0 -33
- llama_stack_api/providers/api.py +0 -16
- llama_stack_api/providers/fastapi_routes.py +0 -57
- llama_stack_api/providers/models.py +0 -24
- llama_stack_api/py.typed +0 -0
- llama_stack_api/rag_tool.py +0 -168
- llama_stack_api/resource.py +0 -37
- llama_stack_api/router_utils.py +0 -160
- llama_stack_api/safety.py +0 -132
- llama_stack_api/schema_utils.py +0 -208
- llama_stack_api/scoring.py +0 -93
- llama_stack_api/scoring_functions.py +0 -211
- llama_stack_api/shields.py +0 -93
- llama_stack_api/tools.py +0 -226
- llama_stack_api/vector_io.py +0 -941
- llama_stack_api/vector_stores.py +0 -51
- llama_stack_api/version.py +0 -9
- {llama_stack-0.4.2.dist-info → llama_stack-0.4.4.dist-info}/entry_points.txt +0 -0
- {llama_stack-0.4.2.dist-info → llama_stack-0.4.4.dist-info}/licenses/LICENSE +0 -0
|
@@ -1,51 +0,0 @@
|
|
|
1
|
-
# Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
2
|
-
# All rights reserved.
|
|
3
|
-
#
|
|
4
|
-
# This source code is licensed under the terms described in the LICENSE file in
|
|
5
|
-
# the root directory of this source tree.
|
|
6
|
-
|
|
7
|
-
from typing import Literal
|
|
8
|
-
|
|
9
|
-
from pydantic import BaseModel
|
|
10
|
-
|
|
11
|
-
from llama_stack_api.resource import Resource, ResourceType
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
# Internal resource type for storing the vector store routing and other information
|
|
15
|
-
class VectorStore(Resource):
|
|
16
|
-
"""Vector database resource for storing and querying vector embeddings.
|
|
17
|
-
|
|
18
|
-
:param type: Type of resource, always 'vector_store' for vector stores
|
|
19
|
-
:param embedding_model: Name of the embedding model to use for vector generation
|
|
20
|
-
:param embedding_dimension: Dimension of the embedding vectors
|
|
21
|
-
"""
|
|
22
|
-
|
|
23
|
-
type: Literal[ResourceType.vector_store] = ResourceType.vector_store
|
|
24
|
-
|
|
25
|
-
embedding_model: str
|
|
26
|
-
embedding_dimension: int
|
|
27
|
-
vector_store_name: str | None = None
|
|
28
|
-
|
|
29
|
-
@property
|
|
30
|
-
def vector_store_id(self) -> str:
|
|
31
|
-
return self.identifier
|
|
32
|
-
|
|
33
|
-
@property
|
|
34
|
-
def provider_vector_store_id(self) -> str | None:
|
|
35
|
-
return self.provider_resource_id
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
class VectorStoreInput(BaseModel):
|
|
39
|
-
"""Input parameters for creating or configuring a vector database.
|
|
40
|
-
|
|
41
|
-
:param vector_store_id: Unique identifier for the vector store
|
|
42
|
-
:param embedding_model: Name of the embedding model to use for vector generation
|
|
43
|
-
:param embedding_dimension: Dimension of the embedding vectors
|
|
44
|
-
:param provider_vector_store_id: (Optional) Provider-specific identifier for the vector store
|
|
45
|
-
"""
|
|
46
|
-
|
|
47
|
-
vector_store_id: str
|
|
48
|
-
embedding_model: str
|
|
49
|
-
embedding_dimension: int
|
|
50
|
-
provider_id: str | None = None
|
|
51
|
-
provider_vector_store_id: str | None = None
|
|
@@ -1,9 +0,0 @@
|
|
|
1
|
-
# Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
2
|
-
# All rights reserved.
|
|
3
|
-
#
|
|
4
|
-
# This source code is licensed under the terms described in the LICENSE file in
|
|
5
|
-
# the root directory of this source tree.
|
|
6
|
-
|
|
7
|
-
LLAMA_STACK_API_V1 = "v1"
|
|
8
|
-
LLAMA_STACK_API_V1BETA = "v1beta"
|
|
9
|
-
LLAMA_STACK_API_V1ALPHA = "v1alpha"
|
llama_stack_api/models.py
DELETED
|
@@ -1,171 +0,0 @@
|
|
|
1
|
-
# Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
2
|
-
# All rights reserved.
|
|
3
|
-
#
|
|
4
|
-
# This source code is licensed under the terms described in the LICENSE file in
|
|
5
|
-
# the root directory of this source tree.
|
|
6
|
-
|
|
7
|
-
from enum import StrEnum
|
|
8
|
-
from typing import Any, Literal, Protocol, runtime_checkable
|
|
9
|
-
|
|
10
|
-
from pydantic import BaseModel, ConfigDict, Field, field_validator
|
|
11
|
-
|
|
12
|
-
from llama_stack_api.resource import Resource, ResourceType
|
|
13
|
-
from llama_stack_api.schema_utils import json_schema_type, webmethod
|
|
14
|
-
from llama_stack_api.version import LLAMA_STACK_API_V1
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
class CommonModelFields(BaseModel):
|
|
18
|
-
metadata: dict[str, Any] = Field(
|
|
19
|
-
default_factory=dict,
|
|
20
|
-
description="Any additional metadata for this model",
|
|
21
|
-
)
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
@json_schema_type
|
|
25
|
-
class ModelType(StrEnum):
|
|
26
|
-
"""Enumeration of supported model types in Llama Stack.
|
|
27
|
-
:cvar llm: Large language model for text generation and completion
|
|
28
|
-
:cvar embedding: Embedding model for converting text to vector representations
|
|
29
|
-
:cvar rerank: Reranking model for reordering documents based on their relevance to a query
|
|
30
|
-
"""
|
|
31
|
-
|
|
32
|
-
llm = "llm"
|
|
33
|
-
embedding = "embedding"
|
|
34
|
-
rerank = "rerank"
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
@json_schema_type
|
|
38
|
-
class Model(CommonModelFields, Resource):
|
|
39
|
-
"""A model resource representing an AI model registered in Llama Stack.
|
|
40
|
-
|
|
41
|
-
:param type: The resource type, always 'model' for model resources
|
|
42
|
-
:param model_type: The type of model (LLM or embedding model)
|
|
43
|
-
:param metadata: Any additional metadata for this model
|
|
44
|
-
:param identifier: Unique identifier for this resource in llama stack
|
|
45
|
-
:param provider_resource_id: Unique identifier for this resource in the provider
|
|
46
|
-
:param provider_id: ID of the provider that owns this resource
|
|
47
|
-
"""
|
|
48
|
-
|
|
49
|
-
type: Literal[ResourceType.model] = ResourceType.model
|
|
50
|
-
|
|
51
|
-
@property
|
|
52
|
-
def model_id(self) -> str:
|
|
53
|
-
return self.identifier
|
|
54
|
-
|
|
55
|
-
@property
|
|
56
|
-
def provider_model_id(self) -> str:
|
|
57
|
-
assert self.provider_resource_id is not None, "Provider resource ID must be set"
|
|
58
|
-
return self.provider_resource_id
|
|
59
|
-
|
|
60
|
-
model_config = ConfigDict(protected_namespaces=())
|
|
61
|
-
|
|
62
|
-
model_type: ModelType = Field(default=ModelType.llm)
|
|
63
|
-
|
|
64
|
-
@field_validator("provider_resource_id")
|
|
65
|
-
@classmethod
|
|
66
|
-
def validate_provider_resource_id(cls, v):
|
|
67
|
-
if v is None:
|
|
68
|
-
raise ValueError("provider_resource_id cannot be None")
|
|
69
|
-
return v
|
|
70
|
-
|
|
71
|
-
|
|
72
|
-
class ModelInput(CommonModelFields):
|
|
73
|
-
model_id: str
|
|
74
|
-
provider_id: str | None = None
|
|
75
|
-
provider_model_id: str | None = None
|
|
76
|
-
model_type: ModelType | None = ModelType.llm
|
|
77
|
-
model_config = ConfigDict(protected_namespaces=())
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
class ListModelsResponse(BaseModel):
|
|
81
|
-
data: list[Model]
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
@json_schema_type
|
|
85
|
-
class OpenAIModel(BaseModel):
|
|
86
|
-
"""A model from OpenAI.
|
|
87
|
-
|
|
88
|
-
:id: The ID of the model
|
|
89
|
-
:object: The object type, which will be "model"
|
|
90
|
-
:created: The Unix timestamp in seconds when the model was created
|
|
91
|
-
:owned_by: The owner of the model
|
|
92
|
-
:custom_metadata: Llama Stack-specific metadata including model_type, provider info, and additional metadata
|
|
93
|
-
"""
|
|
94
|
-
|
|
95
|
-
id: str
|
|
96
|
-
object: Literal["model"] = "model"
|
|
97
|
-
created: int
|
|
98
|
-
owned_by: str
|
|
99
|
-
custom_metadata: dict[str, Any] | None = None
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
@json_schema_type
|
|
103
|
-
class OpenAIListModelsResponse(BaseModel):
|
|
104
|
-
data: list[OpenAIModel]
|
|
105
|
-
|
|
106
|
-
|
|
107
|
-
@runtime_checkable
|
|
108
|
-
class Models(Protocol):
|
|
109
|
-
async def list_models(self) -> ListModelsResponse:
|
|
110
|
-
"""List all models.
|
|
111
|
-
|
|
112
|
-
:returns: A ListModelsResponse.
|
|
113
|
-
"""
|
|
114
|
-
...
|
|
115
|
-
|
|
116
|
-
@webmethod(route="/models", method="GET", level=LLAMA_STACK_API_V1)
|
|
117
|
-
async def openai_list_models(self) -> OpenAIListModelsResponse:
|
|
118
|
-
"""List models using the OpenAI API.
|
|
119
|
-
|
|
120
|
-
:returns: A OpenAIListModelsResponse.
|
|
121
|
-
"""
|
|
122
|
-
...
|
|
123
|
-
|
|
124
|
-
@webmethod(route="/models/{model_id:path}", method="GET", level=LLAMA_STACK_API_V1)
|
|
125
|
-
async def get_model(
|
|
126
|
-
self,
|
|
127
|
-
model_id: str,
|
|
128
|
-
) -> Model:
|
|
129
|
-
"""Get model.
|
|
130
|
-
|
|
131
|
-
Get a model by its identifier.
|
|
132
|
-
|
|
133
|
-
:param model_id: The identifier of the model to get.
|
|
134
|
-
:returns: A Model.
|
|
135
|
-
"""
|
|
136
|
-
...
|
|
137
|
-
|
|
138
|
-
@webmethod(route="/models", method="POST", level=LLAMA_STACK_API_V1, deprecated=True)
|
|
139
|
-
async def register_model(
|
|
140
|
-
self,
|
|
141
|
-
model_id: str,
|
|
142
|
-
provider_model_id: str | None = None,
|
|
143
|
-
provider_id: str | None = None,
|
|
144
|
-
metadata: dict[str, Any] | None = None,
|
|
145
|
-
model_type: ModelType | None = None,
|
|
146
|
-
) -> Model:
|
|
147
|
-
"""Register model.
|
|
148
|
-
|
|
149
|
-
Register a model.
|
|
150
|
-
|
|
151
|
-
:param model_id: The identifier of the model to register.
|
|
152
|
-
:param provider_model_id: The identifier of the model in the provider.
|
|
153
|
-
:param provider_id: The identifier of the provider.
|
|
154
|
-
:param metadata: Any additional metadata for this model.
|
|
155
|
-
:param model_type: The type of model to register.
|
|
156
|
-
:returns: A Model.
|
|
157
|
-
"""
|
|
158
|
-
...
|
|
159
|
-
|
|
160
|
-
@webmethod(route="/models/{model_id:path}", method="DELETE", level=LLAMA_STACK_API_V1, deprecated=True)
|
|
161
|
-
async def unregister_model(
|
|
162
|
-
self,
|
|
163
|
-
model_id: str,
|
|
164
|
-
) -> None:
|
|
165
|
-
"""Unregister model.
|
|
166
|
-
|
|
167
|
-
Unregister a model.
|
|
168
|
-
|
|
169
|
-
:param model_id: The identifier of the model to unregister.
|
|
170
|
-
"""
|
|
171
|
-
...
|