llama-stack 0.4.3__py3-none-any.whl → 0.4.4__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (166) hide show
  1. llama_stack/distributions/dell/doc_template.md +209 -0
  2. llama_stack/distributions/meta-reference-gpu/doc_template.md +119 -0
  3. llama_stack/distributions/nvidia/doc_template.md +170 -0
  4. llama_stack/distributions/oci/doc_template.md +140 -0
  5. llama_stack/models/llama/llama3/dog.jpg +0 -0
  6. llama_stack/models/llama/llama3/pasta.jpeg +0 -0
  7. llama_stack/models/llama/resources/dog.jpg +0 -0
  8. llama_stack/models/llama/resources/pasta.jpeg +0 -0
  9. llama_stack/models/llama/resources/small_dog.jpg +0 -0
  10. llama_stack/providers/inline/agents/meta_reference/responses/openai_responses.py +136 -11
  11. llama_stack/providers/inline/ios/inference/LocalInferenceImpl/LocalInference.h +9 -0
  12. llama_stack/providers/inline/ios/inference/LocalInferenceImpl/LocalInference.swift +189 -0
  13. llama_stack/providers/inline/ios/inference/LocalInferenceImpl/Parsing.swift +238 -0
  14. llama_stack/providers/inline/ios/inference/LocalInferenceImpl/PromptTemplate.swift +12 -0
  15. llama_stack/providers/inline/ios/inference/LocalInferenceImpl/SystemPrompts.swift +89 -0
  16. llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.pbxproj +550 -0
  17. llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.xcworkspace/contents.xcworkspacedata +7 -0
  18. llama_stack/providers/inline/ios/inference/LocalInferenceImpl.xcodeproj/project.xcworkspace/xcshareddata/IDEWorkspaceChecks.plist +8 -0
  19. llama_stack/providers/remote/datasetio/nvidia/README.md +74 -0
  20. llama_stack/providers/remote/eval/nvidia/README.md +134 -0
  21. llama_stack/providers/remote/files/s3/README.md +266 -0
  22. llama_stack/providers/remote/inference/nvidia/NVIDIA.md +203 -0
  23. llama_stack/providers/remote/post_training/nvidia/README.md +151 -0
  24. llama_stack/providers/remote/safety/nvidia/README.md +78 -0
  25. llama_stack/providers/utils/responses/responses_store.py +34 -0
  26. {llama_stack-0.4.3.dist-info → llama_stack-0.4.4.dist-info}/METADATA +2 -2
  27. {llama_stack-0.4.3.dist-info → llama_stack-0.4.4.dist-info}/RECORD +31 -142
  28. llama_stack-0.4.4.dist-info/top_level.txt +1 -0
  29. llama_stack-0.4.3.dist-info/top_level.txt +0 -2
  30. llama_stack_api/__init__.py +0 -945
  31. llama_stack_api/admin/__init__.py +0 -45
  32. llama_stack_api/admin/api.py +0 -72
  33. llama_stack_api/admin/fastapi_routes.py +0 -117
  34. llama_stack_api/admin/models.py +0 -113
  35. llama_stack_api/agents.py +0 -173
  36. llama_stack_api/batches/__init__.py +0 -40
  37. llama_stack_api/batches/api.py +0 -53
  38. llama_stack_api/batches/fastapi_routes.py +0 -113
  39. llama_stack_api/batches/models.py +0 -78
  40. llama_stack_api/benchmarks/__init__.py +0 -43
  41. llama_stack_api/benchmarks/api.py +0 -39
  42. llama_stack_api/benchmarks/fastapi_routes.py +0 -109
  43. llama_stack_api/benchmarks/models.py +0 -109
  44. llama_stack_api/common/__init__.py +0 -5
  45. llama_stack_api/common/content_types.py +0 -101
  46. llama_stack_api/common/errors.py +0 -95
  47. llama_stack_api/common/job_types.py +0 -38
  48. llama_stack_api/common/responses.py +0 -77
  49. llama_stack_api/common/training_types.py +0 -47
  50. llama_stack_api/common/type_system.py +0 -146
  51. llama_stack_api/connectors.py +0 -146
  52. llama_stack_api/conversations.py +0 -270
  53. llama_stack_api/datasetio.py +0 -55
  54. llama_stack_api/datasets/__init__.py +0 -61
  55. llama_stack_api/datasets/api.py +0 -35
  56. llama_stack_api/datasets/fastapi_routes.py +0 -104
  57. llama_stack_api/datasets/models.py +0 -152
  58. llama_stack_api/datatypes.py +0 -373
  59. llama_stack_api/eval.py +0 -137
  60. llama_stack_api/file_processors/__init__.py +0 -27
  61. llama_stack_api/file_processors/api.py +0 -64
  62. llama_stack_api/file_processors/fastapi_routes.py +0 -78
  63. llama_stack_api/file_processors/models.py +0 -42
  64. llama_stack_api/files/__init__.py +0 -35
  65. llama_stack_api/files/api.py +0 -51
  66. llama_stack_api/files/fastapi_routes.py +0 -124
  67. llama_stack_api/files/models.py +0 -107
  68. llama_stack_api/inference.py +0 -1169
  69. llama_stack_api/inspect_api/__init__.py +0 -37
  70. llama_stack_api/inspect_api/api.py +0 -25
  71. llama_stack_api/inspect_api/fastapi_routes.py +0 -76
  72. llama_stack_api/inspect_api/models.py +0 -28
  73. llama_stack_api/internal/__init__.py +0 -9
  74. llama_stack_api/internal/kvstore.py +0 -28
  75. llama_stack_api/internal/sqlstore.py +0 -81
  76. llama_stack_api/llama_stack_api/__init__.py +0 -945
  77. llama_stack_api/llama_stack_api/admin/__init__.py +0 -45
  78. llama_stack_api/llama_stack_api/admin/api.py +0 -72
  79. llama_stack_api/llama_stack_api/admin/fastapi_routes.py +0 -117
  80. llama_stack_api/llama_stack_api/admin/models.py +0 -113
  81. llama_stack_api/llama_stack_api/agents.py +0 -173
  82. llama_stack_api/llama_stack_api/batches/__init__.py +0 -40
  83. llama_stack_api/llama_stack_api/batches/api.py +0 -53
  84. llama_stack_api/llama_stack_api/batches/fastapi_routes.py +0 -113
  85. llama_stack_api/llama_stack_api/batches/models.py +0 -78
  86. llama_stack_api/llama_stack_api/benchmarks/__init__.py +0 -43
  87. llama_stack_api/llama_stack_api/benchmarks/api.py +0 -39
  88. llama_stack_api/llama_stack_api/benchmarks/fastapi_routes.py +0 -109
  89. llama_stack_api/llama_stack_api/benchmarks/models.py +0 -109
  90. llama_stack_api/llama_stack_api/common/__init__.py +0 -5
  91. llama_stack_api/llama_stack_api/common/content_types.py +0 -101
  92. llama_stack_api/llama_stack_api/common/errors.py +0 -95
  93. llama_stack_api/llama_stack_api/common/job_types.py +0 -38
  94. llama_stack_api/llama_stack_api/common/responses.py +0 -77
  95. llama_stack_api/llama_stack_api/common/training_types.py +0 -47
  96. llama_stack_api/llama_stack_api/common/type_system.py +0 -146
  97. llama_stack_api/llama_stack_api/connectors.py +0 -146
  98. llama_stack_api/llama_stack_api/conversations.py +0 -270
  99. llama_stack_api/llama_stack_api/datasetio.py +0 -55
  100. llama_stack_api/llama_stack_api/datasets/__init__.py +0 -61
  101. llama_stack_api/llama_stack_api/datasets/api.py +0 -35
  102. llama_stack_api/llama_stack_api/datasets/fastapi_routes.py +0 -104
  103. llama_stack_api/llama_stack_api/datasets/models.py +0 -152
  104. llama_stack_api/llama_stack_api/datatypes.py +0 -373
  105. llama_stack_api/llama_stack_api/eval.py +0 -137
  106. llama_stack_api/llama_stack_api/file_processors/__init__.py +0 -27
  107. llama_stack_api/llama_stack_api/file_processors/api.py +0 -64
  108. llama_stack_api/llama_stack_api/file_processors/fastapi_routes.py +0 -78
  109. llama_stack_api/llama_stack_api/file_processors/models.py +0 -42
  110. llama_stack_api/llama_stack_api/files/__init__.py +0 -35
  111. llama_stack_api/llama_stack_api/files/api.py +0 -51
  112. llama_stack_api/llama_stack_api/files/fastapi_routes.py +0 -124
  113. llama_stack_api/llama_stack_api/files/models.py +0 -107
  114. llama_stack_api/llama_stack_api/inference.py +0 -1169
  115. llama_stack_api/llama_stack_api/inspect_api/__init__.py +0 -37
  116. llama_stack_api/llama_stack_api/inspect_api/api.py +0 -25
  117. llama_stack_api/llama_stack_api/inspect_api/fastapi_routes.py +0 -76
  118. llama_stack_api/llama_stack_api/inspect_api/models.py +0 -28
  119. llama_stack_api/llama_stack_api/internal/__init__.py +0 -9
  120. llama_stack_api/llama_stack_api/internal/kvstore.py +0 -28
  121. llama_stack_api/llama_stack_api/internal/sqlstore.py +0 -81
  122. llama_stack_api/llama_stack_api/models.py +0 -171
  123. llama_stack_api/llama_stack_api/openai_responses.py +0 -1468
  124. llama_stack_api/llama_stack_api/post_training.py +0 -370
  125. llama_stack_api/llama_stack_api/prompts.py +0 -203
  126. llama_stack_api/llama_stack_api/providers/__init__.py +0 -33
  127. llama_stack_api/llama_stack_api/providers/api.py +0 -16
  128. llama_stack_api/llama_stack_api/providers/fastapi_routes.py +0 -57
  129. llama_stack_api/llama_stack_api/providers/models.py +0 -24
  130. llama_stack_api/llama_stack_api/py.typed +0 -0
  131. llama_stack_api/llama_stack_api/rag_tool.py +0 -168
  132. llama_stack_api/llama_stack_api/resource.py +0 -37
  133. llama_stack_api/llama_stack_api/router_utils.py +0 -160
  134. llama_stack_api/llama_stack_api/safety.py +0 -132
  135. llama_stack_api/llama_stack_api/schema_utils.py +0 -208
  136. llama_stack_api/llama_stack_api/scoring.py +0 -93
  137. llama_stack_api/llama_stack_api/scoring_functions.py +0 -211
  138. llama_stack_api/llama_stack_api/shields.py +0 -93
  139. llama_stack_api/llama_stack_api/tools.py +0 -226
  140. llama_stack_api/llama_stack_api/vector_io.py +0 -941
  141. llama_stack_api/llama_stack_api/vector_stores.py +0 -53
  142. llama_stack_api/llama_stack_api/version.py +0 -9
  143. llama_stack_api/models.py +0 -171
  144. llama_stack_api/openai_responses.py +0 -1468
  145. llama_stack_api/post_training.py +0 -370
  146. llama_stack_api/prompts.py +0 -203
  147. llama_stack_api/providers/__init__.py +0 -33
  148. llama_stack_api/providers/api.py +0 -16
  149. llama_stack_api/providers/fastapi_routes.py +0 -57
  150. llama_stack_api/providers/models.py +0 -24
  151. llama_stack_api/py.typed +0 -0
  152. llama_stack_api/rag_tool.py +0 -168
  153. llama_stack_api/resource.py +0 -37
  154. llama_stack_api/router_utils.py +0 -160
  155. llama_stack_api/safety.py +0 -132
  156. llama_stack_api/schema_utils.py +0 -208
  157. llama_stack_api/scoring.py +0 -93
  158. llama_stack_api/scoring_functions.py +0 -211
  159. llama_stack_api/shields.py +0 -93
  160. llama_stack_api/tools.py +0 -226
  161. llama_stack_api/vector_io.py +0 -941
  162. llama_stack_api/vector_stores.py +0 -53
  163. llama_stack_api/version.py +0 -9
  164. {llama_stack-0.4.3.dist-info → llama_stack-0.4.4.dist-info}/WHEEL +0 -0
  165. {llama_stack-0.4.3.dist-info → llama_stack-0.4.4.dist-info}/entry_points.txt +0 -0
  166. {llama_stack-0.4.3.dist-info → llama_stack-0.4.4.dist-info}/licenses/LICENSE +0 -0
@@ -1,53 +0,0 @@
1
- # Copyright (c) Meta Platforms, Inc. and affiliates.
2
- # All rights reserved.
3
- #
4
- # This source code is licensed under the terms described in the LICENSE file in
5
- # the root directory of this source tree.
6
-
7
- from typing import Literal
8
-
9
- from pydantic import BaseModel
10
-
11
- from llama_stack_api.resource import Resource, ResourceType
12
-
13
-
14
- # Internal resource type for storing the vector store routing and other information
15
- class VectorStore(Resource):
16
- """Vector database resource for storing and querying vector embeddings.
17
-
18
- :param type: Type of resource, always 'vector_store' for vector stores
19
- :param embedding_model: Name of the embedding model to use for vector generation
20
- :param embedding_dimension: Dimension of the embedding vectors
21
- """
22
-
23
- type: Literal[ResourceType.vector_store] = ResourceType.vector_store
24
-
25
- embedding_model: str
26
- embedding_dimension: int
27
- vector_store_name: str | None = None
28
-
29
- @property
30
- def vector_store_id(self) -> str:
31
- return self.identifier
32
-
33
- @property
34
- def provider_vector_store_id(self) -> str | None:
35
- return self.provider_resource_id
36
-
37
-
38
- class VectorStoreInput(BaseModel):
39
- """Input parameters for creating or configuring a vector database.
40
-
41
- :param vector_store_id: Unique identifier for the vector store
42
- :param embedding_model: Name of the embedding model to use for vector generation
43
- :param embedding_dimension: Dimension of the embedding vectors
44
- :param provider_vector_store_id: (Optional) Provider-specific identifier for the vector store
45
- :param vector_store_name: (Optional) Human-readable name for the vector store
46
- """
47
-
48
- vector_store_id: str
49
- embedding_model: str
50
- embedding_dimension: int
51
- provider_id: str | None = None
52
- provider_vector_store_id: str | None = None
53
- vector_store_name: str | None = None
@@ -1,9 +0,0 @@
1
- # Copyright (c) Meta Platforms, Inc. and affiliates.
2
- # All rights reserved.
3
- #
4
- # This source code is licensed under the terms described in the LICENSE file in
5
- # the root directory of this source tree.
6
-
7
- LLAMA_STACK_API_V1 = "v1"
8
- LLAMA_STACK_API_V1BETA = "v1beta"
9
- LLAMA_STACK_API_V1ALPHA = "v1alpha"
llama_stack_api/models.py DELETED
@@ -1,171 +0,0 @@
1
- # Copyright (c) Meta Platforms, Inc. and affiliates.
2
- # All rights reserved.
3
- #
4
- # This source code is licensed under the terms described in the LICENSE file in
5
- # the root directory of this source tree.
6
-
7
- from enum import StrEnum
8
- from typing import Any, Literal, Protocol, runtime_checkable
9
-
10
- from pydantic import BaseModel, ConfigDict, Field, field_validator
11
-
12
- from llama_stack_api.resource import Resource, ResourceType
13
- from llama_stack_api.schema_utils import json_schema_type, webmethod
14
- from llama_stack_api.version import LLAMA_STACK_API_V1
15
-
16
-
17
- class CommonModelFields(BaseModel):
18
- metadata: dict[str, Any] = Field(
19
- default_factory=dict,
20
- description="Any additional metadata for this model",
21
- )
22
-
23
-
24
- @json_schema_type
25
- class ModelType(StrEnum):
26
- """Enumeration of supported model types in Llama Stack.
27
- :cvar llm: Large language model for text generation and completion
28
- :cvar embedding: Embedding model for converting text to vector representations
29
- :cvar rerank: Reranking model for reordering documents based on their relevance to a query
30
- """
31
-
32
- llm = "llm"
33
- embedding = "embedding"
34
- rerank = "rerank"
35
-
36
-
37
- @json_schema_type
38
- class Model(CommonModelFields, Resource):
39
- """A model resource representing an AI model registered in Llama Stack.
40
-
41
- :param type: The resource type, always 'model' for model resources
42
- :param model_type: The type of model (LLM or embedding model)
43
- :param metadata: Any additional metadata for this model
44
- :param identifier: Unique identifier for this resource in llama stack
45
- :param provider_resource_id: Unique identifier for this resource in the provider
46
- :param provider_id: ID of the provider that owns this resource
47
- """
48
-
49
- type: Literal[ResourceType.model] = ResourceType.model
50
-
51
- @property
52
- def model_id(self) -> str:
53
- return self.identifier
54
-
55
- @property
56
- def provider_model_id(self) -> str:
57
- assert self.provider_resource_id is not None, "Provider resource ID must be set"
58
- return self.provider_resource_id
59
-
60
- model_config = ConfigDict(protected_namespaces=())
61
-
62
- model_type: ModelType = Field(default=ModelType.llm)
63
-
64
- @field_validator("provider_resource_id")
65
- @classmethod
66
- def validate_provider_resource_id(cls, v):
67
- if v is None:
68
- raise ValueError("provider_resource_id cannot be None")
69
- return v
70
-
71
-
72
- class ModelInput(CommonModelFields):
73
- model_id: str
74
- provider_id: str | None = None
75
- provider_model_id: str | None = None
76
- model_type: ModelType | None = ModelType.llm
77
- model_config = ConfigDict(protected_namespaces=())
78
-
79
-
80
- class ListModelsResponse(BaseModel):
81
- data: list[Model]
82
-
83
-
84
- @json_schema_type
85
- class OpenAIModel(BaseModel):
86
- """A model from OpenAI.
87
-
88
- :id: The ID of the model
89
- :object: The object type, which will be "model"
90
- :created: The Unix timestamp in seconds when the model was created
91
- :owned_by: The owner of the model
92
- :custom_metadata: Llama Stack-specific metadata including model_type, provider info, and additional metadata
93
- """
94
-
95
- id: str
96
- object: Literal["model"] = "model"
97
- created: int
98
- owned_by: str
99
- custom_metadata: dict[str, Any] | None = None
100
-
101
-
102
- @json_schema_type
103
- class OpenAIListModelsResponse(BaseModel):
104
- data: list[OpenAIModel]
105
-
106
-
107
- @runtime_checkable
108
- class Models(Protocol):
109
- async def list_models(self) -> ListModelsResponse:
110
- """List all models.
111
-
112
- :returns: A ListModelsResponse.
113
- """
114
- ...
115
-
116
- @webmethod(route="/models", method="GET", level=LLAMA_STACK_API_V1)
117
- async def openai_list_models(self) -> OpenAIListModelsResponse:
118
- """List models using the OpenAI API.
119
-
120
- :returns: A OpenAIListModelsResponse.
121
- """
122
- ...
123
-
124
- @webmethod(route="/models/{model_id:path}", method="GET", level=LLAMA_STACK_API_V1)
125
- async def get_model(
126
- self,
127
- model_id: str,
128
- ) -> Model:
129
- """Get model.
130
-
131
- Get a model by its identifier.
132
-
133
- :param model_id: The identifier of the model to get.
134
- :returns: A Model.
135
- """
136
- ...
137
-
138
- @webmethod(route="/models", method="POST", level=LLAMA_STACK_API_V1, deprecated=True)
139
- async def register_model(
140
- self,
141
- model_id: str,
142
- provider_model_id: str | None = None,
143
- provider_id: str | None = None,
144
- metadata: dict[str, Any] | None = None,
145
- model_type: ModelType | None = None,
146
- ) -> Model:
147
- """Register model.
148
-
149
- Register a model.
150
-
151
- :param model_id: The identifier of the model to register.
152
- :param provider_model_id: The identifier of the model in the provider.
153
- :param provider_id: The identifier of the provider.
154
- :param metadata: Any additional metadata for this model.
155
- :param model_type: The type of model to register.
156
- :returns: A Model.
157
- """
158
- ...
159
-
160
- @webmethod(route="/models/{model_id:path}", method="DELETE", level=LLAMA_STACK_API_V1, deprecated=True)
161
- async def unregister_model(
162
- self,
163
- model_id: str,
164
- ) -> None:
165
- """Unregister model.
166
-
167
- Unregister a model.
168
-
169
- :param model_id: The identifier of the model to unregister.
170
- """
171
- ...