llama-stack-api 0.4.1__tar.gz → 0.4.3__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: llama-stack-api
3
- Version: 0.4.1
3
+ Version: 0.4.3
4
4
  Summary: API and Provider specifications for Llama Stack - lightweight package with protocol definitions and provider specs
5
5
  Author-email: Meta Llama <llama-oss@meta.com>
6
6
  License: MIT
@@ -15,6 +15,7 @@ Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
15
15
  Classifier: Topic :: Scientific/Engineering :: Information Analysis
16
16
  Requires-Python: >=3.12
17
17
  Description-Content-Type: text/markdown
18
+ Requires-Dist: openai>=2.5.0
18
19
  Requires-Dist: fastapi<1.0,>=0.115.0
19
20
  Requires-Dist: pydantic>=2.11.9
20
21
  Requires-Dist: jsonschema
@@ -7,7 +7,7 @@ required-version = ">=0.7.0"
7
7
 
8
8
  [project]
9
9
  name = "llama-stack-api"
10
- version = "0.4.1"
10
+ version = "0.4.3"
11
11
  authors = [{ name = "Meta Llama", email = "llama-oss@meta.com" }]
12
12
  description = "API and Provider specifications for Llama Stack - lightweight package with protocol definitions and provider specs"
13
13
  readme = "README.md"
@@ -24,6 +24,7 @@ classifiers = [
24
24
  "Topic :: Scientific/Engineering :: Information Analysis",
25
25
  ]
26
26
  dependencies = [
27
+ "openai>=2.5.0",
27
28
  "fastapi>=0.115.0,<1.0",
28
29
  "pydantic>=2.11.9",
29
30
  "jsonschema",
@@ -35,7 +36,7 @@ dependencies = [
35
36
  Homepage = "https://github.com/llamastack/llama-stack"
36
37
 
37
38
  [tool.setuptools.packages.find]
38
- where = ["."]
39
+ where = [".."]
39
40
  include = ["llama_stack_api", "llama_stack_api.*"]
40
41
 
41
42
  [tool.setuptools.package-data]
@@ -1,126 +0,0 @@
1
- Metadata-Version: 2.4
2
- Name: llama-stack-api
3
- Version: 0.4.1
4
- Summary: API and Provider specifications for Llama Stack - lightweight package with protocol definitions and provider specs
5
- Author-email: Meta Llama <llama-oss@meta.com>
6
- License: MIT
7
- Project-URL: Homepage, https://github.com/llamastack/llama-stack
8
- Classifier: License :: OSI Approved :: MIT License
9
- Classifier: Programming Language :: Python :: 3
10
- Classifier: Operating System :: OS Independent
11
- Classifier: Intended Audience :: Developers
12
- Classifier: Intended Audience :: Information Technology
13
- Classifier: Intended Audience :: Science/Research
14
- Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
15
- Classifier: Topic :: Scientific/Engineering :: Information Analysis
16
- Requires-Python: >=3.12
17
- Description-Content-Type: text/markdown
18
- Requires-Dist: fastapi<1.0,>=0.115.0
19
- Requires-Dist: pydantic>=2.11.9
20
- Requires-Dist: jsonschema
21
- Requires-Dist: opentelemetry-sdk>=1.30.0
22
- Requires-Dist: opentelemetry-exporter-otlp-proto-http>=1.30.0
23
-
24
- # llama-stack-api
25
-
26
- API and Provider specifications for Llama Stack - a lightweight package with protocol definitions and provider specs.
27
-
28
- ## Overview
29
-
30
- `llama-stack-api` is a minimal dependency package that contains:
31
-
32
- - **API Protocol Definitions**: Type-safe protocol definitions for all Llama Stack APIs (inference, agents, safety, etc.)
33
- - **Provider Specifications**: Provider spec definitions for building custom providers
34
- - **Data Types**: Shared data types and models used across the Llama Stack ecosystem
35
- - **Type Utilities**: Strong typing utilities and schema validation
36
-
37
- ## What This Package Does NOT Include
38
-
39
- - Server implementation (see `llama-stack` package)
40
- - Provider implementations (see `llama-stack` package)
41
- - CLI tools (see `llama-stack` package)
42
- - Runtime orchestration (see `llama-stack` package)
43
-
44
- ## Use Cases
45
-
46
- This package is designed for:
47
-
48
- 1. **Third-party Provider Developers**: Build custom providers without depending on the full Llama Stack server
49
- 2. **Client Library Authors**: Use type definitions without server dependencies
50
- 3. **Documentation Generation**: Generate API docs from protocol definitions
51
- 4. **Type Checking**: Validate implementations against the official specs
52
-
53
- ## Installation
54
-
55
- ```bash
56
- pip install llama-stack-api
57
- ```
58
-
59
- Or with uv:
60
-
61
- ```bash
62
- uv pip install llama-stack-api
63
- ```
64
-
65
- ## Dependencies
66
-
67
- Minimal dependencies:
68
- - `pydantic>=2.11.9` - For data validation and serialization
69
- - `jsonschema` - For JSON schema utilities
70
-
71
- ## Versioning
72
-
73
- This package follows semantic versioning independently from the main `llama-stack` package:
74
-
75
- - **Patch versions** (0.1.x): Documentation, internal improvements
76
- - **Minor versions** (0.x.0): New APIs, backward-compatible changes
77
- - **Major versions** (x.0.0): Breaking changes to existing APIs
78
-
79
- Current version: **0.4.0.dev0**
80
-
81
- ## Usage Example
82
-
83
- ```python
84
- from llama_stack_api.inference import Inference, ChatCompletionRequest
85
- from llama_stack_api.providers.datatypes import ProviderSpec, InlineProviderSpec
86
- from llama_stack_api.datatypes import Api
87
-
88
-
89
- # Use protocol definitions for type checking
90
- class MyInferenceProvider(Inference):
91
- async def chat_completion(self, request: ChatCompletionRequest):
92
- # Your implementation
93
- pass
94
-
95
-
96
- # Define provider specifications
97
- my_provider_spec = InlineProviderSpec(
98
- api=Api.inference,
99
- provider_type="inline::my-provider",
100
- pip_packages=["my-dependencies"],
101
- module="my_package.providers.inference",
102
- config_class="my_package.providers.inference.MyConfig",
103
- )
104
- ```
105
-
106
- ## Relationship to llama-stack
107
-
108
- The main `llama-stack` package depends on `llama-stack-api` and provides:
109
- - Full server implementation
110
- - Built-in provider implementations
111
- - CLI tools for running and managing stacks
112
- - Runtime provider resolution and orchestration
113
-
114
- ## Contributing
115
-
116
- See the main [Llama Stack repository](https://github.com/llamastack/llama-stack) for contribution guidelines.
117
-
118
- ## License
119
-
120
- MIT License - see LICENSE file for details.
121
-
122
- ## Links
123
-
124
- - [Main Llama Stack Repository](https://github.com/llamastack/llama-stack)
125
- - [Documentation](https://llamastack.ai/)
126
- - [Client Library](https://pypi.org/project/llama-stack-client/)
@@ -1,7 +0,0 @@
1
- README.md
2
- pyproject.toml
3
- llama_stack_api.egg-info/PKG-INFO
4
- llama_stack_api.egg-info/SOURCES.txt
5
- llama_stack_api.egg-info/dependency_links.txt
6
- llama_stack_api.egg-info/requires.txt
7
- llama_stack_api.egg-info/top_level.txt
@@ -1,5 +0,0 @@
1
- fastapi<1.0,>=0.115.0
2
- pydantic>=2.11.9
3
- jsonschema
4
- opentelemetry-sdk>=1.30.0
5
- opentelemetry-exporter-otlp-proto-http>=1.30.0