arize-phoenix 11.20.0__py3-none-any.whl → 11.21.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of arize-phoenix might be problematic. Click here for more details.
- {arize_phoenix-11.20.0.dist-info → arize_phoenix-11.21.0.dist-info}/METADATA +1 -1
- {arize_phoenix-11.20.0.dist-info → arize_phoenix-11.21.0.dist-info}/RECORD +11 -11
- phoenix/server/api/helpers/playground_clients.py +4 -44
- phoenix/server/api/helpers/prompts/models.py +1 -1
- phoenix/server/api/input_types/PromptVersionInput.py +5 -0
- phoenix/server/cost_tracking/model_cost_manifest.json +264 -0
- phoenix/version.py +1 -1
- {arize_phoenix-11.20.0.dist-info → arize_phoenix-11.21.0.dist-info}/WHEEL +0 -0
- {arize_phoenix-11.20.0.dist-info → arize_phoenix-11.21.0.dist-info}/entry_points.txt +0 -0
- {arize_phoenix-11.20.0.dist-info → arize_phoenix-11.21.0.dist-info}/licenses/IP_NOTICE +0 -0
- {arize_phoenix-11.20.0.dist-info → arize_phoenix-11.21.0.dist-info}/licenses/LICENSE +0 -0
|
@@ -6,7 +6,7 @@ phoenix/exceptions.py,sha256=n2L2KKuecrdflB9MsCdAYCiSEvGJptIsfRkXMoJle7A,169
|
|
|
6
6
|
phoenix/py.typed,sha256=AbpHGcgLb-kRsJGnwFEktk7uzpZOCcBY74-YBdrKVGs,1
|
|
7
7
|
phoenix/services.py,sha256=ngkyKGVatX3cO2WJdo2hKdaVKP-xJCMvqthvga6kJss,5196
|
|
8
8
|
phoenix/settings.py,sha256=2kHfT3BNOVd4dAO1bq-syEQbHSG8oX2-7NhOwK2QREk,896
|
|
9
|
-
phoenix/version.py,sha256=
|
|
9
|
+
phoenix/version.py,sha256=S2TiwHQgEbsL_ldKr5oOsMKO-n7_zvhjBACaqir74do,24
|
|
10
10
|
phoenix/core/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
11
11
|
phoenix/core/embedding_dimension.py,sha256=zKGbcvwOXgLf-yrJBpQyKtd-LEOPRKHnUToyAU8Owis,87
|
|
12
12
|
phoenix/core/model.py,sha256=qBFraOtmwCCnWJltKNP18DDG0mULXigytlFsa6YOz6k,4837
|
|
@@ -176,11 +176,11 @@ phoenix/server/api/helpers/__init__.py,sha256=m2-xaSPqUiSs91k62JaRDjFNfl-1byxBfY
|
|
|
176
176
|
phoenix/server/api/helpers/annotations.py,sha256=9gMXKpMTfWEChoSCnvdWYuyB0hlSnNOp-qUdar9Vono,262
|
|
177
177
|
phoenix/server/api/helpers/dataset_helpers.py,sha256=3bdGBoUzqrtg-sr5p2wpQLOU6dhg_3TKFHNeJj8p0TU,9155
|
|
178
178
|
phoenix/server/api/helpers/experiment_run_filters.py,sha256=DOnVwrmn39eAkk2mwuZP8kIcAnR5jrOgllEwWSjsw94,29893
|
|
179
|
-
phoenix/server/api/helpers/playground_clients.py,sha256=
|
|
179
|
+
phoenix/server/api/helpers/playground_clients.py,sha256=R_lXbm58ejnsdHMvlbHbrasC7XsaLfqh5pZoHaL3DSg,72129
|
|
180
180
|
phoenix/server/api/helpers/playground_registry.py,sha256=n0v4-KnvZJxeaEwOla5qBbnOQjSWznKmMhZnh9ziJt0,2584
|
|
181
181
|
phoenix/server/api/helpers/playground_spans.py,sha256=QpXwPl_fFNwm_iA1A77XApUyXMl1aDmonw8aXuNZ_4k,17132
|
|
182
182
|
phoenix/server/api/helpers/prompts/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
183
|
-
phoenix/server/api/helpers/prompts/models.py,sha256=
|
|
183
|
+
phoenix/server/api/helpers/prompts/models.py,sha256=pEPmloU27j5c4dM0dbDSAeIAzxpfsjPtjNRdF99mlhM,23558
|
|
184
184
|
phoenix/server/api/helpers/prompts/conversions/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
185
185
|
phoenix/server/api/helpers/prompts/conversions/anthropic.py,sha256=ZT--UqBwoGf7QMusajB6aeB7zyWGttaZigb113kgiY8,3571
|
|
186
186
|
phoenix/server/api/helpers/prompts/conversions/aws.py,sha256=6vaT8K13r0bMXB9XHA8qY1MCjVsmR9TO0VIwyBMjQoY,2941
|
|
@@ -220,7 +220,7 @@ phoenix/server/api/input_types/ProjectSessionSort.py,sha256=KZzEtOMpcxtP11brL4fX
|
|
|
220
220
|
phoenix/server/api/input_types/ProjectSort.py,sha256=ZTT-InFB6NvInDwLuivyHe9PxR5nsmQ8aXCHAPjZm9k,329
|
|
221
221
|
phoenix/server/api/input_types/PromptFilter.py,sha256=f2F7fDlYRsNJp_rKkmvpgUt9rfgr_e-dyZxuHX8YfkU,256
|
|
222
222
|
phoenix/server/api/input_types/PromptTemplateOptions.py,sha256=8ZJdH1F9fExcdH9dF8SJ29WycCvtEpK-Z6dZwFO7KgQ,232
|
|
223
|
-
phoenix/server/api/input_types/PromptVersionInput.py,sha256=
|
|
223
|
+
phoenix/server/api/input_types/PromptVersionInput.py,sha256=6iFWf2Ye9K1dwL1810L-wUTLkMtOIHemJTxar4xF4kw,3911
|
|
224
224
|
phoenix/server/api/input_types/SpanAnnotationFilter.py,sha256=-djfIXYCxV6sV3GPOZQUV0SPfiWDhRlTORfeQ7tCBgQ,2671
|
|
225
225
|
phoenix/server/api/input_types/SpanAnnotationSort.py,sha256=T5pAGzmh4MiJp9JMAzNDByFVTczfw02FH4WFWwFezyI,361
|
|
226
226
|
phoenix/server/api/input_types/SpanSort.py,sha256=GReQx9yOo0Kehi2y4AtY69aZhRtcqvcg-9bSIFru69U,7540
|
|
@@ -362,7 +362,7 @@ phoenix/server/cost_tracking/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NM
|
|
|
362
362
|
phoenix/server/cost_tracking/cost_details_calculator.py,sha256=Tt0YcuLhgPuXKWJemWVmYQfG0xQUvH4VziIj6KcDnoA,8945
|
|
363
363
|
phoenix/server/cost_tracking/cost_model_lookup.py,sha256=jhtVdnQBzrTUHeOGPWgOebk-Io5hpJ1vAgWOu8ojeJ4,6801
|
|
364
364
|
phoenix/server/cost_tracking/helpers.py,sha256=Pk6ECjnYreTxrldtRwxnwFcxIPVsvDq_yAwDA_spkOc,2122
|
|
365
|
-
phoenix/server/cost_tracking/model_cost_manifest.json,sha256=
|
|
365
|
+
phoenix/server/cost_tracking/model_cost_manifest.json,sha256=9y3c0YrYSYTZvwerCr6yAQD97YBTZ6GUZ4UykDvwfFU,63485
|
|
366
366
|
phoenix/server/cost_tracking/regex_specificity.py,sha256=9kqWuQ68C-hlwW25hr7BhFlRt5y2Nnpy0Ax3n9UN6Xk,11622
|
|
367
367
|
phoenix/server/cost_tracking/token_cost_calculator.py,sha256=2JEZnvusx2-xbhp8krp9EarjWuyGH2KO4e-ZwJX-K0s,1598
|
|
368
368
|
phoenix/server/daemons/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
@@ -441,9 +441,9 @@ phoenix/utilities/project.py,sha256=auVpARXkDb-JgeX5f2aStyFIkeKvGwN9l7qrFeJMVxI,
|
|
|
441
441
|
phoenix/utilities/re.py,sha256=6YyUWIkv0zc2SigsxfOWIHzdpjKA_TZo2iqKq7zJKvw,2081
|
|
442
442
|
phoenix/utilities/span_store.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
443
443
|
phoenix/utilities/template_formatters.py,sha256=gh9PJD6WEGw7TEYXfSst1UR4pWWwmjxMLrDVQ_CkpkQ,2779
|
|
444
|
-
arize_phoenix-11.
|
|
445
|
-
arize_phoenix-11.
|
|
446
|
-
arize_phoenix-11.
|
|
447
|
-
arize_phoenix-11.
|
|
448
|
-
arize_phoenix-11.
|
|
449
|
-
arize_phoenix-11.
|
|
444
|
+
arize_phoenix-11.21.0.dist-info/METADATA,sha256=YdUThYDa8tZXg7RRzkuImDvvvGjvUtQFuv5tTUzHDWc,30950
|
|
445
|
+
arize_phoenix-11.21.0.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
|
|
446
|
+
arize_phoenix-11.21.0.dist-info/entry_points.txt,sha256=Pgpn8Upxx9P8z8joPXZWl2LlnAlGc3gcQoVchb06X1Q,94
|
|
447
|
+
arize_phoenix-11.21.0.dist-info/licenses/IP_NOTICE,sha256=JBqyyCYYxGDfzQ0TtsQgjts41IJoa-hiwDrBjCb9gHM,469
|
|
448
|
+
arize_phoenix-11.21.0.dist-info/licenses/LICENSE,sha256=HFkW9REuMOkvKRACuwLPT0hRydHb3zNg-fdFt94td18,3794
|
|
449
|
+
arize_phoenix-11.21.0.dist-info/RECORD,,
|
|
@@ -1141,6 +1141,10 @@ class OpenAIStreamingClient(OpenAIBaseStreamingClient):
|
|
|
1141
1141
|
|
|
1142
1142
|
|
|
1143
1143
|
_OPENAI_REASONING_MODELS = [
|
|
1144
|
+
"gpt-5",
|
|
1145
|
+
"gpt-5-mini",
|
|
1146
|
+
"gpt-5-nano",
|
|
1147
|
+
"gpt-5-chat-latest",
|
|
1144
1148
|
"o1",
|
|
1145
1149
|
"o1-pro",
|
|
1146
1150
|
"o1-2024-12-17",
|
|
@@ -1201,50 +1205,6 @@ class OpenAIReasoningNonStreamingClient(
|
|
|
1201
1205
|
OpenAIReasoningReasoningModelsMixin,
|
|
1202
1206
|
OpenAIStreamingClient,
|
|
1203
1207
|
):
|
|
1204
|
-
@override
|
|
1205
|
-
async def chat_completion_create(
|
|
1206
|
-
self,
|
|
1207
|
-
messages: list[
|
|
1208
|
-
tuple[ChatCompletionMessageRole, str, Optional[str], Optional[list[JSONScalarType]]]
|
|
1209
|
-
],
|
|
1210
|
-
tools: list[JSONScalarType],
|
|
1211
|
-
**invocation_parameters: Any,
|
|
1212
|
-
) -> AsyncIterator[ChatCompletionChunk]:
|
|
1213
|
-
from openai import NOT_GIVEN
|
|
1214
|
-
|
|
1215
|
-
# Convert standard messages to OpenAI messages
|
|
1216
|
-
openai_messages = []
|
|
1217
|
-
for message in messages:
|
|
1218
|
-
openai_message = self.to_openai_chat_completion_param(*message)
|
|
1219
|
-
if openai_message is not None:
|
|
1220
|
-
openai_messages.append(openai_message)
|
|
1221
|
-
|
|
1222
|
-
throttled_create = self.rate_limiter._alimit(self.client.chat.completions.create)
|
|
1223
|
-
response = await throttled_create(
|
|
1224
|
-
messages=openai_messages,
|
|
1225
|
-
model=self.model_name,
|
|
1226
|
-
stream=False,
|
|
1227
|
-
tools=tools or NOT_GIVEN,
|
|
1228
|
-
**invocation_parameters,
|
|
1229
|
-
)
|
|
1230
|
-
|
|
1231
|
-
if response.usage is not None:
|
|
1232
|
-
self._attributes.update(dict(self._llm_token_counts(response.usage)))
|
|
1233
|
-
|
|
1234
|
-
choice = response.choices[0]
|
|
1235
|
-
if choice.message.content:
|
|
1236
|
-
yield TextChunk(content=choice.message.content)
|
|
1237
|
-
|
|
1238
|
-
if choice.message.tool_calls:
|
|
1239
|
-
for tool_call in choice.message.tool_calls:
|
|
1240
|
-
yield ToolCallChunk(
|
|
1241
|
-
id=tool_call.id,
|
|
1242
|
-
function=FunctionCallChunk(
|
|
1243
|
-
name=tool_call.function.name,
|
|
1244
|
-
arguments=tool_call.function.arguments,
|
|
1245
|
-
),
|
|
1246
|
-
)
|
|
1247
|
-
|
|
1248
1208
|
def to_openai_chat_completion_param(
|
|
1249
1209
|
self,
|
|
1250
1210
|
role: ChatCompletionMessageRole,
|
|
@@ -329,7 +329,7 @@ class PromptOpenAIInvocationParametersContent(DBBaseModel):
|
|
|
329
329
|
presence_penalty: float = UNDEFINED
|
|
330
330
|
top_p: float = UNDEFINED
|
|
331
331
|
seed: int = UNDEFINED
|
|
332
|
-
reasoning_effort: Literal["low", "medium", "high"] = UNDEFINED
|
|
332
|
+
reasoning_effort: Literal["minimal", "low", "medium", "high"] = UNDEFINED
|
|
333
333
|
|
|
334
334
|
|
|
335
335
|
class PromptOpenAIInvocationParameters(DBBaseModel):
|
|
@@ -83,6 +83,11 @@ class ChatPromptVersionInput:
|
|
|
83
83
|
model_provider: ModelProvider
|
|
84
84
|
model_name: str
|
|
85
85
|
|
|
86
|
+
def __post_init__(self) -> None:
|
|
87
|
+
self.invocation_parameters = {
|
|
88
|
+
k: v for k, v in self.invocation_parameters.items() if v is not None
|
|
89
|
+
}
|
|
90
|
+
|
|
86
91
|
|
|
87
92
|
def to_pydantic_prompt_chat_template_v1(
|
|
88
93
|
prompt_chat_template_input: PromptChatTemplateInput,
|
|
@@ -341,6 +341,60 @@
|
|
|
341
341
|
}
|
|
342
342
|
]
|
|
343
343
|
},
|
|
344
|
+
{
|
|
345
|
+
"name": "claude-opus-4-1",
|
|
346
|
+
"name_pattern": "claude-opus-4-1",
|
|
347
|
+
"source": "litellm",
|
|
348
|
+
"token_prices": [
|
|
349
|
+
{
|
|
350
|
+
"base_rate": 0.000015,
|
|
351
|
+
"is_prompt": true,
|
|
352
|
+
"token_type": "input"
|
|
353
|
+
},
|
|
354
|
+
{
|
|
355
|
+
"base_rate": 0.000075,
|
|
356
|
+
"is_prompt": false,
|
|
357
|
+
"token_type": "output"
|
|
358
|
+
},
|
|
359
|
+
{
|
|
360
|
+
"base_rate": 1.5e-6,
|
|
361
|
+
"is_prompt": true,
|
|
362
|
+
"token_type": "cache_read"
|
|
363
|
+
},
|
|
364
|
+
{
|
|
365
|
+
"base_rate": 0.00001875,
|
|
366
|
+
"is_prompt": true,
|
|
367
|
+
"token_type": "cache_write"
|
|
368
|
+
}
|
|
369
|
+
]
|
|
370
|
+
},
|
|
371
|
+
{
|
|
372
|
+
"name": "claude-opus-4-1-20250805",
|
|
373
|
+
"name_pattern": "claude-opus-4-1-20250805",
|
|
374
|
+
"source": "litellm",
|
|
375
|
+
"token_prices": [
|
|
376
|
+
{
|
|
377
|
+
"base_rate": 0.000015,
|
|
378
|
+
"is_prompt": true,
|
|
379
|
+
"token_type": "input"
|
|
380
|
+
},
|
|
381
|
+
{
|
|
382
|
+
"base_rate": 0.000075,
|
|
383
|
+
"is_prompt": false,
|
|
384
|
+
"token_type": "output"
|
|
385
|
+
},
|
|
386
|
+
{
|
|
387
|
+
"base_rate": 1.5e-6,
|
|
388
|
+
"is_prompt": true,
|
|
389
|
+
"token_type": "cache_read"
|
|
390
|
+
},
|
|
391
|
+
{
|
|
392
|
+
"base_rate": 0.00001875,
|
|
393
|
+
"is_prompt": true,
|
|
394
|
+
"token_type": "cache_write"
|
|
395
|
+
}
|
|
396
|
+
]
|
|
397
|
+
},
|
|
344
398
|
{
|
|
345
399
|
"name": "claude-opus-4-20250514",
|
|
346
400
|
"name_pattern": "claude-opus-4-20250514",
|
|
@@ -2078,6 +2132,182 @@
|
|
|
2078
2132
|
}
|
|
2079
2133
|
]
|
|
2080
2134
|
},
|
|
2135
|
+
{
|
|
2136
|
+
"name": "gpt-5",
|
|
2137
|
+
"name_pattern": "gpt-5",
|
|
2138
|
+
"source": "litellm",
|
|
2139
|
+
"token_prices": [
|
|
2140
|
+
{
|
|
2141
|
+
"base_rate": 1.25e-6,
|
|
2142
|
+
"is_prompt": true,
|
|
2143
|
+
"token_type": "input"
|
|
2144
|
+
},
|
|
2145
|
+
{
|
|
2146
|
+
"base_rate": 0.00001,
|
|
2147
|
+
"is_prompt": false,
|
|
2148
|
+
"token_type": "output"
|
|
2149
|
+
},
|
|
2150
|
+
{
|
|
2151
|
+
"base_rate": 1.25e-7,
|
|
2152
|
+
"is_prompt": true,
|
|
2153
|
+
"token_type": "cache_read"
|
|
2154
|
+
}
|
|
2155
|
+
]
|
|
2156
|
+
},
|
|
2157
|
+
{
|
|
2158
|
+
"name": "gpt-5-2025-08-07",
|
|
2159
|
+
"name_pattern": "gpt-5-2025-08-07",
|
|
2160
|
+
"source": "litellm",
|
|
2161
|
+
"token_prices": [
|
|
2162
|
+
{
|
|
2163
|
+
"base_rate": 1.25e-6,
|
|
2164
|
+
"is_prompt": true,
|
|
2165
|
+
"token_type": "input"
|
|
2166
|
+
},
|
|
2167
|
+
{
|
|
2168
|
+
"base_rate": 0.00001,
|
|
2169
|
+
"is_prompt": false,
|
|
2170
|
+
"token_type": "output"
|
|
2171
|
+
},
|
|
2172
|
+
{
|
|
2173
|
+
"base_rate": 1.25e-7,
|
|
2174
|
+
"is_prompt": true,
|
|
2175
|
+
"token_type": "cache_read"
|
|
2176
|
+
}
|
|
2177
|
+
]
|
|
2178
|
+
},
|
|
2179
|
+
{
|
|
2180
|
+
"name": "gpt-5-chat",
|
|
2181
|
+
"name_pattern": "gpt-5-chat",
|
|
2182
|
+
"source": "litellm",
|
|
2183
|
+
"token_prices": [
|
|
2184
|
+
{
|
|
2185
|
+
"base_rate": 5e-6,
|
|
2186
|
+
"is_prompt": true,
|
|
2187
|
+
"token_type": "input"
|
|
2188
|
+
},
|
|
2189
|
+
{
|
|
2190
|
+
"base_rate": 0.00002,
|
|
2191
|
+
"is_prompt": false,
|
|
2192
|
+
"token_type": "output"
|
|
2193
|
+
},
|
|
2194
|
+
{
|
|
2195
|
+
"base_rate": 1.25e-6,
|
|
2196
|
+
"is_prompt": true,
|
|
2197
|
+
"token_type": "cache_read"
|
|
2198
|
+
}
|
|
2199
|
+
]
|
|
2200
|
+
},
|
|
2201
|
+
{
|
|
2202
|
+
"name": "gpt-5-chat-latest",
|
|
2203
|
+
"name_pattern": "gpt-5-chat-latest",
|
|
2204
|
+
"source": "litellm",
|
|
2205
|
+
"token_prices": [
|
|
2206
|
+
{
|
|
2207
|
+
"base_rate": 1.25e-6,
|
|
2208
|
+
"is_prompt": true,
|
|
2209
|
+
"token_type": "input"
|
|
2210
|
+
},
|
|
2211
|
+
{
|
|
2212
|
+
"base_rate": 0.00001,
|
|
2213
|
+
"is_prompt": false,
|
|
2214
|
+
"token_type": "output"
|
|
2215
|
+
},
|
|
2216
|
+
{
|
|
2217
|
+
"base_rate": 1.25e-7,
|
|
2218
|
+
"is_prompt": true,
|
|
2219
|
+
"token_type": "cache_read"
|
|
2220
|
+
}
|
|
2221
|
+
]
|
|
2222
|
+
},
|
|
2223
|
+
{
|
|
2224
|
+
"name": "gpt-5-mini",
|
|
2225
|
+
"name_pattern": "gpt-5-mini",
|
|
2226
|
+
"source": "litellm",
|
|
2227
|
+
"token_prices": [
|
|
2228
|
+
{
|
|
2229
|
+
"base_rate": 2.5e-7,
|
|
2230
|
+
"is_prompt": true,
|
|
2231
|
+
"token_type": "input"
|
|
2232
|
+
},
|
|
2233
|
+
{
|
|
2234
|
+
"base_rate": 2e-6,
|
|
2235
|
+
"is_prompt": false,
|
|
2236
|
+
"token_type": "output"
|
|
2237
|
+
},
|
|
2238
|
+
{
|
|
2239
|
+
"base_rate": 2.5e-8,
|
|
2240
|
+
"is_prompt": true,
|
|
2241
|
+
"token_type": "cache_read"
|
|
2242
|
+
}
|
|
2243
|
+
]
|
|
2244
|
+
},
|
|
2245
|
+
{
|
|
2246
|
+
"name": "gpt-5-mini-2025-08-07",
|
|
2247
|
+
"name_pattern": "gpt-5-mini-2025-08-07",
|
|
2248
|
+
"source": "litellm",
|
|
2249
|
+
"token_prices": [
|
|
2250
|
+
{
|
|
2251
|
+
"base_rate": 2.5e-7,
|
|
2252
|
+
"is_prompt": true,
|
|
2253
|
+
"token_type": "input"
|
|
2254
|
+
},
|
|
2255
|
+
{
|
|
2256
|
+
"base_rate": 2e-6,
|
|
2257
|
+
"is_prompt": false,
|
|
2258
|
+
"token_type": "output"
|
|
2259
|
+
},
|
|
2260
|
+
{
|
|
2261
|
+
"base_rate": 2.5e-8,
|
|
2262
|
+
"is_prompt": true,
|
|
2263
|
+
"token_type": "cache_read"
|
|
2264
|
+
}
|
|
2265
|
+
]
|
|
2266
|
+
},
|
|
2267
|
+
{
|
|
2268
|
+
"name": "gpt-5-nano",
|
|
2269
|
+
"name_pattern": "gpt-5-nano",
|
|
2270
|
+
"source": "litellm",
|
|
2271
|
+
"token_prices": [
|
|
2272
|
+
{
|
|
2273
|
+
"base_rate": 5e-8,
|
|
2274
|
+
"is_prompt": true,
|
|
2275
|
+
"token_type": "input"
|
|
2276
|
+
},
|
|
2277
|
+
{
|
|
2278
|
+
"base_rate": 4e-7,
|
|
2279
|
+
"is_prompt": false,
|
|
2280
|
+
"token_type": "output"
|
|
2281
|
+
},
|
|
2282
|
+
{
|
|
2283
|
+
"base_rate": 5e-9,
|
|
2284
|
+
"is_prompt": true,
|
|
2285
|
+
"token_type": "cache_read"
|
|
2286
|
+
}
|
|
2287
|
+
]
|
|
2288
|
+
},
|
|
2289
|
+
{
|
|
2290
|
+
"name": "gpt-5-nano-2025-08-07",
|
|
2291
|
+
"name_pattern": "gpt-5-nano-2025-08-07",
|
|
2292
|
+
"source": "litellm",
|
|
2293
|
+
"token_prices": [
|
|
2294
|
+
{
|
|
2295
|
+
"base_rate": 5e-8,
|
|
2296
|
+
"is_prompt": true,
|
|
2297
|
+
"token_type": "input"
|
|
2298
|
+
},
|
|
2299
|
+
{
|
|
2300
|
+
"base_rate": 4e-7,
|
|
2301
|
+
"is_prompt": false,
|
|
2302
|
+
"token_type": "output"
|
|
2303
|
+
},
|
|
2304
|
+
{
|
|
2305
|
+
"base_rate": 5e-9,
|
|
2306
|
+
"is_prompt": true,
|
|
2307
|
+
"token_type": "cache_read"
|
|
2308
|
+
}
|
|
2309
|
+
]
|
|
2310
|
+
},
|
|
2081
2311
|
{
|
|
2082
2312
|
"name": "o1",
|
|
2083
2313
|
"name_pattern": "o1",
|
|
@@ -2497,6 +2727,40 @@
|
|
|
2497
2727
|
"token_type": "cache_read"
|
|
2498
2728
|
}
|
|
2499
2729
|
]
|
|
2730
|
+
},
|
|
2731
|
+
{
|
|
2732
|
+
"name": "openai.gpt-oss-120b-1:0",
|
|
2733
|
+
"name_pattern": "openai\\.gpt-oss-120b-1:0",
|
|
2734
|
+
"source": "litellm",
|
|
2735
|
+
"token_prices": [
|
|
2736
|
+
{
|
|
2737
|
+
"base_rate": 1.5e-7,
|
|
2738
|
+
"is_prompt": true,
|
|
2739
|
+
"token_type": "input"
|
|
2740
|
+
},
|
|
2741
|
+
{
|
|
2742
|
+
"base_rate": 6e-7,
|
|
2743
|
+
"is_prompt": false,
|
|
2744
|
+
"token_type": "output"
|
|
2745
|
+
}
|
|
2746
|
+
]
|
|
2747
|
+
},
|
|
2748
|
+
{
|
|
2749
|
+
"name": "openai.gpt-oss-20b-1:0",
|
|
2750
|
+
"name_pattern": "openai\\.gpt-oss-20b-1:0",
|
|
2751
|
+
"source": "litellm",
|
|
2752
|
+
"token_prices": [
|
|
2753
|
+
{
|
|
2754
|
+
"base_rate": 7e-8,
|
|
2755
|
+
"is_prompt": true,
|
|
2756
|
+
"token_type": "input"
|
|
2757
|
+
},
|
|
2758
|
+
{
|
|
2759
|
+
"base_rate": 3e-7,
|
|
2760
|
+
"is_prompt": false,
|
|
2761
|
+
"token_type": "output"
|
|
2762
|
+
}
|
|
2763
|
+
]
|
|
2500
2764
|
}
|
|
2501
2765
|
]
|
|
2502
2766
|
}
|
phoenix/version.py
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
__version__ = "11.
|
|
1
|
+
__version__ = "11.21.0"
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|