tokencostauto 0.1.355__py3-none-any.whl → 0.1.359__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- tokencostauto/model_prices.json +830 -4
- {tokencostauto-0.1.355.dist-info → tokencostauto-0.1.359.dist-info}/METADATA +1 -1
- tokencostauto-0.1.359.dist-info/RECORD +9 -0
- tokencostauto-0.1.355.dist-info/RECORD +0 -9
- {tokencostauto-0.1.355.dist-info → tokencostauto-0.1.359.dist-info}/WHEEL +0 -0
- {tokencostauto-0.1.355.dist-info → tokencostauto-0.1.359.dist-info}/licenses/LICENSE +0 -0
- {tokencostauto-0.1.355.dist-info → tokencostauto-0.1.359.dist-info}/top_level.txt +0 -0
tokencostauto/model_prices.json
CHANGED
|
@@ -26620,10 +26620,12 @@
|
|
|
26620
26620
|
"supports_web_search": true
|
|
26621
26621
|
},
|
|
26622
26622
|
"gemini-3-pro-preview": {
|
|
26623
|
-
"cache_read_input_token_cost":
|
|
26623
|
+
"cache_read_input_token_cost": 2e-07,
|
|
26624
|
+
"cache_read_input_token_cost_above_200k_tokens": 4e-07,
|
|
26624
26625
|
"cache_creation_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
26625
26626
|
"input_cost_per_token": 2e-06,
|
|
26626
26627
|
"input_cost_per_token_above_200k_tokens": 4e-06,
|
|
26628
|
+
"input_cost_per_token_batches": 1e-06,
|
|
26627
26629
|
"litellm_provider": "vertex_ai-language-models",
|
|
26628
26630
|
"max_audio_length_hours": 8.4,
|
|
26629
26631
|
"max_audio_per_prompt": 1,
|
|
@@ -26637,10 +26639,12 @@
|
|
|
26637
26639
|
"mode": "chat",
|
|
26638
26640
|
"output_cost_per_token": 1.2e-05,
|
|
26639
26641
|
"output_cost_per_token_above_200k_tokens": 1.8e-05,
|
|
26642
|
+
"output_cost_per_token_batches": 6e-06,
|
|
26640
26643
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing",
|
|
26641
26644
|
"supported_endpoints": [
|
|
26642
26645
|
"/v1/chat/completions",
|
|
26643
|
-
"/v1/completions"
|
|
26646
|
+
"/v1/completions",
|
|
26647
|
+
"/v1/batch"
|
|
26644
26648
|
],
|
|
26645
26649
|
"supported_modalities": [
|
|
26646
26650
|
"text",
|
|
@@ -26664,9 +26668,11 @@
|
|
|
26664
26668
|
"supports_web_search": true
|
|
26665
26669
|
},
|
|
26666
26670
|
"gemini/gemini-3-pro-preview": {
|
|
26667
|
-
"cache_read_input_token_cost":
|
|
26671
|
+
"cache_read_input_token_cost": 2e-07,
|
|
26672
|
+
"cache_read_input_token_cost_above_200k_tokens": 4e-07,
|
|
26668
26673
|
"input_cost_per_token": 2e-06,
|
|
26669
26674
|
"input_cost_per_token_above_200k_tokens": 4e-06,
|
|
26675
|
+
"input_cost_per_token_batches": 1e-06,
|
|
26670
26676
|
"litellm_provider": "gemini",
|
|
26671
26677
|
"max_audio_length_hours": 8.4,
|
|
26672
26678
|
"max_audio_per_prompt": 1,
|
|
@@ -26680,11 +26686,13 @@
|
|
|
26680
26686
|
"mode": "chat",
|
|
26681
26687
|
"output_cost_per_token": 1.2e-05,
|
|
26682
26688
|
"output_cost_per_token_above_200k_tokens": 1.8e-05,
|
|
26689
|
+
"output_cost_per_token_batches": 6e-06,
|
|
26683
26690
|
"rpm": 2000,
|
|
26684
26691
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing",
|
|
26685
26692
|
"supported_endpoints": [
|
|
26686
26693
|
"/v1/chat/completions",
|
|
26687
|
-
"/v1/completions"
|
|
26694
|
+
"/v1/completions",
|
|
26695
|
+
"/v1/batch"
|
|
26688
26696
|
],
|
|
26689
26697
|
"supported_modalities": [
|
|
26690
26698
|
"text",
|
|
@@ -27010,5 +27018,823 @@
|
|
|
27010
27018
|
"supports_system_messages": false,
|
|
27011
27019
|
"supports_tool_choice": true,
|
|
27012
27020
|
"supports_vision": true
|
|
27021
|
+
},
|
|
27022
|
+
"azure/eu/gpt-5.1": {
|
|
27023
|
+
"cache_read_input_token_cost": 1.4e-07,
|
|
27024
|
+
"input_cost_per_token": 1.38e-06,
|
|
27025
|
+
"litellm_provider": "azure",
|
|
27026
|
+
"max_input_tokens": 272000,
|
|
27027
|
+
"max_output_tokens": 128000,
|
|
27028
|
+
"max_tokens": 128000,
|
|
27029
|
+
"mode": "chat",
|
|
27030
|
+
"output_cost_per_token": 1.1e-05,
|
|
27031
|
+
"supported_endpoints": [
|
|
27032
|
+
"/v1/chat/completions",
|
|
27033
|
+
"/v1/batch",
|
|
27034
|
+
"/v1/responses"
|
|
27035
|
+
],
|
|
27036
|
+
"supported_modalities": [
|
|
27037
|
+
"text",
|
|
27038
|
+
"image"
|
|
27039
|
+
],
|
|
27040
|
+
"supported_output_modalities": [
|
|
27041
|
+
"text",
|
|
27042
|
+
"image"
|
|
27043
|
+
],
|
|
27044
|
+
"supports_function_calling": true,
|
|
27045
|
+
"supports_native_streaming": true,
|
|
27046
|
+
"supports_parallel_function_calling": true,
|
|
27047
|
+
"supports_pdf_input": true,
|
|
27048
|
+
"supports_prompt_caching": true,
|
|
27049
|
+
"supports_reasoning": true,
|
|
27050
|
+
"supports_response_schema": true,
|
|
27051
|
+
"supports_system_messages": true,
|
|
27052
|
+
"supports_tool_choice": true,
|
|
27053
|
+
"supports_vision": true
|
|
27054
|
+
},
|
|
27055
|
+
"azure/eu/gpt-5.1-chat": {
|
|
27056
|
+
"cache_read_input_token_cost": 1.4e-07,
|
|
27057
|
+
"input_cost_per_token": 1.38e-06,
|
|
27058
|
+
"litellm_provider": "azure",
|
|
27059
|
+
"max_input_tokens": 272000,
|
|
27060
|
+
"max_output_tokens": 128000,
|
|
27061
|
+
"max_tokens": 128000,
|
|
27062
|
+
"mode": "chat",
|
|
27063
|
+
"output_cost_per_token": 1.1e-05,
|
|
27064
|
+
"supported_endpoints": [
|
|
27065
|
+
"/v1/chat/completions",
|
|
27066
|
+
"/v1/batch",
|
|
27067
|
+
"/v1/responses"
|
|
27068
|
+
],
|
|
27069
|
+
"supported_modalities": [
|
|
27070
|
+
"text",
|
|
27071
|
+
"image"
|
|
27072
|
+
],
|
|
27073
|
+
"supported_output_modalities": [
|
|
27074
|
+
"text",
|
|
27075
|
+
"image"
|
|
27076
|
+
],
|
|
27077
|
+
"supports_function_calling": true,
|
|
27078
|
+
"supports_native_streaming": true,
|
|
27079
|
+
"supports_parallel_function_calling": true,
|
|
27080
|
+
"supports_pdf_input": true,
|
|
27081
|
+
"supports_prompt_caching": true,
|
|
27082
|
+
"supports_reasoning": true,
|
|
27083
|
+
"supports_response_schema": true,
|
|
27084
|
+
"supports_system_messages": true,
|
|
27085
|
+
"supports_tool_choice": true,
|
|
27086
|
+
"supports_vision": true
|
|
27087
|
+
},
|
|
27088
|
+
"azure/eu/gpt-5.1-codex": {
|
|
27089
|
+
"cache_read_input_token_cost": 1.4e-07,
|
|
27090
|
+
"input_cost_per_token": 1.38e-06,
|
|
27091
|
+
"litellm_provider": "azure",
|
|
27092
|
+
"max_input_tokens": 272000,
|
|
27093
|
+
"max_output_tokens": 128000,
|
|
27094
|
+
"max_tokens": 128000,
|
|
27095
|
+
"mode": "responses",
|
|
27096
|
+
"output_cost_per_token": 1.1e-05,
|
|
27097
|
+
"supported_endpoints": [
|
|
27098
|
+
"/v1/responses"
|
|
27099
|
+
],
|
|
27100
|
+
"supported_modalities": [
|
|
27101
|
+
"text",
|
|
27102
|
+
"image"
|
|
27103
|
+
],
|
|
27104
|
+
"supported_output_modalities": [
|
|
27105
|
+
"text"
|
|
27106
|
+
],
|
|
27107
|
+
"supports_function_calling": true,
|
|
27108
|
+
"supports_native_streaming": true,
|
|
27109
|
+
"supports_parallel_function_calling": true,
|
|
27110
|
+
"supports_pdf_input": true,
|
|
27111
|
+
"supports_prompt_caching": true,
|
|
27112
|
+
"supports_reasoning": true,
|
|
27113
|
+
"supports_response_schema": true,
|
|
27114
|
+
"supports_system_messages": false,
|
|
27115
|
+
"supports_tool_choice": true,
|
|
27116
|
+
"supports_vision": true
|
|
27117
|
+
},
|
|
27118
|
+
"azure/eu/gpt-5.1-codex-mini": {
|
|
27119
|
+
"cache_read_input_token_cost": 2.8e-08,
|
|
27120
|
+
"input_cost_per_token": 2.75e-07,
|
|
27121
|
+
"litellm_provider": "azure",
|
|
27122
|
+
"max_input_tokens": 272000,
|
|
27123
|
+
"max_output_tokens": 128000,
|
|
27124
|
+
"max_tokens": 128000,
|
|
27125
|
+
"mode": "responses",
|
|
27126
|
+
"output_cost_per_token": 2.2e-06,
|
|
27127
|
+
"supported_endpoints": [
|
|
27128
|
+
"/v1/responses"
|
|
27129
|
+
],
|
|
27130
|
+
"supported_modalities": [
|
|
27131
|
+
"text",
|
|
27132
|
+
"image"
|
|
27133
|
+
],
|
|
27134
|
+
"supported_output_modalities": [
|
|
27135
|
+
"text"
|
|
27136
|
+
],
|
|
27137
|
+
"supports_function_calling": true,
|
|
27138
|
+
"supports_native_streaming": true,
|
|
27139
|
+
"supports_parallel_function_calling": true,
|
|
27140
|
+
"supports_pdf_input": true,
|
|
27141
|
+
"supports_prompt_caching": true,
|
|
27142
|
+
"supports_reasoning": true,
|
|
27143
|
+
"supports_response_schema": true,
|
|
27144
|
+
"supports_system_messages": false,
|
|
27145
|
+
"supports_tool_choice": true,
|
|
27146
|
+
"supports_vision": true
|
|
27147
|
+
},
|
|
27148
|
+
"azure/global/gpt-5.1": {
|
|
27149
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
27150
|
+
"input_cost_per_token": 1.25e-06,
|
|
27151
|
+
"litellm_provider": "azure",
|
|
27152
|
+
"max_input_tokens": 272000,
|
|
27153
|
+
"max_output_tokens": 128000,
|
|
27154
|
+
"max_tokens": 128000,
|
|
27155
|
+
"mode": "chat",
|
|
27156
|
+
"output_cost_per_token": 1e-05,
|
|
27157
|
+
"supported_endpoints": [
|
|
27158
|
+
"/v1/chat/completions",
|
|
27159
|
+
"/v1/batch",
|
|
27160
|
+
"/v1/responses"
|
|
27161
|
+
],
|
|
27162
|
+
"supported_modalities": [
|
|
27163
|
+
"text",
|
|
27164
|
+
"image"
|
|
27165
|
+
],
|
|
27166
|
+
"supported_output_modalities": [
|
|
27167
|
+
"text",
|
|
27168
|
+
"image"
|
|
27169
|
+
],
|
|
27170
|
+
"supports_function_calling": true,
|
|
27171
|
+
"supports_native_streaming": true,
|
|
27172
|
+
"supports_parallel_function_calling": true,
|
|
27173
|
+
"supports_pdf_input": true,
|
|
27174
|
+
"supports_prompt_caching": true,
|
|
27175
|
+
"supports_reasoning": true,
|
|
27176
|
+
"supports_response_schema": true,
|
|
27177
|
+
"supports_system_messages": true,
|
|
27178
|
+
"supports_tool_choice": true,
|
|
27179
|
+
"supports_vision": true
|
|
27180
|
+
},
|
|
27181
|
+
"azure/global/gpt-5.1-chat": {
|
|
27182
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
27183
|
+
"input_cost_per_token": 1.25e-06,
|
|
27184
|
+
"litellm_provider": "azure",
|
|
27185
|
+
"max_input_tokens": 272000,
|
|
27186
|
+
"max_output_tokens": 128000,
|
|
27187
|
+
"max_tokens": 128000,
|
|
27188
|
+
"mode": "chat",
|
|
27189
|
+
"output_cost_per_token": 1e-05,
|
|
27190
|
+
"supported_endpoints": [
|
|
27191
|
+
"/v1/chat/completions",
|
|
27192
|
+
"/v1/batch",
|
|
27193
|
+
"/v1/responses"
|
|
27194
|
+
],
|
|
27195
|
+
"supported_modalities": [
|
|
27196
|
+
"text",
|
|
27197
|
+
"image"
|
|
27198
|
+
],
|
|
27199
|
+
"supported_output_modalities": [
|
|
27200
|
+
"text",
|
|
27201
|
+
"image"
|
|
27202
|
+
],
|
|
27203
|
+
"supports_function_calling": true,
|
|
27204
|
+
"supports_native_streaming": true,
|
|
27205
|
+
"supports_parallel_function_calling": true,
|
|
27206
|
+
"supports_pdf_input": true,
|
|
27207
|
+
"supports_prompt_caching": true,
|
|
27208
|
+
"supports_reasoning": true,
|
|
27209
|
+
"supports_response_schema": true,
|
|
27210
|
+
"supports_system_messages": true,
|
|
27211
|
+
"supports_tool_choice": true,
|
|
27212
|
+
"supports_vision": true
|
|
27213
|
+
},
|
|
27214
|
+
"azure/global/gpt-5.1-codex": {
|
|
27215
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
27216
|
+
"input_cost_per_token": 1.25e-06,
|
|
27217
|
+
"litellm_provider": "azure",
|
|
27218
|
+
"max_input_tokens": 272000,
|
|
27219
|
+
"max_output_tokens": 128000,
|
|
27220
|
+
"max_tokens": 128000,
|
|
27221
|
+
"mode": "responses",
|
|
27222
|
+
"output_cost_per_token": 1e-05,
|
|
27223
|
+
"supported_endpoints": [
|
|
27224
|
+
"/v1/responses"
|
|
27225
|
+
],
|
|
27226
|
+
"supported_modalities": [
|
|
27227
|
+
"text",
|
|
27228
|
+
"image"
|
|
27229
|
+
],
|
|
27230
|
+
"supported_output_modalities": [
|
|
27231
|
+
"text"
|
|
27232
|
+
],
|
|
27233
|
+
"supports_function_calling": true,
|
|
27234
|
+
"supports_native_streaming": true,
|
|
27235
|
+
"supports_parallel_function_calling": true,
|
|
27236
|
+
"supports_pdf_input": true,
|
|
27237
|
+
"supports_prompt_caching": true,
|
|
27238
|
+
"supports_reasoning": true,
|
|
27239
|
+
"supports_response_schema": true,
|
|
27240
|
+
"supports_system_messages": false,
|
|
27241
|
+
"supports_tool_choice": true,
|
|
27242
|
+
"supports_vision": true
|
|
27243
|
+
},
|
|
27244
|
+
"azure/global/gpt-5.1-codex-mini": {
|
|
27245
|
+
"cache_read_input_token_cost": 2.5e-08,
|
|
27246
|
+
"input_cost_per_token": 2.5e-07,
|
|
27247
|
+
"litellm_provider": "azure",
|
|
27248
|
+
"max_input_tokens": 272000,
|
|
27249
|
+
"max_output_tokens": 128000,
|
|
27250
|
+
"max_tokens": 128000,
|
|
27251
|
+
"mode": "responses",
|
|
27252
|
+
"output_cost_per_token": 2e-06,
|
|
27253
|
+
"supported_endpoints": [
|
|
27254
|
+
"/v1/responses"
|
|
27255
|
+
],
|
|
27256
|
+
"supported_modalities": [
|
|
27257
|
+
"text",
|
|
27258
|
+
"image"
|
|
27259
|
+
],
|
|
27260
|
+
"supported_output_modalities": [
|
|
27261
|
+
"text"
|
|
27262
|
+
],
|
|
27263
|
+
"supports_function_calling": true,
|
|
27264
|
+
"supports_native_streaming": true,
|
|
27265
|
+
"supports_parallel_function_calling": true,
|
|
27266
|
+
"supports_pdf_input": true,
|
|
27267
|
+
"supports_prompt_caching": true,
|
|
27268
|
+
"supports_reasoning": true,
|
|
27269
|
+
"supports_response_schema": true,
|
|
27270
|
+
"supports_system_messages": false,
|
|
27271
|
+
"supports_tool_choice": true,
|
|
27272
|
+
"supports_vision": true
|
|
27273
|
+
},
|
|
27274
|
+
"azure/gpt-5.1": {
|
|
27275
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
27276
|
+
"input_cost_per_token": 1.25e-06,
|
|
27277
|
+
"litellm_provider": "azure",
|
|
27278
|
+
"max_input_tokens": 272000,
|
|
27279
|
+
"max_output_tokens": 128000,
|
|
27280
|
+
"max_tokens": 128000,
|
|
27281
|
+
"mode": "chat",
|
|
27282
|
+
"output_cost_per_token": 1e-05,
|
|
27283
|
+
"supported_endpoints": [
|
|
27284
|
+
"/v1/chat/completions",
|
|
27285
|
+
"/v1/batch",
|
|
27286
|
+
"/v1/responses"
|
|
27287
|
+
],
|
|
27288
|
+
"supported_modalities": [
|
|
27289
|
+
"text",
|
|
27290
|
+
"image"
|
|
27291
|
+
],
|
|
27292
|
+
"supported_output_modalities": [
|
|
27293
|
+
"text",
|
|
27294
|
+
"image"
|
|
27295
|
+
],
|
|
27296
|
+
"supports_function_calling": true,
|
|
27297
|
+
"supports_native_streaming": true,
|
|
27298
|
+
"supports_parallel_function_calling": true,
|
|
27299
|
+
"supports_pdf_input": true,
|
|
27300
|
+
"supports_prompt_caching": true,
|
|
27301
|
+
"supports_reasoning": true,
|
|
27302
|
+
"supports_response_schema": true,
|
|
27303
|
+
"supports_system_messages": true,
|
|
27304
|
+
"supports_tool_choice": true,
|
|
27305
|
+
"supports_vision": true
|
|
27306
|
+
},
|
|
27307
|
+
"azure/gpt-5.1-chat": {
|
|
27308
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
27309
|
+
"input_cost_per_token": 1.25e-06,
|
|
27310
|
+
"litellm_provider": "azure",
|
|
27311
|
+
"max_input_tokens": 272000,
|
|
27312
|
+
"max_output_tokens": 128000,
|
|
27313
|
+
"max_tokens": 128000,
|
|
27314
|
+
"mode": "chat",
|
|
27315
|
+
"output_cost_per_token": 1e-05,
|
|
27316
|
+
"supported_endpoints": [
|
|
27317
|
+
"/v1/chat/completions",
|
|
27318
|
+
"/v1/batch",
|
|
27319
|
+
"/v1/responses"
|
|
27320
|
+
],
|
|
27321
|
+
"supported_modalities": [
|
|
27322
|
+
"text",
|
|
27323
|
+
"image"
|
|
27324
|
+
],
|
|
27325
|
+
"supported_output_modalities": [
|
|
27326
|
+
"text",
|
|
27327
|
+
"image"
|
|
27328
|
+
],
|
|
27329
|
+
"supports_function_calling": true,
|
|
27330
|
+
"supports_native_streaming": true,
|
|
27331
|
+
"supports_parallel_function_calling": true,
|
|
27332
|
+
"supports_pdf_input": true,
|
|
27333
|
+
"supports_prompt_caching": true,
|
|
27334
|
+
"supports_reasoning": true,
|
|
27335
|
+
"supports_response_schema": true,
|
|
27336
|
+
"supports_system_messages": true,
|
|
27337
|
+
"supports_tool_choice": true,
|
|
27338
|
+
"supports_vision": true
|
|
27339
|
+
},
|
|
27340
|
+
"azure/gpt-5.1-codex": {
|
|
27341
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
27342
|
+
"input_cost_per_token": 1.25e-06,
|
|
27343
|
+
"litellm_provider": "azure",
|
|
27344
|
+
"max_input_tokens": 272000,
|
|
27345
|
+
"max_output_tokens": 128000,
|
|
27346
|
+
"max_tokens": 128000,
|
|
27347
|
+
"mode": "responses",
|
|
27348
|
+
"output_cost_per_token": 1e-05,
|
|
27349
|
+
"supported_endpoints": [
|
|
27350
|
+
"/v1/responses"
|
|
27351
|
+
],
|
|
27352
|
+
"supported_modalities": [
|
|
27353
|
+
"text",
|
|
27354
|
+
"image"
|
|
27355
|
+
],
|
|
27356
|
+
"supported_output_modalities": [
|
|
27357
|
+
"text"
|
|
27358
|
+
],
|
|
27359
|
+
"supports_function_calling": true,
|
|
27360
|
+
"supports_native_streaming": true,
|
|
27361
|
+
"supports_parallel_function_calling": true,
|
|
27362
|
+
"supports_pdf_input": true,
|
|
27363
|
+
"supports_prompt_caching": true,
|
|
27364
|
+
"supports_reasoning": true,
|
|
27365
|
+
"supports_response_schema": true,
|
|
27366
|
+
"supports_system_messages": false,
|
|
27367
|
+
"supports_tool_choice": true,
|
|
27368
|
+
"supports_vision": true
|
|
27369
|
+
},
|
|
27370
|
+
"azure/gpt-5.1-codex-mini": {
|
|
27371
|
+
"cache_read_input_token_cost": 2.5e-08,
|
|
27372
|
+
"input_cost_per_token": 2.5e-07,
|
|
27373
|
+
"litellm_provider": "azure",
|
|
27374
|
+
"max_input_tokens": 272000,
|
|
27375
|
+
"max_output_tokens": 128000,
|
|
27376
|
+
"max_tokens": 128000,
|
|
27377
|
+
"mode": "responses",
|
|
27378
|
+
"output_cost_per_token": 2e-06,
|
|
27379
|
+
"supported_endpoints": [
|
|
27380
|
+
"/v1/responses"
|
|
27381
|
+
],
|
|
27382
|
+
"supported_modalities": [
|
|
27383
|
+
"text",
|
|
27384
|
+
"image"
|
|
27385
|
+
],
|
|
27386
|
+
"supported_output_modalities": [
|
|
27387
|
+
"text"
|
|
27388
|
+
],
|
|
27389
|
+
"supports_function_calling": true,
|
|
27390
|
+
"supports_native_streaming": true,
|
|
27391
|
+
"supports_parallel_function_calling": true,
|
|
27392
|
+
"supports_pdf_input": true,
|
|
27393
|
+
"supports_prompt_caching": true,
|
|
27394
|
+
"supports_reasoning": true,
|
|
27395
|
+
"supports_response_schema": true,
|
|
27396
|
+
"supports_system_messages": false,
|
|
27397
|
+
"supports_tool_choice": true,
|
|
27398
|
+
"supports_vision": true
|
|
27399
|
+
},
|
|
27400
|
+
"azure/us/gpt-5.1": {
|
|
27401
|
+
"cache_read_input_token_cost": 1.4e-07,
|
|
27402
|
+
"input_cost_per_token": 1.38e-06,
|
|
27403
|
+
"litellm_provider": "azure",
|
|
27404
|
+
"max_input_tokens": 272000,
|
|
27405
|
+
"max_output_tokens": 128000,
|
|
27406
|
+
"max_tokens": 128000,
|
|
27407
|
+
"mode": "chat",
|
|
27408
|
+
"output_cost_per_token": 1.1e-05,
|
|
27409
|
+
"supported_endpoints": [
|
|
27410
|
+
"/v1/chat/completions",
|
|
27411
|
+
"/v1/batch",
|
|
27412
|
+
"/v1/responses"
|
|
27413
|
+
],
|
|
27414
|
+
"supported_modalities": [
|
|
27415
|
+
"text",
|
|
27416
|
+
"image"
|
|
27417
|
+
],
|
|
27418
|
+
"supported_output_modalities": [
|
|
27419
|
+
"text",
|
|
27420
|
+
"image"
|
|
27421
|
+
],
|
|
27422
|
+
"supports_function_calling": true,
|
|
27423
|
+
"supports_native_streaming": true,
|
|
27424
|
+
"supports_parallel_function_calling": true,
|
|
27425
|
+
"supports_pdf_input": true,
|
|
27426
|
+
"supports_prompt_caching": true,
|
|
27427
|
+
"supports_reasoning": true,
|
|
27428
|
+
"supports_response_schema": true,
|
|
27429
|
+
"supports_system_messages": true,
|
|
27430
|
+
"supports_tool_choice": true,
|
|
27431
|
+
"supports_vision": true
|
|
27432
|
+
},
|
|
27433
|
+
"azure/us/gpt-5.1-chat": {
|
|
27434
|
+
"cache_read_input_token_cost": 1.4e-07,
|
|
27435
|
+
"input_cost_per_token": 1.38e-06,
|
|
27436
|
+
"litellm_provider": "azure",
|
|
27437
|
+
"max_input_tokens": 272000,
|
|
27438
|
+
"max_output_tokens": 128000,
|
|
27439
|
+
"max_tokens": 128000,
|
|
27440
|
+
"mode": "chat",
|
|
27441
|
+
"output_cost_per_token": 1.1e-05,
|
|
27442
|
+
"supported_endpoints": [
|
|
27443
|
+
"/v1/chat/completions",
|
|
27444
|
+
"/v1/batch",
|
|
27445
|
+
"/v1/responses"
|
|
27446
|
+
],
|
|
27447
|
+
"supported_modalities": [
|
|
27448
|
+
"text",
|
|
27449
|
+
"image"
|
|
27450
|
+
],
|
|
27451
|
+
"supported_output_modalities": [
|
|
27452
|
+
"text",
|
|
27453
|
+
"image"
|
|
27454
|
+
],
|
|
27455
|
+
"supports_function_calling": true,
|
|
27456
|
+
"supports_native_streaming": true,
|
|
27457
|
+
"supports_parallel_function_calling": true,
|
|
27458
|
+
"supports_pdf_input": true,
|
|
27459
|
+
"supports_prompt_caching": true,
|
|
27460
|
+
"supports_reasoning": true,
|
|
27461
|
+
"supports_response_schema": true,
|
|
27462
|
+
"supports_system_messages": true,
|
|
27463
|
+
"supports_tool_choice": true,
|
|
27464
|
+
"supports_vision": true
|
|
27465
|
+
},
|
|
27466
|
+
"azure/us/gpt-5.1-codex": {
|
|
27467
|
+
"cache_read_input_token_cost": 1.4e-07,
|
|
27468
|
+
"input_cost_per_token": 1.38e-06,
|
|
27469
|
+
"litellm_provider": "azure",
|
|
27470
|
+
"max_input_tokens": 272000,
|
|
27471
|
+
"max_output_tokens": 128000,
|
|
27472
|
+
"max_tokens": 128000,
|
|
27473
|
+
"mode": "responses",
|
|
27474
|
+
"output_cost_per_token": 1.1e-05,
|
|
27475
|
+
"supported_endpoints": [
|
|
27476
|
+
"/v1/responses"
|
|
27477
|
+
],
|
|
27478
|
+
"supported_modalities": [
|
|
27479
|
+
"text",
|
|
27480
|
+
"image"
|
|
27481
|
+
],
|
|
27482
|
+
"supported_output_modalities": [
|
|
27483
|
+
"text"
|
|
27484
|
+
],
|
|
27485
|
+
"supports_function_calling": true,
|
|
27486
|
+
"supports_native_streaming": true,
|
|
27487
|
+
"supports_parallel_function_calling": true,
|
|
27488
|
+
"supports_pdf_input": true,
|
|
27489
|
+
"supports_prompt_caching": true,
|
|
27490
|
+
"supports_reasoning": true,
|
|
27491
|
+
"supports_response_schema": true,
|
|
27492
|
+
"supports_system_messages": false,
|
|
27493
|
+
"supports_tool_choice": true,
|
|
27494
|
+
"supports_vision": true
|
|
27495
|
+
},
|
|
27496
|
+
"azure/us/gpt-5.1-codex-mini": {
|
|
27497
|
+
"cache_read_input_token_cost": 2.8e-08,
|
|
27498
|
+
"input_cost_per_token": 2.75e-07,
|
|
27499
|
+
"litellm_provider": "azure",
|
|
27500
|
+
"max_input_tokens": 272000,
|
|
27501
|
+
"max_output_tokens": 128000,
|
|
27502
|
+
"max_tokens": 128000,
|
|
27503
|
+
"mode": "responses",
|
|
27504
|
+
"output_cost_per_token": 2.2e-06,
|
|
27505
|
+
"supported_endpoints": [
|
|
27506
|
+
"/v1/responses"
|
|
27507
|
+
],
|
|
27508
|
+
"supported_modalities": [
|
|
27509
|
+
"text",
|
|
27510
|
+
"image"
|
|
27511
|
+
],
|
|
27512
|
+
"supported_output_modalities": [
|
|
27513
|
+
"text"
|
|
27514
|
+
],
|
|
27515
|
+
"supports_function_calling": true,
|
|
27516
|
+
"supports_native_streaming": true,
|
|
27517
|
+
"supports_parallel_function_calling": true,
|
|
27518
|
+
"supports_pdf_input": true,
|
|
27519
|
+
"supports_prompt_caching": true,
|
|
27520
|
+
"supports_reasoning": true,
|
|
27521
|
+
"supports_response_schema": true,
|
|
27522
|
+
"supports_system_messages": false,
|
|
27523
|
+
"supports_tool_choice": true,
|
|
27524
|
+
"supports_vision": true
|
|
27525
|
+
},
|
|
27526
|
+
"vertex_ai/gemini-3-pro-preview": {
|
|
27527
|
+
"cache_read_input_token_cost": 2e-07,
|
|
27528
|
+
"cache_read_input_token_cost_above_200k_tokens": 4e-07,
|
|
27529
|
+
"cache_creation_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
27530
|
+
"input_cost_per_token": 2e-06,
|
|
27531
|
+
"input_cost_per_token_above_200k_tokens": 4e-06,
|
|
27532
|
+
"input_cost_per_token_batches": 1e-06,
|
|
27533
|
+
"litellm_provider": "vertex_ai",
|
|
27534
|
+
"max_audio_length_hours": 8.4,
|
|
27535
|
+
"max_audio_per_prompt": 1,
|
|
27536
|
+
"max_images_per_prompt": 3000,
|
|
27537
|
+
"max_input_tokens": 1048576,
|
|
27538
|
+
"max_output_tokens": 65535,
|
|
27539
|
+
"max_pdf_size_mb": 30,
|
|
27540
|
+
"max_tokens": 65535,
|
|
27541
|
+
"max_video_length": 1,
|
|
27542
|
+
"max_videos_per_prompt": 10,
|
|
27543
|
+
"mode": "chat",
|
|
27544
|
+
"output_cost_per_token": 1.2e-05,
|
|
27545
|
+
"output_cost_per_token_above_200k_tokens": 1.8e-05,
|
|
27546
|
+
"output_cost_per_token_batches": 6e-06,
|
|
27547
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing",
|
|
27548
|
+
"supported_endpoints": [
|
|
27549
|
+
"/v1/chat/completions",
|
|
27550
|
+
"/v1/completions",
|
|
27551
|
+
"/v1/batch"
|
|
27552
|
+
],
|
|
27553
|
+
"supported_modalities": [
|
|
27554
|
+
"text",
|
|
27555
|
+
"image",
|
|
27556
|
+
"audio",
|
|
27557
|
+
"video"
|
|
27558
|
+
],
|
|
27559
|
+
"supported_output_modalities": [
|
|
27560
|
+
"text"
|
|
27561
|
+
],
|
|
27562
|
+
"supports_audio_input": true,
|
|
27563
|
+
"supports_function_calling": true,
|
|
27564
|
+
"supports_pdf_input": true,
|
|
27565
|
+
"supports_prompt_caching": true,
|
|
27566
|
+
"supports_reasoning": true,
|
|
27567
|
+
"supports_response_schema": true,
|
|
27568
|
+
"supports_system_messages": true,
|
|
27569
|
+
"supports_tool_choice": true,
|
|
27570
|
+
"supports_video_input": true,
|
|
27571
|
+
"supports_vision": true,
|
|
27572
|
+
"supports_web_search": true
|
|
27573
|
+
},
|
|
27574
|
+
"vertex_ai/gemini-2.5-flash-image": {
|
|
27575
|
+
"litellm_provider": "vertex_ai-language-models",
|
|
27576
|
+
"mode": "image_generation",
|
|
27577
|
+
"output_cost_per_image": 0.039,
|
|
27578
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/multimodal/image-generation#edit-an-image"
|
|
27579
|
+
},
|
|
27580
|
+
"vertex_ai/imagen-3.0-capability-001": {
|
|
27581
|
+
"litellm_provider": "vertex_ai-image-models",
|
|
27582
|
+
"mode": "image_generation",
|
|
27583
|
+
"output_cost_per_image": 0.04,
|
|
27584
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/image/edit-insert-objects"
|
|
27585
|
+
},
|
|
27586
|
+
"cerebras/gpt-oss-120b": {
|
|
27587
|
+
"input_cost_per_token": 2.5e-07,
|
|
27588
|
+
"litellm_provider": "cerebras",
|
|
27589
|
+
"max_input_tokens": 131072,
|
|
27590
|
+
"max_output_tokens": 32768,
|
|
27591
|
+
"max_tokens": 32768,
|
|
27592
|
+
"mode": "chat",
|
|
27593
|
+
"output_cost_per_token": 6.9e-07,
|
|
27594
|
+
"source": "https://www.cerebras.ai/blog/openai-gpt-oss-120b-runs-fastest-on-cerebras",
|
|
27595
|
+
"supports_function_calling": true,
|
|
27596
|
+
"supports_parallel_function_calling": true,
|
|
27597
|
+
"supports_reasoning": true,
|
|
27598
|
+
"supports_response_schema": true,
|
|
27599
|
+
"supports_tool_choice": true
|
|
27600
|
+
},
|
|
27601
|
+
"gemini-3-pro-image-preview": {
|
|
27602
|
+
"input_cost_per_image": 0.0011,
|
|
27603
|
+
"input_cost_per_token": 2e-06,
|
|
27604
|
+
"input_cost_per_token_batches": 1e-06,
|
|
27605
|
+
"litellm_provider": "vertex_ai-language-models",
|
|
27606
|
+
"max_input_tokens": 65536,
|
|
27607
|
+
"max_output_tokens": 32768,
|
|
27608
|
+
"max_tokens": 65536,
|
|
27609
|
+
"mode": "image_generation",
|
|
27610
|
+
"output_cost_per_image": 0.134,
|
|
27611
|
+
"output_cost_per_token": 1.2e-05,
|
|
27612
|
+
"output_cost_per_token_batches": 6e-06,
|
|
27613
|
+
"source": "https://ai.google.dev/gemini-api/docs/pricing",
|
|
27614
|
+
"supported_endpoints": [
|
|
27615
|
+
"/v1/chat/completions",
|
|
27616
|
+
"/v1/completions",
|
|
27617
|
+
"/v1/batch"
|
|
27618
|
+
],
|
|
27619
|
+
"supported_modalities": [
|
|
27620
|
+
"text",
|
|
27621
|
+
"image"
|
|
27622
|
+
],
|
|
27623
|
+
"supported_output_modalities": [
|
|
27624
|
+
"text",
|
|
27625
|
+
"image"
|
|
27626
|
+
],
|
|
27627
|
+
"supports_function_calling": false,
|
|
27628
|
+
"supports_prompt_caching": true,
|
|
27629
|
+
"supports_response_schema": true,
|
|
27630
|
+
"supports_system_messages": true,
|
|
27631
|
+
"supports_vision": true,
|
|
27632
|
+
"supports_web_search": true
|
|
27633
|
+
},
|
|
27634
|
+
"gemini/gemini-3-pro-image-preview": {
|
|
27635
|
+
"input_cost_per_image": 0.0011,
|
|
27636
|
+
"input_cost_per_token": 2e-06,
|
|
27637
|
+
"input_cost_per_token_batches": 1e-06,
|
|
27638
|
+
"litellm_provider": "gemini",
|
|
27639
|
+
"max_input_tokens": 65536,
|
|
27640
|
+
"max_output_tokens": 32768,
|
|
27641
|
+
"max_tokens": 65536,
|
|
27642
|
+
"mode": "image_generation",
|
|
27643
|
+
"output_cost_per_image": 0.134,
|
|
27644
|
+
"output_cost_per_token": 1.2e-05,
|
|
27645
|
+
"output_cost_per_token_batches": 6e-06,
|
|
27646
|
+
"source": "https://ai.google.dev/gemini-api/docs/pricing",
|
|
27647
|
+
"supported_endpoints": [
|
|
27648
|
+
"/v1/chat/completions",
|
|
27649
|
+
"/v1/completions",
|
|
27650
|
+
"/v1/batch"
|
|
27651
|
+
],
|
|
27652
|
+
"supported_modalities": [
|
|
27653
|
+
"text",
|
|
27654
|
+
"image"
|
|
27655
|
+
],
|
|
27656
|
+
"supported_output_modalities": [
|
|
27657
|
+
"text",
|
|
27658
|
+
"image"
|
|
27659
|
+
],
|
|
27660
|
+
"supports_function_calling": false,
|
|
27661
|
+
"supports_prompt_caching": true,
|
|
27662
|
+
"supports_response_schema": true,
|
|
27663
|
+
"supports_system_messages": true,
|
|
27664
|
+
"supports_vision": true,
|
|
27665
|
+
"supports_web_search": true
|
|
27666
|
+
},
|
|
27667
|
+
"openrouter/google/gemini-3-pro-preview": {
|
|
27668
|
+
"cache_read_input_token_cost": 2e-07,
|
|
27669
|
+
"cache_read_input_token_cost_above_200k_tokens": 4e-07,
|
|
27670
|
+
"cache_creation_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
27671
|
+
"input_cost_per_token": 2e-06,
|
|
27672
|
+
"input_cost_per_token_above_200k_tokens": 4e-06,
|
|
27673
|
+
"input_cost_per_token_batches": 1e-06,
|
|
27674
|
+
"litellm_provider": "openrouter",
|
|
27675
|
+
"max_audio_length_hours": 8.4,
|
|
27676
|
+
"max_audio_per_prompt": 1,
|
|
27677
|
+
"max_images_per_prompt": 3000,
|
|
27678
|
+
"max_input_tokens": 1048576,
|
|
27679
|
+
"max_output_tokens": 65535,
|
|
27680
|
+
"max_pdf_size_mb": 30,
|
|
27681
|
+
"max_tokens": 65535,
|
|
27682
|
+
"max_video_length": 1,
|
|
27683
|
+
"max_videos_per_prompt": 10,
|
|
27684
|
+
"mode": "chat",
|
|
27685
|
+
"output_cost_per_token": 1.2e-05,
|
|
27686
|
+
"output_cost_per_token_above_200k_tokens": 1.8e-05,
|
|
27687
|
+
"output_cost_per_token_batches": 6e-06,
|
|
27688
|
+
"supported_endpoints": [
|
|
27689
|
+
"/v1/chat/completions",
|
|
27690
|
+
"/v1/completions",
|
|
27691
|
+
"/v1/batch"
|
|
27692
|
+
],
|
|
27693
|
+
"supported_modalities": [
|
|
27694
|
+
"text",
|
|
27695
|
+
"image",
|
|
27696
|
+
"audio",
|
|
27697
|
+
"video"
|
|
27698
|
+
],
|
|
27699
|
+
"supported_output_modalities": [
|
|
27700
|
+
"text"
|
|
27701
|
+
],
|
|
27702
|
+
"supports_audio_input": true,
|
|
27703
|
+
"supports_function_calling": true,
|
|
27704
|
+
"supports_pdf_input": true,
|
|
27705
|
+
"supports_prompt_caching": true,
|
|
27706
|
+
"supports_reasoning": true,
|
|
27707
|
+
"supports_response_schema": true,
|
|
27708
|
+
"supports_system_messages": true,
|
|
27709
|
+
"supports_tool_choice": true,
|
|
27710
|
+
"supports_video_input": true,
|
|
27711
|
+
"supports_vision": true,
|
|
27712
|
+
"supports_web_search": true
|
|
27713
|
+
},
|
|
27714
|
+
"together_ai/zai-org/GLM-4.6": {
|
|
27715
|
+
"input_cost_per_token": 6e-07,
|
|
27716
|
+
"litellm_provider": "together_ai",
|
|
27717
|
+
"max_input_tokens": 200000,
|
|
27718
|
+
"max_output_tokens": 200000,
|
|
27719
|
+
"max_tokens": 200000,
|
|
27720
|
+
"mode": "chat",
|
|
27721
|
+
"output_cost_per_token": 2.2e-06,
|
|
27722
|
+
"source": "https://www.together.ai/models/glm-4-6",
|
|
27723
|
+
"supports_function_calling": true,
|
|
27724
|
+
"supports_parallel_function_calling": true,
|
|
27725
|
+
"supports_reasoning": true,
|
|
27726
|
+
"supports_tool_choice": true
|
|
27727
|
+
},
|
|
27728
|
+
"vertex_ai/gemini-3-pro-image-preview": {
|
|
27729
|
+
"input_cost_per_image": 0.0011,
|
|
27730
|
+
"input_cost_per_token": 2e-06,
|
|
27731
|
+
"input_cost_per_token_batches": 1e-06,
|
|
27732
|
+
"litellm_provider": "vertex_ai-language-models",
|
|
27733
|
+
"max_input_tokens": 65536,
|
|
27734
|
+
"max_output_tokens": 32768,
|
|
27735
|
+
"max_tokens": 65536,
|
|
27736
|
+
"mode": "image_generation",
|
|
27737
|
+
"output_cost_per_image": 0.134,
|
|
27738
|
+
"output_cost_per_token": 1.2e-05,
|
|
27739
|
+
"output_cost_per_token_batches": 6e-06,
|
|
27740
|
+
"source": "https://docs.cloud.google.com/vertex-ai/generative-ai/docs/models/gemini/3-pro-image"
|
|
27741
|
+
},
|
|
27742
|
+
"xai/grok-4-1-fast": {
|
|
27743
|
+
"cache_read_input_token_cost": 5e-08,
|
|
27744
|
+
"input_cost_per_token": 2e-07,
|
|
27745
|
+
"input_cost_per_token_above_128k_tokens": 4e-07,
|
|
27746
|
+
"litellm_provider": "xai",
|
|
27747
|
+
"max_input_tokens": 2000000.0,
|
|
27748
|
+
"max_output_tokens": 2000000.0,
|
|
27749
|
+
"max_tokens": 2000000.0,
|
|
27750
|
+
"mode": "chat",
|
|
27751
|
+
"output_cost_per_token": 5e-07,
|
|
27752
|
+
"output_cost_per_token_above_128k_tokens": 1e-06,
|
|
27753
|
+
"source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning",
|
|
27754
|
+
"supports_audio_input": true,
|
|
27755
|
+
"supports_function_calling": true,
|
|
27756
|
+
"supports_reasoning": true,
|
|
27757
|
+
"supports_response_schema": true,
|
|
27758
|
+
"supports_tool_choice": true,
|
|
27759
|
+
"supports_vision": true,
|
|
27760
|
+
"supports_web_search": true
|
|
27761
|
+
},
|
|
27762
|
+
"xai/grok-4-1-fast-reasoning": {
|
|
27763
|
+
"cache_read_input_token_cost": 5e-08,
|
|
27764
|
+
"input_cost_per_token": 2e-07,
|
|
27765
|
+
"input_cost_per_token_above_128k_tokens": 4e-07,
|
|
27766
|
+
"litellm_provider": "xai",
|
|
27767
|
+
"max_input_tokens": 2000000.0,
|
|
27768
|
+
"max_output_tokens": 2000000.0,
|
|
27769
|
+
"max_tokens": 2000000.0,
|
|
27770
|
+
"mode": "chat",
|
|
27771
|
+
"output_cost_per_token": 5e-07,
|
|
27772
|
+
"output_cost_per_token_above_128k_tokens": 1e-06,
|
|
27773
|
+
"source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning",
|
|
27774
|
+
"supports_audio_input": true,
|
|
27775
|
+
"supports_function_calling": true,
|
|
27776
|
+
"supports_reasoning": true,
|
|
27777
|
+
"supports_response_schema": true,
|
|
27778
|
+
"supports_tool_choice": true,
|
|
27779
|
+
"supports_vision": true,
|
|
27780
|
+
"supports_web_search": true
|
|
27781
|
+
},
|
|
27782
|
+
"xai/grok-4-1-fast-reasoning-latest": {
|
|
27783
|
+
"cache_read_input_token_cost": 5e-08,
|
|
27784
|
+
"input_cost_per_token": 2e-07,
|
|
27785
|
+
"input_cost_per_token_above_128k_tokens": 4e-07,
|
|
27786
|
+
"litellm_provider": "xai",
|
|
27787
|
+
"max_input_tokens": 2000000.0,
|
|
27788
|
+
"max_output_tokens": 2000000.0,
|
|
27789
|
+
"max_tokens": 2000000.0,
|
|
27790
|
+
"mode": "chat",
|
|
27791
|
+
"output_cost_per_token": 5e-07,
|
|
27792
|
+
"output_cost_per_token_above_128k_tokens": 1e-06,
|
|
27793
|
+
"source": "https://docs.x.ai/docs/models/grok-4-1-fast-reasoning",
|
|
27794
|
+
"supports_audio_input": true,
|
|
27795
|
+
"supports_function_calling": true,
|
|
27796
|
+
"supports_reasoning": true,
|
|
27797
|
+
"supports_response_schema": true,
|
|
27798
|
+
"supports_tool_choice": true,
|
|
27799
|
+
"supports_vision": true,
|
|
27800
|
+
"supports_web_search": true
|
|
27801
|
+
},
|
|
27802
|
+
"xai/grok-4-1-fast-non-reasoning": {
|
|
27803
|
+
"cache_read_input_token_cost": 5e-08,
|
|
27804
|
+
"input_cost_per_token": 2e-07,
|
|
27805
|
+
"input_cost_per_token_above_128k_tokens": 4e-07,
|
|
27806
|
+
"litellm_provider": "xai",
|
|
27807
|
+
"max_input_tokens": 2000000.0,
|
|
27808
|
+
"max_output_tokens": 2000000.0,
|
|
27809
|
+
"max_tokens": 2000000.0,
|
|
27810
|
+
"mode": "chat",
|
|
27811
|
+
"output_cost_per_token": 5e-07,
|
|
27812
|
+
"output_cost_per_token_above_128k_tokens": 1e-06,
|
|
27813
|
+
"source": "https://docs.x.ai/docs/models/grok-4-1-fast-non-reasoning",
|
|
27814
|
+
"supports_audio_input": true,
|
|
27815
|
+
"supports_function_calling": true,
|
|
27816
|
+
"supports_response_schema": true,
|
|
27817
|
+
"supports_tool_choice": true,
|
|
27818
|
+
"supports_vision": true,
|
|
27819
|
+
"supports_web_search": true
|
|
27820
|
+
},
|
|
27821
|
+
"xai/grok-4-1-fast-non-reasoning-latest": {
|
|
27822
|
+
"cache_read_input_token_cost": 5e-08,
|
|
27823
|
+
"input_cost_per_token": 2e-07,
|
|
27824
|
+
"input_cost_per_token_above_128k_tokens": 4e-07,
|
|
27825
|
+
"litellm_provider": "xai",
|
|
27826
|
+
"max_input_tokens": 2000000.0,
|
|
27827
|
+
"max_output_tokens": 2000000.0,
|
|
27828
|
+
"max_tokens": 2000000.0,
|
|
27829
|
+
"mode": "chat",
|
|
27830
|
+
"output_cost_per_token": 5e-07,
|
|
27831
|
+
"output_cost_per_token_above_128k_tokens": 1e-06,
|
|
27832
|
+
"source": "https://docs.x.ai/docs/models/grok-4-1-fast-non-reasoning",
|
|
27833
|
+
"supports_audio_input": true,
|
|
27834
|
+
"supports_function_calling": true,
|
|
27835
|
+
"supports_response_schema": true,
|
|
27836
|
+
"supports_tool_choice": true,
|
|
27837
|
+
"supports_vision": true,
|
|
27838
|
+
"supports_web_search": true
|
|
27013
27839
|
}
|
|
27014
27840
|
}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: tokencostauto
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.359
|
|
4
4
|
Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
|
|
5
5
|
Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
|
|
6
6
|
Project-URL: Homepage, https://github.com/madpin/tokencostaudo
|
|
@@ -0,0 +1,9 @@
|
|
|
1
|
+
tokencostauto/__init__.py,sha256=-4d_ryFH62SgNXPXA8vGPFZoAKtOBjnsg37EB_RkZG8,289
|
|
2
|
+
tokencostauto/constants.py,sha256=_82MlTkTrdrwzyRosQD7d3JdgNP9KAUM-cZo8DE00P0,3395
|
|
3
|
+
tokencostauto/costs.py,sha256=tXsgrTypq-dCHaHtoXcg2XepezWsAvZpl9gEsv_53iE,10679
|
|
4
|
+
tokencostauto/model_prices.json,sha256=xKn3KixDcaBUPEiu4bU-SsGyKuKSzKrO3bShsv2_fN4,988335
|
|
5
|
+
tokencostauto-0.1.359.dist-info/licenses/LICENSE,sha256=4PLv_CD6Ughnsvg_nM2XeTqGwVK6lQVR77kVWbPq-0U,1065
|
|
6
|
+
tokencostauto-0.1.359.dist-info/METADATA,sha256=0EhAJ8xBit7MZQWQd4XLYjVJNa-0LKq8DnBIRLmmDl8,204076
|
|
7
|
+
tokencostauto-0.1.359.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
8
|
+
tokencostauto-0.1.359.dist-info/top_level.txt,sha256=szZQTUJRotfIaeZCDsOgvofIkLt2ak88RP13oI51-TU,14
|
|
9
|
+
tokencostauto-0.1.359.dist-info/RECORD,,
|
|
@@ -1,9 +0,0 @@
|
|
|
1
|
-
tokencostauto/__init__.py,sha256=-4d_ryFH62SgNXPXA8vGPFZoAKtOBjnsg37EB_RkZG8,289
|
|
2
|
-
tokencostauto/constants.py,sha256=_82MlTkTrdrwzyRosQD7d3JdgNP9KAUM-cZo8DE00P0,3395
|
|
3
|
-
tokencostauto/costs.py,sha256=tXsgrTypq-dCHaHtoXcg2XepezWsAvZpl9gEsv_53iE,10679
|
|
4
|
-
tokencostauto/model_prices.json,sha256=InjFY52gPQIEmewG9tXwCb1Np4Vd9NN9NvLe9Jo2aaw,959967
|
|
5
|
-
tokencostauto-0.1.355.dist-info/licenses/LICENSE,sha256=4PLv_CD6Ughnsvg_nM2XeTqGwVK6lQVR77kVWbPq-0U,1065
|
|
6
|
-
tokencostauto-0.1.355.dist-info/METADATA,sha256=5BQJ1JzFqsHoGiozHKRroLwZJuQow__oO0ex97FASfY,204076
|
|
7
|
-
tokencostauto-0.1.355.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
8
|
-
tokencostauto-0.1.355.dist-info/top_level.txt,sha256=szZQTUJRotfIaeZCDsOgvofIkLt2ak88RP13oI51-TU,14
|
|
9
|
-
tokencostauto-0.1.355.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|