tokencostauto 0.1.264__py3-none-any.whl → 0.1.333__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- tokencostauto/model_prices.json +2454 -290
- {tokencostauto-0.1.264.dist-info → tokencostauto-0.1.333.dist-info}/METADATA +1 -1
- tokencostauto-0.1.333.dist-info/RECORD +9 -0
- tokencostauto-0.1.264.dist-info/RECORD +0 -9
- {tokencostauto-0.1.264.dist-info → tokencostauto-0.1.333.dist-info}/WHEEL +0 -0
- {tokencostauto-0.1.264.dist-info → tokencostauto-0.1.333.dist-info}/licenses/LICENSE +0 -0
- {tokencostauto-0.1.264.dist-info → tokencostauto-0.1.333.dist-info}/top_level.txt +0 -0
tokencostauto/model_prices.json
CHANGED
|
@@ -33,6 +33,7 @@
|
|
|
33
33
|
"supports_response_schema": true,
|
|
34
34
|
"supports_system_messages": true,
|
|
35
35
|
"supports_tool_choice": true,
|
|
36
|
+
"supports_service_tier": true,
|
|
36
37
|
"supports_vision": true
|
|
37
38
|
},
|
|
38
39
|
"gpt-4o-audio-preview": {
|
|
@@ -90,6 +91,7 @@
|
|
|
90
91
|
"supports_response_schema": true,
|
|
91
92
|
"supports_system_messages": true,
|
|
92
93
|
"supports_tool_choice": true,
|
|
94
|
+
"supports_service_tier": true,
|
|
93
95
|
"supports_vision": true
|
|
94
96
|
},
|
|
95
97
|
"gpt-4o-mini-2024-07-18": {
|
|
@@ -115,6 +117,7 @@
|
|
|
115
117
|
"supports_response_schema": true,
|
|
116
118
|
"supports_system_messages": true,
|
|
117
119
|
"supports_tool_choice": true,
|
|
120
|
+
"supports_service_tier": true,
|
|
118
121
|
"supports_vision": true
|
|
119
122
|
},
|
|
120
123
|
"o1-mini": {
|
|
@@ -131,6 +134,7 @@
|
|
|
131
134
|
"supports_vision": true
|
|
132
135
|
},
|
|
133
136
|
"o1-mini-2024-09-12": {
|
|
137
|
+
"deprecation_date": "2025-10-27",
|
|
134
138
|
"cache_read_input_token_cost": 1.5e-06,
|
|
135
139
|
"input_cost_per_token": 3e-06,
|
|
136
140
|
"litellm_provider": "openai",
|
|
@@ -226,6 +230,7 @@
|
|
|
226
230
|
"supports_response_schema": true,
|
|
227
231
|
"supports_system_messages": true,
|
|
228
232
|
"supports_tool_choice": true,
|
|
233
|
+
"supports_service_tier": true,
|
|
229
234
|
"supports_vision": true
|
|
230
235
|
},
|
|
231
236
|
"gpt-4-turbo-preview": {
|
|
@@ -338,6 +343,7 @@
|
|
|
338
343
|
"supports_vision": true
|
|
339
344
|
},
|
|
340
345
|
"gpt-4-1106-preview": {
|
|
346
|
+
"deprecation_date": "2026-03-26",
|
|
341
347
|
"input_cost_per_token": 1e-05,
|
|
342
348
|
"litellm_provider": "openai",
|
|
343
349
|
"max_input_tokens": 128000,
|
|
@@ -352,6 +358,7 @@
|
|
|
352
358
|
"supports_tool_choice": true
|
|
353
359
|
},
|
|
354
360
|
"gpt-4-0125-preview": {
|
|
361
|
+
"deprecation_date": "2026-03-26",
|
|
355
362
|
"input_cost_per_token": 1e-05,
|
|
356
363
|
"litellm_provider": "openai",
|
|
357
364
|
"max_input_tokens": 128000,
|
|
@@ -434,6 +441,7 @@
|
|
|
434
441
|
"supports_tool_choice": true
|
|
435
442
|
},
|
|
436
443
|
"gpt-3.5-turbo-1106": {
|
|
444
|
+
"deprecation_date": "2026-09-28",
|
|
437
445
|
"input_cost_per_token": 1e-06,
|
|
438
446
|
"litellm_provider": "openai",
|
|
439
447
|
"max_input_tokens": 16385,
|
|
@@ -845,6 +853,7 @@
|
|
|
845
853
|
"supports_vision": true
|
|
846
854
|
},
|
|
847
855
|
"azure/gpt-4o-2024-08-06": {
|
|
856
|
+
"deprecation_date": "2026-02-27",
|
|
848
857
|
"cache_read_input_token_cost": 1.25e-06,
|
|
849
858
|
"input_cost_per_token": 2.5e-06,
|
|
850
859
|
"litellm_provider": "azure",
|
|
@@ -876,7 +885,7 @@
|
|
|
876
885
|
},
|
|
877
886
|
"azure/global-standard/gpt-4o-2024-08-06": {
|
|
878
887
|
"cache_read_input_token_cost": 1.25e-06,
|
|
879
|
-
"deprecation_date": "
|
|
888
|
+
"deprecation_date": "2026-02-27",
|
|
880
889
|
"input_cost_per_token": 2.5e-06,
|
|
881
890
|
"litellm_provider": "azure",
|
|
882
891
|
"max_input_tokens": 128000,
|
|
@@ -1184,6 +1193,7 @@
|
|
|
1184
1193
|
"output_cost_per_token": 0.0
|
|
1185
1194
|
},
|
|
1186
1195
|
"azure/text-embedding-3-small": {
|
|
1196
|
+
"deprecation_date": "2026-04-30",
|
|
1187
1197
|
"input_cost_per_token": 2e-08,
|
|
1188
1198
|
"litellm_provider": "azure",
|
|
1189
1199
|
"max_input_tokens": 8191,
|
|
@@ -1942,7 +1952,6 @@
|
|
|
1942
1952
|
"cache_creation_input_token_cost": 3e-07,
|
|
1943
1953
|
"cache_creation_input_token_cost_above_1hr": 6e-06,
|
|
1944
1954
|
"cache_read_input_token_cost": 3e-08,
|
|
1945
|
-
"deprecation_date": "2025-03-01",
|
|
1946
1955
|
"input_cost_per_token": 2.5e-07,
|
|
1947
1956
|
"litellm_provider": "anthropic",
|
|
1948
1957
|
"max_input_tokens": 200000,
|
|
@@ -1978,7 +1987,7 @@
|
|
|
1978
1987
|
"cache_creation_input_token_cost": 1.875e-05,
|
|
1979
1988
|
"cache_creation_input_token_cost_above_1hr": 6e-06,
|
|
1980
1989
|
"cache_read_input_token_cost": 1.5e-06,
|
|
1981
|
-
"deprecation_date": "
|
|
1990
|
+
"deprecation_date": "2026-05-01",
|
|
1982
1991
|
"input_cost_per_token": 1.5e-05,
|
|
1983
1992
|
"litellm_provider": "anthropic",
|
|
1984
1993
|
"max_input_tokens": 200000,
|
|
@@ -6433,13 +6442,13 @@
|
|
|
6433
6442
|
"supports_tool_choice": false
|
|
6434
6443
|
},
|
|
6435
6444
|
"deepinfra/Gryphe/MythoMax-L2-13b": {
|
|
6436
|
-
"
|
|
6437
|
-
"litellm_provider": "deepinfra",
|
|
6445
|
+
"max_tokens": 4096,
|
|
6438
6446
|
"max_input_tokens": 4096,
|
|
6439
6447
|
"max_output_tokens": 4096,
|
|
6440
|
-
"
|
|
6448
|
+
"input_cost_per_token": 8e-08,
|
|
6449
|
+
"output_cost_per_token": 9e-08,
|
|
6450
|
+
"litellm_provider": "deepinfra",
|
|
6441
6451
|
"mode": "chat",
|
|
6442
|
-
"output_cost_per_token": 7.2e-08,
|
|
6443
6452
|
"supports_tool_choice": true
|
|
6444
6453
|
},
|
|
6445
6454
|
"deepinfra/mistralai/Mistral-7B-Instruct-v0.1": {
|
|
@@ -6502,13 +6511,13 @@
|
|
|
6502
6511
|
"supports_tool_choice": true
|
|
6503
6512
|
},
|
|
6504
6513
|
"deepinfra/mistralai/Mixtral-8x7B-Instruct-v0.1": {
|
|
6505
|
-
"
|
|
6506
|
-
"litellm_provider": "deepinfra",
|
|
6514
|
+
"max_tokens": 32768,
|
|
6507
6515
|
"max_input_tokens": 32768,
|
|
6508
6516
|
"max_output_tokens": 32768,
|
|
6509
|
-
"
|
|
6517
|
+
"input_cost_per_token": 4e-07,
|
|
6518
|
+
"output_cost_per_token": 4e-07,
|
|
6519
|
+
"litellm_provider": "deepinfra",
|
|
6510
6520
|
"mode": "chat",
|
|
6511
|
-
"output_cost_per_token": 2.4e-07,
|
|
6512
6521
|
"supports_tool_choice": true
|
|
6513
6522
|
},
|
|
6514
6523
|
"deepinfra/deepinfra/airoboros-70b": {
|
|
@@ -6581,13 +6590,13 @@
|
|
|
6581
6590
|
"supports_tool_choice": true
|
|
6582
6591
|
},
|
|
6583
6592
|
"deepinfra/meta-llama/Meta-Llama-3-8B-Instruct": {
|
|
6584
|
-
"
|
|
6585
|
-
"litellm_provider": "deepinfra",
|
|
6593
|
+
"max_tokens": 8192,
|
|
6586
6594
|
"max_input_tokens": 8192,
|
|
6587
6595
|
"max_output_tokens": 8192,
|
|
6588
|
-
"
|
|
6589
|
-
"mode": "chat",
|
|
6596
|
+
"input_cost_per_token": 3e-08,
|
|
6590
6597
|
"output_cost_per_token": 6e-08,
|
|
6598
|
+
"litellm_provider": "deepinfra",
|
|
6599
|
+
"mode": "chat",
|
|
6591
6600
|
"supports_tool_choice": true
|
|
6592
6601
|
},
|
|
6593
6602
|
"deepinfra/meta-llama/Meta-Llama-3-70B-Instruct": {
|
|
@@ -7720,6 +7729,7 @@
|
|
|
7720
7729
|
"supports_response_schema": true,
|
|
7721
7730
|
"supports_system_messages": true,
|
|
7722
7731
|
"supports_tool_choice": true,
|
|
7732
|
+
"supports_service_tier": true,
|
|
7723
7733
|
"supports_vision": true
|
|
7724
7734
|
},
|
|
7725
7735
|
"ft:gpt-4o-2024-11-20": {
|
|
@@ -7741,6 +7751,7 @@
|
|
|
7741
7751
|
"supports_vision": true
|
|
7742
7752
|
},
|
|
7743
7753
|
"azure/gpt-4o-2024-11-20": {
|
|
7754
|
+
"deprecation_date": "2026-03-01",
|
|
7744
7755
|
"cache_read_input_token_cost": 1.25e-06,
|
|
7745
7756
|
"input_cost_per_token": 2.75e-06,
|
|
7746
7757
|
"litellm_provider": "azure",
|
|
@@ -7758,7 +7769,7 @@
|
|
|
7758
7769
|
},
|
|
7759
7770
|
"azure/global-standard/gpt-4o-2024-11-20": {
|
|
7760
7771
|
"cache_read_input_token_cost": 1.25e-06,
|
|
7761
|
-
"deprecation_date": "
|
|
7772
|
+
"deprecation_date": "2026-03-01",
|
|
7762
7773
|
"input_cost_per_token": 2.5e-06,
|
|
7763
7774
|
"litellm_provider": "azure",
|
|
7764
7775
|
"max_input_tokens": 128000,
|
|
@@ -9795,7 +9806,7 @@
|
|
|
9795
9806
|
"cache_creation_input_token_cost": 3.75e-06,
|
|
9796
9807
|
"cache_creation_input_token_cost_above_1hr": 6e-06,
|
|
9797
9808
|
"cache_read_input_token_cost": 3e-07,
|
|
9798
|
-
"deprecation_date": "2026-02-
|
|
9809
|
+
"deprecation_date": "2026-02-19",
|
|
9799
9810
|
"input_cost_per_token": 3e-06,
|
|
9800
9811
|
"litellm_provider": "anthropic",
|
|
9801
9812
|
"max_input_tokens": 200000,
|
|
@@ -10204,6 +10215,7 @@
|
|
|
10204
10215
|
"supports_vision": false
|
|
10205
10216
|
},
|
|
10206
10217
|
"azure/us/gpt-4o-2024-11-20": {
|
|
10218
|
+
"deprecation_date": "2026-03-01",
|
|
10207
10219
|
"cache_creation_input_token_cost": 1.38e-06,
|
|
10208
10220
|
"input_cost_per_token": 2.75e-06,
|
|
10209
10221
|
"litellm_provider": "azure",
|
|
@@ -10219,6 +10231,7 @@
|
|
|
10219
10231
|
"supports_vision": true
|
|
10220
10232
|
},
|
|
10221
10233
|
"azure/eu/gpt-4o-2024-11-20": {
|
|
10234
|
+
"deprecation_date": "2026-03-01",
|
|
10222
10235
|
"cache_creation_input_token_cost": 1.38e-06,
|
|
10223
10236
|
"input_cost_per_token": 2.75e-06,
|
|
10224
10237
|
"litellm_provider": "azure",
|
|
@@ -10234,6 +10247,7 @@
|
|
|
10234
10247
|
"supports_vision": true
|
|
10235
10248
|
},
|
|
10236
10249
|
"azure/us/gpt-4o-2024-08-06": {
|
|
10250
|
+
"deprecation_date": "2026-02-27",
|
|
10237
10251
|
"cache_read_input_token_cost": 1.375e-06,
|
|
10238
10252
|
"input_cost_per_token": 2.75e-06,
|
|
10239
10253
|
"litellm_provider": "azure",
|
|
@@ -10250,6 +10264,7 @@
|
|
|
10250
10264
|
"supports_vision": true
|
|
10251
10265
|
},
|
|
10252
10266
|
"azure/eu/gpt-4o-2024-08-06": {
|
|
10267
|
+
"deprecation_date": "2026-02-27",
|
|
10253
10268
|
"cache_read_input_token_cost": 1.375e-06,
|
|
10254
10269
|
"input_cost_per_token": 2.75e-06,
|
|
10255
10270
|
"litellm_provider": "azure",
|
|
@@ -10746,6 +10761,7 @@
|
|
|
10746
10761
|
"mode": "chat"
|
|
10747
10762
|
},
|
|
10748
10763
|
"azure/global/gpt-4o-2024-11-20": {
|
|
10764
|
+
"deprecation_date": "2026-03-01",
|
|
10749
10765
|
"cache_read_input_token_cost": 1.25e-06,
|
|
10750
10766
|
"input_cost_per_token": 2.5e-06,
|
|
10751
10767
|
"litellm_provider": "azure",
|
|
@@ -10762,6 +10778,7 @@
|
|
|
10762
10778
|
"supports_vision": true
|
|
10763
10779
|
},
|
|
10764
10780
|
"azure/global/gpt-4o-2024-08-06": {
|
|
10781
|
+
"deprecation_date": "2026-02-27",
|
|
10765
10782
|
"cache_read_input_token_cost": 1.25e-06,
|
|
10766
10783
|
"input_cost_per_token": 2.5e-06,
|
|
10767
10784
|
"litellm_provider": "azure",
|
|
@@ -11009,6 +11026,7 @@
|
|
|
11009
11026
|
"supports_response_schema": true,
|
|
11010
11027
|
"supports_system_messages": true,
|
|
11011
11028
|
"supports_tool_choice": true,
|
|
11029
|
+
"supports_service_tier": true,
|
|
11012
11030
|
"supports_vision": true
|
|
11013
11031
|
},
|
|
11014
11032
|
"gpt-4.1-2025-04-14": {
|
|
@@ -11042,6 +11060,7 @@
|
|
|
11042
11060
|
"supports_response_schema": true,
|
|
11043
11061
|
"supports_system_messages": true,
|
|
11044
11062
|
"supports_tool_choice": true,
|
|
11063
|
+
"supports_service_tier": true,
|
|
11045
11064
|
"supports_vision": true
|
|
11046
11065
|
},
|
|
11047
11066
|
"gpt-4.1-mini": {
|
|
@@ -11078,6 +11097,7 @@
|
|
|
11078
11097
|
"supports_response_schema": true,
|
|
11079
11098
|
"supports_system_messages": true,
|
|
11080
11099
|
"supports_tool_choice": true,
|
|
11100
|
+
"supports_service_tier": true,
|
|
11081
11101
|
"supports_vision": true
|
|
11082
11102
|
},
|
|
11083
11103
|
"gpt-4.1-mini-2025-04-14": {
|
|
@@ -11111,6 +11131,7 @@
|
|
|
11111
11131
|
"supports_response_schema": true,
|
|
11112
11132
|
"supports_system_messages": true,
|
|
11113
11133
|
"supports_tool_choice": true,
|
|
11134
|
+
"supports_service_tier": true,
|
|
11114
11135
|
"supports_vision": true
|
|
11115
11136
|
},
|
|
11116
11137
|
"gpt-4.1-nano": {
|
|
@@ -11147,6 +11168,7 @@
|
|
|
11147
11168
|
"supports_response_schema": true,
|
|
11148
11169
|
"supports_system_messages": true,
|
|
11149
11170
|
"supports_tool_choice": true,
|
|
11171
|
+
"supports_service_tier": true,
|
|
11150
11172
|
"supports_vision": true
|
|
11151
11173
|
},
|
|
11152
11174
|
"gpt-4.1-nano-2025-04-14": {
|
|
@@ -11180,16 +11202,17 @@
|
|
|
11180
11202
|
"supports_response_schema": true,
|
|
11181
11203
|
"supports_system_messages": true,
|
|
11182
11204
|
"supports_tool_choice": true,
|
|
11205
|
+
"supports_service_tier": true,
|
|
11183
11206
|
"supports_vision": true
|
|
11184
11207
|
},
|
|
11185
11208
|
"watsonx/ibm/granite-3-8b-instruct": {
|
|
11186
|
-
"input_cost_per_token":
|
|
11209
|
+
"input_cost_per_token": 2e-07,
|
|
11187
11210
|
"litellm_provider": "watsonx",
|
|
11188
11211
|
"max_input_tokens": 8192,
|
|
11189
11212
|
"max_output_tokens": 1024,
|
|
11190
11213
|
"max_tokens": 8192,
|
|
11191
11214
|
"mode": "chat",
|
|
11192
|
-
"output_cost_per_token":
|
|
11215
|
+
"output_cost_per_token": 2e-07,
|
|
11193
11216
|
"supports_audio_input": false,
|
|
11194
11217
|
"supports_audio_output": false,
|
|
11195
11218
|
"supports_function_calling": true,
|
|
@@ -11262,6 +11285,7 @@
|
|
|
11262
11285
|
"supports_reasoning": true,
|
|
11263
11286
|
"supports_response_schema": true,
|
|
11264
11287
|
"supports_tool_choice": true,
|
|
11288
|
+
"supports_service_tier": true,
|
|
11265
11289
|
"supports_vision": true
|
|
11266
11290
|
},
|
|
11267
11291
|
"o3-2025-04-16": {
|
|
@@ -11293,6 +11317,7 @@
|
|
|
11293
11317
|
"supports_reasoning": true,
|
|
11294
11318
|
"supports_response_schema": true,
|
|
11295
11319
|
"supports_tool_choice": true,
|
|
11320
|
+
"supports_service_tier": true,
|
|
11296
11321
|
"supports_vision": true
|
|
11297
11322
|
},
|
|
11298
11323
|
"o4-mini": {
|
|
@@ -11317,6 +11342,7 @@
|
|
|
11317
11342
|
"supports_reasoning": true,
|
|
11318
11343
|
"supports_response_schema": true,
|
|
11319
11344
|
"supports_tool_choice": true,
|
|
11345
|
+
"supports_service_tier": true,
|
|
11320
11346
|
"supports_vision": true
|
|
11321
11347
|
},
|
|
11322
11348
|
"o4-mini-2025-04-16": {
|
|
@@ -11335,6 +11361,7 @@
|
|
|
11335
11361
|
"supports_reasoning": true,
|
|
11336
11362
|
"supports_response_schema": true,
|
|
11337
11363
|
"supports_tool_choice": true,
|
|
11364
|
+
"supports_service_tier": true,
|
|
11338
11365
|
"supports_vision": true
|
|
11339
11366
|
},
|
|
11340
11367
|
"gpt-image-1": {
|
|
@@ -11592,6 +11619,7 @@
|
|
|
11592
11619
|
"supports_web_search": false
|
|
11593
11620
|
},
|
|
11594
11621
|
"azure/gpt-4.1-2025-04-14": {
|
|
11622
|
+
"deprecation_date": "2026-11-04",
|
|
11595
11623
|
"cache_read_input_token_cost": 5e-07,
|
|
11596
11624
|
"input_cost_per_token": 2e-06,
|
|
11597
11625
|
"input_cost_per_token_batches": 1e-06,
|
|
@@ -11658,6 +11686,7 @@
|
|
|
11658
11686
|
"supports_web_search": false
|
|
11659
11687
|
},
|
|
11660
11688
|
"azure/gpt-4.1-mini-2025-04-14": {
|
|
11689
|
+
"deprecation_date": "2026-11-04",
|
|
11661
11690
|
"cache_read_input_token_cost": 1e-07,
|
|
11662
11691
|
"input_cost_per_token": 4e-07,
|
|
11663
11692
|
"input_cost_per_token_batches": 2e-07,
|
|
@@ -11723,6 +11752,7 @@
|
|
|
11723
11752
|
"supports_vision": true
|
|
11724
11753
|
},
|
|
11725
11754
|
"azure/gpt-4.1-nano-2025-04-14": {
|
|
11755
|
+
"deprecation_date": "2026-11-04",
|
|
11726
11756
|
"cache_read_input_token_cost": 2.5e-08,
|
|
11727
11757
|
"input_cost_per_token": 1e-07,
|
|
11728
11758
|
"input_cost_per_token_batches": 5e-08,
|
|
@@ -11784,6 +11814,7 @@
|
|
|
11784
11814
|
"supports_vision": true
|
|
11785
11815
|
},
|
|
11786
11816
|
"azure/o3-2025-04-16": {
|
|
11817
|
+
"deprecation_date": "2026-04-16",
|
|
11787
11818
|
"cache_read_input_token_cost": 2.5e-06,
|
|
11788
11819
|
"input_cost_per_token": 1e-05,
|
|
11789
11820
|
"litellm_provider": "azure",
|
|
@@ -13948,6 +13979,7 @@
|
|
|
13948
13979
|
"cache_creation_input_token_cost_above_1hr": 3e-05,
|
|
13949
13980
|
"cache_read_input_token_cost": 1.5e-06,
|
|
13950
13981
|
"input_cost_per_token": 1.5e-05,
|
|
13982
|
+
"deprecation_date": "2026-05-14",
|
|
13951
13983
|
"litellm_provider": "anthropic",
|
|
13952
13984
|
"max_input_tokens": 200000,
|
|
13953
13985
|
"max_output_tokens": 32000,
|
|
@@ -13971,6 +14003,7 @@
|
|
|
13971
14003
|
"tool_use_system_prompt_tokens": 159
|
|
13972
14004
|
},
|
|
13973
14005
|
"claude-sonnet-4-20250514": {
|
|
14006
|
+
"deprecation_date": "2026-05-14",
|
|
13974
14007
|
"cache_creation_input_token_cost": 3.75e-06,
|
|
13975
14008
|
"cache_creation_input_token_cost_above_1hr": 6e-06,
|
|
13976
14009
|
"cache_read_input_token_cost": 3e-07,
|
|
@@ -14368,7 +14401,7 @@
|
|
|
14368
14401
|
"input_cost_per_token_above_200k_tokens": 6e-06,
|
|
14369
14402
|
"litellm_provider": "anthropic",
|
|
14370
14403
|
"max_input_tokens": 1000000,
|
|
14371
|
-
"max_output_tokens":
|
|
14404
|
+
"max_output_tokens": 64000,
|
|
14372
14405
|
"max_tokens": 1000000,
|
|
14373
14406
|
"mode": "chat",
|
|
14374
14407
|
"output_cost_per_token": 1.5e-05,
|
|
@@ -15396,7 +15429,7 @@
|
|
|
15396
15429
|
"tpm": 800000
|
|
15397
15430
|
},
|
|
15398
15431
|
"gemini-2.5-flash": {
|
|
15399
|
-
"cache_read_input_token_cost":
|
|
15432
|
+
"cache_read_input_token_cost": 3e-08,
|
|
15400
15433
|
"input_cost_per_audio_token": 1e-06,
|
|
15401
15434
|
"input_cost_per_token": 3e-07,
|
|
15402
15435
|
"litellm_provider": "vertex_ai-language-models",
|
|
@@ -15546,6 +15579,10 @@
|
|
|
15546
15579
|
},
|
|
15547
15580
|
"openrouter/anthropic/claude-sonnet-4": {
|
|
15548
15581
|
"input_cost_per_image": 0.0048,
|
|
15582
|
+
"cache_creation_input_token_cost": 3.75e-06,
|
|
15583
|
+
"cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
|
|
15584
|
+
"cache_read_input_token_cost": 3e-07,
|
|
15585
|
+
"cache_read_input_token_cost_above_200k_tokens": 6e-07,
|
|
15549
15586
|
"input_cost_per_token": 3e-06,
|
|
15550
15587
|
"input_cost_per_token_above_200k_tokens": 6e-06,
|
|
15551
15588
|
"output_cost_per_token_above_200k_tokens": 2.25e-05,
|
|
@@ -15558,6 +15595,7 @@
|
|
|
15558
15595
|
"supports_assistant_prefill": true,
|
|
15559
15596
|
"supports_computer_use": true,
|
|
15560
15597
|
"supports_function_calling": true,
|
|
15598
|
+
"supports_prompt_caching": true,
|
|
15561
15599
|
"supports_reasoning": true,
|
|
15562
15600
|
"supports_tool_choice": true,
|
|
15563
15601
|
"supports_vision": true,
|
|
@@ -15658,7 +15696,8 @@
|
|
|
15658
15696
|
"tpm": 250000
|
|
15659
15697
|
},
|
|
15660
15698
|
"gemini-2.5-pro": {
|
|
15661
|
-
"cache_read_input_token_cost":
|
|
15699
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
15700
|
+
"cache_creation_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
15662
15701
|
"input_cost_per_token": 1.25e-06,
|
|
15663
15702
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
15664
15703
|
"litellm_provider": "vertex_ai-language-models",
|
|
@@ -16228,35 +16267,36 @@
|
|
|
16228
16267
|
"output_cost_per_token": 1.5e-05,
|
|
16229
16268
|
"source": "https://docs.x.ai/docs/models",
|
|
16230
16269
|
"supports_function_calling": true,
|
|
16231
|
-
"supports_reasoning": true,
|
|
16232
16270
|
"supports_tool_choice": true,
|
|
16233
16271
|
"supports_web_search": true
|
|
16234
16272
|
},
|
|
16235
16273
|
"xai/grok-4-0709": {
|
|
16236
16274
|
"input_cost_per_token": 3e-06,
|
|
16275
|
+
"input_cost_per_token_above_128k_tokens": 6e-06,
|
|
16237
16276
|
"litellm_provider": "xai",
|
|
16238
16277
|
"max_input_tokens": 256000,
|
|
16239
16278
|
"max_output_tokens": 256000,
|
|
16240
16279
|
"max_tokens": 256000,
|
|
16241
16280
|
"mode": "chat",
|
|
16242
16281
|
"output_cost_per_token": 1.5e-05,
|
|
16282
|
+
"output_cost_per_token_above_128k_tokens": 3e-05,
|
|
16243
16283
|
"source": "https://docs.x.ai/docs/models",
|
|
16244
16284
|
"supports_function_calling": true,
|
|
16245
|
-
"supports_reasoning": true,
|
|
16246
16285
|
"supports_tool_choice": true,
|
|
16247
16286
|
"supports_web_search": true
|
|
16248
16287
|
},
|
|
16249
16288
|
"xai/grok-4-latest": {
|
|
16250
16289
|
"input_cost_per_token": 3e-06,
|
|
16290
|
+
"input_cost_per_token_above_128k_tokens": 6e-06,
|
|
16251
16291
|
"litellm_provider": "xai",
|
|
16252
16292
|
"max_input_tokens": 256000,
|
|
16253
16293
|
"max_output_tokens": 256000,
|
|
16254
16294
|
"max_tokens": 256000,
|
|
16255
16295
|
"mode": "chat",
|
|
16256
16296
|
"output_cost_per_token": 1.5e-05,
|
|
16297
|
+
"output_cost_per_token_above_128k_tokens": 3e-05,
|
|
16257
16298
|
"source": "https://docs.x.ai/docs/models",
|
|
16258
16299
|
"supports_function_calling": true,
|
|
16259
|
-
"supports_reasoning": true,
|
|
16260
16300
|
"supports_tool_choice": true,
|
|
16261
16301
|
"supports_web_search": true
|
|
16262
16302
|
},
|
|
@@ -16977,7 +17017,8 @@
|
|
|
16977
17017
|
"max_tokens": 8192,
|
|
16978
17018
|
"mode": "chat",
|
|
16979
17019
|
"output_cost_per_token": 6.3e-07,
|
|
16980
|
-
"supports_tool_choice": true
|
|
17020
|
+
"supports_tool_choice": true,
|
|
17021
|
+
"supports_vision": true
|
|
16981
17022
|
},
|
|
16982
17023
|
"fireworks_ai/accounts/fireworks/models/kimi-k2-instruct": {
|
|
16983
17024
|
"input_cost_per_token": 6e-07,
|
|
@@ -17836,6 +17877,7 @@
|
|
|
17836
17877
|
"cache_creation_input_token_cost_above_1hr": 3e-05,
|
|
17837
17878
|
"cache_read_input_token_cost": 1.5e-06,
|
|
17838
17879
|
"input_cost_per_token": 1.5e-05,
|
|
17880
|
+
"deprecation_date": "2026-08-05",
|
|
17839
17881
|
"litellm_provider": "anthropic",
|
|
17840
17882
|
"max_input_tokens": 200000,
|
|
17841
17883
|
"max_output_tokens": 32000,
|
|
@@ -17995,8 +18037,8 @@
|
|
|
17995
18037
|
"input_cost_per_token_batches": 7.5e-06,
|
|
17996
18038
|
"litellm_provider": "vertex_ai-anthropic_models",
|
|
17997
18039
|
"max_input_tokens": 200000,
|
|
17998
|
-
"max_output_tokens":
|
|
17999
|
-
"max_tokens":
|
|
18040
|
+
"max_output_tokens": 32000,
|
|
18041
|
+
"max_tokens": 32000,
|
|
18000
18042
|
"mode": "chat",
|
|
18001
18043
|
"output_cost_per_token": 7.5e-05,
|
|
18002
18044
|
"output_cost_per_token_batches": 3.75e-05,
|
|
@@ -18012,8 +18054,8 @@
|
|
|
18012
18054
|
"input_cost_per_token_batches": 7.5e-06,
|
|
18013
18055
|
"litellm_provider": "vertex_ai-anthropic_models",
|
|
18014
18056
|
"max_input_tokens": 200000,
|
|
18015
|
-
"max_output_tokens":
|
|
18016
|
-
"max_tokens":
|
|
18057
|
+
"max_output_tokens": 32000,
|
|
18058
|
+
"max_tokens": 32000,
|
|
18017
18059
|
"mode": "chat",
|
|
18018
18060
|
"output_cost_per_token": 7.5e-05,
|
|
18019
18061
|
"output_cost_per_token_batches": 3.75e-05,
|
|
@@ -18084,6 +18126,7 @@
|
|
|
18084
18126
|
"supports_response_schema": true,
|
|
18085
18127
|
"supports_system_messages": true,
|
|
18086
18128
|
"supports_tool_choice": true,
|
|
18129
|
+
"supports_service_tier": true,
|
|
18087
18130
|
"supports_vision": true
|
|
18088
18131
|
},
|
|
18089
18132
|
"gpt-5-mini": {
|
|
@@ -18122,6 +18165,7 @@
|
|
|
18122
18165
|
"supports_response_schema": true,
|
|
18123
18166
|
"supports_system_messages": true,
|
|
18124
18167
|
"supports_tool_choice": true,
|
|
18168
|
+
"supports_service_tier": true,
|
|
18125
18169
|
"supports_vision": true
|
|
18126
18170
|
},
|
|
18127
18171
|
"gpt-5-nano": {
|
|
@@ -18196,9 +18240,9 @@
|
|
|
18196
18240
|
"cache_read_input_token_cost": 1.25e-07,
|
|
18197
18241
|
"input_cost_per_token": 1.25e-06,
|
|
18198
18242
|
"litellm_provider": "openai",
|
|
18199
|
-
"max_input_tokens":
|
|
18200
|
-
"max_output_tokens":
|
|
18201
|
-
"max_tokens":
|
|
18243
|
+
"max_input_tokens": 128000,
|
|
18244
|
+
"max_output_tokens": 16384,
|
|
18245
|
+
"max_tokens": 16384,
|
|
18202
18246
|
"mode": "chat",
|
|
18203
18247
|
"output_cost_per_token": 1e-05,
|
|
18204
18248
|
"supported_endpoints": [
|
|
@@ -18299,6 +18343,7 @@
|
|
|
18299
18343
|
"supports_response_schema": true,
|
|
18300
18344
|
"supports_system_messages": true,
|
|
18301
18345
|
"supports_tool_choice": true,
|
|
18346
|
+
"supports_service_tier": true,
|
|
18302
18347
|
"supports_vision": true
|
|
18303
18348
|
},
|
|
18304
18349
|
"gpt-5-nano-2025-08-07": {
|
|
@@ -18565,9 +18610,9 @@
|
|
|
18565
18610
|
"cache_read_input_token_cost": 1.25e-07,
|
|
18566
18611
|
"input_cost_per_token": 1.25e-06,
|
|
18567
18612
|
"litellm_provider": "azure",
|
|
18568
|
-
"max_input_tokens":
|
|
18569
|
-
"max_output_tokens":
|
|
18570
|
-
"max_tokens":
|
|
18613
|
+
"max_input_tokens": 128000,
|
|
18614
|
+
"max_output_tokens": 16384,
|
|
18615
|
+
"max_tokens": 16384,
|
|
18571
18616
|
"mode": "chat",
|
|
18572
18617
|
"output_cost_per_token": 1e-05,
|
|
18573
18618
|
"supported_endpoints": [
|
|
@@ -19211,23 +19256,23 @@
|
|
|
19211
19256
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing"
|
|
19212
19257
|
},
|
|
19213
19258
|
"deepinfra/deepseek-ai/DeepSeek-V3": {
|
|
19214
|
-
"
|
|
19215
|
-
"litellm_provider": "deepinfra",
|
|
19259
|
+
"max_tokens": 163840,
|
|
19216
19260
|
"max_input_tokens": 163840,
|
|
19217
19261
|
"max_output_tokens": 163840,
|
|
19218
|
-
"
|
|
19219
|
-
"mode": "chat",
|
|
19262
|
+
"input_cost_per_token": 3.8e-07,
|
|
19220
19263
|
"output_cost_per_token": 8.9e-07,
|
|
19264
|
+
"litellm_provider": "deepinfra",
|
|
19265
|
+
"mode": "chat",
|
|
19221
19266
|
"supports_tool_choice": true
|
|
19222
19267
|
},
|
|
19223
19268
|
"deepinfra/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo": {
|
|
19224
|
-
"
|
|
19225
|
-
"litellm_provider": "deepinfra",
|
|
19269
|
+
"max_tokens": 131072,
|
|
19226
19270
|
"max_input_tokens": 131072,
|
|
19227
19271
|
"max_output_tokens": 131072,
|
|
19228
|
-
"
|
|
19272
|
+
"input_cost_per_token": 2e-08,
|
|
19273
|
+
"output_cost_per_token": 3e-08,
|
|
19274
|
+
"litellm_provider": "deepinfra",
|
|
19229
19275
|
"mode": "chat",
|
|
19230
|
-
"output_cost_per_token": 2e-08,
|
|
19231
19276
|
"supports_tool_choice": true
|
|
19232
19277
|
},
|
|
19233
19278
|
"deepinfra/google/gemma-2-9b-it": {
|
|
@@ -19241,13 +19286,13 @@
|
|
|
19241
19286
|
"supports_tool_choice": false
|
|
19242
19287
|
},
|
|
19243
19288
|
"deepinfra/deepseek-ai/DeepSeek-R1-0528-Turbo": {
|
|
19244
|
-
"
|
|
19245
|
-
"litellm_provider": "deepinfra",
|
|
19289
|
+
"max_tokens": 32768,
|
|
19246
19290
|
"max_input_tokens": 32768,
|
|
19247
19291
|
"max_output_tokens": 32768,
|
|
19248
|
-
"
|
|
19249
|
-
"mode": "chat",
|
|
19292
|
+
"input_cost_per_token": 1e-06,
|
|
19250
19293
|
"output_cost_per_token": 3e-06,
|
|
19294
|
+
"litellm_provider": "deepinfra",
|
|
19295
|
+
"mode": "chat",
|
|
19251
19296
|
"supports_tool_choice": true
|
|
19252
19297
|
},
|
|
19253
19298
|
"deepinfra/Qwen/Qwen2-7B-Instruct": {
|
|
@@ -19271,13 +19316,13 @@
|
|
|
19271
19316
|
"supports_tool_choice": false
|
|
19272
19317
|
},
|
|
19273
19318
|
"deepinfra/meta-llama/Llama-3.3-70B-Instruct": {
|
|
19274
|
-
"
|
|
19275
|
-
"litellm_provider": "deepinfra",
|
|
19319
|
+
"max_tokens": 131072,
|
|
19276
19320
|
"max_input_tokens": 131072,
|
|
19277
19321
|
"max_output_tokens": 131072,
|
|
19278
|
-
"
|
|
19279
|
-
"mode": "chat",
|
|
19322
|
+
"input_cost_per_token": 2.3e-07,
|
|
19280
19323
|
"output_cost_per_token": 4e-07,
|
|
19324
|
+
"litellm_provider": "deepinfra",
|
|
19325
|
+
"mode": "chat",
|
|
19281
19326
|
"supports_tool_choice": true
|
|
19282
19327
|
},
|
|
19283
19328
|
"deepinfra/microsoft/Phi-4-multimodal-instruct": {
|
|
@@ -19321,35 +19366,34 @@
|
|
|
19321
19366
|
"supports_tool_choice": false
|
|
19322
19367
|
},
|
|
19323
19368
|
"deepinfra/mistralai/Mistral-Small-3.2-24B-Instruct-2506": {
|
|
19324
|
-
"
|
|
19325
|
-
"litellm_provider": "deepinfra",
|
|
19369
|
+
"max_tokens": 128000,
|
|
19326
19370
|
"max_input_tokens": 128000,
|
|
19327
19371
|
"max_output_tokens": 128000,
|
|
19328
|
-
"
|
|
19372
|
+
"input_cost_per_token": 7.5e-08,
|
|
19373
|
+
"output_cost_per_token": 2e-07,
|
|
19374
|
+
"litellm_provider": "deepinfra",
|
|
19329
19375
|
"mode": "chat",
|
|
19330
|
-
"output_cost_per_token": 1e-07,
|
|
19331
19376
|
"supports_tool_choice": true
|
|
19332
19377
|
},
|
|
19333
19378
|
"deepinfra/deepseek-ai/DeepSeek-V3-0324": {
|
|
19334
|
-
"
|
|
19335
|
-
"input_cost_per_token": 2.8e-07,
|
|
19336
|
-
"litellm_provider": "deepinfra",
|
|
19379
|
+
"max_tokens": 163840,
|
|
19337
19380
|
"max_input_tokens": 163840,
|
|
19338
19381
|
"max_output_tokens": 163840,
|
|
19339
|
-
"
|
|
19340
|
-
"mode": "chat",
|
|
19382
|
+
"input_cost_per_token": 2.5e-07,
|
|
19341
19383
|
"output_cost_per_token": 8.8e-07,
|
|
19384
|
+
"litellm_provider": "deepinfra",
|
|
19385
|
+
"mode": "chat",
|
|
19342
19386
|
"supports_tool_choice": true
|
|
19343
19387
|
},
|
|
19344
19388
|
"deepinfra/anthropic/claude-3-7-sonnet-latest": {
|
|
19345
|
-
"
|
|
19346
|
-
"input_cost_per_token": 3.3e-06,
|
|
19347
|
-
"litellm_provider": "deepinfra",
|
|
19389
|
+
"max_tokens": 200000,
|
|
19348
19390
|
"max_input_tokens": 200000,
|
|
19349
19391
|
"max_output_tokens": 200000,
|
|
19350
|
-
"
|
|
19351
|
-
"mode": "chat",
|
|
19392
|
+
"input_cost_per_token": 3.3e-06,
|
|
19352
19393
|
"output_cost_per_token": 1.65e-05,
|
|
19394
|
+
"cache_read_input_token_cost": 3.3e-07,
|
|
19395
|
+
"litellm_provider": "deepinfra",
|
|
19396
|
+
"mode": "chat",
|
|
19353
19397
|
"supports_tool_choice": true
|
|
19354
19398
|
},
|
|
19355
19399
|
"deepinfra/cognitivecomputations/dolphin-2.9.1-llama-3-70b": {
|
|
@@ -19373,13 +19417,13 @@
|
|
|
19373
19417
|
"supports_tool_choice": false
|
|
19374
19418
|
},
|
|
19375
19419
|
"deepinfra/Qwen/Qwen3-235B-A22B": {
|
|
19376
|
-
"
|
|
19377
|
-
"litellm_provider": "deepinfra",
|
|
19420
|
+
"max_tokens": 40960,
|
|
19378
19421
|
"max_input_tokens": 40960,
|
|
19379
19422
|
"max_output_tokens": 40960,
|
|
19380
|
-
"
|
|
19423
|
+
"input_cost_per_token": 1.8e-07,
|
|
19424
|
+
"output_cost_per_token": 5.4e-07,
|
|
19425
|
+
"litellm_provider": "deepinfra",
|
|
19381
19426
|
"mode": "chat",
|
|
19382
|
-
"output_cost_per_token": 6e-07,
|
|
19383
19427
|
"supports_tool_choice": true
|
|
19384
19428
|
},
|
|
19385
19429
|
"deepinfra/deepseek-ai/DeepSeek-V3-0324-Turbo": {
|
|
@@ -19394,33 +19438,33 @@
|
|
|
19394
19438
|
"supports_reasoning": true
|
|
19395
19439
|
},
|
|
19396
19440
|
"deepinfra/microsoft/WizardLM-2-8x22B": {
|
|
19397
|
-
"
|
|
19398
|
-
"litellm_provider": "deepinfra",
|
|
19441
|
+
"max_tokens": 65536,
|
|
19399
19442
|
"max_input_tokens": 65536,
|
|
19400
19443
|
"max_output_tokens": 65536,
|
|
19401
|
-
"
|
|
19402
|
-
"mode": "chat",
|
|
19444
|
+
"input_cost_per_token": 4.8e-07,
|
|
19403
19445
|
"output_cost_per_token": 4.8e-07,
|
|
19446
|
+
"litellm_provider": "deepinfra",
|
|
19447
|
+
"mode": "chat",
|
|
19404
19448
|
"supports_tool_choice": false
|
|
19405
19449
|
},
|
|
19406
19450
|
"deepinfra/Sao10K/L3-8B-Lunaris-v1-Turbo": {
|
|
19407
|
-
"
|
|
19408
|
-
"litellm_provider": "deepinfra",
|
|
19451
|
+
"max_tokens": 8192,
|
|
19409
19452
|
"max_input_tokens": 8192,
|
|
19410
19453
|
"max_output_tokens": 8192,
|
|
19411
|
-
"
|
|
19412
|
-
"mode": "chat",
|
|
19454
|
+
"input_cost_per_token": 4e-08,
|
|
19413
19455
|
"output_cost_per_token": 5e-08,
|
|
19456
|
+
"litellm_provider": "deepinfra",
|
|
19457
|
+
"mode": "chat",
|
|
19414
19458
|
"supports_tool_choice": false
|
|
19415
19459
|
},
|
|
19416
19460
|
"deepinfra/meta-llama/Llama-Guard-4-12B": {
|
|
19417
|
-
"
|
|
19418
|
-
"litellm_provider": "deepinfra",
|
|
19461
|
+
"max_tokens": 163840,
|
|
19419
19462
|
"max_input_tokens": 163840,
|
|
19420
19463
|
"max_output_tokens": 163840,
|
|
19421
|
-
"
|
|
19422
|
-
"mode": "chat",
|
|
19464
|
+
"input_cost_per_token": 1.8e-07,
|
|
19423
19465
|
"output_cost_per_token": 1.8e-07,
|
|
19466
|
+
"litellm_provider": "deepinfra",
|
|
19467
|
+
"mode": "chat",
|
|
19424
19468
|
"supports_tool_choice": false
|
|
19425
19469
|
},
|
|
19426
19470
|
"deepinfra/meta-llama/Llama-3.2-1B-Instruct": {
|
|
@@ -19444,23 +19488,24 @@
|
|
|
19444
19488
|
"supports_tool_choice": false
|
|
19445
19489
|
},
|
|
19446
19490
|
"deepinfra/Qwen/Qwen2.5-VL-32B-Instruct": {
|
|
19447
|
-
"
|
|
19448
|
-
"litellm_provider": "deepinfra",
|
|
19491
|
+
"max_tokens": 128000,
|
|
19449
19492
|
"max_input_tokens": 128000,
|
|
19450
19493
|
"max_output_tokens": 128000,
|
|
19451
|
-
"
|
|
19452
|
-
"mode": "chat",
|
|
19494
|
+
"input_cost_per_token": 2e-07,
|
|
19453
19495
|
"output_cost_per_token": 6e-07,
|
|
19454
|
-
"
|
|
19496
|
+
"litellm_provider": "deepinfra",
|
|
19497
|
+
"mode": "chat",
|
|
19498
|
+
"supports_tool_choice": true,
|
|
19499
|
+
"supports_vision": true
|
|
19455
19500
|
},
|
|
19456
19501
|
"deepinfra/nvidia/Llama-3.1-Nemotron-70B-Instruct": {
|
|
19457
|
-
"
|
|
19458
|
-
"litellm_provider": "deepinfra",
|
|
19502
|
+
"max_tokens": 131072,
|
|
19459
19503
|
"max_input_tokens": 131072,
|
|
19460
19504
|
"max_output_tokens": 131072,
|
|
19461
|
-
"
|
|
19505
|
+
"input_cost_per_token": 6e-07,
|
|
19506
|
+
"output_cost_per_token": 6e-07,
|
|
19507
|
+
"litellm_provider": "deepinfra",
|
|
19462
19508
|
"mode": "chat",
|
|
19463
|
-
"output_cost_per_token": 3e-07,
|
|
19464
19509
|
"supports_tool_choice": true
|
|
19465
19510
|
},
|
|
19466
19511
|
"deepinfra/mistralai/Mixtral-8x22B-Instruct-v0.1": {
|
|
@@ -19474,13 +19519,13 @@
|
|
|
19474
19519
|
"supports_tool_choice": true
|
|
19475
19520
|
},
|
|
19476
19521
|
"deepinfra/Qwen/Qwen2.5-7B-Instruct": {
|
|
19477
|
-
"
|
|
19478
|
-
"litellm_provider": "deepinfra",
|
|
19522
|
+
"max_tokens": 32768,
|
|
19479
19523
|
"max_input_tokens": 32768,
|
|
19480
19524
|
"max_output_tokens": 32768,
|
|
19481
|
-
"
|
|
19482
|
-
"mode": "chat",
|
|
19525
|
+
"input_cost_per_token": 4e-08,
|
|
19483
19526
|
"output_cost_per_token": 1e-07,
|
|
19527
|
+
"litellm_provider": "deepinfra",
|
|
19528
|
+
"mode": "chat",
|
|
19484
19529
|
"supports_tool_choice": false
|
|
19485
19530
|
},
|
|
19486
19531
|
"deepinfra/google/gemini-1.5-flash-8b": {
|
|
@@ -19494,53 +19539,53 @@
|
|
|
19494
19539
|
"supports_tool_choice": true
|
|
19495
19540
|
},
|
|
19496
19541
|
"deepinfra/NousResearch/Hermes-3-Llama-3.1-70B": {
|
|
19497
|
-
"
|
|
19498
|
-
"litellm_provider": "deepinfra",
|
|
19542
|
+
"max_tokens": 131072,
|
|
19499
19543
|
"max_input_tokens": 131072,
|
|
19500
19544
|
"max_output_tokens": 131072,
|
|
19501
|
-
"
|
|
19545
|
+
"input_cost_per_token": 3e-07,
|
|
19546
|
+
"output_cost_per_token": 3e-07,
|
|
19547
|
+
"litellm_provider": "deepinfra",
|
|
19502
19548
|
"mode": "chat",
|
|
19503
|
-
"output_cost_per_token": 2.8e-07,
|
|
19504
19549
|
"supports_tool_choice": false
|
|
19505
19550
|
},
|
|
19506
19551
|
"deepinfra/deepseek-ai/DeepSeek-R1-Distill-Llama-70B": {
|
|
19507
|
-
"
|
|
19508
|
-
"litellm_provider": "deepinfra",
|
|
19552
|
+
"max_tokens": 131072,
|
|
19509
19553
|
"max_input_tokens": 131072,
|
|
19510
19554
|
"max_output_tokens": 131072,
|
|
19511
|
-
"
|
|
19555
|
+
"input_cost_per_token": 2e-07,
|
|
19556
|
+
"output_cost_per_token": 6e-07,
|
|
19557
|
+
"litellm_provider": "deepinfra",
|
|
19512
19558
|
"mode": "chat",
|
|
19513
|
-
"output_cost_per_token": 4e-07,
|
|
19514
19559
|
"supports_tool_choice": false
|
|
19515
19560
|
},
|
|
19516
19561
|
"deepinfra/meta-llama/Llama-Guard-3-8B": {
|
|
19517
|
-
"
|
|
19518
|
-
"litellm_provider": "deepinfra",
|
|
19562
|
+
"max_tokens": 131072,
|
|
19519
19563
|
"max_input_tokens": 131072,
|
|
19520
19564
|
"max_output_tokens": 131072,
|
|
19521
|
-
"
|
|
19522
|
-
"mode": "chat",
|
|
19565
|
+
"input_cost_per_token": 5.5e-08,
|
|
19523
19566
|
"output_cost_per_token": 5.5e-08,
|
|
19567
|
+
"litellm_provider": "deepinfra",
|
|
19568
|
+
"mode": "chat",
|
|
19524
19569
|
"supports_tool_choice": false
|
|
19525
19570
|
},
|
|
19526
19571
|
"deepinfra/mistralai/Mistral-Small-24B-Instruct-2501": {
|
|
19527
|
-
"
|
|
19528
|
-
"litellm_provider": "deepinfra",
|
|
19572
|
+
"max_tokens": 32768,
|
|
19529
19573
|
"max_input_tokens": 32768,
|
|
19530
19574
|
"max_output_tokens": 32768,
|
|
19531
|
-
"
|
|
19532
|
-
"mode": "chat",
|
|
19575
|
+
"input_cost_per_token": 5e-08,
|
|
19533
19576
|
"output_cost_per_token": 8e-08,
|
|
19577
|
+
"litellm_provider": "deepinfra",
|
|
19578
|
+
"mode": "chat",
|
|
19534
19579
|
"supports_tool_choice": true
|
|
19535
19580
|
},
|
|
19536
19581
|
"deepinfra/anthropic/claude-4-opus": {
|
|
19537
|
-
"
|
|
19538
|
-
"litellm_provider": "deepinfra",
|
|
19582
|
+
"max_tokens": 200000,
|
|
19539
19583
|
"max_input_tokens": 200000,
|
|
19540
19584
|
"max_output_tokens": 200000,
|
|
19541
|
-
"
|
|
19542
|
-
"mode": "chat",
|
|
19585
|
+
"input_cost_per_token": 1.65e-05,
|
|
19543
19586
|
"output_cost_per_token": 8.25e-05,
|
|
19587
|
+
"litellm_provider": "deepinfra",
|
|
19588
|
+
"mode": "chat",
|
|
19544
19589
|
"supports_tool_choice": true
|
|
19545
19590
|
},
|
|
19546
19591
|
"deepinfra/openchat/openchat-3.6-8b": {
|
|
@@ -19554,13 +19599,13 @@
|
|
|
19554
19599
|
"supports_tool_choice": false
|
|
19555
19600
|
},
|
|
19556
19601
|
"deepinfra/google/gemma-3-27b-it": {
|
|
19557
|
-
"
|
|
19558
|
-
"litellm_provider": "deepinfra",
|
|
19602
|
+
"max_tokens": 131072,
|
|
19559
19603
|
"max_input_tokens": 131072,
|
|
19560
19604
|
"max_output_tokens": 131072,
|
|
19561
|
-
"
|
|
19605
|
+
"input_cost_per_token": 9e-08,
|
|
19606
|
+
"output_cost_per_token": 1.6e-07,
|
|
19607
|
+
"litellm_provider": "deepinfra",
|
|
19562
19608
|
"mode": "chat",
|
|
19563
|
-
"output_cost_per_token": 1.7e-07,
|
|
19564
19609
|
"supports_tool_choice": true
|
|
19565
19610
|
},
|
|
19566
19611
|
"deepinfra/Austism/chronos-hermes-13b-v2": {
|
|
@@ -19574,13 +19619,13 @@
|
|
|
19574
19619
|
"supports_tool_choice": true
|
|
19575
19620
|
},
|
|
19576
19621
|
"deepinfra/Sao10K/L3.1-70B-Euryale-v2.2": {
|
|
19577
|
-
"
|
|
19578
|
-
"litellm_provider": "deepinfra",
|
|
19622
|
+
"max_tokens": 131072,
|
|
19579
19623
|
"max_input_tokens": 131072,
|
|
19580
19624
|
"max_output_tokens": 131072,
|
|
19581
|
-
"
|
|
19582
|
-
"mode": "chat",
|
|
19625
|
+
"input_cost_per_token": 6.5e-07,
|
|
19583
19626
|
"output_cost_per_token": 7.5e-07,
|
|
19627
|
+
"litellm_provider": "deepinfra",
|
|
19628
|
+
"mode": "chat",
|
|
19584
19629
|
"supports_tool_choice": false
|
|
19585
19630
|
},
|
|
19586
19631
|
"deepinfra/Qwen/QwQ-32B-Preview": {
|
|
@@ -19594,13 +19639,13 @@
|
|
|
19594
19639
|
"supports_tool_choice": false
|
|
19595
19640
|
},
|
|
19596
19641
|
"deepinfra/anthropic/claude-4-sonnet": {
|
|
19597
|
-
"
|
|
19598
|
-
"litellm_provider": "deepinfra",
|
|
19642
|
+
"max_tokens": 200000,
|
|
19599
19643
|
"max_input_tokens": 200000,
|
|
19600
19644
|
"max_output_tokens": 200000,
|
|
19601
|
-
"
|
|
19602
|
-
"mode": "chat",
|
|
19645
|
+
"input_cost_per_token": 3.3e-06,
|
|
19603
19646
|
"output_cost_per_token": 1.65e-05,
|
|
19647
|
+
"litellm_provider": "deepinfra",
|
|
19648
|
+
"mode": "chat",
|
|
19604
19649
|
"supports_tool_choice": true
|
|
19605
19650
|
},
|
|
19606
19651
|
"deepinfra/microsoft/Phi-3-medium-4k-instruct": {
|
|
@@ -19624,36 +19669,36 @@
|
|
|
19624
19669
|
"supports_tool_choice": false
|
|
19625
19670
|
},
|
|
19626
19671
|
"deepinfra/Sao10K/L3.3-70B-Euryale-v2.3": {
|
|
19627
|
-
"
|
|
19628
|
-
"litellm_provider": "deepinfra",
|
|
19672
|
+
"max_tokens": 131072,
|
|
19629
19673
|
"max_input_tokens": 131072,
|
|
19630
19674
|
"max_output_tokens": 131072,
|
|
19631
|
-
"
|
|
19632
|
-
"mode": "chat",
|
|
19675
|
+
"input_cost_per_token": 6.5e-07,
|
|
19633
19676
|
"output_cost_per_token": 7.5e-07,
|
|
19677
|
+
"litellm_provider": "deepinfra",
|
|
19678
|
+
"mode": "chat",
|
|
19634
19679
|
"supports_tool_choice": false
|
|
19635
19680
|
},
|
|
19636
19681
|
"deepinfra/meta-llama/Meta-Llama-3.1-70B-Instruct": {
|
|
19637
|
-
"
|
|
19638
|
-
"litellm_provider": "deepinfra",
|
|
19682
|
+
"max_tokens": 131072,
|
|
19639
19683
|
"max_input_tokens": 131072,
|
|
19640
19684
|
"max_output_tokens": 131072,
|
|
19641
|
-
"
|
|
19642
|
-
"mode": "chat",
|
|
19685
|
+
"input_cost_per_token": 4e-07,
|
|
19643
19686
|
"output_cost_per_token": 4e-07,
|
|
19687
|
+
"litellm_provider": "deepinfra",
|
|
19688
|
+
"mode": "chat",
|
|
19644
19689
|
"supports_tool_choice": true
|
|
19645
19690
|
},
|
|
19646
19691
|
"deepinfra/deepseek-ai/DeepSeek-V3.1": {
|
|
19647
|
-
"
|
|
19648
|
-
"input_cost_per_token": 2.7e-07,
|
|
19649
|
-
"litellm_provider": "deepinfra",
|
|
19692
|
+
"max_tokens": 163840,
|
|
19650
19693
|
"max_input_tokens": 163840,
|
|
19651
19694
|
"max_output_tokens": 163840,
|
|
19652
|
-
"
|
|
19653
|
-
"mode": "chat",
|
|
19695
|
+
"input_cost_per_token": 2.7e-07,
|
|
19654
19696
|
"output_cost_per_token": 1e-06,
|
|
19655
|
-
"
|
|
19656
|
-
"
|
|
19697
|
+
"cache_read_input_token_cost": 2.16e-07,
|
|
19698
|
+
"litellm_provider": "deepinfra",
|
|
19699
|
+
"mode": "chat",
|
|
19700
|
+
"supports_tool_choice": true,
|
|
19701
|
+
"supports_reasoning": true
|
|
19657
19702
|
},
|
|
19658
19703
|
"deepinfra/Qwen/Qwen2.5-Coder-7B": {
|
|
19659
19704
|
"max_tokens": 32768,
|
|
@@ -19677,23 +19722,23 @@
|
|
|
19677
19722
|
"supports_reasoning": true
|
|
19678
19723
|
},
|
|
19679
19724
|
"deepinfra/zai-org/GLM-4.5": {
|
|
19680
|
-
"
|
|
19681
|
-
"litellm_provider": "deepinfra",
|
|
19725
|
+
"max_tokens": 131072,
|
|
19682
19726
|
"max_input_tokens": 131072,
|
|
19683
19727
|
"max_output_tokens": 131072,
|
|
19684
|
-
"
|
|
19728
|
+
"input_cost_per_token": 4e-07,
|
|
19729
|
+
"output_cost_per_token": 1.6e-06,
|
|
19730
|
+
"litellm_provider": "deepinfra",
|
|
19685
19731
|
"mode": "chat",
|
|
19686
|
-
"output_cost_per_token": 2e-06,
|
|
19687
19732
|
"supports_tool_choice": true
|
|
19688
19733
|
},
|
|
19689
19734
|
"deepinfra/meta-llama/Llama-3.2-3B-Instruct": {
|
|
19690
|
-
"
|
|
19691
|
-
"litellm_provider": "deepinfra",
|
|
19735
|
+
"max_tokens": 131072,
|
|
19692
19736
|
"max_input_tokens": 131072,
|
|
19693
19737
|
"max_output_tokens": 131072,
|
|
19694
|
-
"
|
|
19738
|
+
"input_cost_per_token": 2e-08,
|
|
19739
|
+
"output_cost_per_token": 2e-08,
|
|
19740
|
+
"litellm_provider": "deepinfra",
|
|
19695
19741
|
"mode": "chat",
|
|
19696
|
-
"output_cost_per_token": 2.4e-08,
|
|
19697
19742
|
"supports_tool_choice": true
|
|
19698
19743
|
},
|
|
19699
19744
|
"deepinfra/google/gemini-1.5-flash": {
|
|
@@ -19717,43 +19762,43 @@
|
|
|
19717
19762
|
"supports_tool_choice": true
|
|
19718
19763
|
},
|
|
19719
19764
|
"deepinfra/google/gemini-2.5-pro": {
|
|
19720
|
-
"
|
|
19721
|
-
"litellm_provider": "deepinfra",
|
|
19765
|
+
"max_tokens": 1000000,
|
|
19722
19766
|
"max_input_tokens": 1000000,
|
|
19723
19767
|
"max_output_tokens": 1000000,
|
|
19724
|
-
"
|
|
19768
|
+
"input_cost_per_token": 1.25e-06,
|
|
19769
|
+
"output_cost_per_token": 1e-05,
|
|
19770
|
+
"litellm_provider": "deepinfra",
|
|
19725
19771
|
"mode": "chat",
|
|
19726
|
-
"output_cost_per_token": 7e-06,
|
|
19727
19772
|
"supports_tool_choice": true
|
|
19728
19773
|
},
|
|
19729
19774
|
"deepinfra/Qwen/Qwen3-30B-A3B": {
|
|
19730
|
-
"
|
|
19731
|
-
"litellm_provider": "deepinfra",
|
|
19775
|
+
"max_tokens": 40960,
|
|
19732
19776
|
"max_input_tokens": 40960,
|
|
19733
19777
|
"max_output_tokens": 40960,
|
|
19734
|
-
"
|
|
19735
|
-
"mode": "chat",
|
|
19778
|
+
"input_cost_per_token": 8e-08,
|
|
19736
19779
|
"output_cost_per_token": 2.9e-07,
|
|
19780
|
+
"litellm_provider": "deepinfra",
|
|
19781
|
+
"mode": "chat",
|
|
19737
19782
|
"supports_tool_choice": true
|
|
19738
19783
|
},
|
|
19739
19784
|
"deepinfra/Qwen/QwQ-32B": {
|
|
19740
|
-
"
|
|
19741
|
-
"litellm_provider": "deepinfra",
|
|
19785
|
+
"max_tokens": 131072,
|
|
19742
19786
|
"max_input_tokens": 131072,
|
|
19743
19787
|
"max_output_tokens": 131072,
|
|
19744
|
-
"
|
|
19745
|
-
"mode": "chat",
|
|
19788
|
+
"input_cost_per_token": 1.5e-07,
|
|
19746
19789
|
"output_cost_per_token": 4e-07,
|
|
19790
|
+
"litellm_provider": "deepinfra",
|
|
19791
|
+
"mode": "chat",
|
|
19747
19792
|
"supports_tool_choice": true
|
|
19748
19793
|
},
|
|
19749
19794
|
"deepinfra/moonshotai/Kimi-K2-Instruct": {
|
|
19750
|
-
"
|
|
19751
|
-
"litellm_provider": "deepinfra",
|
|
19795
|
+
"max_tokens": 131072,
|
|
19752
19796
|
"max_input_tokens": 131072,
|
|
19753
19797
|
"max_output_tokens": 131072,
|
|
19754
|
-
"
|
|
19755
|
-
"mode": "chat",
|
|
19798
|
+
"input_cost_per_token": 5e-07,
|
|
19756
19799
|
"output_cost_per_token": 2e-06,
|
|
19800
|
+
"litellm_provider": "deepinfra",
|
|
19801
|
+
"mode": "chat",
|
|
19757
19802
|
"supports_tool_choice": true
|
|
19758
19803
|
},
|
|
19759
19804
|
"deepinfra/Sao10K/L3-70B-Euryale-v2.1": {
|
|
@@ -19777,33 +19822,33 @@
|
|
|
19777
19822
|
"supports_tool_choice": false
|
|
19778
19823
|
},
|
|
19779
19824
|
"deepinfra/google/gemma-3-12b-it": {
|
|
19780
|
-
"
|
|
19781
|
-
"litellm_provider": "deepinfra",
|
|
19825
|
+
"max_tokens": 131072,
|
|
19782
19826
|
"max_input_tokens": 131072,
|
|
19783
19827
|
"max_output_tokens": 131072,
|
|
19784
|
-
"
|
|
19785
|
-
"mode": "chat",
|
|
19828
|
+
"input_cost_per_token": 5e-08,
|
|
19786
19829
|
"output_cost_per_token": 1e-07,
|
|
19830
|
+
"litellm_provider": "deepinfra",
|
|
19831
|
+
"mode": "chat",
|
|
19787
19832
|
"supports_tool_choice": true
|
|
19788
19833
|
},
|
|
19789
19834
|
"deepinfra/google/gemini-2.5-flash": {
|
|
19790
|
-
"
|
|
19791
|
-
"litellm_provider": "deepinfra",
|
|
19835
|
+
"max_tokens": 1000000,
|
|
19792
19836
|
"max_input_tokens": 1000000,
|
|
19793
19837
|
"max_output_tokens": 1000000,
|
|
19794
|
-
"
|
|
19838
|
+
"input_cost_per_token": 3e-07,
|
|
19839
|
+
"output_cost_per_token": 2.5e-06,
|
|
19840
|
+
"litellm_provider": "deepinfra",
|
|
19795
19841
|
"mode": "chat",
|
|
19796
|
-
"output_cost_per_token": 1.75e-06,
|
|
19797
19842
|
"supports_tool_choice": true
|
|
19798
19843
|
},
|
|
19799
19844
|
"deepinfra/deepseek-ai/DeepSeek-R1": {
|
|
19800
|
-
"
|
|
19801
|
-
"litellm_provider": "deepinfra",
|
|
19845
|
+
"max_tokens": 163840,
|
|
19802
19846
|
"max_input_tokens": 163840,
|
|
19803
19847
|
"max_output_tokens": 163840,
|
|
19804
|
-
"
|
|
19805
|
-
"mode": "chat",
|
|
19848
|
+
"input_cost_per_token": 7e-07,
|
|
19806
19849
|
"output_cost_per_token": 2.4e-06,
|
|
19850
|
+
"litellm_provider": "deepinfra",
|
|
19851
|
+
"mode": "chat",
|
|
19807
19852
|
"supports_tool_choice": true
|
|
19808
19853
|
},
|
|
19809
19854
|
"deepinfra/mistralai/Mistral-7B-Instruct-v0.3": {
|
|
@@ -19817,63 +19862,63 @@
|
|
|
19817
19862
|
"supports_tool_choice": true
|
|
19818
19863
|
},
|
|
19819
19864
|
"deepinfra/Qwen/Qwen2.5-72B-Instruct": {
|
|
19820
|
-
"
|
|
19821
|
-
"litellm_provider": "deepinfra",
|
|
19865
|
+
"max_tokens": 32768,
|
|
19822
19866
|
"max_input_tokens": 32768,
|
|
19823
19867
|
"max_output_tokens": 32768,
|
|
19824
|
-
"
|
|
19825
|
-
"mode": "chat",
|
|
19868
|
+
"input_cost_per_token": 1.2e-07,
|
|
19826
19869
|
"output_cost_per_token": 3.9e-07,
|
|
19870
|
+
"litellm_provider": "deepinfra",
|
|
19871
|
+
"mode": "chat",
|
|
19827
19872
|
"supports_tool_choice": true
|
|
19828
19873
|
},
|
|
19829
19874
|
"deepinfra/Qwen/Qwen3-14B": {
|
|
19830
|
-
"
|
|
19831
|
-
"litellm_provider": "deepinfra",
|
|
19875
|
+
"max_tokens": 40960,
|
|
19832
19876
|
"max_input_tokens": 40960,
|
|
19833
19877
|
"max_output_tokens": 40960,
|
|
19834
|
-
"
|
|
19835
|
-
"mode": "chat",
|
|
19878
|
+
"input_cost_per_token": 6e-08,
|
|
19836
19879
|
"output_cost_per_token": 2.4e-07,
|
|
19880
|
+
"litellm_provider": "deepinfra",
|
|
19881
|
+
"mode": "chat",
|
|
19837
19882
|
"supports_tool_choice": true
|
|
19838
19883
|
},
|
|
19839
19884
|
"deepinfra/allenai/olmOCR-7B-0725-FP8": {
|
|
19840
|
-
"
|
|
19841
|
-
"litellm_provider": "deepinfra",
|
|
19885
|
+
"max_tokens": 16384,
|
|
19842
19886
|
"max_input_tokens": 16384,
|
|
19843
19887
|
"max_output_tokens": 16384,
|
|
19844
|
-
"
|
|
19845
|
-
"mode": "chat",
|
|
19888
|
+
"input_cost_per_token": 2.7e-07,
|
|
19846
19889
|
"output_cost_per_token": 1.5e-06,
|
|
19890
|
+
"litellm_provider": "deepinfra",
|
|
19891
|
+
"mode": "chat",
|
|
19847
19892
|
"supports_tool_choice": false
|
|
19848
19893
|
},
|
|
19849
19894
|
"deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct": {
|
|
19850
|
-
"
|
|
19851
|
-
"litellm_provider": "deepinfra",
|
|
19895
|
+
"max_tokens": 262144,
|
|
19852
19896
|
"max_input_tokens": 262144,
|
|
19853
19897
|
"max_output_tokens": 262144,
|
|
19854
|
-
"
|
|
19855
|
-
"mode": "chat",
|
|
19898
|
+
"input_cost_per_token": 4e-07,
|
|
19856
19899
|
"output_cost_per_token": 1.6e-06,
|
|
19900
|
+
"litellm_provider": "deepinfra",
|
|
19901
|
+
"mode": "chat",
|
|
19857
19902
|
"supports_tool_choice": true
|
|
19858
19903
|
},
|
|
19859
19904
|
"deepinfra/microsoft/phi-4": {
|
|
19860
|
-
"
|
|
19861
|
-
"litellm_provider": "deepinfra",
|
|
19905
|
+
"max_tokens": 16384,
|
|
19862
19906
|
"max_input_tokens": 16384,
|
|
19863
19907
|
"max_output_tokens": 16384,
|
|
19864
|
-
"
|
|
19865
|
-
"mode": "chat",
|
|
19908
|
+
"input_cost_per_token": 7e-08,
|
|
19866
19909
|
"output_cost_per_token": 1.4e-07,
|
|
19910
|
+
"litellm_provider": "deepinfra",
|
|
19911
|
+
"mode": "chat",
|
|
19867
19912
|
"supports_tool_choice": true
|
|
19868
19913
|
},
|
|
19869
19914
|
"deepinfra/NousResearch/Hermes-3-Llama-3.1-405B": {
|
|
19870
|
-
"
|
|
19871
|
-
"litellm_provider": "deepinfra",
|
|
19915
|
+
"max_tokens": 131072,
|
|
19872
19916
|
"max_input_tokens": 131072,
|
|
19873
19917
|
"max_output_tokens": 131072,
|
|
19874
|
-
"
|
|
19918
|
+
"input_cost_per_token": 1e-06,
|
|
19919
|
+
"output_cost_per_token": 1e-06,
|
|
19920
|
+
"litellm_provider": "deepinfra",
|
|
19875
19921
|
"mode": "chat",
|
|
19876
|
-
"output_cost_per_token": 8e-07,
|
|
19877
19922
|
"supports_tool_choice": true
|
|
19878
19923
|
},
|
|
19879
19924
|
"deepinfra/zai-org/GLM-4.5-Air": {
|
|
@@ -19887,23 +19932,23 @@
|
|
|
19887
19932
|
"supports_tool_choice": true
|
|
19888
19933
|
},
|
|
19889
19934
|
"deepinfra/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B": {
|
|
19890
|
-
"
|
|
19891
|
-
"litellm_provider": "deepinfra",
|
|
19935
|
+
"max_tokens": 131072,
|
|
19892
19936
|
"max_input_tokens": 131072,
|
|
19893
19937
|
"max_output_tokens": 131072,
|
|
19894
|
-
"
|
|
19938
|
+
"input_cost_per_token": 2.7e-07,
|
|
19939
|
+
"output_cost_per_token": 2.7e-07,
|
|
19940
|
+
"litellm_provider": "deepinfra",
|
|
19895
19941
|
"mode": "chat",
|
|
19896
|
-
"output_cost_per_token": 1.5e-07,
|
|
19897
19942
|
"supports_tool_choice": true
|
|
19898
19943
|
},
|
|
19899
19944
|
"deepinfra/openai/gpt-oss-120b": {
|
|
19900
|
-
"
|
|
19901
|
-
"litellm_provider": "deepinfra",
|
|
19945
|
+
"max_tokens": 131072,
|
|
19902
19946
|
"max_input_tokens": 131072,
|
|
19903
19947
|
"max_output_tokens": 131072,
|
|
19904
|
-
"
|
|
19905
|
-
"mode": "chat",
|
|
19948
|
+
"input_cost_per_token": 5e-08,
|
|
19906
19949
|
"output_cost_per_token": 4.5e-07,
|
|
19950
|
+
"litellm_provider": "deepinfra",
|
|
19951
|
+
"mode": "chat",
|
|
19907
19952
|
"supports_tool_choice": true
|
|
19908
19953
|
},
|
|
19909
19954
|
"deepinfra/google/codegemma-7b-it": {
|
|
@@ -19917,24 +19962,23 @@
|
|
|
19917
19962
|
"supports_tool_choice": false
|
|
19918
19963
|
},
|
|
19919
19964
|
"deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo": {
|
|
19920
|
-
"
|
|
19921
|
-
"input_cost_per_token": 3e-07,
|
|
19922
|
-
"litellm_provider": "deepinfra",
|
|
19965
|
+
"max_tokens": 262144,
|
|
19923
19966
|
"max_input_tokens": 262144,
|
|
19924
19967
|
"max_output_tokens": 262144,
|
|
19925
|
-
"
|
|
19926
|
-
"mode": "chat",
|
|
19968
|
+
"input_cost_per_token": 2.9e-07,
|
|
19927
19969
|
"output_cost_per_token": 1.2e-06,
|
|
19970
|
+
"litellm_provider": "deepinfra",
|
|
19971
|
+
"mode": "chat",
|
|
19928
19972
|
"supports_tool_choice": true
|
|
19929
19973
|
},
|
|
19930
19974
|
"deepinfra/mistralai/Mistral-Nemo-Instruct-2407": {
|
|
19931
|
-
"
|
|
19932
|
-
"litellm_provider": "deepinfra",
|
|
19975
|
+
"max_tokens": 131072,
|
|
19933
19976
|
"max_input_tokens": 131072,
|
|
19934
19977
|
"max_output_tokens": 131072,
|
|
19935
|
-
"
|
|
19936
|
-
"mode": "chat",
|
|
19978
|
+
"input_cost_per_token": 2e-08,
|
|
19937
19979
|
"output_cost_per_token": 4e-08,
|
|
19980
|
+
"litellm_provider": "deepinfra",
|
|
19981
|
+
"mode": "chat",
|
|
19938
19982
|
"supports_tool_choice": true
|
|
19939
19983
|
},
|
|
19940
19984
|
"deepinfra/openbmb/MiniCPM-Llama3-V-2_5": {
|
|
@@ -19958,33 +20002,33 @@
|
|
|
19958
20002
|
"supports_tool_choice": false
|
|
19959
20003
|
},
|
|
19960
20004
|
"deepinfra/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
|
|
19961
|
-
"
|
|
19962
|
-
"litellm_provider": "deepinfra",
|
|
20005
|
+
"max_tokens": 1048576,
|
|
19963
20006
|
"max_input_tokens": 1048576,
|
|
19964
20007
|
"max_output_tokens": 1048576,
|
|
19965
|
-
"
|
|
19966
|
-
"mode": "chat",
|
|
20008
|
+
"input_cost_per_token": 1.5e-07,
|
|
19967
20009
|
"output_cost_per_token": 6e-07,
|
|
20010
|
+
"litellm_provider": "deepinfra",
|
|
20011
|
+
"mode": "chat",
|
|
19968
20012
|
"supports_tool_choice": true
|
|
19969
20013
|
},
|
|
19970
20014
|
"deepinfra/meta-llama/Llama-4-Scout-17B-16E-Instruct": {
|
|
19971
|
-
"
|
|
19972
|
-
"litellm_provider": "deepinfra",
|
|
20015
|
+
"max_tokens": 327680,
|
|
19973
20016
|
"max_input_tokens": 327680,
|
|
19974
20017
|
"max_output_tokens": 327680,
|
|
19975
|
-
"
|
|
19976
|
-
"mode": "chat",
|
|
20018
|
+
"input_cost_per_token": 8e-08,
|
|
19977
20019
|
"output_cost_per_token": 3e-07,
|
|
20020
|
+
"litellm_provider": "deepinfra",
|
|
20021
|
+
"mode": "chat",
|
|
19978
20022
|
"supports_tool_choice": true
|
|
19979
20023
|
},
|
|
19980
20024
|
"deepinfra/google/gemini-2.0-flash-001": {
|
|
19981
|
-
"
|
|
19982
|
-
"litellm_provider": "deepinfra",
|
|
20025
|
+
"max_tokens": 1000000,
|
|
19983
20026
|
"max_input_tokens": 1000000,
|
|
19984
20027
|
"max_output_tokens": 1000000,
|
|
19985
|
-
"
|
|
19986
|
-
"mode": "chat",
|
|
20028
|
+
"input_cost_per_token": 1e-07,
|
|
19987
20029
|
"output_cost_per_token": 4e-07,
|
|
20030
|
+
"litellm_provider": "deepinfra",
|
|
20031
|
+
"mode": "chat",
|
|
19988
20032
|
"supports_tool_choice": true
|
|
19989
20033
|
},
|
|
19990
20034
|
"deepinfra/Gryphe/MythoMax-L2-13b-turbo": {
|
|
@@ -20008,43 +20052,43 @@
|
|
|
20008
20052
|
"supports_tool_choice": true
|
|
20009
20053
|
},
|
|
20010
20054
|
"deepinfra/meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo": {
|
|
20011
|
-
"
|
|
20012
|
-
"litellm_provider": "deepinfra",
|
|
20055
|
+
"max_tokens": 131072,
|
|
20013
20056
|
"max_input_tokens": 131072,
|
|
20014
20057
|
"max_output_tokens": 131072,
|
|
20015
|
-
"
|
|
20016
|
-
"mode": "chat",
|
|
20058
|
+
"input_cost_per_token": 1e-07,
|
|
20017
20059
|
"output_cost_per_token": 2.8e-07,
|
|
20060
|
+
"litellm_provider": "deepinfra",
|
|
20061
|
+
"mode": "chat",
|
|
20018
20062
|
"supports_tool_choice": true
|
|
20019
20063
|
},
|
|
20020
20064
|
"deepinfra/meta-llama/Meta-Llama-3.1-8B-Instruct": {
|
|
20021
|
-
"
|
|
20022
|
-
"litellm_provider": "deepinfra",
|
|
20065
|
+
"max_tokens": 131072,
|
|
20023
20066
|
"max_input_tokens": 131072,
|
|
20024
20067
|
"max_output_tokens": 131072,
|
|
20025
|
-
"
|
|
20026
|
-
"mode": "chat",
|
|
20068
|
+
"input_cost_per_token": 3e-08,
|
|
20027
20069
|
"output_cost_per_token": 5e-08,
|
|
20070
|
+
"litellm_provider": "deepinfra",
|
|
20071
|
+
"mode": "chat",
|
|
20028
20072
|
"supports_tool_choice": true
|
|
20029
20073
|
},
|
|
20030
20074
|
"deepinfra/Qwen/Qwen3-32B": {
|
|
20031
|
-
"
|
|
20032
|
-
"litellm_provider": "deepinfra",
|
|
20075
|
+
"max_tokens": 40960,
|
|
20033
20076
|
"max_input_tokens": 40960,
|
|
20034
20077
|
"max_output_tokens": 40960,
|
|
20035
|
-
"
|
|
20078
|
+
"input_cost_per_token": 1e-07,
|
|
20079
|
+
"output_cost_per_token": 2.8e-07,
|
|
20080
|
+
"litellm_provider": "deepinfra",
|
|
20036
20081
|
"mode": "chat",
|
|
20037
|
-
"output_cost_per_token": 3e-07,
|
|
20038
20082
|
"supports_tool_choice": true
|
|
20039
20083
|
},
|
|
20040
20084
|
"deepinfra/Qwen/Qwen3-235B-A22B-Thinking-2507": {
|
|
20041
|
-
"
|
|
20042
|
-
"litellm_provider": "deepinfra",
|
|
20085
|
+
"max_tokens": 262144,
|
|
20043
20086
|
"max_input_tokens": 262144,
|
|
20044
20087
|
"max_output_tokens": 262144,
|
|
20045
|
-
"
|
|
20088
|
+
"input_cost_per_token": 3e-07,
|
|
20089
|
+
"output_cost_per_token": 2.9e-06,
|
|
20090
|
+
"litellm_provider": "deepinfra",
|
|
20046
20091
|
"mode": "chat",
|
|
20047
|
-
"output_cost_per_token": 6e-07,
|
|
20048
20092
|
"supports_tool_choice": true
|
|
20049
20093
|
},
|
|
20050
20094
|
"deepinfra/nvidia/Nemotron-4-340B-Instruct": {
|
|
@@ -20058,24 +20102,24 @@
|
|
|
20058
20102
|
"supports_tool_choice": true
|
|
20059
20103
|
},
|
|
20060
20104
|
"deepinfra/deepseek-ai/DeepSeek-R1-0528": {
|
|
20061
|
-
"
|
|
20062
|
-
"input_cost_per_token": 5e-07,
|
|
20063
|
-
"litellm_provider": "deepinfra",
|
|
20105
|
+
"max_tokens": 163840,
|
|
20064
20106
|
"max_input_tokens": 163840,
|
|
20065
20107
|
"max_output_tokens": 163840,
|
|
20066
|
-
"
|
|
20067
|
-
"mode": "chat",
|
|
20108
|
+
"input_cost_per_token": 5e-07,
|
|
20068
20109
|
"output_cost_per_token": 2.15e-06,
|
|
20110
|
+
"cache_read_input_token_cost": 4e-07,
|
|
20111
|
+
"litellm_provider": "deepinfra",
|
|
20112
|
+
"mode": "chat",
|
|
20069
20113
|
"supports_tool_choice": true
|
|
20070
20114
|
},
|
|
20071
20115
|
"deepinfra/deepseek-ai/DeepSeek-R1-Turbo": {
|
|
20072
|
-
"
|
|
20073
|
-
"litellm_provider": "deepinfra",
|
|
20116
|
+
"max_tokens": 40960,
|
|
20074
20117
|
"max_input_tokens": 40960,
|
|
20075
20118
|
"max_output_tokens": 40960,
|
|
20076
|
-
"
|
|
20077
|
-
"mode": "chat",
|
|
20119
|
+
"input_cost_per_token": 1e-06,
|
|
20078
20120
|
"output_cost_per_token": 3e-06,
|
|
20121
|
+
"litellm_provider": "deepinfra",
|
|
20122
|
+
"mode": "chat",
|
|
20079
20123
|
"supports_tool_choice": true
|
|
20080
20124
|
},
|
|
20081
20125
|
"deepinfra/NovaSky-AI/Sky-T1-32B-Preview": {
|
|
@@ -20089,13 +20133,13 @@
|
|
|
20089
20133
|
"supports_tool_choice": false
|
|
20090
20134
|
},
|
|
20091
20135
|
"deepinfra/Qwen/Qwen3-235B-A22B-Instruct-2507": {
|
|
20092
|
-
"
|
|
20093
|
-
"litellm_provider": "deepinfra",
|
|
20136
|
+
"max_tokens": 262144,
|
|
20094
20137
|
"max_input_tokens": 262144,
|
|
20095
20138
|
"max_output_tokens": 262144,
|
|
20096
|
-
"
|
|
20097
|
-
"mode": "chat",
|
|
20139
|
+
"input_cost_per_token": 9e-08,
|
|
20098
20140
|
"output_cost_per_token": 6e-07,
|
|
20141
|
+
"litellm_provider": "deepinfra",
|
|
20142
|
+
"mode": "chat",
|
|
20099
20143
|
"supports_tool_choice": true
|
|
20100
20144
|
},
|
|
20101
20145
|
"deepinfra/mistralai/Mistral-Small-3.1-24B-Instruct-2503": {
|
|
@@ -20139,13 +20183,13 @@
|
|
|
20139
20183
|
"supports_tool_choice": false
|
|
20140
20184
|
},
|
|
20141
20185
|
"deepinfra/google/gemma-3-4b-it": {
|
|
20142
|
-
"
|
|
20143
|
-
"litellm_provider": "deepinfra",
|
|
20186
|
+
"max_tokens": 131072,
|
|
20144
20187
|
"max_input_tokens": 131072,
|
|
20145
20188
|
"max_output_tokens": 131072,
|
|
20146
|
-
"
|
|
20147
|
-
"mode": "chat",
|
|
20189
|
+
"input_cost_per_token": 4e-08,
|
|
20148
20190
|
"output_cost_per_token": 8e-08,
|
|
20191
|
+
"litellm_provider": "deepinfra",
|
|
20192
|
+
"mode": "chat",
|
|
20149
20193
|
"supports_tool_choice": true
|
|
20150
20194
|
},
|
|
20151
20195
|
"deepinfra/mistralai/Mistral-7B-Instruct-v0.2": {
|
|
@@ -20159,13 +20203,13 @@
|
|
|
20159
20203
|
"supports_tool_choice": false
|
|
20160
20204
|
},
|
|
20161
20205
|
"deepinfra/meta-llama/Llama-3.3-70B-Instruct-Turbo": {
|
|
20162
|
-
"
|
|
20163
|
-
"litellm_provider": "deepinfra",
|
|
20206
|
+
"max_tokens": 131072,
|
|
20164
20207
|
"max_input_tokens": 131072,
|
|
20165
20208
|
"max_output_tokens": 131072,
|
|
20166
|
-
"
|
|
20209
|
+
"input_cost_per_token": 1.3e-07,
|
|
20210
|
+
"output_cost_per_token": 3.9e-07,
|
|
20211
|
+
"litellm_provider": "deepinfra",
|
|
20167
20212
|
"mode": "chat",
|
|
20168
|
-
"output_cost_per_token": 1.2e-07,
|
|
20169
20213
|
"supports_tool_choice": true
|
|
20170
20214
|
},
|
|
20171
20215
|
"deepinfra/mistralai/Devstral-Small-2505": {
|
|
@@ -20179,23 +20223,23 @@
|
|
|
20179
20223
|
"supports_tool_choice": true
|
|
20180
20224
|
},
|
|
20181
20225
|
"deepinfra/meta-llama/Llama-3.2-11B-Vision-Instruct": {
|
|
20182
|
-
"
|
|
20183
|
-
"litellm_provider": "deepinfra",
|
|
20226
|
+
"max_tokens": 131072,
|
|
20184
20227
|
"max_input_tokens": 131072,
|
|
20185
20228
|
"max_output_tokens": 131072,
|
|
20186
|
-
"
|
|
20187
|
-
"mode": "chat",
|
|
20229
|
+
"input_cost_per_token": 4.9e-08,
|
|
20188
20230
|
"output_cost_per_token": 4.9e-08,
|
|
20231
|
+
"litellm_provider": "deepinfra",
|
|
20232
|
+
"mode": "chat",
|
|
20189
20233
|
"supports_tool_choice": false
|
|
20190
20234
|
},
|
|
20191
20235
|
"deepinfra/openai/gpt-oss-20b": {
|
|
20192
|
-
"
|
|
20193
|
-
"litellm_provider": "deepinfra",
|
|
20236
|
+
"max_tokens": 131072,
|
|
20194
20237
|
"max_input_tokens": 131072,
|
|
20195
20238
|
"max_output_tokens": 131072,
|
|
20196
|
-
"
|
|
20239
|
+
"input_cost_per_token": 4e-08,
|
|
20240
|
+
"output_cost_per_token": 1.5e-07,
|
|
20241
|
+
"litellm_provider": "deepinfra",
|
|
20197
20242
|
"mode": "chat",
|
|
20198
|
-
"output_cost_per_token": 1.6e-07,
|
|
20199
20243
|
"supports_tool_choice": true
|
|
20200
20244
|
},
|
|
20201
20245
|
"voyage/voyage-context-3": {
|
|
@@ -20431,6 +20475,8 @@
|
|
|
20431
20475
|
},
|
|
20432
20476
|
"openrouter/anthropic/claude-opus-4": {
|
|
20433
20477
|
"input_cost_per_image": 0.0048,
|
|
20478
|
+
"cache_creation_input_token_cost": 1.875e-05,
|
|
20479
|
+
"cache_read_input_token_cost": 1.5e-06,
|
|
20434
20480
|
"input_cost_per_token": 1.5e-05,
|
|
20435
20481
|
"litellm_provider": "openrouter",
|
|
20436
20482
|
"max_input_tokens": 200000,
|
|
@@ -20441,6 +20487,7 @@
|
|
|
20441
20487
|
"supports_assistant_prefill": true,
|
|
20442
20488
|
"supports_computer_use": true,
|
|
20443
20489
|
"supports_function_calling": true,
|
|
20490
|
+
"supports_prompt_caching": true,
|
|
20444
20491
|
"supports_reasoning": true,
|
|
20445
20492
|
"supports_tool_choice": true,
|
|
20446
20493
|
"supports_vision": true,
|
|
@@ -20448,6 +20495,9 @@
|
|
|
20448
20495
|
},
|
|
20449
20496
|
"openrouter/anthropic/claude-opus-4.1": {
|
|
20450
20497
|
"input_cost_per_image": 0.0048,
|
|
20498
|
+
"cache_creation_input_token_cost": 1.875e-05,
|
|
20499
|
+
"cache_creation_input_token_cost_above_1hr": 3e-05,
|
|
20500
|
+
"cache_read_input_token_cost": 1.5e-06,
|
|
20451
20501
|
"input_cost_per_token": 1.5e-05,
|
|
20452
20502
|
"litellm_provider": "openrouter",
|
|
20453
20503
|
"max_input_tokens": 200000,
|
|
@@ -20458,6 +20508,7 @@
|
|
|
20458
20508
|
"supports_assistant_prefill": true,
|
|
20459
20509
|
"supports_computer_use": true,
|
|
20460
20510
|
"supports_function_calling": true,
|
|
20511
|
+
"supports_prompt_caching": true,
|
|
20461
20512
|
"supports_reasoning": true,
|
|
20462
20513
|
"supports_tool_choice": true,
|
|
20463
20514
|
"supports_vision": true,
|
|
@@ -22719,11 +22770,12 @@
|
|
|
22719
22770
|
"max_tokens": 2000000.0,
|
|
22720
22771
|
"mode": "chat",
|
|
22721
22772
|
"input_cost_per_token": 2e-07,
|
|
22773
|
+
"input_cost_per_token_above_128k_tokens": 4e-07,
|
|
22722
22774
|
"output_cost_per_token": 5e-07,
|
|
22775
|
+
"output_cost_per_token_above_128k_tokens": 1e-06,
|
|
22723
22776
|
"cache_read_input_token_cost": 5e-08,
|
|
22724
22777
|
"source": "https://docs.x.ai/docs/models",
|
|
22725
22778
|
"supports_function_calling": true,
|
|
22726
|
-
"supports_reasoning": true,
|
|
22727
22779
|
"supports_tool_choice": true,
|
|
22728
22780
|
"supports_web_search": true
|
|
22729
22781
|
},
|
|
@@ -22735,7 +22787,9 @@
|
|
|
22735
22787
|
"max_tokens": 2000000.0,
|
|
22736
22788
|
"mode": "chat",
|
|
22737
22789
|
"input_cost_per_token": 2e-07,
|
|
22790
|
+
"input_cost_per_token_above_128k_tokens": 4e-07,
|
|
22738
22791
|
"output_cost_per_token": 5e-07,
|
|
22792
|
+
"output_cost_per_token_above_128k_tokens": 1e-06,
|
|
22739
22793
|
"source": "https://docs.x.ai/docs/models",
|
|
22740
22794
|
"supports_function_calling": true,
|
|
22741
22795
|
"supports_tool_choice": true,
|
|
@@ -22825,11 +22879,11 @@
|
|
|
22825
22879
|
"text"
|
|
22826
22880
|
],
|
|
22827
22881
|
"supports_function_calling": true,
|
|
22828
|
-
"supports_native_streaming":
|
|
22882
|
+
"supports_native_streaming": true,
|
|
22829
22883
|
"supports_parallel_function_calling": true,
|
|
22830
22884
|
"supports_pdf_input": true,
|
|
22831
22885
|
"supports_prompt_caching": true,
|
|
22832
|
-
"supports_reasoning":
|
|
22886
|
+
"supports_reasoning": true,
|
|
22833
22887
|
"supports_response_schema": true,
|
|
22834
22888
|
"supports_system_messages": false,
|
|
22835
22889
|
"supports_tool_choice": true,
|
|
@@ -23363,10 +23417,14 @@
|
|
|
23363
23417
|
"cache_creation_input_token_cost": 3.75e-06,
|
|
23364
23418
|
"cache_read_input_token_cost": 3e-07,
|
|
23365
23419
|
"input_cost_per_token": 3e-06,
|
|
23420
|
+
"input_cost_per_token_above_200k_tokens": 6e-06,
|
|
23421
|
+
"output_cost_per_token_above_200k_tokens": 2.25e-05,
|
|
23422
|
+
"cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
|
|
23423
|
+
"cache_read_input_token_cost_above_200k_tokens": 6e-07,
|
|
23366
23424
|
"litellm_provider": "anthropic",
|
|
23367
23425
|
"max_input_tokens": 200000,
|
|
23368
23426
|
"max_output_tokens": 64000,
|
|
23369
|
-
"max_tokens":
|
|
23427
|
+
"max_tokens": 64000,
|
|
23370
23428
|
"mode": "chat",
|
|
23371
23429
|
"output_cost_per_token": 1.5e-05,
|
|
23372
23430
|
"search_context_cost_per_query": {
|
|
@@ -23383,6 +23441,7 @@
|
|
|
23383
23441
|
"supports_response_schema": true,
|
|
23384
23442
|
"supports_tool_choice": true,
|
|
23385
23443
|
"supports_vision": true,
|
|
23444
|
+
"supports_web_search": true,
|
|
23386
23445
|
"tool_use_system_prompt_tokens": 346
|
|
23387
23446
|
},
|
|
23388
23447
|
"ollama/deepseek-v3.1:671b-cloud": {
|
|
@@ -23426,15 +23485,19 @@
|
|
|
23426
23485
|
"supports_function_calling": true
|
|
23427
23486
|
},
|
|
23428
23487
|
"us.anthropic.claude-sonnet-4-5-20250929-v1:0": {
|
|
23429
|
-
"cache_creation_input_token_cost":
|
|
23430
|
-
"cache_read_input_token_cost": 3e-07,
|
|
23431
|
-
"input_cost_per_token": 3e-06,
|
|
23488
|
+
"cache_creation_input_token_cost": 4.125e-06,
|
|
23489
|
+
"cache_read_input_token_cost": 3.3e-07,
|
|
23490
|
+
"input_cost_per_token": 3.3e-06,
|
|
23491
|
+
"input_cost_per_token_above_200k_tokens": 6.6e-06,
|
|
23492
|
+
"output_cost_per_token_above_200k_tokens": 2.475e-05,
|
|
23493
|
+
"cache_creation_input_token_cost_above_200k_tokens": 8.25e-06,
|
|
23494
|
+
"cache_read_input_token_cost_above_200k_tokens": 6.6e-07,
|
|
23432
23495
|
"litellm_provider": "bedrock_converse",
|
|
23433
23496
|
"max_input_tokens": 200000,
|
|
23434
23497
|
"max_output_tokens": 64000,
|
|
23435
|
-
"max_tokens":
|
|
23498
|
+
"max_tokens": 64000,
|
|
23436
23499
|
"mode": "chat",
|
|
23437
|
-
"output_cost_per_token": 1.
|
|
23500
|
+
"output_cost_per_token": 1.65e-05,
|
|
23438
23501
|
"search_context_cost_per_query": {
|
|
23439
23502
|
"search_context_size_high": 0.01,
|
|
23440
23503
|
"search_context_size_low": 0.01,
|
|
@@ -23455,11 +23518,15 @@
|
|
|
23455
23518
|
"cache_creation_input_token_cost": 3.75e-06,
|
|
23456
23519
|
"cache_read_input_token_cost": 3e-07,
|
|
23457
23520
|
"input_cost_per_token": 3e-06,
|
|
23521
|
+
"input_cost_per_token_above_200k_tokens": 6e-06,
|
|
23522
|
+
"output_cost_per_token_above_200k_tokens": 2.25e-05,
|
|
23523
|
+
"cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
|
|
23524
|
+
"cache_read_input_token_cost_above_200k_tokens": 6e-07,
|
|
23458
23525
|
"input_cost_per_token_batches": 1.5e-06,
|
|
23459
23526
|
"litellm_provider": "vertex_ai-anthropic_models",
|
|
23460
23527
|
"max_input_tokens": 200000,
|
|
23461
23528
|
"max_output_tokens": 64000,
|
|
23462
|
-
"max_tokens":
|
|
23529
|
+
"max_tokens": 64000,
|
|
23463
23530
|
"mode": "chat",
|
|
23464
23531
|
"output_cost_per_token": 1.5e-05,
|
|
23465
23532
|
"output_cost_per_token_batches": 7.5e-06,
|
|
@@ -23477,11 +23544,15 @@
|
|
|
23477
23544
|
"cache_creation_input_token_cost": 3.75e-06,
|
|
23478
23545
|
"cache_read_input_token_cost": 3e-07,
|
|
23479
23546
|
"input_cost_per_token": 3e-06,
|
|
23547
|
+
"input_cost_per_token_above_200k_tokens": 6e-06,
|
|
23548
|
+
"output_cost_per_token_above_200k_tokens": 2.25e-05,
|
|
23549
|
+
"cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
|
|
23550
|
+
"cache_read_input_token_cost_above_200k_tokens": 6e-07,
|
|
23480
23551
|
"input_cost_per_token_batches": 1.5e-06,
|
|
23481
23552
|
"litellm_provider": "vertex_ai-anthropic_models",
|
|
23482
23553
|
"max_input_tokens": 200000,
|
|
23483
23554
|
"max_output_tokens": 64000,
|
|
23484
|
-
"max_tokens":
|
|
23555
|
+
"max_tokens": 64000,
|
|
23485
23556
|
"mode": "chat",
|
|
23486
23557
|
"output_cost_per_token": 1.5e-05,
|
|
23487
23558
|
"output_cost_per_token_batches": 7.5e-06,
|
|
@@ -23499,10 +23570,14 @@
|
|
|
23499
23570
|
"cache_creation_input_token_cost": 3.75e-06,
|
|
23500
23571
|
"cache_read_input_token_cost": 3e-07,
|
|
23501
23572
|
"input_cost_per_token": 3e-06,
|
|
23573
|
+
"input_cost_per_token_above_200k_tokens": 6e-06,
|
|
23574
|
+
"output_cost_per_token_above_200k_tokens": 2.25e-05,
|
|
23575
|
+
"cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
|
|
23576
|
+
"cache_read_input_token_cost_above_200k_tokens": 6e-07,
|
|
23502
23577
|
"litellm_provider": "anthropic",
|
|
23503
23578
|
"max_input_tokens": 200000,
|
|
23504
23579
|
"max_output_tokens": 64000,
|
|
23505
|
-
"max_tokens":
|
|
23580
|
+
"max_tokens": 64000,
|
|
23506
23581
|
"mode": "chat",
|
|
23507
23582
|
"output_cost_per_token": 1.5e-05,
|
|
23508
23583
|
"search_context_cost_per_query": {
|
|
@@ -23524,8 +23599,8 @@
|
|
|
23524
23599
|
"lemonade/Qwen3-Coder-30B-A3B-Instruct-GGUF": {
|
|
23525
23600
|
"input_cost_per_token": 0,
|
|
23526
23601
|
"litellm_provider": "lemonade",
|
|
23527
|
-
"max_tokens":
|
|
23528
|
-
"max_input_tokens":
|
|
23602
|
+
"max_tokens": 262144,
|
|
23603
|
+
"max_input_tokens": 262144,
|
|
23529
23604
|
"max_output_tokens": 32768,
|
|
23530
23605
|
"mode": "chat",
|
|
23531
23606
|
"output_cost_per_token": 0,
|
|
@@ -23545,5 +23620,2094 @@
|
|
|
23545
23620
|
"supports_function_calling": true,
|
|
23546
23621
|
"supports_response_schema": true,
|
|
23547
23622
|
"supports_tool_choice": true
|
|
23623
|
+
},
|
|
23624
|
+
"azure_ai/grok-4": {
|
|
23625
|
+
"input_cost_per_token": 5.5e-06,
|
|
23626
|
+
"litellm_provider": "azure_ai",
|
|
23627
|
+
"max_input_tokens": 131072,
|
|
23628
|
+
"max_output_tokens": 131072,
|
|
23629
|
+
"max_tokens": 131072,
|
|
23630
|
+
"mode": "chat",
|
|
23631
|
+
"output_cost_per_token": 2.75e-05,
|
|
23632
|
+
"source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/",
|
|
23633
|
+
"supports_function_calling": true,
|
|
23634
|
+
"supports_response_schema": true,
|
|
23635
|
+
"supports_tool_choice": true,
|
|
23636
|
+
"supports_web_search": true
|
|
23637
|
+
},
|
|
23638
|
+
"azure_ai/grok-4-fast-non-reasoning": {
|
|
23639
|
+
"input_cost_per_token": 4.3e-07,
|
|
23640
|
+
"output_cost_per_token": 1.73e-06,
|
|
23641
|
+
"litellm_provider": "azure_ai",
|
|
23642
|
+
"max_input_tokens": 131072,
|
|
23643
|
+
"max_output_tokens": 131072,
|
|
23644
|
+
"max_tokens": 131072,
|
|
23645
|
+
"mode": "chat",
|
|
23646
|
+
"supports_function_calling": true,
|
|
23647
|
+
"supports_response_schema": true,
|
|
23648
|
+
"supports_tool_choice": true,
|
|
23649
|
+
"supports_web_search": true
|
|
23650
|
+
},
|
|
23651
|
+
"azure_ai/grok-4-fast-reasoning": {
|
|
23652
|
+
"input_cost_per_token": 4.3e-07,
|
|
23653
|
+
"output_cost_per_token": 1.73e-06,
|
|
23654
|
+
"litellm_provider": "azure_ai",
|
|
23655
|
+
"max_input_tokens": 131072,
|
|
23656
|
+
"max_output_tokens": 131072,
|
|
23657
|
+
"max_tokens": 131072,
|
|
23658
|
+
"mode": "chat",
|
|
23659
|
+
"source": "https://techcommunity.microsoft.com/blog/azure-ai-foundry-blog/announcing-the-grok-4-fast-models-from-xai-now-available-in-azure-ai-foundry/4456701",
|
|
23660
|
+
"supports_function_calling": true,
|
|
23661
|
+
"supports_response_schema": true,
|
|
23662
|
+
"supports_tool_choice": true,
|
|
23663
|
+
"supports_web_search": true
|
|
23664
|
+
},
|
|
23665
|
+
"azure_ai/grok-code-fast-1": {
|
|
23666
|
+
"input_cost_per_token": 3.5e-06,
|
|
23667
|
+
"litellm_provider": "azure_ai",
|
|
23668
|
+
"max_input_tokens": 131072,
|
|
23669
|
+
"max_output_tokens": 131072,
|
|
23670
|
+
"max_tokens": 131072,
|
|
23671
|
+
"mode": "chat",
|
|
23672
|
+
"output_cost_per_token": 1.75e-05,
|
|
23673
|
+
"source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/",
|
|
23674
|
+
"supports_function_calling": true,
|
|
23675
|
+
"supports_response_schema": true,
|
|
23676
|
+
"supports_tool_choice": true,
|
|
23677
|
+
"supports_web_search": true
|
|
23678
|
+
},
|
|
23679
|
+
"eu.anthropic.claude-sonnet-4-5-20250929-v1:0": {
|
|
23680
|
+
"cache_creation_input_token_cost": 4.125e-06,
|
|
23681
|
+
"cache_read_input_token_cost": 3.3e-07,
|
|
23682
|
+
"input_cost_per_token": 3.3e-06,
|
|
23683
|
+
"input_cost_per_token_above_200k_tokens": 6.6e-06,
|
|
23684
|
+
"output_cost_per_token_above_200k_tokens": 2.475e-05,
|
|
23685
|
+
"cache_creation_input_token_cost_above_200k_tokens": 8.25e-06,
|
|
23686
|
+
"cache_read_input_token_cost_above_200k_tokens": 6.6e-07,
|
|
23687
|
+
"litellm_provider": "bedrock_converse",
|
|
23688
|
+
"max_input_tokens": 200000,
|
|
23689
|
+
"max_output_tokens": 64000,
|
|
23690
|
+
"max_tokens": 64000,
|
|
23691
|
+
"mode": "chat",
|
|
23692
|
+
"output_cost_per_token": 1.65e-05,
|
|
23693
|
+
"search_context_cost_per_query": {
|
|
23694
|
+
"search_context_size_high": 0.01,
|
|
23695
|
+
"search_context_size_low": 0.01,
|
|
23696
|
+
"search_context_size_medium": 0.01
|
|
23697
|
+
},
|
|
23698
|
+
"supports_assistant_prefill": true,
|
|
23699
|
+
"supports_computer_use": true,
|
|
23700
|
+
"supports_function_calling": true,
|
|
23701
|
+
"supports_pdf_input": true,
|
|
23702
|
+
"supports_prompt_caching": true,
|
|
23703
|
+
"supports_reasoning": true,
|
|
23704
|
+
"supports_response_schema": true,
|
|
23705
|
+
"supports_tool_choice": true,
|
|
23706
|
+
"supports_vision": true,
|
|
23707
|
+
"tool_use_system_prompt_tokens": 346
|
|
23708
|
+
},
|
|
23709
|
+
"nvidia_nim/nvidia/nv-rerankqa-mistral-4b-v3": {
|
|
23710
|
+
"input_cost_per_query": 0.0,
|
|
23711
|
+
"input_cost_per_token": 0.0,
|
|
23712
|
+
"litellm_provider": "nvidia_nim",
|
|
23713
|
+
"mode": "rerank",
|
|
23714
|
+
"output_cost_per_token": 0.0
|
|
23715
|
+
},
|
|
23716
|
+
"nvidia_nim/nvidia/llama-3_2-nv-rerankqa-1b-v2": {
|
|
23717
|
+
"input_cost_per_query": 0.0,
|
|
23718
|
+
"input_cost_per_token": 0.0,
|
|
23719
|
+
"litellm_provider": "nvidia_nim",
|
|
23720
|
+
"mode": "rerank",
|
|
23721
|
+
"output_cost_per_token": 0.0
|
|
23722
|
+
},
|
|
23723
|
+
"deepinfra/Qwen/Qwen3-Next-80B-A3B-Instruct": {
|
|
23724
|
+
"max_tokens": 262144,
|
|
23725
|
+
"max_input_tokens": 262144,
|
|
23726
|
+
"max_output_tokens": 262144,
|
|
23727
|
+
"input_cost_per_token": 1.4e-07,
|
|
23728
|
+
"output_cost_per_token": 1.4e-06,
|
|
23729
|
+
"litellm_provider": "deepinfra",
|
|
23730
|
+
"mode": "chat",
|
|
23731
|
+
"supports_tool_choice": true
|
|
23732
|
+
},
|
|
23733
|
+
"deepinfra/Qwen/Qwen3-Next-80B-A3B-Thinking": {
|
|
23734
|
+
"max_tokens": 262144,
|
|
23735
|
+
"max_input_tokens": 262144,
|
|
23736
|
+
"max_output_tokens": 262144,
|
|
23737
|
+
"input_cost_per_token": 1.4e-07,
|
|
23738
|
+
"output_cost_per_token": 1.4e-06,
|
|
23739
|
+
"litellm_provider": "deepinfra",
|
|
23740
|
+
"mode": "chat",
|
|
23741
|
+
"supports_tool_choice": true
|
|
23742
|
+
},
|
|
23743
|
+
"deepinfra/deepseek-ai/DeepSeek-V3.1-Terminus": {
|
|
23744
|
+
"max_tokens": 163840,
|
|
23745
|
+
"max_input_tokens": 163840,
|
|
23746
|
+
"max_output_tokens": 163840,
|
|
23747
|
+
"input_cost_per_token": 2.7e-07,
|
|
23748
|
+
"output_cost_per_token": 1e-06,
|
|
23749
|
+
"cache_read_input_token_cost": 2.16e-07,
|
|
23750
|
+
"litellm_provider": "deepinfra",
|
|
23751
|
+
"mode": "chat",
|
|
23752
|
+
"supports_tool_choice": true
|
|
23753
|
+
},
|
|
23754
|
+
"deepinfra/moonshotai/Kimi-K2-Instruct-0905": {
|
|
23755
|
+
"max_tokens": 262144,
|
|
23756
|
+
"max_input_tokens": 262144,
|
|
23757
|
+
"max_output_tokens": 262144,
|
|
23758
|
+
"input_cost_per_token": 5e-07,
|
|
23759
|
+
"output_cost_per_token": 2e-06,
|
|
23760
|
+
"cache_read_input_token_cost": 4e-07,
|
|
23761
|
+
"litellm_provider": "deepinfra",
|
|
23762
|
+
"mode": "chat",
|
|
23763
|
+
"supports_tool_choice": true
|
|
23764
|
+
},
|
|
23765
|
+
"deepinfra/nvidia/Llama-3.3-Nemotron-Super-49B-v1.5": {
|
|
23766
|
+
"max_tokens": 131072,
|
|
23767
|
+
"max_input_tokens": 131072,
|
|
23768
|
+
"max_output_tokens": 131072,
|
|
23769
|
+
"input_cost_per_token": 1e-07,
|
|
23770
|
+
"output_cost_per_token": 4e-07,
|
|
23771
|
+
"litellm_provider": "deepinfra",
|
|
23772
|
+
"mode": "chat",
|
|
23773
|
+
"supports_tool_choice": true
|
|
23774
|
+
},
|
|
23775
|
+
"deepinfra/nvidia/NVIDIA-Nemotron-Nano-9B-v2": {
|
|
23776
|
+
"max_tokens": 131072,
|
|
23777
|
+
"max_input_tokens": 131072,
|
|
23778
|
+
"max_output_tokens": 131072,
|
|
23779
|
+
"input_cost_per_token": 4e-08,
|
|
23780
|
+
"output_cost_per_token": 1.6e-07,
|
|
23781
|
+
"litellm_provider": "deepinfra",
|
|
23782
|
+
"mode": "chat",
|
|
23783
|
+
"supports_tool_choice": true
|
|
23784
|
+
},
|
|
23785
|
+
"jp.anthropic.claude-sonnet-4-5-20250929-v1:0": {
|
|
23786
|
+
"cache_creation_input_token_cost": 4.125e-06,
|
|
23787
|
+
"cache_read_input_token_cost": 3.3e-07,
|
|
23788
|
+
"input_cost_per_token": 3.3e-06,
|
|
23789
|
+
"input_cost_per_token_above_200k_tokens": 6.6e-06,
|
|
23790
|
+
"output_cost_per_token_above_200k_tokens": 2.475e-05,
|
|
23791
|
+
"cache_creation_input_token_cost_above_200k_tokens": 8.25e-06,
|
|
23792
|
+
"cache_read_input_token_cost_above_200k_tokens": 6.6e-07,
|
|
23793
|
+
"litellm_provider": "bedrock_converse",
|
|
23794
|
+
"max_input_tokens": 200000,
|
|
23795
|
+
"max_output_tokens": 64000,
|
|
23796
|
+
"max_tokens": 64000,
|
|
23797
|
+
"mode": "chat",
|
|
23798
|
+
"output_cost_per_token": 1.65e-05,
|
|
23799
|
+
"search_context_cost_per_query": {
|
|
23800
|
+
"search_context_size_high": 0.01,
|
|
23801
|
+
"search_context_size_low": 0.01,
|
|
23802
|
+
"search_context_size_medium": 0.01
|
|
23803
|
+
},
|
|
23804
|
+
"supports_assistant_prefill": true,
|
|
23805
|
+
"supports_computer_use": true,
|
|
23806
|
+
"supports_function_calling": true,
|
|
23807
|
+
"supports_pdf_input": true,
|
|
23808
|
+
"supports_prompt_caching": true,
|
|
23809
|
+
"supports_reasoning": true,
|
|
23810
|
+
"supports_response_schema": true,
|
|
23811
|
+
"supports_tool_choice": true,
|
|
23812
|
+
"supports_vision": true,
|
|
23813
|
+
"tool_use_system_prompt_tokens": 346
|
|
23814
|
+
},
|
|
23815
|
+
"watsonx/bigscience/mt0-xxl-13b": {
|
|
23816
|
+
"max_tokens": 8192,
|
|
23817
|
+
"max_input_tokens": 8192,
|
|
23818
|
+
"max_output_tokens": 8192,
|
|
23819
|
+
"input_cost_per_token": 0.0005,
|
|
23820
|
+
"output_cost_per_token": 0.002,
|
|
23821
|
+
"litellm_provider": "watsonx",
|
|
23822
|
+
"mode": "chat",
|
|
23823
|
+
"supports_function_calling": false,
|
|
23824
|
+
"supports_parallel_function_calling": false,
|
|
23825
|
+
"supports_vision": false
|
|
23826
|
+
},
|
|
23827
|
+
"watsonx/core42/jais-13b-chat": {
|
|
23828
|
+
"max_tokens": 8192,
|
|
23829
|
+
"max_input_tokens": 8192,
|
|
23830
|
+
"max_output_tokens": 8192,
|
|
23831
|
+
"input_cost_per_token": 0.0005,
|
|
23832
|
+
"output_cost_per_token": 0.002,
|
|
23833
|
+
"litellm_provider": "watsonx",
|
|
23834
|
+
"mode": "chat",
|
|
23835
|
+
"supports_function_calling": false,
|
|
23836
|
+
"supports_parallel_function_calling": false,
|
|
23837
|
+
"supports_vision": false
|
|
23838
|
+
},
|
|
23839
|
+
"watsonx/google/flan-t5-xl-3b": {
|
|
23840
|
+
"max_tokens": 8192,
|
|
23841
|
+
"max_input_tokens": 8192,
|
|
23842
|
+
"max_output_tokens": 8192,
|
|
23843
|
+
"input_cost_per_token": 6e-07,
|
|
23844
|
+
"output_cost_per_token": 6e-07,
|
|
23845
|
+
"litellm_provider": "watsonx",
|
|
23846
|
+
"mode": "chat",
|
|
23847
|
+
"supports_function_calling": false,
|
|
23848
|
+
"supports_parallel_function_calling": false,
|
|
23849
|
+
"supports_vision": false
|
|
23850
|
+
},
|
|
23851
|
+
"watsonx/ibm/granite-13b-chat-v2": {
|
|
23852
|
+
"max_tokens": 8192,
|
|
23853
|
+
"max_input_tokens": 8192,
|
|
23854
|
+
"max_output_tokens": 8192,
|
|
23855
|
+
"input_cost_per_token": 6e-07,
|
|
23856
|
+
"output_cost_per_token": 6e-07,
|
|
23857
|
+
"litellm_provider": "watsonx",
|
|
23858
|
+
"mode": "chat",
|
|
23859
|
+
"supports_function_calling": false,
|
|
23860
|
+
"supports_parallel_function_calling": false,
|
|
23861
|
+
"supports_vision": false
|
|
23862
|
+
},
|
|
23863
|
+
"watsonx/ibm/granite-13b-instruct-v2": {
|
|
23864
|
+
"max_tokens": 8192,
|
|
23865
|
+
"max_input_tokens": 8192,
|
|
23866
|
+
"max_output_tokens": 8192,
|
|
23867
|
+
"input_cost_per_token": 6e-07,
|
|
23868
|
+
"output_cost_per_token": 6e-07,
|
|
23869
|
+
"litellm_provider": "watsonx",
|
|
23870
|
+
"mode": "chat",
|
|
23871
|
+
"supports_function_calling": false,
|
|
23872
|
+
"supports_parallel_function_calling": false,
|
|
23873
|
+
"supports_vision": false
|
|
23874
|
+
},
|
|
23875
|
+
"watsonx/ibm/granite-3-3-8b-instruct": {
|
|
23876
|
+
"max_tokens": 8192,
|
|
23877
|
+
"max_input_tokens": 8192,
|
|
23878
|
+
"max_output_tokens": 8192,
|
|
23879
|
+
"input_cost_per_token": 2e-07,
|
|
23880
|
+
"output_cost_per_token": 2e-07,
|
|
23881
|
+
"litellm_provider": "watsonx",
|
|
23882
|
+
"mode": "chat",
|
|
23883
|
+
"supports_function_calling": true,
|
|
23884
|
+
"supports_parallel_function_calling": true,
|
|
23885
|
+
"supports_vision": false
|
|
23886
|
+
},
|
|
23887
|
+
"watsonx/ibm/granite-4-h-small": {
|
|
23888
|
+
"max_tokens": 20480,
|
|
23889
|
+
"max_input_tokens": 20480,
|
|
23890
|
+
"max_output_tokens": 20480,
|
|
23891
|
+
"input_cost_per_token": 6e-08,
|
|
23892
|
+
"output_cost_per_token": 2.5e-07,
|
|
23893
|
+
"litellm_provider": "watsonx",
|
|
23894
|
+
"mode": "chat",
|
|
23895
|
+
"supports_function_calling": true,
|
|
23896
|
+
"supports_parallel_function_calling": true,
|
|
23897
|
+
"supports_vision": false
|
|
23898
|
+
},
|
|
23899
|
+
"watsonx/ibm/granite-guardian-3-2-2b": {
|
|
23900
|
+
"max_tokens": 8192,
|
|
23901
|
+
"max_input_tokens": 8192,
|
|
23902
|
+
"max_output_tokens": 8192,
|
|
23903
|
+
"input_cost_per_token": 1e-07,
|
|
23904
|
+
"output_cost_per_token": 1e-07,
|
|
23905
|
+
"litellm_provider": "watsonx",
|
|
23906
|
+
"mode": "chat",
|
|
23907
|
+
"supports_function_calling": false,
|
|
23908
|
+
"supports_parallel_function_calling": false,
|
|
23909
|
+
"supports_vision": false
|
|
23910
|
+
},
|
|
23911
|
+
"watsonx/ibm/granite-guardian-3-3-8b": {
|
|
23912
|
+
"max_tokens": 8192,
|
|
23913
|
+
"max_input_tokens": 8192,
|
|
23914
|
+
"max_output_tokens": 8192,
|
|
23915
|
+
"input_cost_per_token": 2e-07,
|
|
23916
|
+
"output_cost_per_token": 2e-07,
|
|
23917
|
+
"litellm_provider": "watsonx",
|
|
23918
|
+
"mode": "chat",
|
|
23919
|
+
"supports_function_calling": false,
|
|
23920
|
+
"supports_parallel_function_calling": false,
|
|
23921
|
+
"supports_vision": false
|
|
23922
|
+
},
|
|
23923
|
+
"watsonx/ibm/granite-ttm-1024-96-r2": {
|
|
23924
|
+
"max_tokens": 512,
|
|
23925
|
+
"max_input_tokens": 512,
|
|
23926
|
+
"max_output_tokens": 512,
|
|
23927
|
+
"input_cost_per_token": 3.8e-07,
|
|
23928
|
+
"output_cost_per_token": 3.8e-07,
|
|
23929
|
+
"litellm_provider": "watsonx",
|
|
23930
|
+
"mode": "chat",
|
|
23931
|
+
"supports_function_calling": false,
|
|
23932
|
+
"supports_parallel_function_calling": false,
|
|
23933
|
+
"supports_vision": false
|
|
23934
|
+
},
|
|
23935
|
+
"watsonx/ibm/granite-ttm-1536-96-r2": {
|
|
23936
|
+
"max_tokens": 512,
|
|
23937
|
+
"max_input_tokens": 512,
|
|
23938
|
+
"max_output_tokens": 512,
|
|
23939
|
+
"input_cost_per_token": 3.8e-07,
|
|
23940
|
+
"output_cost_per_token": 3.8e-07,
|
|
23941
|
+
"litellm_provider": "watsonx",
|
|
23942
|
+
"mode": "chat",
|
|
23943
|
+
"supports_function_calling": false,
|
|
23944
|
+
"supports_parallel_function_calling": false,
|
|
23945
|
+
"supports_vision": false
|
|
23946
|
+
},
|
|
23947
|
+
"watsonx/ibm/granite-ttm-512-96-r2": {
|
|
23948
|
+
"max_tokens": 512,
|
|
23949
|
+
"max_input_tokens": 512,
|
|
23950
|
+
"max_output_tokens": 512,
|
|
23951
|
+
"input_cost_per_token": 3.8e-07,
|
|
23952
|
+
"output_cost_per_token": 3.8e-07,
|
|
23953
|
+
"litellm_provider": "watsonx",
|
|
23954
|
+
"mode": "chat",
|
|
23955
|
+
"supports_function_calling": false,
|
|
23956
|
+
"supports_parallel_function_calling": false,
|
|
23957
|
+
"supports_vision": false
|
|
23958
|
+
},
|
|
23959
|
+
"watsonx/ibm/granite-vision-3-2-2b": {
|
|
23960
|
+
"max_tokens": 8192,
|
|
23961
|
+
"max_input_tokens": 8192,
|
|
23962
|
+
"max_output_tokens": 8192,
|
|
23963
|
+
"input_cost_per_token": 1e-07,
|
|
23964
|
+
"output_cost_per_token": 1e-07,
|
|
23965
|
+
"litellm_provider": "watsonx",
|
|
23966
|
+
"mode": "chat",
|
|
23967
|
+
"supports_function_calling": false,
|
|
23968
|
+
"supports_parallel_function_calling": false,
|
|
23969
|
+
"supports_vision": true
|
|
23970
|
+
},
|
|
23971
|
+
"watsonx/meta-llama/llama-3-2-11b-vision-instruct": {
|
|
23972
|
+
"max_tokens": 128000,
|
|
23973
|
+
"max_input_tokens": 128000,
|
|
23974
|
+
"max_output_tokens": 128000,
|
|
23975
|
+
"input_cost_per_token": 3.5e-07,
|
|
23976
|
+
"output_cost_per_token": 3.5e-07,
|
|
23977
|
+
"litellm_provider": "watsonx",
|
|
23978
|
+
"mode": "chat",
|
|
23979
|
+
"supports_function_calling": true,
|
|
23980
|
+
"supports_parallel_function_calling": true,
|
|
23981
|
+
"supports_vision": true
|
|
23982
|
+
},
|
|
23983
|
+
"watsonx/meta-llama/llama-3-2-1b-instruct": {
|
|
23984
|
+
"max_tokens": 128000,
|
|
23985
|
+
"max_input_tokens": 128000,
|
|
23986
|
+
"max_output_tokens": 128000,
|
|
23987
|
+
"input_cost_per_token": 1e-07,
|
|
23988
|
+
"output_cost_per_token": 1e-07,
|
|
23989
|
+
"litellm_provider": "watsonx",
|
|
23990
|
+
"mode": "chat",
|
|
23991
|
+
"supports_function_calling": true,
|
|
23992
|
+
"supports_parallel_function_calling": true,
|
|
23993
|
+
"supports_vision": false
|
|
23994
|
+
},
|
|
23995
|
+
"watsonx/meta-llama/llama-3-2-3b-instruct": {
|
|
23996
|
+
"max_tokens": 128000,
|
|
23997
|
+
"max_input_tokens": 128000,
|
|
23998
|
+
"max_output_tokens": 128000,
|
|
23999
|
+
"input_cost_per_token": 1.5e-07,
|
|
24000
|
+
"output_cost_per_token": 1.5e-07,
|
|
24001
|
+
"litellm_provider": "watsonx",
|
|
24002
|
+
"mode": "chat",
|
|
24003
|
+
"supports_function_calling": true,
|
|
24004
|
+
"supports_parallel_function_calling": true,
|
|
24005
|
+
"supports_vision": false
|
|
24006
|
+
},
|
|
24007
|
+
"watsonx/meta-llama/llama-3-2-90b-vision-instruct": {
|
|
24008
|
+
"max_tokens": 128000,
|
|
24009
|
+
"max_input_tokens": 128000,
|
|
24010
|
+
"max_output_tokens": 128000,
|
|
24011
|
+
"input_cost_per_token": 2e-06,
|
|
24012
|
+
"output_cost_per_token": 2e-06,
|
|
24013
|
+
"litellm_provider": "watsonx",
|
|
24014
|
+
"mode": "chat",
|
|
24015
|
+
"supports_function_calling": true,
|
|
24016
|
+
"supports_parallel_function_calling": true,
|
|
24017
|
+
"supports_vision": true
|
|
24018
|
+
},
|
|
24019
|
+
"watsonx/meta-llama/llama-3-3-70b-instruct": {
|
|
24020
|
+
"max_tokens": 128000,
|
|
24021
|
+
"max_input_tokens": 128000,
|
|
24022
|
+
"max_output_tokens": 128000,
|
|
24023
|
+
"input_cost_per_token": 7.1e-07,
|
|
24024
|
+
"output_cost_per_token": 7.1e-07,
|
|
24025
|
+
"litellm_provider": "watsonx",
|
|
24026
|
+
"mode": "chat",
|
|
24027
|
+
"supports_function_calling": true,
|
|
24028
|
+
"supports_parallel_function_calling": true,
|
|
24029
|
+
"supports_vision": false
|
|
24030
|
+
},
|
|
24031
|
+
"watsonx/meta-llama/llama-4-maverick-17b": {
|
|
24032
|
+
"max_tokens": 128000,
|
|
24033
|
+
"max_input_tokens": 128000,
|
|
24034
|
+
"max_output_tokens": 128000,
|
|
24035
|
+
"input_cost_per_token": 3.5e-07,
|
|
24036
|
+
"output_cost_per_token": 1.4e-06,
|
|
24037
|
+
"litellm_provider": "watsonx",
|
|
24038
|
+
"mode": "chat",
|
|
24039
|
+
"supports_function_calling": true,
|
|
24040
|
+
"supports_parallel_function_calling": true,
|
|
24041
|
+
"supports_vision": false
|
|
24042
|
+
},
|
|
24043
|
+
"watsonx/meta-llama/llama-guard-3-11b-vision": {
|
|
24044
|
+
"max_tokens": 128000,
|
|
24045
|
+
"max_input_tokens": 128000,
|
|
24046
|
+
"max_output_tokens": 128000,
|
|
24047
|
+
"input_cost_per_token": 3.5e-07,
|
|
24048
|
+
"output_cost_per_token": 3.5e-07,
|
|
24049
|
+
"litellm_provider": "watsonx",
|
|
24050
|
+
"mode": "chat",
|
|
24051
|
+
"supports_function_calling": false,
|
|
24052
|
+
"supports_parallel_function_calling": false,
|
|
24053
|
+
"supports_vision": true
|
|
24054
|
+
},
|
|
24055
|
+
"watsonx/mistralai/mistral-medium-2505": {
|
|
24056
|
+
"max_tokens": 128000,
|
|
24057
|
+
"max_input_tokens": 128000,
|
|
24058
|
+
"max_output_tokens": 128000,
|
|
24059
|
+
"input_cost_per_token": 3e-06,
|
|
24060
|
+
"output_cost_per_token": 1e-05,
|
|
24061
|
+
"litellm_provider": "watsonx",
|
|
24062
|
+
"mode": "chat",
|
|
24063
|
+
"supports_function_calling": true,
|
|
24064
|
+
"supports_parallel_function_calling": true,
|
|
24065
|
+
"supports_vision": false
|
|
24066
|
+
},
|
|
24067
|
+
"watsonx/mistralai/mistral-small-2503": {
|
|
24068
|
+
"max_tokens": 32000,
|
|
24069
|
+
"max_input_tokens": 32000,
|
|
24070
|
+
"max_output_tokens": 32000,
|
|
24071
|
+
"input_cost_per_token": 1e-07,
|
|
24072
|
+
"output_cost_per_token": 3e-07,
|
|
24073
|
+
"litellm_provider": "watsonx",
|
|
24074
|
+
"mode": "chat",
|
|
24075
|
+
"supports_function_calling": true,
|
|
24076
|
+
"supports_parallel_function_calling": true,
|
|
24077
|
+
"supports_vision": false
|
|
24078
|
+
},
|
|
24079
|
+
"watsonx/mistralai/pixtral-12b-2409": {
|
|
24080
|
+
"max_tokens": 128000,
|
|
24081
|
+
"max_input_tokens": 128000,
|
|
24082
|
+
"max_output_tokens": 128000,
|
|
24083
|
+
"input_cost_per_token": 3.5e-07,
|
|
24084
|
+
"output_cost_per_token": 3.5e-07,
|
|
24085
|
+
"litellm_provider": "watsonx",
|
|
24086
|
+
"mode": "chat",
|
|
24087
|
+
"supports_function_calling": false,
|
|
24088
|
+
"supports_parallel_function_calling": false,
|
|
24089
|
+
"supports_vision": true
|
|
24090
|
+
},
|
|
24091
|
+
"watsonx/openai/gpt-oss-120b": {
|
|
24092
|
+
"max_tokens": 8192,
|
|
24093
|
+
"max_input_tokens": 8192,
|
|
24094
|
+
"max_output_tokens": 8192,
|
|
24095
|
+
"input_cost_per_token": 1.5e-07,
|
|
24096
|
+
"output_cost_per_token": 6e-07,
|
|
24097
|
+
"litellm_provider": "watsonx",
|
|
24098
|
+
"mode": "chat",
|
|
24099
|
+
"supports_function_calling": false,
|
|
24100
|
+
"supports_parallel_function_calling": false,
|
|
24101
|
+
"supports_vision": false
|
|
24102
|
+
},
|
|
24103
|
+
"watsonx/sdaia/allam-1-13b-instruct": {
|
|
24104
|
+
"max_tokens": 8192,
|
|
24105
|
+
"max_input_tokens": 8192,
|
|
24106
|
+
"max_output_tokens": 8192,
|
|
24107
|
+
"input_cost_per_token": 1.8e-06,
|
|
24108
|
+
"output_cost_per_token": 1.8e-06,
|
|
24109
|
+
"litellm_provider": "watsonx",
|
|
24110
|
+
"mode": "chat",
|
|
24111
|
+
"supports_function_calling": false,
|
|
24112
|
+
"supports_parallel_function_calling": false,
|
|
24113
|
+
"supports_vision": false
|
|
24114
|
+
},
|
|
24115
|
+
"global.anthropic.claude-sonnet-4-5-20250929-v1:0": {
|
|
24116
|
+
"cache_creation_input_token_cost": 3.75e-06,
|
|
24117
|
+
"cache_read_input_token_cost": 3e-07,
|
|
24118
|
+
"input_cost_per_token": 3e-06,
|
|
24119
|
+
"input_cost_per_token_above_200k_tokens": 6e-06,
|
|
24120
|
+
"output_cost_per_token_above_200k_tokens": 2.25e-05,
|
|
24121
|
+
"cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
|
|
24122
|
+
"cache_read_input_token_cost_above_200k_tokens": 6e-07,
|
|
24123
|
+
"litellm_provider": "bedrock_converse",
|
|
24124
|
+
"max_input_tokens": 200000,
|
|
24125
|
+
"max_output_tokens": 64000,
|
|
24126
|
+
"max_tokens": 64000,
|
|
24127
|
+
"mode": "chat",
|
|
24128
|
+
"output_cost_per_token": 1.5e-05,
|
|
24129
|
+
"search_context_cost_per_query": {
|
|
24130
|
+
"search_context_size_high": 0.01,
|
|
24131
|
+
"search_context_size_low": 0.01,
|
|
24132
|
+
"search_context_size_medium": 0.01
|
|
24133
|
+
},
|
|
24134
|
+
"supports_assistant_prefill": true,
|
|
24135
|
+
"supports_computer_use": true,
|
|
24136
|
+
"supports_function_calling": true,
|
|
24137
|
+
"supports_pdf_input": true,
|
|
24138
|
+
"supports_prompt_caching": true,
|
|
24139
|
+
"supports_reasoning": true,
|
|
24140
|
+
"supports_response_schema": true,
|
|
24141
|
+
"supports_tool_choice": true,
|
|
24142
|
+
"supports_vision": true,
|
|
24143
|
+
"tool_use_system_prompt_tokens": 346
|
|
24144
|
+
},
|
|
24145
|
+
"global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
24146
|
+
"cache_creation_input_token_cost": 3.75e-06,
|
|
24147
|
+
"cache_read_input_token_cost": 3e-07,
|
|
24148
|
+
"input_cost_per_token": 3e-06,
|
|
24149
|
+
"input_cost_per_token_above_200k_tokens": 6e-06,
|
|
24150
|
+
"output_cost_per_token_above_200k_tokens": 2.25e-05,
|
|
24151
|
+
"cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
|
|
24152
|
+
"cache_read_input_token_cost_above_200k_tokens": 6e-07,
|
|
24153
|
+
"litellm_provider": "bedrock_converse",
|
|
24154
|
+
"max_input_tokens": 1000000,
|
|
24155
|
+
"max_output_tokens": 64000,
|
|
24156
|
+
"max_tokens": 64000,
|
|
24157
|
+
"mode": "chat",
|
|
24158
|
+
"output_cost_per_token": 1.5e-05,
|
|
24159
|
+
"search_context_cost_per_query": {
|
|
24160
|
+
"search_context_size_high": 0.01,
|
|
24161
|
+
"search_context_size_low": 0.01,
|
|
24162
|
+
"search_context_size_medium": 0.01
|
|
24163
|
+
},
|
|
24164
|
+
"supports_assistant_prefill": true,
|
|
24165
|
+
"supports_computer_use": true,
|
|
24166
|
+
"supports_function_calling": true,
|
|
24167
|
+
"supports_pdf_input": true,
|
|
24168
|
+
"supports_prompt_caching": true,
|
|
24169
|
+
"supports_reasoning": true,
|
|
24170
|
+
"supports_response_schema": true,
|
|
24171
|
+
"supports_tool_choice": true,
|
|
24172
|
+
"supports_vision": true,
|
|
24173
|
+
"tool_use_system_prompt_tokens": 159
|
|
24174
|
+
},
|
|
24175
|
+
"gpt-5-pro": {
|
|
24176
|
+
"input_cost_per_token": 1.5e-05,
|
|
24177
|
+
"input_cost_per_token_batches": 7.5e-06,
|
|
24178
|
+
"litellm_provider": "openai",
|
|
24179
|
+
"max_input_tokens": 400000,
|
|
24180
|
+
"max_output_tokens": 272000,
|
|
24181
|
+
"max_tokens": 272000,
|
|
24182
|
+
"mode": "responses",
|
|
24183
|
+
"output_cost_per_token": 0.00012,
|
|
24184
|
+
"output_cost_per_token_batches": 6e-05,
|
|
24185
|
+
"supported_endpoints": [
|
|
24186
|
+
"/v1/batch",
|
|
24187
|
+
"/v1/responses"
|
|
24188
|
+
],
|
|
24189
|
+
"supported_modalities": [
|
|
24190
|
+
"text",
|
|
24191
|
+
"image"
|
|
24192
|
+
],
|
|
24193
|
+
"supported_output_modalities": [
|
|
24194
|
+
"text"
|
|
24195
|
+
],
|
|
24196
|
+
"supports_function_calling": true,
|
|
24197
|
+
"supports_native_streaming": false,
|
|
24198
|
+
"supports_parallel_function_calling": true,
|
|
24199
|
+
"supports_pdf_input": true,
|
|
24200
|
+
"supports_prompt_caching": true,
|
|
24201
|
+
"supports_reasoning": true,
|
|
24202
|
+
"supports_response_schema": true,
|
|
24203
|
+
"supports_system_messages": true,
|
|
24204
|
+
"supports_tool_choice": true,
|
|
24205
|
+
"supports_vision": true,
|
|
24206
|
+
"supports_web_search": true
|
|
24207
|
+
},
|
|
24208
|
+
"gpt-image-1-mini": {
|
|
24209
|
+
"cache_read_input_image_token_cost": 2.5e-07,
|
|
24210
|
+
"cache_read_input_token_cost": 2e-07,
|
|
24211
|
+
"input_cost_per_image_token": 2.5e-06,
|
|
24212
|
+
"input_cost_per_token": 2e-06,
|
|
24213
|
+
"litellm_provider": "openai",
|
|
24214
|
+
"mode": "chat",
|
|
24215
|
+
"output_cost_per_image_token": 8e-06,
|
|
24216
|
+
"supported_endpoints": [
|
|
24217
|
+
"/v1/images/generations",
|
|
24218
|
+
"/v1/images/edits"
|
|
24219
|
+
]
|
|
24220
|
+
},
|
|
24221
|
+
"gpt-realtime-mini": {
|
|
24222
|
+
"cache_creation_input_audio_token_cost": 3e-07,
|
|
24223
|
+
"cache_read_input_audio_token_cost": 3e-07,
|
|
24224
|
+
"input_cost_per_audio_token": 1e-05,
|
|
24225
|
+
"input_cost_per_token": 6e-07,
|
|
24226
|
+
"litellm_provider": "openai",
|
|
24227
|
+
"max_input_tokens": 128000,
|
|
24228
|
+
"max_output_tokens": 4096,
|
|
24229
|
+
"max_tokens": 4096,
|
|
24230
|
+
"mode": "chat",
|
|
24231
|
+
"output_cost_per_audio_token": 2e-05,
|
|
24232
|
+
"output_cost_per_token": 2.4e-06,
|
|
24233
|
+
"supported_endpoints": [
|
|
24234
|
+
"/v1/realtime"
|
|
24235
|
+
],
|
|
24236
|
+
"supported_modalities": [
|
|
24237
|
+
"text",
|
|
24238
|
+
"image",
|
|
24239
|
+
"audio"
|
|
24240
|
+
],
|
|
24241
|
+
"supported_output_modalities": [
|
|
24242
|
+
"text",
|
|
24243
|
+
"audio"
|
|
24244
|
+
],
|
|
24245
|
+
"supports_audio_input": true,
|
|
24246
|
+
"supports_audio_output": true,
|
|
24247
|
+
"supports_function_calling": true,
|
|
24248
|
+
"supports_parallel_function_calling": true,
|
|
24249
|
+
"supports_system_messages": true,
|
|
24250
|
+
"supports_tool_choice": true
|
|
24251
|
+
},
|
|
24252
|
+
"low/1024-x-1024/gpt-image-1-mini": {
|
|
24253
|
+
"input_cost_per_image": 0.005,
|
|
24254
|
+
"litellm_provider": "openai",
|
|
24255
|
+
"mode": "image_generation",
|
|
24256
|
+
"supported_endpoints": [
|
|
24257
|
+
"/v1/images/generations"
|
|
24258
|
+
]
|
|
24259
|
+
},
|
|
24260
|
+
"low/1024-x-1536/gpt-image-1-mini": {
|
|
24261
|
+
"input_cost_per_image": 0.006,
|
|
24262
|
+
"litellm_provider": "openai",
|
|
24263
|
+
"mode": "image_generation",
|
|
24264
|
+
"supported_endpoints": [
|
|
24265
|
+
"/v1/images/generations"
|
|
24266
|
+
]
|
|
24267
|
+
},
|
|
24268
|
+
"low/1536-x-1024/gpt-image-1-mini": {
|
|
24269
|
+
"input_cost_per_image": 0.006,
|
|
24270
|
+
"litellm_provider": "openai",
|
|
24271
|
+
"mode": "image_generation",
|
|
24272
|
+
"supported_endpoints": [
|
|
24273
|
+
"/v1/images/generations"
|
|
24274
|
+
]
|
|
24275
|
+
},
|
|
24276
|
+
"medium/1024-x-1024/gpt-image-1-mini": {
|
|
24277
|
+
"input_cost_per_image": 0.011,
|
|
24278
|
+
"litellm_provider": "openai",
|
|
24279
|
+
"mode": "image_generation",
|
|
24280
|
+
"supported_endpoints": [
|
|
24281
|
+
"/v1/images/generations"
|
|
24282
|
+
]
|
|
24283
|
+
},
|
|
24284
|
+
"medium/1024-x-1536/gpt-image-1-mini": {
|
|
24285
|
+
"input_cost_per_image": 0.015,
|
|
24286
|
+
"litellm_provider": "openai",
|
|
24287
|
+
"mode": "image_generation",
|
|
24288
|
+
"supported_endpoints": [
|
|
24289
|
+
"/v1/images/generations"
|
|
24290
|
+
]
|
|
24291
|
+
},
|
|
24292
|
+
"medium/1536-x-1024/gpt-image-1-mini": {
|
|
24293
|
+
"input_cost_per_image": 0.015,
|
|
24294
|
+
"litellm_provider": "openai",
|
|
24295
|
+
"mode": "image_generation",
|
|
24296
|
+
"supported_endpoints": [
|
|
24297
|
+
"/v1/images/generations"
|
|
24298
|
+
]
|
|
24299
|
+
},
|
|
24300
|
+
"cohere.embed-v4:0": {
|
|
24301
|
+
"input_cost_per_token": 1.2e-07,
|
|
24302
|
+
"litellm_provider": "bedrock",
|
|
24303
|
+
"max_input_tokens": 128000,
|
|
24304
|
+
"max_tokens": 128000,
|
|
24305
|
+
"mode": "embedding",
|
|
24306
|
+
"output_cost_per_token": 0.0,
|
|
24307
|
+
"output_vector_size": 1536,
|
|
24308
|
+
"supports_embedding_image_input": true
|
|
24309
|
+
},
|
|
24310
|
+
"gpt-5-pro-2025-10-06": {
|
|
24311
|
+
"input_cost_per_token": 1.5e-05,
|
|
24312
|
+
"input_cost_per_token_batches": 7.5e-06,
|
|
24313
|
+
"litellm_provider": "openai",
|
|
24314
|
+
"max_input_tokens": 400000,
|
|
24315
|
+
"max_output_tokens": 272000,
|
|
24316
|
+
"max_tokens": 272000,
|
|
24317
|
+
"mode": "responses",
|
|
24318
|
+
"output_cost_per_token": 0.00012,
|
|
24319
|
+
"output_cost_per_token_batches": 6e-05,
|
|
24320
|
+
"supported_endpoints": [
|
|
24321
|
+
"/v1/batch",
|
|
24322
|
+
"/v1/responses"
|
|
24323
|
+
],
|
|
24324
|
+
"supported_modalities": [
|
|
24325
|
+
"text",
|
|
24326
|
+
"image"
|
|
24327
|
+
],
|
|
24328
|
+
"supported_output_modalities": [
|
|
24329
|
+
"text"
|
|
24330
|
+
],
|
|
24331
|
+
"supports_function_calling": true,
|
|
24332
|
+
"supports_native_streaming": false,
|
|
24333
|
+
"supports_parallel_function_calling": true,
|
|
24334
|
+
"supports_pdf_input": true,
|
|
24335
|
+
"supports_prompt_caching": true,
|
|
24336
|
+
"supports_reasoning": true,
|
|
24337
|
+
"supports_response_schema": true,
|
|
24338
|
+
"supports_system_messages": true,
|
|
24339
|
+
"supports_tool_choice": true,
|
|
24340
|
+
"supports_vision": true,
|
|
24341
|
+
"supports_web_search": true
|
|
24342
|
+
},
|
|
24343
|
+
"azure_ai/Phi-4-mini-reasoning": {
|
|
24344
|
+
"input_cost_per_token": 8e-08,
|
|
24345
|
+
"litellm_provider": "azure_ai",
|
|
24346
|
+
"max_input_tokens": 131072,
|
|
24347
|
+
"max_output_tokens": 4096,
|
|
24348
|
+
"max_tokens": 4096,
|
|
24349
|
+
"mode": "chat",
|
|
24350
|
+
"output_cost_per_token": 3.2e-07,
|
|
24351
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/microsoft/",
|
|
24352
|
+
"supports_function_calling": true
|
|
24353
|
+
},
|
|
24354
|
+
"azure_ai/Phi-4-reasoning": {
|
|
24355
|
+
"input_cost_per_token": 1.25e-07,
|
|
24356
|
+
"litellm_provider": "azure_ai",
|
|
24357
|
+
"max_input_tokens": 32768,
|
|
24358
|
+
"max_output_tokens": 4096,
|
|
24359
|
+
"max_tokens": 4096,
|
|
24360
|
+
"mode": "chat",
|
|
24361
|
+
"output_cost_per_token": 5e-07,
|
|
24362
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/microsoft/",
|
|
24363
|
+
"supports_function_calling": true,
|
|
24364
|
+
"supports_tool_choice": true,
|
|
24365
|
+
"supports_reasoning": true
|
|
24366
|
+
},
|
|
24367
|
+
"azure_ai/MAI-DS-R1": {
|
|
24368
|
+
"input_cost_per_token": 1.35e-06,
|
|
24369
|
+
"litellm_provider": "azure_ai",
|
|
24370
|
+
"max_input_tokens": 128000,
|
|
24371
|
+
"max_output_tokens": 8192,
|
|
24372
|
+
"max_tokens": 8192,
|
|
24373
|
+
"mode": "chat",
|
|
24374
|
+
"output_cost_per_token": 5.4e-06,
|
|
24375
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/microsoft/",
|
|
24376
|
+
"supports_reasoning": true,
|
|
24377
|
+
"supports_tool_choice": true
|
|
24378
|
+
},
|
|
24379
|
+
"au.anthropic.claude-sonnet-4-5-20250929-v1:0": {
|
|
24380
|
+
"cache_creation_input_token_cost": 4.125e-06,
|
|
24381
|
+
"cache_read_input_token_cost": 3.3e-07,
|
|
24382
|
+
"input_cost_per_token": 3.3e-06,
|
|
24383
|
+
"input_cost_per_token_above_200k_tokens": 6.6e-06,
|
|
24384
|
+
"output_cost_per_token_above_200k_tokens": 2.475e-05,
|
|
24385
|
+
"cache_creation_input_token_cost_above_200k_tokens": 8.25e-06,
|
|
24386
|
+
"cache_read_input_token_cost_above_200k_tokens": 6.6e-07,
|
|
24387
|
+
"litellm_provider": "bedrock_converse",
|
|
24388
|
+
"max_input_tokens": 200000,
|
|
24389
|
+
"max_output_tokens": 64000,
|
|
24390
|
+
"max_tokens": 64000,
|
|
24391
|
+
"mode": "chat",
|
|
24392
|
+
"output_cost_per_token": 1.65e-05,
|
|
24393
|
+
"search_context_cost_per_query": {
|
|
24394
|
+
"search_context_size_high": 0.01,
|
|
24395
|
+
"search_context_size_low": 0.01,
|
|
24396
|
+
"search_context_size_medium": 0.01
|
|
24397
|
+
},
|
|
24398
|
+
"supports_assistant_prefill": true,
|
|
24399
|
+
"supports_computer_use": true,
|
|
24400
|
+
"supports_function_calling": true,
|
|
24401
|
+
"supports_pdf_input": true,
|
|
24402
|
+
"supports_prompt_caching": true,
|
|
24403
|
+
"supports_reasoning": true,
|
|
24404
|
+
"supports_response_schema": true,
|
|
24405
|
+
"supports_tool_choice": true,
|
|
24406
|
+
"supports_vision": true,
|
|
24407
|
+
"tool_use_system_prompt_tokens": 346
|
|
24408
|
+
},
|
|
24409
|
+
"oci/cohere.command-latest": {
|
|
24410
|
+
"input_cost_per_token": 1.56e-06,
|
|
24411
|
+
"litellm_provider": "oci",
|
|
24412
|
+
"max_input_tokens": 128000,
|
|
24413
|
+
"max_output_tokens": 4000,
|
|
24414
|
+
"max_tokens": 128000,
|
|
24415
|
+
"mode": "chat",
|
|
24416
|
+
"output_cost_per_token": 1.56e-06,
|
|
24417
|
+
"source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/",
|
|
24418
|
+
"supports_function_calling": true,
|
|
24419
|
+
"supports_response_schema": false
|
|
24420
|
+
},
|
|
24421
|
+
"oci/cohere.command-a-03-2025": {
|
|
24422
|
+
"input_cost_per_token": 1.56e-06,
|
|
24423
|
+
"litellm_provider": "oci",
|
|
24424
|
+
"max_input_tokens": 256000,
|
|
24425
|
+
"max_output_tokens": 4000,
|
|
24426
|
+
"max_tokens": 256000,
|
|
24427
|
+
"mode": "chat",
|
|
24428
|
+
"output_cost_per_token": 1.56e-06,
|
|
24429
|
+
"source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/",
|
|
24430
|
+
"supports_function_calling": true,
|
|
24431
|
+
"supports_response_schema": false
|
|
24432
|
+
},
|
|
24433
|
+
"oci/cohere.command-plus-latest": {
|
|
24434
|
+
"input_cost_per_token": 1.56e-06,
|
|
24435
|
+
"litellm_provider": "oci",
|
|
24436
|
+
"max_input_tokens": 128000,
|
|
24437
|
+
"max_output_tokens": 4000,
|
|
24438
|
+
"max_tokens": 128000,
|
|
24439
|
+
"mode": "chat",
|
|
24440
|
+
"output_cost_per_token": 1.56e-06,
|
|
24441
|
+
"source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/",
|
|
24442
|
+
"supports_function_calling": true,
|
|
24443
|
+
"supports_response_schema": false
|
|
24444
|
+
},
|
|
24445
|
+
"together_ai/moonshotai/Kimi-K2-Instruct-0905": {
|
|
24446
|
+
"input_cost_per_token": 1e-06,
|
|
24447
|
+
"litellm_provider": "together_ai",
|
|
24448
|
+
"max_input_tokens": 262144,
|
|
24449
|
+
"mode": "chat",
|
|
24450
|
+
"output_cost_per_token": 3e-06,
|
|
24451
|
+
"source": "https://www.together.ai/models/kimi-k2-0905",
|
|
24452
|
+
"supports_function_calling": true,
|
|
24453
|
+
"supports_parallel_function_calling": true,
|
|
24454
|
+
"supports_tool_choice": true
|
|
24455
|
+
},
|
|
24456
|
+
"together_ai/Qwen/Qwen3-Next-80B-A3B-Instruct": {
|
|
24457
|
+
"input_cost_per_token": 1.5e-07,
|
|
24458
|
+
"litellm_provider": "together_ai",
|
|
24459
|
+
"max_input_tokens": 262144,
|
|
24460
|
+
"mode": "chat",
|
|
24461
|
+
"output_cost_per_token": 1.5e-06,
|
|
24462
|
+
"source": "https://www.together.ai/models/qwen3-next-80b-a3b-instruct",
|
|
24463
|
+
"supports_function_calling": true,
|
|
24464
|
+
"supports_parallel_function_calling": true,
|
|
24465
|
+
"supports_tool_choice": true
|
|
24466
|
+
},
|
|
24467
|
+
"together_ai/Qwen/Qwen3-Next-80B-A3B-Thinking": {
|
|
24468
|
+
"input_cost_per_token": 1.5e-07,
|
|
24469
|
+
"litellm_provider": "together_ai",
|
|
24470
|
+
"max_input_tokens": 262144,
|
|
24471
|
+
"mode": "chat",
|
|
24472
|
+
"output_cost_per_token": 1.5e-06,
|
|
24473
|
+
"source": "https://www.together.ai/models/qwen3-next-80b-a3b-thinking",
|
|
24474
|
+
"supports_function_calling": true,
|
|
24475
|
+
"supports_parallel_function_calling": true,
|
|
24476
|
+
"supports_tool_choice": true
|
|
24477
|
+
},
|
|
24478
|
+
"together_ai/baai/bge-base-en-v1.5": {
|
|
24479
|
+
"input_cost_per_token": 8e-09,
|
|
24480
|
+
"litellm_provider": "together_ai",
|
|
24481
|
+
"max_input_tokens": 512,
|
|
24482
|
+
"mode": "embedding",
|
|
24483
|
+
"output_cost_per_token": 0.0,
|
|
24484
|
+
"output_vector_size": 768
|
|
24485
|
+
},
|
|
24486
|
+
"together_ai/BAAI/bge-base-en-v1.5": {
|
|
24487
|
+
"input_cost_per_token": 8e-09,
|
|
24488
|
+
"litellm_provider": "together_ai",
|
|
24489
|
+
"max_input_tokens": 512,
|
|
24490
|
+
"mode": "embedding",
|
|
24491
|
+
"output_cost_per_token": 0.0,
|
|
24492
|
+
"output_vector_size": 768
|
|
24493
|
+
},
|
|
24494
|
+
"openrouter/anthropic/claude-sonnet-4.5": {
|
|
24495
|
+
"input_cost_per_image": 0.0048,
|
|
24496
|
+
"cache_creation_input_token_cost": 3.75e-06,
|
|
24497
|
+
"cache_read_input_token_cost": 3e-07,
|
|
24498
|
+
"input_cost_per_token": 3e-06,
|
|
24499
|
+
"input_cost_per_token_above_200k_tokens": 6e-06,
|
|
24500
|
+
"output_cost_per_token_above_200k_tokens": 2.25e-05,
|
|
24501
|
+
"cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
|
|
24502
|
+
"cache_read_input_token_cost_above_200k_tokens": 6e-07,
|
|
24503
|
+
"litellm_provider": "openrouter",
|
|
24504
|
+
"max_input_tokens": 1000000,
|
|
24505
|
+
"max_output_tokens": 1000000,
|
|
24506
|
+
"max_tokens": 1000000,
|
|
24507
|
+
"mode": "chat",
|
|
24508
|
+
"output_cost_per_token": 1.5e-05,
|
|
24509
|
+
"supports_assistant_prefill": true,
|
|
24510
|
+
"supports_computer_use": true,
|
|
24511
|
+
"supports_function_calling": true,
|
|
24512
|
+
"supports_prompt_caching": true,
|
|
24513
|
+
"supports_reasoning": true,
|
|
24514
|
+
"supports_tool_choice": true,
|
|
24515
|
+
"supports_vision": true,
|
|
24516
|
+
"tool_use_system_prompt_tokens": 159
|
|
24517
|
+
},
|
|
24518
|
+
"anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
24519
|
+
"cache_creation_input_token_cost": 1.25e-06,
|
|
24520
|
+
"cache_read_input_token_cost": 1e-07,
|
|
24521
|
+
"input_cost_per_token": 1e-06,
|
|
24522
|
+
"litellm_provider": "bedrock",
|
|
24523
|
+
"max_input_tokens": 200000,
|
|
24524
|
+
"max_output_tokens": 8192,
|
|
24525
|
+
"max_tokens": 8192,
|
|
24526
|
+
"mode": "chat",
|
|
24527
|
+
"output_cost_per_token": 5e-06,
|
|
24528
|
+
"source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
|
|
24529
|
+
"supports_assistant_prefill": true,
|
|
24530
|
+
"supports_function_calling": true,
|
|
24531
|
+
"supports_pdf_input": true,
|
|
24532
|
+
"supports_prompt_caching": true,
|
|
24533
|
+
"supports_reasoning": true,
|
|
24534
|
+
"supports_response_schema": true,
|
|
24535
|
+
"supports_tool_choice": true
|
|
24536
|
+
},
|
|
24537
|
+
"anthropic.claude-haiku-4-5@20251001": {
|
|
24538
|
+
"cache_creation_input_token_cost": 1.25e-06,
|
|
24539
|
+
"cache_read_input_token_cost": 1e-07,
|
|
24540
|
+
"input_cost_per_token": 1e-06,
|
|
24541
|
+
"litellm_provider": "bedrock",
|
|
24542
|
+
"max_input_tokens": 200000,
|
|
24543
|
+
"max_output_tokens": 8192,
|
|
24544
|
+
"max_tokens": 8192,
|
|
24545
|
+
"mode": "chat",
|
|
24546
|
+
"output_cost_per_token": 5e-06,
|
|
24547
|
+
"source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
|
|
24548
|
+
"supports_assistant_prefill": true,
|
|
24549
|
+
"supports_function_calling": true,
|
|
24550
|
+
"supports_pdf_input": true,
|
|
24551
|
+
"supports_prompt_caching": true,
|
|
24552
|
+
"supports_reasoning": true,
|
|
24553
|
+
"supports_response_schema": true,
|
|
24554
|
+
"supports_tool_choice": true
|
|
24555
|
+
},
|
|
24556
|
+
"apac.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
24557
|
+
"cache_creation_input_token_cost": 1.375e-06,
|
|
24558
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
24559
|
+
"input_cost_per_token": 1.1e-06,
|
|
24560
|
+
"litellm_provider": "bedrock",
|
|
24561
|
+
"max_input_tokens": 200000,
|
|
24562
|
+
"max_output_tokens": 8192,
|
|
24563
|
+
"max_tokens": 8192,
|
|
24564
|
+
"mode": "chat",
|
|
24565
|
+
"output_cost_per_token": 5.5e-06,
|
|
24566
|
+
"source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
|
|
24567
|
+
"supports_assistant_prefill": true,
|
|
24568
|
+
"supports_function_calling": true,
|
|
24569
|
+
"supports_pdf_input": true,
|
|
24570
|
+
"supports_prompt_caching": true,
|
|
24571
|
+
"supports_reasoning": true,
|
|
24572
|
+
"supports_response_schema": true,
|
|
24573
|
+
"supports_tool_choice": true
|
|
24574
|
+
},
|
|
24575
|
+
"claude-haiku-4-5-20251001": {
|
|
24576
|
+
"cache_creation_input_token_cost": 1.25e-06,
|
|
24577
|
+
"cache_creation_input_token_cost_above_1hr": 2e-06,
|
|
24578
|
+
"cache_read_input_token_cost": 1e-07,
|
|
24579
|
+
"input_cost_per_token": 1e-06,
|
|
24580
|
+
"litellm_provider": "anthropic",
|
|
24581
|
+
"max_input_tokens": 200000,
|
|
24582
|
+
"max_output_tokens": 64000,
|
|
24583
|
+
"max_tokens": 64000,
|
|
24584
|
+
"mode": "chat",
|
|
24585
|
+
"output_cost_per_token": 5e-06,
|
|
24586
|
+
"supports_assistant_prefill": true,
|
|
24587
|
+
"supports_function_calling": true,
|
|
24588
|
+
"supports_computer_use": true,
|
|
24589
|
+
"supports_pdf_input": true,
|
|
24590
|
+
"supports_prompt_caching": true,
|
|
24591
|
+
"supports_reasoning": true,
|
|
24592
|
+
"supports_response_schema": true,
|
|
24593
|
+
"supports_tool_choice": true,
|
|
24594
|
+
"supports_vision": true
|
|
24595
|
+
},
|
|
24596
|
+
"claude-haiku-4-5": {
|
|
24597
|
+
"cache_creation_input_token_cost": 1.25e-06,
|
|
24598
|
+
"cache_creation_input_token_cost_above_1hr": 2e-06,
|
|
24599
|
+
"cache_read_input_token_cost": 1e-07,
|
|
24600
|
+
"input_cost_per_token": 1e-06,
|
|
24601
|
+
"litellm_provider": "anthropic",
|
|
24602
|
+
"max_input_tokens": 200000,
|
|
24603
|
+
"max_output_tokens": 64000,
|
|
24604
|
+
"max_tokens": 64000,
|
|
24605
|
+
"mode": "chat",
|
|
24606
|
+
"output_cost_per_token": 5e-06,
|
|
24607
|
+
"supports_assistant_prefill": true,
|
|
24608
|
+
"supports_function_calling": true,
|
|
24609
|
+
"supports_computer_use": true,
|
|
24610
|
+
"supports_pdf_input": true,
|
|
24611
|
+
"supports_prompt_caching": true,
|
|
24612
|
+
"supports_reasoning": true,
|
|
24613
|
+
"supports_response_schema": true,
|
|
24614
|
+
"supports_tool_choice": true,
|
|
24615
|
+
"supports_vision": true
|
|
24616
|
+
},
|
|
24617
|
+
"eu.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
24618
|
+
"cache_creation_input_token_cost": 1.375e-06,
|
|
24619
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
24620
|
+
"input_cost_per_token": 1.1e-06,
|
|
24621
|
+
"deprecation_date": "2026-10-15",
|
|
24622
|
+
"litellm_provider": "bedrock",
|
|
24623
|
+
"max_input_tokens": 200000,
|
|
24624
|
+
"max_output_tokens": 8192,
|
|
24625
|
+
"max_tokens": 8192,
|
|
24626
|
+
"mode": "chat",
|
|
24627
|
+
"output_cost_per_token": 5.5e-06,
|
|
24628
|
+
"source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
|
|
24629
|
+
"supports_assistant_prefill": true,
|
|
24630
|
+
"supports_function_calling": true,
|
|
24631
|
+
"supports_pdf_input": true,
|
|
24632
|
+
"supports_prompt_caching": true,
|
|
24633
|
+
"supports_reasoning": true,
|
|
24634
|
+
"supports_response_schema": true,
|
|
24635
|
+
"supports_tool_choice": true
|
|
24636
|
+
},
|
|
24637
|
+
"gemini-2.5-flash-image": {
|
|
24638
|
+
"cache_read_input_token_cost": 3e-08,
|
|
24639
|
+
"input_cost_per_audio_token": 1e-06,
|
|
24640
|
+
"input_cost_per_token": 3e-07,
|
|
24641
|
+
"litellm_provider": "vertex_ai-language-models",
|
|
24642
|
+
"max_audio_length_hours": 8.4,
|
|
24643
|
+
"max_audio_per_prompt": 1,
|
|
24644
|
+
"max_images_per_prompt": 3000,
|
|
24645
|
+
"max_input_tokens": 32768,
|
|
24646
|
+
"max_output_tokens": 32768,
|
|
24647
|
+
"max_tokens": 32768,
|
|
24648
|
+
"max_pdf_size_mb": 30,
|
|
24649
|
+
"max_video_length": 1,
|
|
24650
|
+
"max_videos_per_prompt": 10,
|
|
24651
|
+
"mode": "image_generation",
|
|
24652
|
+
"output_cost_per_image": 0.039,
|
|
24653
|
+
"output_cost_per_reasoning_token": 2.5e-06,
|
|
24654
|
+
"output_cost_per_token": 2.5e-06,
|
|
24655
|
+
"rpm": 100000,
|
|
24656
|
+
"source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-flash-image",
|
|
24657
|
+
"supported_endpoints": [
|
|
24658
|
+
"/v1/chat/completions",
|
|
24659
|
+
"/v1/completions",
|
|
24660
|
+
"/v1/batch"
|
|
24661
|
+
],
|
|
24662
|
+
"supported_modalities": [
|
|
24663
|
+
"text",
|
|
24664
|
+
"image",
|
|
24665
|
+
"audio",
|
|
24666
|
+
"video"
|
|
24667
|
+
],
|
|
24668
|
+
"supported_output_modalities": [
|
|
24669
|
+
"text",
|
|
24670
|
+
"image"
|
|
24671
|
+
],
|
|
24672
|
+
"supports_audio_output": false,
|
|
24673
|
+
"supports_function_calling": true,
|
|
24674
|
+
"supports_parallel_function_calling": true,
|
|
24675
|
+
"supports_pdf_input": true,
|
|
24676
|
+
"supports_prompt_caching": true,
|
|
24677
|
+
"supports_response_schema": true,
|
|
24678
|
+
"supports_system_messages": true,
|
|
24679
|
+
"supports_tool_choice": true,
|
|
24680
|
+
"supports_url_context": true,
|
|
24681
|
+
"supports_vision": true,
|
|
24682
|
+
"supports_web_search": false,
|
|
24683
|
+
"tpm": 8000000
|
|
24684
|
+
},
|
|
24685
|
+
"gemini/gemini-2.5-flash-image": {
|
|
24686
|
+
"cache_read_input_token_cost": 3e-08,
|
|
24687
|
+
"input_cost_per_audio_token": 1e-06,
|
|
24688
|
+
"input_cost_per_token": 3e-07,
|
|
24689
|
+
"litellm_provider": "vertex_ai-language-models",
|
|
24690
|
+
"max_audio_length_hours": 8.4,
|
|
24691
|
+
"max_audio_per_prompt": 1,
|
|
24692
|
+
"max_images_per_prompt": 3000,
|
|
24693
|
+
"max_input_tokens": 32768,
|
|
24694
|
+
"max_output_tokens": 32768,
|
|
24695
|
+
"max_tokens": 32768,
|
|
24696
|
+
"max_pdf_size_mb": 30,
|
|
24697
|
+
"max_video_length": 1,
|
|
24698
|
+
"max_videos_per_prompt": 10,
|
|
24699
|
+
"mode": "image_generation",
|
|
24700
|
+
"output_cost_per_image": 0.039,
|
|
24701
|
+
"output_cost_per_reasoning_token": 2.5e-06,
|
|
24702
|
+
"output_cost_per_token": 2.5e-06,
|
|
24703
|
+
"rpm": 100000,
|
|
24704
|
+
"source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-flash-image",
|
|
24705
|
+
"supported_endpoints": [
|
|
24706
|
+
"/v1/chat/completions",
|
|
24707
|
+
"/v1/completions",
|
|
24708
|
+
"/v1/batch"
|
|
24709
|
+
],
|
|
24710
|
+
"supported_modalities": [
|
|
24711
|
+
"text",
|
|
24712
|
+
"image",
|
|
24713
|
+
"audio",
|
|
24714
|
+
"video"
|
|
24715
|
+
],
|
|
24716
|
+
"supported_output_modalities": [
|
|
24717
|
+
"text",
|
|
24718
|
+
"image"
|
|
24719
|
+
],
|
|
24720
|
+
"supports_audio_output": false,
|
|
24721
|
+
"supports_function_calling": true,
|
|
24722
|
+
"supports_parallel_function_calling": true,
|
|
24723
|
+
"supports_pdf_input": true,
|
|
24724
|
+
"supports_prompt_caching": true,
|
|
24725
|
+
"supports_response_schema": true,
|
|
24726
|
+
"supports_system_messages": true,
|
|
24727
|
+
"supports_tool_choice": true,
|
|
24728
|
+
"supports_url_context": true,
|
|
24729
|
+
"supports_vision": true,
|
|
24730
|
+
"supports_web_search": true,
|
|
24731
|
+
"tpm": 8000000
|
|
24732
|
+
},
|
|
24733
|
+
"lemonade/gpt-oss-20b-mxfp4-GGUF": {
|
|
24734
|
+
"input_cost_per_token": 0,
|
|
24735
|
+
"litellm_provider": "lemonade",
|
|
24736
|
+
"max_tokens": 131072,
|
|
24737
|
+
"max_input_tokens": 131072,
|
|
24738
|
+
"max_output_tokens": 32768,
|
|
24739
|
+
"mode": "chat",
|
|
24740
|
+
"output_cost_per_token": 0,
|
|
24741
|
+
"supports_function_calling": true,
|
|
24742
|
+
"supports_response_schema": true,
|
|
24743
|
+
"supports_tool_choice": true
|
|
24744
|
+
},
|
|
24745
|
+
"lemonade/gpt-oss-120b-mxfp-GGUF": {
|
|
24746
|
+
"input_cost_per_token": 0,
|
|
24747
|
+
"litellm_provider": "lemonade",
|
|
24748
|
+
"max_tokens": 131072,
|
|
24749
|
+
"max_input_tokens": 131072,
|
|
24750
|
+
"max_output_tokens": 32768,
|
|
24751
|
+
"mode": "chat",
|
|
24752
|
+
"output_cost_per_token": 0,
|
|
24753
|
+
"supports_function_calling": true,
|
|
24754
|
+
"supports_response_schema": true,
|
|
24755
|
+
"supports_tool_choice": true
|
|
24756
|
+
},
|
|
24757
|
+
"lemonade/Gemma-3-4b-it-GGUF": {
|
|
24758
|
+
"input_cost_per_token": 0,
|
|
24759
|
+
"litellm_provider": "lemonade",
|
|
24760
|
+
"max_tokens": 128000,
|
|
24761
|
+
"max_input_tokens": 128000,
|
|
24762
|
+
"max_output_tokens": 8192,
|
|
24763
|
+
"mode": "chat",
|
|
24764
|
+
"output_cost_per_token": 0,
|
|
24765
|
+
"supports_function_calling": true,
|
|
24766
|
+
"supports_response_schema": true,
|
|
24767
|
+
"supports_tool_choice": true
|
|
24768
|
+
},
|
|
24769
|
+
"lemonade/Qwen3-4B-Instruct-2507-GGUF": {
|
|
24770
|
+
"input_cost_per_token": 0,
|
|
24771
|
+
"litellm_provider": "lemonade",
|
|
24772
|
+
"max_tokens": 262144,
|
|
24773
|
+
"max_input_tokens": 262144,
|
|
24774
|
+
"max_output_tokens": 32768,
|
|
24775
|
+
"mode": "chat",
|
|
24776
|
+
"output_cost_per_token": 0,
|
|
24777
|
+
"supports_function_calling": true,
|
|
24778
|
+
"supports_response_schema": true,
|
|
24779
|
+
"supports_tool_choice": true
|
|
24780
|
+
},
|
|
24781
|
+
"jp.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
24782
|
+
"cache_creation_input_token_cost": 1.375e-06,
|
|
24783
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
24784
|
+
"input_cost_per_token": 1.1e-06,
|
|
24785
|
+
"litellm_provider": "bedrock",
|
|
24786
|
+
"max_input_tokens": 200000,
|
|
24787
|
+
"max_output_tokens": 8192,
|
|
24788
|
+
"max_tokens": 8192,
|
|
24789
|
+
"mode": "chat",
|
|
24790
|
+
"output_cost_per_token": 5.5e-06,
|
|
24791
|
+
"source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
|
|
24792
|
+
"supports_assistant_prefill": true,
|
|
24793
|
+
"supports_function_calling": true,
|
|
24794
|
+
"supports_pdf_input": true,
|
|
24795
|
+
"supports_prompt_caching": true,
|
|
24796
|
+
"supports_reasoning": true,
|
|
24797
|
+
"supports_response_schema": true,
|
|
24798
|
+
"supports_tool_choice": true
|
|
24799
|
+
},
|
|
24800
|
+
"us.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
24801
|
+
"cache_creation_input_token_cost": 1.375e-06,
|
|
24802
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
24803
|
+
"input_cost_per_token": 1.1e-06,
|
|
24804
|
+
"litellm_provider": "bedrock",
|
|
24805
|
+
"max_input_tokens": 200000,
|
|
24806
|
+
"max_output_tokens": 8192,
|
|
24807
|
+
"max_tokens": 8192,
|
|
24808
|
+
"mode": "chat",
|
|
24809
|
+
"output_cost_per_token": 5.5e-06,
|
|
24810
|
+
"source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
|
|
24811
|
+
"supports_assistant_prefill": true,
|
|
24812
|
+
"supports_function_calling": true,
|
|
24813
|
+
"supports_pdf_input": true,
|
|
24814
|
+
"supports_prompt_caching": true,
|
|
24815
|
+
"supports_reasoning": true,
|
|
24816
|
+
"supports_response_schema": true,
|
|
24817
|
+
"supports_tool_choice": true
|
|
24818
|
+
},
|
|
24819
|
+
"vertex_ai/claude-haiku-4-5@20251001": {
|
|
24820
|
+
"cache_creation_input_token_cost": 1.25e-06,
|
|
24821
|
+
"cache_read_input_token_cost": 1e-07,
|
|
24822
|
+
"input_cost_per_token": 1e-06,
|
|
24823
|
+
"litellm_provider": "vertex_ai-anthropic_models",
|
|
24824
|
+
"max_input_tokens": 200000,
|
|
24825
|
+
"max_output_tokens": 8192,
|
|
24826
|
+
"max_tokens": 8192,
|
|
24827
|
+
"mode": "chat",
|
|
24828
|
+
"output_cost_per_token": 5e-06,
|
|
24829
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/partner-models/claude/haiku-4-5",
|
|
24830
|
+
"supports_assistant_prefill": true,
|
|
24831
|
+
"supports_function_calling": true,
|
|
24832
|
+
"supports_pdf_input": true,
|
|
24833
|
+
"supports_prompt_caching": true,
|
|
24834
|
+
"supports_reasoning": true,
|
|
24835
|
+
"supports_response_schema": true,
|
|
24836
|
+
"supports_tool_choice": true
|
|
24837
|
+
},
|
|
24838
|
+
"azure_ai/mistral-document-ai-2505": {
|
|
24839
|
+
"litellm_provider": "azure_ai",
|
|
24840
|
+
"ocr_cost_per_page": 0.003,
|
|
24841
|
+
"mode": "ocr",
|
|
24842
|
+
"supported_endpoints": [
|
|
24843
|
+
"/v1/ocr"
|
|
24844
|
+
],
|
|
24845
|
+
"source": "https://devblogs.microsoft.com/foundry/whats-new-in-azure-ai-foundry-august-2025/#mistral-document-ai-(ocr)-%E2%80%94-serverless-in-foundry"
|
|
24846
|
+
},
|
|
24847
|
+
"mistral/mistral-ocr-latest": {
|
|
24848
|
+
"litellm_provider": "mistral",
|
|
24849
|
+
"ocr_cost_per_page": 0.001,
|
|
24850
|
+
"annotation_cost_per_page": 0.003,
|
|
24851
|
+
"mode": "ocr",
|
|
24852
|
+
"supported_endpoints": [
|
|
24853
|
+
"/v1/ocr"
|
|
24854
|
+
],
|
|
24855
|
+
"source": "https://mistral.ai/pricing#api-pricing"
|
|
24856
|
+
},
|
|
24857
|
+
"mistral/mistral-ocr-2505-completion": {
|
|
24858
|
+
"litellm_provider": "mistral",
|
|
24859
|
+
"ocr_cost_per_page": 0.001,
|
|
24860
|
+
"annotation_cost_per_page": 0.003,
|
|
24861
|
+
"mode": "ocr",
|
|
24862
|
+
"supported_endpoints": [
|
|
24863
|
+
"/v1/ocr"
|
|
24864
|
+
],
|
|
24865
|
+
"source": "https://mistral.ai/pricing#api-pricing"
|
|
24866
|
+
},
|
|
24867
|
+
"watsonx/mistralai/mistral-small-3-1-24b-instruct-2503": {
|
|
24868
|
+
"max_tokens": 32000,
|
|
24869
|
+
"max_input_tokens": 32000,
|
|
24870
|
+
"max_output_tokens": 32000,
|
|
24871
|
+
"input_cost_per_token": 1e-07,
|
|
24872
|
+
"output_cost_per_token": 3e-07,
|
|
24873
|
+
"litellm_provider": "watsonx",
|
|
24874
|
+
"mode": "chat",
|
|
24875
|
+
"supports_function_calling": true,
|
|
24876
|
+
"supports_parallel_function_calling": true,
|
|
24877
|
+
"supports_vision": false
|
|
24878
|
+
},
|
|
24879
|
+
"global.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
24880
|
+
"cache_creation_input_token_cost": 1.375e-06,
|
|
24881
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
24882
|
+
"input_cost_per_token": 1.1e-06,
|
|
24883
|
+
"litellm_provider": "bedrock_converse",
|
|
24884
|
+
"max_input_tokens": 200000,
|
|
24885
|
+
"max_output_tokens": 8192,
|
|
24886
|
+
"max_tokens": 8192,
|
|
24887
|
+
"mode": "chat",
|
|
24888
|
+
"output_cost_per_token": 5.5e-06,
|
|
24889
|
+
"source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
|
|
24890
|
+
"supports_assistant_prefill": true,
|
|
24891
|
+
"supports_function_calling": true,
|
|
24892
|
+
"supports_pdf_input": true,
|
|
24893
|
+
"supports_prompt_caching": true,
|
|
24894
|
+
"supports_reasoning": true,
|
|
24895
|
+
"supports_response_schema": true,
|
|
24896
|
+
"supports_tool_choice": true,
|
|
24897
|
+
"supports_vision": true,
|
|
24898
|
+
"tool_use_system_prompt_tokens": 346
|
|
24899
|
+
},
|
|
24900
|
+
"au.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
24901
|
+
"cache_creation_input_token_cost": 1.375e-06,
|
|
24902
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
24903
|
+
"input_cost_per_token": 1.1e-06,
|
|
24904
|
+
"litellm_provider": "bedrock_converse",
|
|
24905
|
+
"max_input_tokens": 200000,
|
|
24906
|
+
"max_output_tokens": 8192,
|
|
24907
|
+
"max_tokens": 8192,
|
|
24908
|
+
"mode": "chat",
|
|
24909
|
+
"output_cost_per_token": 5.5e-06,
|
|
24910
|
+
"supports_assistant_prefill": true,
|
|
24911
|
+
"supports_function_calling": true,
|
|
24912
|
+
"supports_pdf_input": true,
|
|
24913
|
+
"supports_prompt_caching": true,
|
|
24914
|
+
"supports_response_schema": true,
|
|
24915
|
+
"supports_tool_choice": true,
|
|
24916
|
+
"supports_vision": true,
|
|
24917
|
+
"tool_use_system_prompt_tokens": 346
|
|
24918
|
+
},
|
|
24919
|
+
"vercel_ai_gateway/glm-4.6": {
|
|
24920
|
+
"litellm_provider": "vercel_ai_gateway",
|
|
24921
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
24922
|
+
"input_cost_per_token": 6e-07,
|
|
24923
|
+
"max_input_tokens": 200000,
|
|
24924
|
+
"max_output_tokens": 200000,
|
|
24925
|
+
"max_tokens": 200000,
|
|
24926
|
+
"mode": "chat",
|
|
24927
|
+
"output_cost_per_token": 2.2e-06
|
|
24928
|
+
},
|
|
24929
|
+
"azure/speech/azure-tts": {
|
|
24930
|
+
"input_cost_per_character": 1.5e-05,
|
|
24931
|
+
"litellm_provider": "azure",
|
|
24932
|
+
"mode": "audio_speech",
|
|
24933
|
+
"source": "https://azure.microsoft.com/en-us/pricing/calculator/"
|
|
24934
|
+
},
|
|
24935
|
+
"azure/speech/azure-tts-hd": {
|
|
24936
|
+
"input_cost_per_character": 3e-05,
|
|
24937
|
+
"litellm_provider": "azure",
|
|
24938
|
+
"mode": "audio_speech",
|
|
24939
|
+
"source": "https://azure.microsoft.com/en-us/pricing/calculator/"
|
|
24940
|
+
},
|
|
24941
|
+
"bedrock/us-gov-west-1/anthropic.claude-3-7-sonnet-20240620-v1:0": {
|
|
24942
|
+
"cache_creation_input_token_cost": 4.5e-06,
|
|
24943
|
+
"cache_read_input_token_cost": 3.6e-07,
|
|
24944
|
+
"input_cost_per_token": 3.6e-06,
|
|
24945
|
+
"litellm_provider": "bedrock",
|
|
24946
|
+
"max_input_tokens": 200000,
|
|
24947
|
+
"max_output_tokens": 8192,
|
|
24948
|
+
"max_tokens": 8192,
|
|
24949
|
+
"mode": "chat",
|
|
24950
|
+
"output_cost_per_token": 1.8e-05,
|
|
24951
|
+
"supports_assistant_prefill": true,
|
|
24952
|
+
"supports_computer_use": true,
|
|
24953
|
+
"supports_function_calling": true,
|
|
24954
|
+
"supports_pdf_input": true,
|
|
24955
|
+
"supports_prompt_caching": true,
|
|
24956
|
+
"supports_reasoning": true,
|
|
24957
|
+
"supports_response_schema": true,
|
|
24958
|
+
"supports_tool_choice": true,
|
|
24959
|
+
"supports_vision": true
|
|
24960
|
+
},
|
|
24961
|
+
"anthropic.claude-3-7-sonnet-20240620-v1:0": {
|
|
24962
|
+
"cache_creation_input_token_cost": 4.5e-06,
|
|
24963
|
+
"cache_read_input_token_cost": 3.6e-07,
|
|
24964
|
+
"input_cost_per_token": 3.6e-06,
|
|
24965
|
+
"litellm_provider": "bedrock",
|
|
24966
|
+
"max_input_tokens": 200000,
|
|
24967
|
+
"max_output_tokens": 8192,
|
|
24968
|
+
"max_tokens": 8192,
|
|
24969
|
+
"mode": "chat",
|
|
24970
|
+
"output_cost_per_token": 1.8e-05,
|
|
24971
|
+
"supports_assistant_prefill": true,
|
|
24972
|
+
"supports_computer_use": true,
|
|
24973
|
+
"supports_function_calling": true,
|
|
24974
|
+
"supports_pdf_input": true,
|
|
24975
|
+
"supports_prompt_caching": true,
|
|
24976
|
+
"supports_reasoning": true,
|
|
24977
|
+
"supports_response_schema": true,
|
|
24978
|
+
"supports_tool_choice": true,
|
|
24979
|
+
"supports_vision": true
|
|
24980
|
+
},
|
|
24981
|
+
"bedrock/us-gov-west-1/anthropic.claude-3-7-sonnet-20250219-v1:0": {
|
|
24982
|
+
"cache_creation_input_token_cost": 4.5e-06,
|
|
24983
|
+
"cache_read_input_token_cost": 3.6e-07,
|
|
24984
|
+
"input_cost_per_token": 3.6e-06,
|
|
24985
|
+
"litellm_provider": "bedrock",
|
|
24986
|
+
"max_input_tokens": 200000,
|
|
24987
|
+
"max_output_tokens": 8192,
|
|
24988
|
+
"max_tokens": 8192,
|
|
24989
|
+
"mode": "chat",
|
|
24990
|
+
"output_cost_per_token": 1.8e-05,
|
|
24991
|
+
"supports_assistant_prefill": true,
|
|
24992
|
+
"supports_computer_use": true,
|
|
24993
|
+
"supports_function_calling": true,
|
|
24994
|
+
"supports_pdf_input": true,
|
|
24995
|
+
"supports_prompt_caching": true,
|
|
24996
|
+
"supports_reasoning": true,
|
|
24997
|
+
"supports_response_schema": true,
|
|
24998
|
+
"supports_tool_choice": true,
|
|
24999
|
+
"supports_vision": true
|
|
25000
|
+
},
|
|
25001
|
+
"dataforseo/search": {
|
|
25002
|
+
"input_cost_per_query": 0.003,
|
|
25003
|
+
"litellm_provider": "dataforseo",
|
|
25004
|
+
"mode": "search"
|
|
25005
|
+
},
|
|
25006
|
+
"exa_ai/search": {
|
|
25007
|
+
"litellm_provider": "exa_ai",
|
|
25008
|
+
"mode": "search",
|
|
25009
|
+
"tiered_pricing": [
|
|
25010
|
+
{
|
|
25011
|
+
"input_cost_per_query": 0.005,
|
|
25012
|
+
"max_results_range": [
|
|
25013
|
+
0,
|
|
25014
|
+
25
|
|
25015
|
+
]
|
|
25016
|
+
},
|
|
25017
|
+
{
|
|
25018
|
+
"input_cost_per_query": 0.025,
|
|
25019
|
+
"max_results_range": [
|
|
25020
|
+
26,
|
|
25021
|
+
100
|
|
25022
|
+
]
|
|
25023
|
+
}
|
|
25024
|
+
]
|
|
25025
|
+
},
|
|
25026
|
+
"perplexity/search": {
|
|
25027
|
+
"input_cost_per_query": 0.005,
|
|
25028
|
+
"litellm_provider": "perplexity",
|
|
25029
|
+
"mode": "search"
|
|
25030
|
+
},
|
|
25031
|
+
"google_pse/search": {
|
|
25032
|
+
"input_cost_per_query": 0.005,
|
|
25033
|
+
"litellm_provider": "google_pse",
|
|
25034
|
+
"mode": "search"
|
|
25035
|
+
},
|
|
25036
|
+
"parallel_ai/search": {
|
|
25037
|
+
"input_cost_per_query": 0.004,
|
|
25038
|
+
"litellm_provider": "parallel_ai",
|
|
25039
|
+
"mode": "search"
|
|
25040
|
+
},
|
|
25041
|
+
"parallel_ai/search-pro": {
|
|
25042
|
+
"input_cost_per_query": 0.009,
|
|
25043
|
+
"litellm_provider": "parallel_ai",
|
|
25044
|
+
"mode": "search"
|
|
25045
|
+
},
|
|
25046
|
+
"tavily/search": {
|
|
25047
|
+
"input_cost_per_query": 0.008,
|
|
25048
|
+
"litellm_provider": "tavily",
|
|
25049
|
+
"mode": "search"
|
|
25050
|
+
},
|
|
25051
|
+
"tavily/search-advanced": {
|
|
25052
|
+
"input_cost_per_query": 0.016,
|
|
25053
|
+
"litellm_provider": "tavily",
|
|
25054
|
+
"mode": "search"
|
|
25055
|
+
},
|
|
25056
|
+
"vertex_ai/mistralai/codestral-2@001": {
|
|
25057
|
+
"input_cost_per_token": 3e-07,
|
|
25058
|
+
"litellm_provider": "vertex_ai-mistral_models",
|
|
25059
|
+
"max_input_tokens": 128000,
|
|
25060
|
+
"max_output_tokens": 128000,
|
|
25061
|
+
"max_tokens": 128000,
|
|
25062
|
+
"mode": "chat",
|
|
25063
|
+
"output_cost_per_token": 9e-07,
|
|
25064
|
+
"supports_function_calling": true,
|
|
25065
|
+
"supports_tool_choice": true
|
|
25066
|
+
},
|
|
25067
|
+
"vertex_ai/codestral-2": {
|
|
25068
|
+
"input_cost_per_token": 3e-07,
|
|
25069
|
+
"litellm_provider": "vertex_ai-mistral_models",
|
|
25070
|
+
"max_input_tokens": 128000,
|
|
25071
|
+
"max_output_tokens": 128000,
|
|
25072
|
+
"max_tokens": 128000,
|
|
25073
|
+
"mode": "chat",
|
|
25074
|
+
"output_cost_per_token": 9e-07,
|
|
25075
|
+
"supports_function_calling": true,
|
|
25076
|
+
"supports_tool_choice": true
|
|
25077
|
+
},
|
|
25078
|
+
"vertex_ai/codestral-2@001": {
|
|
25079
|
+
"input_cost_per_token": 3e-07,
|
|
25080
|
+
"litellm_provider": "vertex_ai-mistral_models",
|
|
25081
|
+
"max_input_tokens": 128000,
|
|
25082
|
+
"max_output_tokens": 128000,
|
|
25083
|
+
"max_tokens": 128000,
|
|
25084
|
+
"mode": "chat",
|
|
25085
|
+
"output_cost_per_token": 9e-07,
|
|
25086
|
+
"supports_function_calling": true,
|
|
25087
|
+
"supports_tool_choice": true
|
|
25088
|
+
},
|
|
25089
|
+
"vertex_ai/mistralai/codestral-2": {
|
|
25090
|
+
"input_cost_per_token": 3e-07,
|
|
25091
|
+
"litellm_provider": "vertex_ai-mistral_models",
|
|
25092
|
+
"max_input_tokens": 128000,
|
|
25093
|
+
"max_output_tokens": 128000,
|
|
25094
|
+
"max_tokens": 128000,
|
|
25095
|
+
"mode": "chat",
|
|
25096
|
+
"output_cost_per_token": 9e-07,
|
|
25097
|
+
"supports_function_calling": true,
|
|
25098
|
+
"supports_tool_choice": true
|
|
25099
|
+
},
|
|
25100
|
+
"vertex_ai/mistral-medium-3": {
|
|
25101
|
+
"input_cost_per_token": 4e-07,
|
|
25102
|
+
"litellm_provider": "vertex_ai-mistral_models",
|
|
25103
|
+
"max_input_tokens": 128000,
|
|
25104
|
+
"max_output_tokens": 8191,
|
|
25105
|
+
"max_tokens": 8191,
|
|
25106
|
+
"mode": "chat",
|
|
25107
|
+
"output_cost_per_token": 2e-06,
|
|
25108
|
+
"supports_function_calling": true,
|
|
25109
|
+
"supports_tool_choice": true
|
|
25110
|
+
},
|
|
25111
|
+
"vertex_ai/mistral-medium-3@001": {
|
|
25112
|
+
"input_cost_per_token": 4e-07,
|
|
25113
|
+
"litellm_provider": "vertex_ai-mistral_models",
|
|
25114
|
+
"max_input_tokens": 128000,
|
|
25115
|
+
"max_output_tokens": 8191,
|
|
25116
|
+
"max_tokens": 8191,
|
|
25117
|
+
"mode": "chat",
|
|
25118
|
+
"output_cost_per_token": 2e-06,
|
|
25119
|
+
"supports_function_calling": true,
|
|
25120
|
+
"supports_tool_choice": true
|
|
25121
|
+
},
|
|
25122
|
+
"vertex_ai/mistralai/mistral-medium-3": {
|
|
25123
|
+
"input_cost_per_token": 4e-07,
|
|
25124
|
+
"litellm_provider": "vertex_ai-mistral_models",
|
|
25125
|
+
"max_input_tokens": 128000,
|
|
25126
|
+
"max_output_tokens": 8191,
|
|
25127
|
+
"max_tokens": 8191,
|
|
25128
|
+
"mode": "chat",
|
|
25129
|
+
"output_cost_per_token": 2e-06,
|
|
25130
|
+
"supports_function_calling": true,
|
|
25131
|
+
"supports_tool_choice": true
|
|
25132
|
+
},
|
|
25133
|
+
"vertex_ai/mistralai/mistral-medium-3@001": {
|
|
25134
|
+
"input_cost_per_token": 4e-07,
|
|
25135
|
+
"litellm_provider": "vertex_ai-mistral_models",
|
|
25136
|
+
"max_input_tokens": 128000,
|
|
25137
|
+
"max_output_tokens": 8191,
|
|
25138
|
+
"max_tokens": 8191,
|
|
25139
|
+
"mode": "chat",
|
|
25140
|
+
"output_cost_per_token": 2e-06,
|
|
25141
|
+
"supports_function_calling": true,
|
|
25142
|
+
"supports_tool_choice": true
|
|
25143
|
+
},
|
|
25144
|
+
"openai/sora-2": {
|
|
25145
|
+
"litellm_provider": "openai",
|
|
25146
|
+
"mode": "video_generation",
|
|
25147
|
+
"output_cost_per_video_per_second": 0.1,
|
|
25148
|
+
"source": "https://platform.openai.com/docs/api-reference/videos",
|
|
25149
|
+
"supported_modalities": [
|
|
25150
|
+
"text",
|
|
25151
|
+
"image"
|
|
25152
|
+
],
|
|
25153
|
+
"supported_output_modalities": [
|
|
25154
|
+
"video"
|
|
25155
|
+
],
|
|
25156
|
+
"supported_resolutions": [
|
|
25157
|
+
"720x1280",
|
|
25158
|
+
"1280x720"
|
|
25159
|
+
]
|
|
25160
|
+
},
|
|
25161
|
+
"azure/sora-2": {
|
|
25162
|
+
"litellm_provider": "azure",
|
|
25163
|
+
"mode": "video_generation",
|
|
25164
|
+
"output_cost_per_video_per_second": 0.1,
|
|
25165
|
+
"source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation",
|
|
25166
|
+
"supported_modalities": [
|
|
25167
|
+
"text"
|
|
25168
|
+
],
|
|
25169
|
+
"supported_output_modalities": [
|
|
25170
|
+
"video"
|
|
25171
|
+
],
|
|
25172
|
+
"supported_resolutions": [
|
|
25173
|
+
"720x1280",
|
|
25174
|
+
"1280x720"
|
|
25175
|
+
]
|
|
25176
|
+
},
|
|
25177
|
+
"azure/sora-2-pro": {
|
|
25178
|
+
"litellm_provider": "azure",
|
|
25179
|
+
"mode": "video_generation",
|
|
25180
|
+
"output_cost_per_video_per_second": 0.3,
|
|
25181
|
+
"source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation",
|
|
25182
|
+
"supported_modalities": [
|
|
25183
|
+
"text"
|
|
25184
|
+
],
|
|
25185
|
+
"supported_output_modalities": [
|
|
25186
|
+
"video"
|
|
25187
|
+
],
|
|
25188
|
+
"supported_resolutions": [
|
|
25189
|
+
"720x1280",
|
|
25190
|
+
"1280x720"
|
|
25191
|
+
]
|
|
25192
|
+
},
|
|
25193
|
+
"azure/sora-2-pro-high-res": {
|
|
25194
|
+
"litellm_provider": "azure",
|
|
25195
|
+
"mode": "video_generation",
|
|
25196
|
+
"output_cost_per_video_per_second": 0.5,
|
|
25197
|
+
"source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation",
|
|
25198
|
+
"supported_modalities": [
|
|
25199
|
+
"text"
|
|
25200
|
+
],
|
|
25201
|
+
"supported_output_modalities": [
|
|
25202
|
+
"video"
|
|
25203
|
+
],
|
|
25204
|
+
"supported_resolutions": [
|
|
25205
|
+
"1024x1792",
|
|
25206
|
+
"1792x1024"
|
|
25207
|
+
]
|
|
25208
|
+
},
|
|
25209
|
+
"amazon.titan-image-generator-v1": {
|
|
25210
|
+
"input_cost_per_image": 0.0,
|
|
25211
|
+
"output_cost_per_image": 0.008,
|
|
25212
|
+
"output_cost_per_image_premium_image": 0.01,
|
|
25213
|
+
"output_cost_per_image_above_512_and_512_pixels": 0.01,
|
|
25214
|
+
"output_cost_per_image_above_512_and_512_pixels_and_premium_image": 0.012,
|
|
25215
|
+
"litellm_provider": "bedrock",
|
|
25216
|
+
"mode": "image_generation"
|
|
25217
|
+
},
|
|
25218
|
+
"amazon.titan-image-generator-v2": {
|
|
25219
|
+
"input_cost_per_image": 0.0,
|
|
25220
|
+
"output_cost_per_image": 0.008,
|
|
25221
|
+
"output_cost_per_image_premium_image": 0.01,
|
|
25222
|
+
"output_cost_per_image_above_1024_and_1024_pixels": 0.01,
|
|
25223
|
+
"output_cost_per_image_above_1024_and_1024_pixels_and_premium_image": 0.012,
|
|
25224
|
+
"litellm_provider": "bedrock",
|
|
25225
|
+
"mode": "image_generation"
|
|
25226
|
+
},
|
|
25227
|
+
"vertex_ai/search_api": {
|
|
25228
|
+
"input_cost_per_query": 0.0015,
|
|
25229
|
+
"litellm_provider": "vertex_ai",
|
|
25230
|
+
"mode": "vector_store"
|
|
25231
|
+
},
|
|
25232
|
+
"openai/sora-2-pro": {
|
|
25233
|
+
"litellm_provider": "openai",
|
|
25234
|
+
"mode": "video_generation",
|
|
25235
|
+
"output_cost_per_video_per_second": 0.3,
|
|
25236
|
+
"source": "https://platform.openai.com/docs/api-reference/videos",
|
|
25237
|
+
"supported_modalities": [
|
|
25238
|
+
"text",
|
|
25239
|
+
"image"
|
|
25240
|
+
],
|
|
25241
|
+
"supported_output_modalities": [
|
|
25242
|
+
"video"
|
|
25243
|
+
],
|
|
25244
|
+
"supported_resolutions": [
|
|
25245
|
+
"720x1280",
|
|
25246
|
+
"1280x720"
|
|
25247
|
+
]
|
|
25248
|
+
},
|
|
25249
|
+
"openrouter/anthropic/claude-haiku-4.5": {
|
|
25250
|
+
"cache_creation_input_token_cost": 1.25e-06,
|
|
25251
|
+
"cache_read_input_token_cost": 1e-07,
|
|
25252
|
+
"input_cost_per_token": 1e-06,
|
|
25253
|
+
"litellm_provider": "openrouter",
|
|
25254
|
+
"max_input_tokens": 200000,
|
|
25255
|
+
"max_output_tokens": 200000,
|
|
25256
|
+
"max_tokens": 200000,
|
|
25257
|
+
"mode": "chat",
|
|
25258
|
+
"output_cost_per_token": 5e-06,
|
|
25259
|
+
"supports_assistant_prefill": true,
|
|
25260
|
+
"supports_computer_use": true,
|
|
25261
|
+
"supports_function_calling": true,
|
|
25262
|
+
"supports_prompt_caching": true,
|
|
25263
|
+
"supports_reasoning": true,
|
|
25264
|
+
"supports_tool_choice": true,
|
|
25265
|
+
"supports_vision": true,
|
|
25266
|
+
"tool_use_system_prompt_tokens": 346
|
|
25267
|
+
},
|
|
25268
|
+
"fal_ai/bria/text-to-image/3.2": {
|
|
25269
|
+
"litellm_provider": "fal_ai",
|
|
25270
|
+
"mode": "image_generation",
|
|
25271
|
+
"output_cost_per_image": 0.0398,
|
|
25272
|
+
"supported_endpoints": [
|
|
25273
|
+
"/v1/images/generations"
|
|
25274
|
+
]
|
|
25275
|
+
},
|
|
25276
|
+
"fal_ai/fal-ai/flux-pro/v1.1-ultra": {
|
|
25277
|
+
"litellm_provider": "fal_ai",
|
|
25278
|
+
"mode": "image_generation",
|
|
25279
|
+
"output_cost_per_image": 0.0398,
|
|
25280
|
+
"supported_endpoints": [
|
|
25281
|
+
"/v1/images/generations"
|
|
25282
|
+
]
|
|
25283
|
+
},
|
|
25284
|
+
"fal_ai/fal-ai/imagen4/preview": {
|
|
25285
|
+
"litellm_provider": "fal_ai",
|
|
25286
|
+
"mode": "image_generation",
|
|
25287
|
+
"output_cost_per_image": 0.0398,
|
|
25288
|
+
"supported_endpoints": [
|
|
25289
|
+
"/v1/images/generations"
|
|
25290
|
+
]
|
|
25291
|
+
},
|
|
25292
|
+
"fal_ai/fal-ai/recraft/v3/text-to-image": {
|
|
25293
|
+
"litellm_provider": "fal_ai",
|
|
25294
|
+
"mode": "image_generation",
|
|
25295
|
+
"output_cost_per_image": 0.0398,
|
|
25296
|
+
"supported_endpoints": [
|
|
25297
|
+
"/v1/images/generations"
|
|
25298
|
+
]
|
|
25299
|
+
},
|
|
25300
|
+
"fal_ai/fal-ai/stable-diffusion-v35-medium": {
|
|
25301
|
+
"litellm_provider": "fal_ai",
|
|
25302
|
+
"mode": "image_generation",
|
|
25303
|
+
"output_cost_per_image": 0.0398,
|
|
25304
|
+
"supported_endpoints": [
|
|
25305
|
+
"/v1/images/generations"
|
|
25306
|
+
]
|
|
25307
|
+
},
|
|
25308
|
+
"mistral/codestral-embed": {
|
|
25309
|
+
"input_cost_per_token": 1.5e-07,
|
|
25310
|
+
"litellm_provider": "mistral",
|
|
25311
|
+
"max_input_tokens": 8192,
|
|
25312
|
+
"max_tokens": 8192,
|
|
25313
|
+
"mode": "embedding"
|
|
25314
|
+
},
|
|
25315
|
+
"mistral/codestral-embed-2505": {
|
|
25316
|
+
"input_cost_per_token": 1.5e-07,
|
|
25317
|
+
"litellm_provider": "mistral",
|
|
25318
|
+
"max_input_tokens": 8192,
|
|
25319
|
+
"max_tokens": 8192,
|
|
25320
|
+
"mode": "embedding"
|
|
25321
|
+
},
|
|
25322
|
+
"gemini/gemini-embedding-001": {
|
|
25323
|
+
"input_cost_per_token": 1.5e-07,
|
|
25324
|
+
"litellm_provider": "gemini",
|
|
25325
|
+
"max_input_tokens": 2048,
|
|
25326
|
+
"max_tokens": 2048,
|
|
25327
|
+
"mode": "embedding",
|
|
25328
|
+
"output_cost_per_token": 0,
|
|
25329
|
+
"output_vector_size": 3072,
|
|
25330
|
+
"rpm": 10000,
|
|
25331
|
+
"source": "https://ai.google.dev/gemini-api/docs/embeddings#model-versions",
|
|
25332
|
+
"tpm": 10000000
|
|
25333
|
+
},
|
|
25334
|
+
"vercel_ai_gateway/zai/glm-4.6": {
|
|
25335
|
+
"litellm_provider": "vercel_ai_gateway",
|
|
25336
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
25337
|
+
"input_cost_per_token": 4.5e-07,
|
|
25338
|
+
"max_input_tokens": 200000,
|
|
25339
|
+
"max_output_tokens": 200000,
|
|
25340
|
+
"max_tokens": 200000,
|
|
25341
|
+
"mode": "chat",
|
|
25342
|
+
"output_cost_per_token": 1.8e-06,
|
|
25343
|
+
"source": "https://vercel.com/ai-gateway/models/glm-4.6",
|
|
25344
|
+
"supports_function_calling": true,
|
|
25345
|
+
"supports_parallel_function_calling": true,
|
|
25346
|
+
"supports_tool_choice": true
|
|
25347
|
+
},
|
|
25348
|
+
"openrouter/nvidia/nemotron-nano-9b-v2:free": {
|
|
25349
|
+
"input_cost_per_token": 0,
|
|
25350
|
+
"litellm_provider": "openrouter",
|
|
25351
|
+
"max_tokens": 128000,
|
|
25352
|
+
"mode": "chat",
|
|
25353
|
+
"output_cost_per_token": 0,
|
|
25354
|
+
"source": "https://openrouter.ai/nvidia/nemotron-nano-9b-v2:free",
|
|
25355
|
+
"supports_tool_choice": true
|
|
25356
|
+
},
|
|
25357
|
+
"openai/container": {
|
|
25358
|
+
"code_interpreter_cost_per_session": 0.03,
|
|
25359
|
+
"litellm_provider": "openai",
|
|
25360
|
+
"mode": "chat"
|
|
25361
|
+
},
|
|
25362
|
+
"azure/container": {
|
|
25363
|
+
"code_interpreter_cost_per_session": 0.03,
|
|
25364
|
+
"litellm_provider": "azure",
|
|
25365
|
+
"mode": "chat"
|
|
25366
|
+
},
|
|
25367
|
+
"vertex_ai/mistral-ocr-2505": {
|
|
25368
|
+
"litellm_provider": "vertex_ai",
|
|
25369
|
+
"mode": "ocr",
|
|
25370
|
+
"ocr_cost_per_page": 0.0005,
|
|
25371
|
+
"supported_endpoints": [
|
|
25372
|
+
"/v1/ocr"
|
|
25373
|
+
],
|
|
25374
|
+
"source": "https://cloud.google.com/generative-ai-app-builder/pricing"
|
|
25375
|
+
},
|
|
25376
|
+
"azure_ai/doc-intelligence/prebuilt-read": {
|
|
25377
|
+
"litellm_provider": "azure_ai",
|
|
25378
|
+
"ocr_cost_per_page": 0.0015,
|
|
25379
|
+
"mode": "ocr",
|
|
25380
|
+
"supported_endpoints": [
|
|
25381
|
+
"/v1/ocr"
|
|
25382
|
+
],
|
|
25383
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/"
|
|
25384
|
+
},
|
|
25385
|
+
"azure_ai/doc-intelligence/prebuilt-layout": {
|
|
25386
|
+
"litellm_provider": "azure_ai",
|
|
25387
|
+
"ocr_cost_per_page": 0.01,
|
|
25388
|
+
"mode": "ocr",
|
|
25389
|
+
"supported_endpoints": [
|
|
25390
|
+
"/v1/ocr"
|
|
25391
|
+
],
|
|
25392
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/"
|
|
25393
|
+
},
|
|
25394
|
+
"azure_ai/doc-intelligence/prebuilt-document": {
|
|
25395
|
+
"litellm_provider": "azure_ai",
|
|
25396
|
+
"ocr_cost_per_page": 0.01,
|
|
25397
|
+
"mode": "ocr",
|
|
25398
|
+
"supported_endpoints": [
|
|
25399
|
+
"/v1/ocr"
|
|
25400
|
+
],
|
|
25401
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/"
|
|
25402
|
+
},
|
|
25403
|
+
"gemini-live-2.5-flash-preview-native-audio-09-2025": {
|
|
25404
|
+
"cache_read_input_token_cost": 7.5e-08,
|
|
25405
|
+
"input_cost_per_audio_token": 3e-06,
|
|
25406
|
+
"input_cost_per_token": 3e-07,
|
|
25407
|
+
"litellm_provider": "vertex_ai-language-models",
|
|
25408
|
+
"max_audio_length_hours": 8.4,
|
|
25409
|
+
"max_audio_per_prompt": 1,
|
|
25410
|
+
"max_images_per_prompt": 3000,
|
|
25411
|
+
"max_input_tokens": 1048576,
|
|
25412
|
+
"max_output_tokens": 65535,
|
|
25413
|
+
"max_pdf_size_mb": 30,
|
|
25414
|
+
"max_tokens": 65535,
|
|
25415
|
+
"max_video_length": 1,
|
|
25416
|
+
"max_videos_per_prompt": 10,
|
|
25417
|
+
"mode": "chat",
|
|
25418
|
+
"output_cost_per_audio_token": 1.2e-05,
|
|
25419
|
+
"output_cost_per_token": 2e-06,
|
|
25420
|
+
"source": "https://ai.google.dev/gemini-api/docs/pricing",
|
|
25421
|
+
"supported_endpoints": [
|
|
25422
|
+
"/v1/chat/completions",
|
|
25423
|
+
"/v1/completions"
|
|
25424
|
+
],
|
|
25425
|
+
"supported_modalities": [
|
|
25426
|
+
"text",
|
|
25427
|
+
"image",
|
|
25428
|
+
"audio",
|
|
25429
|
+
"video"
|
|
25430
|
+
],
|
|
25431
|
+
"supported_output_modalities": [
|
|
25432
|
+
"text",
|
|
25433
|
+
"audio"
|
|
25434
|
+
],
|
|
25435
|
+
"supports_audio_input": true,
|
|
25436
|
+
"supports_audio_output": true,
|
|
25437
|
+
"supports_function_calling": true,
|
|
25438
|
+
"supports_parallel_function_calling": true,
|
|
25439
|
+
"supports_pdf_input": true,
|
|
25440
|
+
"supports_prompt_caching": true,
|
|
25441
|
+
"supports_response_schema": true,
|
|
25442
|
+
"supports_system_messages": true,
|
|
25443
|
+
"supports_tool_choice": true,
|
|
25444
|
+
"supports_url_context": true,
|
|
25445
|
+
"supports_vision": true,
|
|
25446
|
+
"supports_web_search": true
|
|
25447
|
+
},
|
|
25448
|
+
"gemini/gemini-live-2.5-flash-preview-native-audio-09-2025": {
|
|
25449
|
+
"cache_read_input_token_cost": 7.5e-08,
|
|
25450
|
+
"input_cost_per_audio_token": 3e-06,
|
|
25451
|
+
"input_cost_per_token": 3e-07,
|
|
25452
|
+
"litellm_provider": "gemini",
|
|
25453
|
+
"max_audio_length_hours": 8.4,
|
|
25454
|
+
"max_audio_per_prompt": 1,
|
|
25455
|
+
"max_images_per_prompt": 3000,
|
|
25456
|
+
"max_input_tokens": 1048576,
|
|
25457
|
+
"max_output_tokens": 65535,
|
|
25458
|
+
"max_pdf_size_mb": 30,
|
|
25459
|
+
"max_tokens": 65535,
|
|
25460
|
+
"max_video_length": 1,
|
|
25461
|
+
"max_videos_per_prompt": 10,
|
|
25462
|
+
"mode": "chat",
|
|
25463
|
+
"output_cost_per_audio_token": 1.2e-05,
|
|
25464
|
+
"output_cost_per_token": 2e-06,
|
|
25465
|
+
"rpm": 100000,
|
|
25466
|
+
"source": "https://ai.google.dev/gemini-api/docs/pricing",
|
|
25467
|
+
"supported_endpoints": [
|
|
25468
|
+
"/v1/chat/completions",
|
|
25469
|
+
"/v1/completions"
|
|
25470
|
+
],
|
|
25471
|
+
"supported_modalities": [
|
|
25472
|
+
"text",
|
|
25473
|
+
"image",
|
|
25474
|
+
"audio",
|
|
25475
|
+
"video"
|
|
25476
|
+
],
|
|
25477
|
+
"supported_output_modalities": [
|
|
25478
|
+
"text",
|
|
25479
|
+
"audio"
|
|
25480
|
+
],
|
|
25481
|
+
"supports_audio_input": true,
|
|
25482
|
+
"supports_audio_output": true,
|
|
25483
|
+
"supports_function_calling": true,
|
|
25484
|
+
"supports_parallel_function_calling": true,
|
|
25485
|
+
"supports_pdf_input": true,
|
|
25486
|
+
"supports_prompt_caching": true,
|
|
25487
|
+
"supports_response_schema": true,
|
|
25488
|
+
"supports_system_messages": true,
|
|
25489
|
+
"supports_tool_choice": true,
|
|
25490
|
+
"supports_url_context": true,
|
|
25491
|
+
"supports_vision": true,
|
|
25492
|
+
"supports_web_search": true,
|
|
25493
|
+
"tpm": 8000000
|
|
25494
|
+
},
|
|
25495
|
+
"azure/gpt-image-1-mini": {
|
|
25496
|
+
"input_cost_per_pixel": 8.0566406e-09,
|
|
25497
|
+
"litellm_provider": "azure",
|
|
25498
|
+
"mode": "image_generation",
|
|
25499
|
+
"output_cost_per_pixel": 0.0,
|
|
25500
|
+
"supported_endpoints": [
|
|
25501
|
+
"/v1/images/generations"
|
|
25502
|
+
]
|
|
25503
|
+
},
|
|
25504
|
+
"azure/low/1024-x-1024/gpt-image-1-mini": {
|
|
25505
|
+
"input_cost_per_pixel": 2.0751953125e-09,
|
|
25506
|
+
"litellm_provider": "azure",
|
|
25507
|
+
"mode": "image_generation",
|
|
25508
|
+
"output_cost_per_pixel": 0.0,
|
|
25509
|
+
"supported_endpoints": [
|
|
25510
|
+
"/v1/images/generations"
|
|
25511
|
+
]
|
|
25512
|
+
},
|
|
25513
|
+
"azure/low/1024-x-1536/gpt-image-1-mini": {
|
|
25514
|
+
"input_cost_per_pixel": 2.0751953125e-09,
|
|
25515
|
+
"litellm_provider": "azure",
|
|
25516
|
+
"mode": "image_generation",
|
|
25517
|
+
"output_cost_per_pixel": 0.0,
|
|
25518
|
+
"supported_endpoints": [
|
|
25519
|
+
"/v1/images/generations"
|
|
25520
|
+
]
|
|
25521
|
+
},
|
|
25522
|
+
"azure/low/1536-x-1024/gpt-image-1-mini": {
|
|
25523
|
+
"input_cost_per_pixel": 2.0345052083e-09,
|
|
25524
|
+
"litellm_provider": "azure",
|
|
25525
|
+
"mode": "image_generation",
|
|
25526
|
+
"output_cost_per_pixel": 0.0,
|
|
25527
|
+
"supported_endpoints": [
|
|
25528
|
+
"/v1/images/generations"
|
|
25529
|
+
]
|
|
25530
|
+
},
|
|
25531
|
+
"azure/medium/1024-x-1024/gpt-image-1-mini": {
|
|
25532
|
+
"input_cost_per_pixel": 8.056640625e-09,
|
|
25533
|
+
"litellm_provider": "azure",
|
|
25534
|
+
"mode": "image_generation",
|
|
25535
|
+
"output_cost_per_pixel": 0.0,
|
|
25536
|
+
"supported_endpoints": [
|
|
25537
|
+
"/v1/images/generations"
|
|
25538
|
+
]
|
|
25539
|
+
},
|
|
25540
|
+
"azure/medium/1024-x-1536/gpt-image-1-mini": {
|
|
25541
|
+
"input_cost_per_pixel": 8.056640625e-09,
|
|
25542
|
+
"litellm_provider": "azure",
|
|
25543
|
+
"mode": "image_generation",
|
|
25544
|
+
"output_cost_per_pixel": 0.0,
|
|
25545
|
+
"supported_endpoints": [
|
|
25546
|
+
"/v1/images/generations"
|
|
25547
|
+
]
|
|
25548
|
+
},
|
|
25549
|
+
"azure/medium/1536-x-1024/gpt-image-1-mini": {
|
|
25550
|
+
"input_cost_per_pixel": 7.9752604167e-09,
|
|
25551
|
+
"litellm_provider": "azure",
|
|
25552
|
+
"mode": "image_generation",
|
|
25553
|
+
"output_cost_per_pixel": 0.0,
|
|
25554
|
+
"supported_endpoints": [
|
|
25555
|
+
"/v1/images/generations"
|
|
25556
|
+
]
|
|
25557
|
+
},
|
|
25558
|
+
"azure/high/1024-x-1024/gpt-image-1-mini": {
|
|
25559
|
+
"input_cost_per_pixel": 3.173828125e-08,
|
|
25560
|
+
"litellm_provider": "azure",
|
|
25561
|
+
"mode": "image_generation",
|
|
25562
|
+
"output_cost_per_pixel": 0.0,
|
|
25563
|
+
"supported_endpoints": [
|
|
25564
|
+
"/v1/images/generations"
|
|
25565
|
+
]
|
|
25566
|
+
},
|
|
25567
|
+
"azure/high/1024-x-1536/gpt-image-1-mini": {
|
|
25568
|
+
"input_cost_per_pixel": 3.173828125e-08,
|
|
25569
|
+
"litellm_provider": "azure",
|
|
25570
|
+
"mode": "image_generation",
|
|
25571
|
+
"output_cost_per_pixel": 0.0,
|
|
25572
|
+
"supported_endpoints": [
|
|
25573
|
+
"/v1/images/generations"
|
|
25574
|
+
]
|
|
25575
|
+
},
|
|
25576
|
+
"azure/high/1536-x-1024/gpt-image-1-mini": {
|
|
25577
|
+
"input_cost_per_pixel": 3.1575520833e-08,
|
|
25578
|
+
"litellm_provider": "azure",
|
|
25579
|
+
"mode": "image_generation",
|
|
25580
|
+
"output_cost_per_pixel": 0.0,
|
|
25581
|
+
"supported_endpoints": [
|
|
25582
|
+
"/v1/images/generations"
|
|
25583
|
+
]
|
|
25584
|
+
},
|
|
25585
|
+
"firecrawl/search": {
|
|
25586
|
+
"litellm_provider": "firecrawl",
|
|
25587
|
+
"mode": "search",
|
|
25588
|
+
"tiered_pricing": [
|
|
25589
|
+
{
|
|
25590
|
+
"input_cost_per_query": 0.00166,
|
|
25591
|
+
"max_results_range": [
|
|
25592
|
+
1,
|
|
25593
|
+
10
|
|
25594
|
+
]
|
|
25595
|
+
},
|
|
25596
|
+
{
|
|
25597
|
+
"input_cost_per_query": 0.00332,
|
|
25598
|
+
"max_results_range": [
|
|
25599
|
+
11,
|
|
25600
|
+
20
|
|
25601
|
+
]
|
|
25602
|
+
},
|
|
25603
|
+
{
|
|
25604
|
+
"input_cost_per_query": 0.00498,
|
|
25605
|
+
"max_results_range": [
|
|
25606
|
+
21,
|
|
25607
|
+
30
|
|
25608
|
+
]
|
|
25609
|
+
},
|
|
25610
|
+
{
|
|
25611
|
+
"input_cost_per_query": 0.00664,
|
|
25612
|
+
"max_results_range": [
|
|
25613
|
+
31,
|
|
25614
|
+
40
|
|
25615
|
+
]
|
|
25616
|
+
},
|
|
25617
|
+
{
|
|
25618
|
+
"input_cost_per_query": 0.0083,
|
|
25619
|
+
"max_results_range": [
|
|
25620
|
+
41,
|
|
25621
|
+
50
|
|
25622
|
+
]
|
|
25623
|
+
},
|
|
25624
|
+
{
|
|
25625
|
+
"input_cost_per_query": 0.00996,
|
|
25626
|
+
"max_results_range": [
|
|
25627
|
+
51,
|
|
25628
|
+
60
|
|
25629
|
+
]
|
|
25630
|
+
},
|
|
25631
|
+
{
|
|
25632
|
+
"input_cost_per_query": 0.01162,
|
|
25633
|
+
"max_results_range": [
|
|
25634
|
+
61,
|
|
25635
|
+
70
|
|
25636
|
+
]
|
|
25637
|
+
},
|
|
25638
|
+
{
|
|
25639
|
+
"input_cost_per_query": 0.01328,
|
|
25640
|
+
"max_results_range": [
|
|
25641
|
+
71,
|
|
25642
|
+
80
|
|
25643
|
+
]
|
|
25644
|
+
},
|
|
25645
|
+
{
|
|
25646
|
+
"input_cost_per_query": 0.01494,
|
|
25647
|
+
"max_results_range": [
|
|
25648
|
+
81,
|
|
25649
|
+
90
|
|
25650
|
+
]
|
|
25651
|
+
},
|
|
25652
|
+
{
|
|
25653
|
+
"input_cost_per_query": 0.0166,
|
|
25654
|
+
"max_results_range": [
|
|
25655
|
+
91,
|
|
25656
|
+
100
|
|
25657
|
+
]
|
|
25658
|
+
}
|
|
25659
|
+
],
|
|
25660
|
+
"metadata": {
|
|
25661
|
+
"notes": "Firecrawl search pricing: $83 for 100,000 credits, 2 credits per 10 results. Cost = ceiling(limit/10) * 2 * $0.00083"
|
|
25662
|
+
}
|
|
25663
|
+
},
|
|
25664
|
+
"searxng/search": {
|
|
25665
|
+
"litellm_provider": "searxng",
|
|
25666
|
+
"mode": "search",
|
|
25667
|
+
"input_cost_per_query": 0.0,
|
|
25668
|
+
"metadata": {
|
|
25669
|
+
"notes": "SearXNG is an open-source metasearch engine. Free to use when self-hosted or using public instances."
|
|
25670
|
+
}
|
|
25671
|
+
},
|
|
25672
|
+
"azure/gpt-5-pro": {
|
|
25673
|
+
"input_cost_per_token": 1.5e-05,
|
|
25674
|
+
"litellm_provider": "azure",
|
|
25675
|
+
"max_input_tokens": 272000,
|
|
25676
|
+
"max_output_tokens": 128000,
|
|
25677
|
+
"max_tokens": 400000,
|
|
25678
|
+
"mode": "responses",
|
|
25679
|
+
"output_cost_per_token": 0.00012,
|
|
25680
|
+
"source": "https://learn.microsoft.com/en-us/azure/ai-foundry/foundry-models/concepts/models-sold-directly-by-azure?pivots=azure-openai&tabs=global-standard-aoai%2Cstandard-chat-completions%2Cglobal-standard#gpt-5",
|
|
25681
|
+
"supported_endpoints": [
|
|
25682
|
+
"/v1/responses"
|
|
25683
|
+
],
|
|
25684
|
+
"supported_modalities": [
|
|
25685
|
+
"text",
|
|
25686
|
+
"image"
|
|
25687
|
+
],
|
|
25688
|
+
"supported_output_modalities": [
|
|
25689
|
+
"text"
|
|
25690
|
+
],
|
|
25691
|
+
"supports_function_calling": true,
|
|
25692
|
+
"supports_parallel_function_calling": true,
|
|
25693
|
+
"supports_pdf_input": true,
|
|
25694
|
+
"supports_prompt_caching": true,
|
|
25695
|
+
"supports_reasoning": true,
|
|
25696
|
+
"supports_response_schema": true,
|
|
25697
|
+
"supports_system_messages": true,
|
|
25698
|
+
"supports_tool_choice": true,
|
|
25699
|
+
"supports_vision": true
|
|
25700
|
+
},
|
|
25701
|
+
"vertex_ai/minimaxai/minimax-m2-maas": {
|
|
25702
|
+
"input_cost_per_token": 3e-07,
|
|
25703
|
+
"litellm_provider": "vertex_ai-minimax_models",
|
|
25704
|
+
"max_input_tokens": 196608,
|
|
25705
|
+
"max_output_tokens": 196608,
|
|
25706
|
+
"max_tokens": 196608,
|
|
25707
|
+
"mode": "chat",
|
|
25708
|
+
"output_cost_per_token": 1.2e-06,
|
|
25709
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
|
|
25710
|
+
"supports_function_calling": true,
|
|
25711
|
+
"supports_tool_choice": true
|
|
23548
25712
|
}
|
|
23549
25713
|
}
|