tokencostauto 0.1.264__py3-none-any.whl → 0.1.333__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -33,6 +33,7 @@
33
33
  "supports_response_schema": true,
34
34
  "supports_system_messages": true,
35
35
  "supports_tool_choice": true,
36
+ "supports_service_tier": true,
36
37
  "supports_vision": true
37
38
  },
38
39
  "gpt-4o-audio-preview": {
@@ -90,6 +91,7 @@
90
91
  "supports_response_schema": true,
91
92
  "supports_system_messages": true,
92
93
  "supports_tool_choice": true,
94
+ "supports_service_tier": true,
93
95
  "supports_vision": true
94
96
  },
95
97
  "gpt-4o-mini-2024-07-18": {
@@ -115,6 +117,7 @@
115
117
  "supports_response_schema": true,
116
118
  "supports_system_messages": true,
117
119
  "supports_tool_choice": true,
120
+ "supports_service_tier": true,
118
121
  "supports_vision": true
119
122
  },
120
123
  "o1-mini": {
@@ -131,6 +134,7 @@
131
134
  "supports_vision": true
132
135
  },
133
136
  "o1-mini-2024-09-12": {
137
+ "deprecation_date": "2025-10-27",
134
138
  "cache_read_input_token_cost": 1.5e-06,
135
139
  "input_cost_per_token": 3e-06,
136
140
  "litellm_provider": "openai",
@@ -226,6 +230,7 @@
226
230
  "supports_response_schema": true,
227
231
  "supports_system_messages": true,
228
232
  "supports_tool_choice": true,
233
+ "supports_service_tier": true,
229
234
  "supports_vision": true
230
235
  },
231
236
  "gpt-4-turbo-preview": {
@@ -338,6 +343,7 @@
338
343
  "supports_vision": true
339
344
  },
340
345
  "gpt-4-1106-preview": {
346
+ "deprecation_date": "2026-03-26",
341
347
  "input_cost_per_token": 1e-05,
342
348
  "litellm_provider": "openai",
343
349
  "max_input_tokens": 128000,
@@ -352,6 +358,7 @@
352
358
  "supports_tool_choice": true
353
359
  },
354
360
  "gpt-4-0125-preview": {
361
+ "deprecation_date": "2026-03-26",
355
362
  "input_cost_per_token": 1e-05,
356
363
  "litellm_provider": "openai",
357
364
  "max_input_tokens": 128000,
@@ -434,6 +441,7 @@
434
441
  "supports_tool_choice": true
435
442
  },
436
443
  "gpt-3.5-turbo-1106": {
444
+ "deprecation_date": "2026-09-28",
437
445
  "input_cost_per_token": 1e-06,
438
446
  "litellm_provider": "openai",
439
447
  "max_input_tokens": 16385,
@@ -845,6 +853,7 @@
845
853
  "supports_vision": true
846
854
  },
847
855
  "azure/gpt-4o-2024-08-06": {
856
+ "deprecation_date": "2026-02-27",
848
857
  "cache_read_input_token_cost": 1.25e-06,
849
858
  "input_cost_per_token": 2.5e-06,
850
859
  "litellm_provider": "azure",
@@ -876,7 +885,7 @@
876
885
  },
877
886
  "azure/global-standard/gpt-4o-2024-08-06": {
878
887
  "cache_read_input_token_cost": 1.25e-06,
879
- "deprecation_date": "2025-08-20",
888
+ "deprecation_date": "2026-02-27",
880
889
  "input_cost_per_token": 2.5e-06,
881
890
  "litellm_provider": "azure",
882
891
  "max_input_tokens": 128000,
@@ -1184,6 +1193,7 @@
1184
1193
  "output_cost_per_token": 0.0
1185
1194
  },
1186
1195
  "azure/text-embedding-3-small": {
1196
+ "deprecation_date": "2026-04-30",
1187
1197
  "input_cost_per_token": 2e-08,
1188
1198
  "litellm_provider": "azure",
1189
1199
  "max_input_tokens": 8191,
@@ -1942,7 +1952,6 @@
1942
1952
  "cache_creation_input_token_cost": 3e-07,
1943
1953
  "cache_creation_input_token_cost_above_1hr": 6e-06,
1944
1954
  "cache_read_input_token_cost": 3e-08,
1945
- "deprecation_date": "2025-03-01",
1946
1955
  "input_cost_per_token": 2.5e-07,
1947
1956
  "litellm_provider": "anthropic",
1948
1957
  "max_input_tokens": 200000,
@@ -1978,7 +1987,7 @@
1978
1987
  "cache_creation_input_token_cost": 1.875e-05,
1979
1988
  "cache_creation_input_token_cost_above_1hr": 6e-06,
1980
1989
  "cache_read_input_token_cost": 1.5e-06,
1981
- "deprecation_date": "2025-03-01",
1990
+ "deprecation_date": "2026-05-01",
1982
1991
  "input_cost_per_token": 1.5e-05,
1983
1992
  "litellm_provider": "anthropic",
1984
1993
  "max_input_tokens": 200000,
@@ -6433,13 +6442,13 @@
6433
6442
  "supports_tool_choice": false
6434
6443
  },
6435
6444
  "deepinfra/Gryphe/MythoMax-L2-13b": {
6436
- "input_cost_per_token": 7.2e-08,
6437
- "litellm_provider": "deepinfra",
6445
+ "max_tokens": 4096,
6438
6446
  "max_input_tokens": 4096,
6439
6447
  "max_output_tokens": 4096,
6440
- "max_tokens": 4096,
6448
+ "input_cost_per_token": 8e-08,
6449
+ "output_cost_per_token": 9e-08,
6450
+ "litellm_provider": "deepinfra",
6441
6451
  "mode": "chat",
6442
- "output_cost_per_token": 7.2e-08,
6443
6452
  "supports_tool_choice": true
6444
6453
  },
6445
6454
  "deepinfra/mistralai/Mistral-7B-Instruct-v0.1": {
@@ -6502,13 +6511,13 @@
6502
6511
  "supports_tool_choice": true
6503
6512
  },
6504
6513
  "deepinfra/mistralai/Mixtral-8x7B-Instruct-v0.1": {
6505
- "input_cost_per_token": 8e-08,
6506
- "litellm_provider": "deepinfra",
6514
+ "max_tokens": 32768,
6507
6515
  "max_input_tokens": 32768,
6508
6516
  "max_output_tokens": 32768,
6509
- "max_tokens": 32768,
6517
+ "input_cost_per_token": 4e-07,
6518
+ "output_cost_per_token": 4e-07,
6519
+ "litellm_provider": "deepinfra",
6510
6520
  "mode": "chat",
6511
- "output_cost_per_token": 2.4e-07,
6512
6521
  "supports_tool_choice": true
6513
6522
  },
6514
6523
  "deepinfra/deepinfra/airoboros-70b": {
@@ -6581,13 +6590,13 @@
6581
6590
  "supports_tool_choice": true
6582
6591
  },
6583
6592
  "deepinfra/meta-llama/Meta-Llama-3-8B-Instruct": {
6584
- "input_cost_per_token": 3e-08,
6585
- "litellm_provider": "deepinfra",
6593
+ "max_tokens": 8192,
6586
6594
  "max_input_tokens": 8192,
6587
6595
  "max_output_tokens": 8192,
6588
- "max_tokens": 8192,
6589
- "mode": "chat",
6596
+ "input_cost_per_token": 3e-08,
6590
6597
  "output_cost_per_token": 6e-08,
6598
+ "litellm_provider": "deepinfra",
6599
+ "mode": "chat",
6591
6600
  "supports_tool_choice": true
6592
6601
  },
6593
6602
  "deepinfra/meta-llama/Meta-Llama-3-70B-Instruct": {
@@ -7720,6 +7729,7 @@
7720
7729
  "supports_response_schema": true,
7721
7730
  "supports_system_messages": true,
7722
7731
  "supports_tool_choice": true,
7732
+ "supports_service_tier": true,
7723
7733
  "supports_vision": true
7724
7734
  },
7725
7735
  "ft:gpt-4o-2024-11-20": {
@@ -7741,6 +7751,7 @@
7741
7751
  "supports_vision": true
7742
7752
  },
7743
7753
  "azure/gpt-4o-2024-11-20": {
7754
+ "deprecation_date": "2026-03-01",
7744
7755
  "cache_read_input_token_cost": 1.25e-06,
7745
7756
  "input_cost_per_token": 2.75e-06,
7746
7757
  "litellm_provider": "azure",
@@ -7758,7 +7769,7 @@
7758
7769
  },
7759
7770
  "azure/global-standard/gpt-4o-2024-11-20": {
7760
7771
  "cache_read_input_token_cost": 1.25e-06,
7761
- "deprecation_date": "2025-12-20",
7772
+ "deprecation_date": "2026-03-01",
7762
7773
  "input_cost_per_token": 2.5e-06,
7763
7774
  "litellm_provider": "azure",
7764
7775
  "max_input_tokens": 128000,
@@ -9795,7 +9806,7 @@
9795
9806
  "cache_creation_input_token_cost": 3.75e-06,
9796
9807
  "cache_creation_input_token_cost_above_1hr": 6e-06,
9797
9808
  "cache_read_input_token_cost": 3e-07,
9798
- "deprecation_date": "2026-02-01",
9809
+ "deprecation_date": "2026-02-19",
9799
9810
  "input_cost_per_token": 3e-06,
9800
9811
  "litellm_provider": "anthropic",
9801
9812
  "max_input_tokens": 200000,
@@ -10204,6 +10215,7 @@
10204
10215
  "supports_vision": false
10205
10216
  },
10206
10217
  "azure/us/gpt-4o-2024-11-20": {
10218
+ "deprecation_date": "2026-03-01",
10207
10219
  "cache_creation_input_token_cost": 1.38e-06,
10208
10220
  "input_cost_per_token": 2.75e-06,
10209
10221
  "litellm_provider": "azure",
@@ -10219,6 +10231,7 @@
10219
10231
  "supports_vision": true
10220
10232
  },
10221
10233
  "azure/eu/gpt-4o-2024-11-20": {
10234
+ "deprecation_date": "2026-03-01",
10222
10235
  "cache_creation_input_token_cost": 1.38e-06,
10223
10236
  "input_cost_per_token": 2.75e-06,
10224
10237
  "litellm_provider": "azure",
@@ -10234,6 +10247,7 @@
10234
10247
  "supports_vision": true
10235
10248
  },
10236
10249
  "azure/us/gpt-4o-2024-08-06": {
10250
+ "deprecation_date": "2026-02-27",
10237
10251
  "cache_read_input_token_cost": 1.375e-06,
10238
10252
  "input_cost_per_token": 2.75e-06,
10239
10253
  "litellm_provider": "azure",
@@ -10250,6 +10264,7 @@
10250
10264
  "supports_vision": true
10251
10265
  },
10252
10266
  "azure/eu/gpt-4o-2024-08-06": {
10267
+ "deprecation_date": "2026-02-27",
10253
10268
  "cache_read_input_token_cost": 1.375e-06,
10254
10269
  "input_cost_per_token": 2.75e-06,
10255
10270
  "litellm_provider": "azure",
@@ -10746,6 +10761,7 @@
10746
10761
  "mode": "chat"
10747
10762
  },
10748
10763
  "azure/global/gpt-4o-2024-11-20": {
10764
+ "deprecation_date": "2026-03-01",
10749
10765
  "cache_read_input_token_cost": 1.25e-06,
10750
10766
  "input_cost_per_token": 2.5e-06,
10751
10767
  "litellm_provider": "azure",
@@ -10762,6 +10778,7 @@
10762
10778
  "supports_vision": true
10763
10779
  },
10764
10780
  "azure/global/gpt-4o-2024-08-06": {
10781
+ "deprecation_date": "2026-02-27",
10765
10782
  "cache_read_input_token_cost": 1.25e-06,
10766
10783
  "input_cost_per_token": 2.5e-06,
10767
10784
  "litellm_provider": "azure",
@@ -11009,6 +11026,7 @@
11009
11026
  "supports_response_schema": true,
11010
11027
  "supports_system_messages": true,
11011
11028
  "supports_tool_choice": true,
11029
+ "supports_service_tier": true,
11012
11030
  "supports_vision": true
11013
11031
  },
11014
11032
  "gpt-4.1-2025-04-14": {
@@ -11042,6 +11060,7 @@
11042
11060
  "supports_response_schema": true,
11043
11061
  "supports_system_messages": true,
11044
11062
  "supports_tool_choice": true,
11063
+ "supports_service_tier": true,
11045
11064
  "supports_vision": true
11046
11065
  },
11047
11066
  "gpt-4.1-mini": {
@@ -11078,6 +11097,7 @@
11078
11097
  "supports_response_schema": true,
11079
11098
  "supports_system_messages": true,
11080
11099
  "supports_tool_choice": true,
11100
+ "supports_service_tier": true,
11081
11101
  "supports_vision": true
11082
11102
  },
11083
11103
  "gpt-4.1-mini-2025-04-14": {
@@ -11111,6 +11131,7 @@
11111
11131
  "supports_response_schema": true,
11112
11132
  "supports_system_messages": true,
11113
11133
  "supports_tool_choice": true,
11134
+ "supports_service_tier": true,
11114
11135
  "supports_vision": true
11115
11136
  },
11116
11137
  "gpt-4.1-nano": {
@@ -11147,6 +11168,7 @@
11147
11168
  "supports_response_schema": true,
11148
11169
  "supports_system_messages": true,
11149
11170
  "supports_tool_choice": true,
11171
+ "supports_service_tier": true,
11150
11172
  "supports_vision": true
11151
11173
  },
11152
11174
  "gpt-4.1-nano-2025-04-14": {
@@ -11180,16 +11202,17 @@
11180
11202
  "supports_response_schema": true,
11181
11203
  "supports_system_messages": true,
11182
11204
  "supports_tool_choice": true,
11205
+ "supports_service_tier": true,
11183
11206
  "supports_vision": true
11184
11207
  },
11185
11208
  "watsonx/ibm/granite-3-8b-instruct": {
11186
- "input_cost_per_token": 0.0002,
11209
+ "input_cost_per_token": 2e-07,
11187
11210
  "litellm_provider": "watsonx",
11188
11211
  "max_input_tokens": 8192,
11189
11212
  "max_output_tokens": 1024,
11190
11213
  "max_tokens": 8192,
11191
11214
  "mode": "chat",
11192
- "output_cost_per_token": 0.0002,
11215
+ "output_cost_per_token": 2e-07,
11193
11216
  "supports_audio_input": false,
11194
11217
  "supports_audio_output": false,
11195
11218
  "supports_function_calling": true,
@@ -11262,6 +11285,7 @@
11262
11285
  "supports_reasoning": true,
11263
11286
  "supports_response_schema": true,
11264
11287
  "supports_tool_choice": true,
11288
+ "supports_service_tier": true,
11265
11289
  "supports_vision": true
11266
11290
  },
11267
11291
  "o3-2025-04-16": {
@@ -11293,6 +11317,7 @@
11293
11317
  "supports_reasoning": true,
11294
11318
  "supports_response_schema": true,
11295
11319
  "supports_tool_choice": true,
11320
+ "supports_service_tier": true,
11296
11321
  "supports_vision": true
11297
11322
  },
11298
11323
  "o4-mini": {
@@ -11317,6 +11342,7 @@
11317
11342
  "supports_reasoning": true,
11318
11343
  "supports_response_schema": true,
11319
11344
  "supports_tool_choice": true,
11345
+ "supports_service_tier": true,
11320
11346
  "supports_vision": true
11321
11347
  },
11322
11348
  "o4-mini-2025-04-16": {
@@ -11335,6 +11361,7 @@
11335
11361
  "supports_reasoning": true,
11336
11362
  "supports_response_schema": true,
11337
11363
  "supports_tool_choice": true,
11364
+ "supports_service_tier": true,
11338
11365
  "supports_vision": true
11339
11366
  },
11340
11367
  "gpt-image-1": {
@@ -11592,6 +11619,7 @@
11592
11619
  "supports_web_search": false
11593
11620
  },
11594
11621
  "azure/gpt-4.1-2025-04-14": {
11622
+ "deprecation_date": "2026-11-04",
11595
11623
  "cache_read_input_token_cost": 5e-07,
11596
11624
  "input_cost_per_token": 2e-06,
11597
11625
  "input_cost_per_token_batches": 1e-06,
@@ -11658,6 +11686,7 @@
11658
11686
  "supports_web_search": false
11659
11687
  },
11660
11688
  "azure/gpt-4.1-mini-2025-04-14": {
11689
+ "deprecation_date": "2026-11-04",
11661
11690
  "cache_read_input_token_cost": 1e-07,
11662
11691
  "input_cost_per_token": 4e-07,
11663
11692
  "input_cost_per_token_batches": 2e-07,
@@ -11723,6 +11752,7 @@
11723
11752
  "supports_vision": true
11724
11753
  },
11725
11754
  "azure/gpt-4.1-nano-2025-04-14": {
11755
+ "deprecation_date": "2026-11-04",
11726
11756
  "cache_read_input_token_cost": 2.5e-08,
11727
11757
  "input_cost_per_token": 1e-07,
11728
11758
  "input_cost_per_token_batches": 5e-08,
@@ -11784,6 +11814,7 @@
11784
11814
  "supports_vision": true
11785
11815
  },
11786
11816
  "azure/o3-2025-04-16": {
11817
+ "deprecation_date": "2026-04-16",
11787
11818
  "cache_read_input_token_cost": 2.5e-06,
11788
11819
  "input_cost_per_token": 1e-05,
11789
11820
  "litellm_provider": "azure",
@@ -13948,6 +13979,7 @@
13948
13979
  "cache_creation_input_token_cost_above_1hr": 3e-05,
13949
13980
  "cache_read_input_token_cost": 1.5e-06,
13950
13981
  "input_cost_per_token": 1.5e-05,
13982
+ "deprecation_date": "2026-05-14",
13951
13983
  "litellm_provider": "anthropic",
13952
13984
  "max_input_tokens": 200000,
13953
13985
  "max_output_tokens": 32000,
@@ -13971,6 +14003,7 @@
13971
14003
  "tool_use_system_prompt_tokens": 159
13972
14004
  },
13973
14005
  "claude-sonnet-4-20250514": {
14006
+ "deprecation_date": "2026-05-14",
13974
14007
  "cache_creation_input_token_cost": 3.75e-06,
13975
14008
  "cache_creation_input_token_cost_above_1hr": 6e-06,
13976
14009
  "cache_read_input_token_cost": 3e-07,
@@ -14368,7 +14401,7 @@
14368
14401
  "input_cost_per_token_above_200k_tokens": 6e-06,
14369
14402
  "litellm_provider": "anthropic",
14370
14403
  "max_input_tokens": 1000000,
14371
- "max_output_tokens": 1000000,
14404
+ "max_output_tokens": 64000,
14372
14405
  "max_tokens": 1000000,
14373
14406
  "mode": "chat",
14374
14407
  "output_cost_per_token": 1.5e-05,
@@ -15396,7 +15429,7 @@
15396
15429
  "tpm": 800000
15397
15430
  },
15398
15431
  "gemini-2.5-flash": {
15399
- "cache_read_input_token_cost": 7.5e-08,
15432
+ "cache_read_input_token_cost": 3e-08,
15400
15433
  "input_cost_per_audio_token": 1e-06,
15401
15434
  "input_cost_per_token": 3e-07,
15402
15435
  "litellm_provider": "vertex_ai-language-models",
@@ -15546,6 +15579,10 @@
15546
15579
  },
15547
15580
  "openrouter/anthropic/claude-sonnet-4": {
15548
15581
  "input_cost_per_image": 0.0048,
15582
+ "cache_creation_input_token_cost": 3.75e-06,
15583
+ "cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
15584
+ "cache_read_input_token_cost": 3e-07,
15585
+ "cache_read_input_token_cost_above_200k_tokens": 6e-07,
15549
15586
  "input_cost_per_token": 3e-06,
15550
15587
  "input_cost_per_token_above_200k_tokens": 6e-06,
15551
15588
  "output_cost_per_token_above_200k_tokens": 2.25e-05,
@@ -15558,6 +15595,7 @@
15558
15595
  "supports_assistant_prefill": true,
15559
15596
  "supports_computer_use": true,
15560
15597
  "supports_function_calling": true,
15598
+ "supports_prompt_caching": true,
15561
15599
  "supports_reasoning": true,
15562
15600
  "supports_tool_choice": true,
15563
15601
  "supports_vision": true,
@@ -15658,7 +15696,8 @@
15658
15696
  "tpm": 250000
15659
15697
  },
15660
15698
  "gemini-2.5-pro": {
15661
- "cache_read_input_token_cost": 3.125e-07,
15699
+ "cache_read_input_token_cost": 1.25e-07,
15700
+ "cache_creation_input_token_cost_above_200k_tokens": 2.5e-07,
15662
15701
  "input_cost_per_token": 1.25e-06,
15663
15702
  "input_cost_per_token_above_200k_tokens": 2.5e-06,
15664
15703
  "litellm_provider": "vertex_ai-language-models",
@@ -16228,35 +16267,36 @@
16228
16267
  "output_cost_per_token": 1.5e-05,
16229
16268
  "source": "https://docs.x.ai/docs/models",
16230
16269
  "supports_function_calling": true,
16231
- "supports_reasoning": true,
16232
16270
  "supports_tool_choice": true,
16233
16271
  "supports_web_search": true
16234
16272
  },
16235
16273
  "xai/grok-4-0709": {
16236
16274
  "input_cost_per_token": 3e-06,
16275
+ "input_cost_per_token_above_128k_tokens": 6e-06,
16237
16276
  "litellm_provider": "xai",
16238
16277
  "max_input_tokens": 256000,
16239
16278
  "max_output_tokens": 256000,
16240
16279
  "max_tokens": 256000,
16241
16280
  "mode": "chat",
16242
16281
  "output_cost_per_token": 1.5e-05,
16282
+ "output_cost_per_token_above_128k_tokens": 3e-05,
16243
16283
  "source": "https://docs.x.ai/docs/models",
16244
16284
  "supports_function_calling": true,
16245
- "supports_reasoning": true,
16246
16285
  "supports_tool_choice": true,
16247
16286
  "supports_web_search": true
16248
16287
  },
16249
16288
  "xai/grok-4-latest": {
16250
16289
  "input_cost_per_token": 3e-06,
16290
+ "input_cost_per_token_above_128k_tokens": 6e-06,
16251
16291
  "litellm_provider": "xai",
16252
16292
  "max_input_tokens": 256000,
16253
16293
  "max_output_tokens": 256000,
16254
16294
  "max_tokens": 256000,
16255
16295
  "mode": "chat",
16256
16296
  "output_cost_per_token": 1.5e-05,
16297
+ "output_cost_per_token_above_128k_tokens": 3e-05,
16257
16298
  "source": "https://docs.x.ai/docs/models",
16258
16299
  "supports_function_calling": true,
16259
- "supports_reasoning": true,
16260
16300
  "supports_tool_choice": true,
16261
16301
  "supports_web_search": true
16262
16302
  },
@@ -16977,7 +17017,8 @@
16977
17017
  "max_tokens": 8192,
16978
17018
  "mode": "chat",
16979
17019
  "output_cost_per_token": 6.3e-07,
16980
- "supports_tool_choice": true
17020
+ "supports_tool_choice": true,
17021
+ "supports_vision": true
16981
17022
  },
16982
17023
  "fireworks_ai/accounts/fireworks/models/kimi-k2-instruct": {
16983
17024
  "input_cost_per_token": 6e-07,
@@ -17836,6 +17877,7 @@
17836
17877
  "cache_creation_input_token_cost_above_1hr": 3e-05,
17837
17878
  "cache_read_input_token_cost": 1.5e-06,
17838
17879
  "input_cost_per_token": 1.5e-05,
17880
+ "deprecation_date": "2026-08-05",
17839
17881
  "litellm_provider": "anthropic",
17840
17882
  "max_input_tokens": 200000,
17841
17883
  "max_output_tokens": 32000,
@@ -17995,8 +18037,8 @@
17995
18037
  "input_cost_per_token_batches": 7.5e-06,
17996
18038
  "litellm_provider": "vertex_ai-anthropic_models",
17997
18039
  "max_input_tokens": 200000,
17998
- "max_output_tokens": 4096,
17999
- "max_tokens": 4096,
18040
+ "max_output_tokens": 32000,
18041
+ "max_tokens": 32000,
18000
18042
  "mode": "chat",
18001
18043
  "output_cost_per_token": 7.5e-05,
18002
18044
  "output_cost_per_token_batches": 3.75e-05,
@@ -18012,8 +18054,8 @@
18012
18054
  "input_cost_per_token_batches": 7.5e-06,
18013
18055
  "litellm_provider": "vertex_ai-anthropic_models",
18014
18056
  "max_input_tokens": 200000,
18015
- "max_output_tokens": 4096,
18016
- "max_tokens": 4096,
18057
+ "max_output_tokens": 32000,
18058
+ "max_tokens": 32000,
18017
18059
  "mode": "chat",
18018
18060
  "output_cost_per_token": 7.5e-05,
18019
18061
  "output_cost_per_token_batches": 3.75e-05,
@@ -18084,6 +18126,7 @@
18084
18126
  "supports_response_schema": true,
18085
18127
  "supports_system_messages": true,
18086
18128
  "supports_tool_choice": true,
18129
+ "supports_service_tier": true,
18087
18130
  "supports_vision": true
18088
18131
  },
18089
18132
  "gpt-5-mini": {
@@ -18122,6 +18165,7 @@
18122
18165
  "supports_response_schema": true,
18123
18166
  "supports_system_messages": true,
18124
18167
  "supports_tool_choice": true,
18168
+ "supports_service_tier": true,
18125
18169
  "supports_vision": true
18126
18170
  },
18127
18171
  "gpt-5-nano": {
@@ -18196,9 +18240,9 @@
18196
18240
  "cache_read_input_token_cost": 1.25e-07,
18197
18241
  "input_cost_per_token": 1.25e-06,
18198
18242
  "litellm_provider": "openai",
18199
- "max_input_tokens": 272000,
18200
- "max_output_tokens": 128000,
18201
- "max_tokens": 128000,
18243
+ "max_input_tokens": 128000,
18244
+ "max_output_tokens": 16384,
18245
+ "max_tokens": 16384,
18202
18246
  "mode": "chat",
18203
18247
  "output_cost_per_token": 1e-05,
18204
18248
  "supported_endpoints": [
@@ -18299,6 +18343,7 @@
18299
18343
  "supports_response_schema": true,
18300
18344
  "supports_system_messages": true,
18301
18345
  "supports_tool_choice": true,
18346
+ "supports_service_tier": true,
18302
18347
  "supports_vision": true
18303
18348
  },
18304
18349
  "gpt-5-nano-2025-08-07": {
@@ -18565,9 +18610,9 @@
18565
18610
  "cache_read_input_token_cost": 1.25e-07,
18566
18611
  "input_cost_per_token": 1.25e-06,
18567
18612
  "litellm_provider": "azure",
18568
- "max_input_tokens": 272000,
18569
- "max_output_tokens": 128000,
18570
- "max_tokens": 128000,
18613
+ "max_input_tokens": 128000,
18614
+ "max_output_tokens": 16384,
18615
+ "max_tokens": 16384,
18571
18616
  "mode": "chat",
18572
18617
  "output_cost_per_token": 1e-05,
18573
18618
  "supported_endpoints": [
@@ -19211,23 +19256,23 @@
19211
19256
  "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing"
19212
19257
  },
19213
19258
  "deepinfra/deepseek-ai/DeepSeek-V3": {
19214
- "input_cost_per_token": 3.8e-07,
19215
- "litellm_provider": "deepinfra",
19259
+ "max_tokens": 163840,
19216
19260
  "max_input_tokens": 163840,
19217
19261
  "max_output_tokens": 163840,
19218
- "max_tokens": 163840,
19219
- "mode": "chat",
19262
+ "input_cost_per_token": 3.8e-07,
19220
19263
  "output_cost_per_token": 8.9e-07,
19264
+ "litellm_provider": "deepinfra",
19265
+ "mode": "chat",
19221
19266
  "supports_tool_choice": true
19222
19267
  },
19223
19268
  "deepinfra/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo": {
19224
- "input_cost_per_token": 1.5e-08,
19225
- "litellm_provider": "deepinfra",
19269
+ "max_tokens": 131072,
19226
19270
  "max_input_tokens": 131072,
19227
19271
  "max_output_tokens": 131072,
19228
- "max_tokens": 131072,
19272
+ "input_cost_per_token": 2e-08,
19273
+ "output_cost_per_token": 3e-08,
19274
+ "litellm_provider": "deepinfra",
19229
19275
  "mode": "chat",
19230
- "output_cost_per_token": 2e-08,
19231
19276
  "supports_tool_choice": true
19232
19277
  },
19233
19278
  "deepinfra/google/gemma-2-9b-it": {
@@ -19241,13 +19286,13 @@
19241
19286
  "supports_tool_choice": false
19242
19287
  },
19243
19288
  "deepinfra/deepseek-ai/DeepSeek-R1-0528-Turbo": {
19244
- "input_cost_per_token": 1e-06,
19245
- "litellm_provider": "deepinfra",
19289
+ "max_tokens": 32768,
19246
19290
  "max_input_tokens": 32768,
19247
19291
  "max_output_tokens": 32768,
19248
- "max_tokens": 32768,
19249
- "mode": "chat",
19292
+ "input_cost_per_token": 1e-06,
19250
19293
  "output_cost_per_token": 3e-06,
19294
+ "litellm_provider": "deepinfra",
19295
+ "mode": "chat",
19251
19296
  "supports_tool_choice": true
19252
19297
  },
19253
19298
  "deepinfra/Qwen/Qwen2-7B-Instruct": {
@@ -19271,13 +19316,13 @@
19271
19316
  "supports_tool_choice": false
19272
19317
  },
19273
19318
  "deepinfra/meta-llama/Llama-3.3-70B-Instruct": {
19274
- "input_cost_per_token": 2.3e-07,
19275
- "litellm_provider": "deepinfra",
19319
+ "max_tokens": 131072,
19276
19320
  "max_input_tokens": 131072,
19277
19321
  "max_output_tokens": 131072,
19278
- "max_tokens": 131072,
19279
- "mode": "chat",
19322
+ "input_cost_per_token": 2.3e-07,
19280
19323
  "output_cost_per_token": 4e-07,
19324
+ "litellm_provider": "deepinfra",
19325
+ "mode": "chat",
19281
19326
  "supports_tool_choice": true
19282
19327
  },
19283
19328
  "deepinfra/microsoft/Phi-4-multimodal-instruct": {
@@ -19321,35 +19366,34 @@
19321
19366
  "supports_tool_choice": false
19322
19367
  },
19323
19368
  "deepinfra/mistralai/Mistral-Small-3.2-24B-Instruct-2506": {
19324
- "input_cost_per_token": 5e-08,
19325
- "litellm_provider": "deepinfra",
19369
+ "max_tokens": 128000,
19326
19370
  "max_input_tokens": 128000,
19327
19371
  "max_output_tokens": 128000,
19328
- "max_tokens": 128000,
19372
+ "input_cost_per_token": 7.5e-08,
19373
+ "output_cost_per_token": 2e-07,
19374
+ "litellm_provider": "deepinfra",
19329
19375
  "mode": "chat",
19330
- "output_cost_per_token": 1e-07,
19331
19376
  "supports_tool_choice": true
19332
19377
  },
19333
19378
  "deepinfra/deepseek-ai/DeepSeek-V3-0324": {
19334
- "cache_read_input_token_cost": 2.24e-07,
19335
- "input_cost_per_token": 2.8e-07,
19336
- "litellm_provider": "deepinfra",
19379
+ "max_tokens": 163840,
19337
19380
  "max_input_tokens": 163840,
19338
19381
  "max_output_tokens": 163840,
19339
- "max_tokens": 163840,
19340
- "mode": "chat",
19382
+ "input_cost_per_token": 2.5e-07,
19341
19383
  "output_cost_per_token": 8.8e-07,
19384
+ "litellm_provider": "deepinfra",
19385
+ "mode": "chat",
19342
19386
  "supports_tool_choice": true
19343
19387
  },
19344
19388
  "deepinfra/anthropic/claude-3-7-sonnet-latest": {
19345
- "cache_read_input_token_cost": 3.3e-07,
19346
- "input_cost_per_token": 3.3e-06,
19347
- "litellm_provider": "deepinfra",
19389
+ "max_tokens": 200000,
19348
19390
  "max_input_tokens": 200000,
19349
19391
  "max_output_tokens": 200000,
19350
- "max_tokens": 200000,
19351
- "mode": "chat",
19392
+ "input_cost_per_token": 3.3e-06,
19352
19393
  "output_cost_per_token": 1.65e-05,
19394
+ "cache_read_input_token_cost": 3.3e-07,
19395
+ "litellm_provider": "deepinfra",
19396
+ "mode": "chat",
19353
19397
  "supports_tool_choice": true
19354
19398
  },
19355
19399
  "deepinfra/cognitivecomputations/dolphin-2.9.1-llama-3-70b": {
@@ -19373,13 +19417,13 @@
19373
19417
  "supports_tool_choice": false
19374
19418
  },
19375
19419
  "deepinfra/Qwen/Qwen3-235B-A22B": {
19376
- "input_cost_per_token": 1.3e-07,
19377
- "litellm_provider": "deepinfra",
19420
+ "max_tokens": 40960,
19378
19421
  "max_input_tokens": 40960,
19379
19422
  "max_output_tokens": 40960,
19380
- "max_tokens": 40960,
19423
+ "input_cost_per_token": 1.8e-07,
19424
+ "output_cost_per_token": 5.4e-07,
19425
+ "litellm_provider": "deepinfra",
19381
19426
  "mode": "chat",
19382
- "output_cost_per_token": 6e-07,
19383
19427
  "supports_tool_choice": true
19384
19428
  },
19385
19429
  "deepinfra/deepseek-ai/DeepSeek-V3-0324-Turbo": {
@@ -19394,33 +19438,33 @@
19394
19438
  "supports_reasoning": true
19395
19439
  },
19396
19440
  "deepinfra/microsoft/WizardLM-2-8x22B": {
19397
- "input_cost_per_token": 4.8e-07,
19398
- "litellm_provider": "deepinfra",
19441
+ "max_tokens": 65536,
19399
19442
  "max_input_tokens": 65536,
19400
19443
  "max_output_tokens": 65536,
19401
- "max_tokens": 65536,
19402
- "mode": "chat",
19444
+ "input_cost_per_token": 4.8e-07,
19403
19445
  "output_cost_per_token": 4.8e-07,
19446
+ "litellm_provider": "deepinfra",
19447
+ "mode": "chat",
19404
19448
  "supports_tool_choice": false
19405
19449
  },
19406
19450
  "deepinfra/Sao10K/L3-8B-Lunaris-v1-Turbo": {
19407
- "input_cost_per_token": 2e-08,
19408
- "litellm_provider": "deepinfra",
19451
+ "max_tokens": 8192,
19409
19452
  "max_input_tokens": 8192,
19410
19453
  "max_output_tokens": 8192,
19411
- "max_tokens": 8192,
19412
- "mode": "chat",
19454
+ "input_cost_per_token": 4e-08,
19413
19455
  "output_cost_per_token": 5e-08,
19456
+ "litellm_provider": "deepinfra",
19457
+ "mode": "chat",
19414
19458
  "supports_tool_choice": false
19415
19459
  },
19416
19460
  "deepinfra/meta-llama/Llama-Guard-4-12B": {
19417
- "input_cost_per_token": 1.8e-07,
19418
- "litellm_provider": "deepinfra",
19461
+ "max_tokens": 163840,
19419
19462
  "max_input_tokens": 163840,
19420
19463
  "max_output_tokens": 163840,
19421
- "max_tokens": 163840,
19422
- "mode": "chat",
19464
+ "input_cost_per_token": 1.8e-07,
19423
19465
  "output_cost_per_token": 1.8e-07,
19466
+ "litellm_provider": "deepinfra",
19467
+ "mode": "chat",
19424
19468
  "supports_tool_choice": false
19425
19469
  },
19426
19470
  "deepinfra/meta-llama/Llama-3.2-1B-Instruct": {
@@ -19444,23 +19488,24 @@
19444
19488
  "supports_tool_choice": false
19445
19489
  },
19446
19490
  "deepinfra/Qwen/Qwen2.5-VL-32B-Instruct": {
19447
- "input_cost_per_token": 2e-07,
19448
- "litellm_provider": "deepinfra",
19491
+ "max_tokens": 128000,
19449
19492
  "max_input_tokens": 128000,
19450
19493
  "max_output_tokens": 128000,
19451
- "max_tokens": 128000,
19452
- "mode": "chat",
19494
+ "input_cost_per_token": 2e-07,
19453
19495
  "output_cost_per_token": 6e-07,
19454
- "supports_tool_choice": true
19496
+ "litellm_provider": "deepinfra",
19497
+ "mode": "chat",
19498
+ "supports_tool_choice": true,
19499
+ "supports_vision": true
19455
19500
  },
19456
19501
  "deepinfra/nvidia/Llama-3.1-Nemotron-70B-Instruct": {
19457
- "input_cost_per_token": 1.2e-07,
19458
- "litellm_provider": "deepinfra",
19502
+ "max_tokens": 131072,
19459
19503
  "max_input_tokens": 131072,
19460
19504
  "max_output_tokens": 131072,
19461
- "max_tokens": 131072,
19505
+ "input_cost_per_token": 6e-07,
19506
+ "output_cost_per_token": 6e-07,
19507
+ "litellm_provider": "deepinfra",
19462
19508
  "mode": "chat",
19463
- "output_cost_per_token": 3e-07,
19464
19509
  "supports_tool_choice": true
19465
19510
  },
19466
19511
  "deepinfra/mistralai/Mixtral-8x22B-Instruct-v0.1": {
@@ -19474,13 +19519,13 @@
19474
19519
  "supports_tool_choice": true
19475
19520
  },
19476
19521
  "deepinfra/Qwen/Qwen2.5-7B-Instruct": {
19477
- "input_cost_per_token": 4e-08,
19478
- "litellm_provider": "deepinfra",
19522
+ "max_tokens": 32768,
19479
19523
  "max_input_tokens": 32768,
19480
19524
  "max_output_tokens": 32768,
19481
- "max_tokens": 32768,
19482
- "mode": "chat",
19525
+ "input_cost_per_token": 4e-08,
19483
19526
  "output_cost_per_token": 1e-07,
19527
+ "litellm_provider": "deepinfra",
19528
+ "mode": "chat",
19484
19529
  "supports_tool_choice": false
19485
19530
  },
19486
19531
  "deepinfra/google/gemini-1.5-flash-8b": {
@@ -19494,53 +19539,53 @@
19494
19539
  "supports_tool_choice": true
19495
19540
  },
19496
19541
  "deepinfra/NousResearch/Hermes-3-Llama-3.1-70B": {
19497
- "input_cost_per_token": 1e-07,
19498
- "litellm_provider": "deepinfra",
19542
+ "max_tokens": 131072,
19499
19543
  "max_input_tokens": 131072,
19500
19544
  "max_output_tokens": 131072,
19501
- "max_tokens": 131072,
19545
+ "input_cost_per_token": 3e-07,
19546
+ "output_cost_per_token": 3e-07,
19547
+ "litellm_provider": "deepinfra",
19502
19548
  "mode": "chat",
19503
- "output_cost_per_token": 2.8e-07,
19504
19549
  "supports_tool_choice": false
19505
19550
  },
19506
19551
  "deepinfra/deepseek-ai/DeepSeek-R1-Distill-Llama-70B": {
19507
- "input_cost_per_token": 1e-07,
19508
- "litellm_provider": "deepinfra",
19552
+ "max_tokens": 131072,
19509
19553
  "max_input_tokens": 131072,
19510
19554
  "max_output_tokens": 131072,
19511
- "max_tokens": 131072,
19555
+ "input_cost_per_token": 2e-07,
19556
+ "output_cost_per_token": 6e-07,
19557
+ "litellm_provider": "deepinfra",
19512
19558
  "mode": "chat",
19513
- "output_cost_per_token": 4e-07,
19514
19559
  "supports_tool_choice": false
19515
19560
  },
19516
19561
  "deepinfra/meta-llama/Llama-Guard-3-8B": {
19517
- "input_cost_per_token": 5.5e-08,
19518
- "litellm_provider": "deepinfra",
19562
+ "max_tokens": 131072,
19519
19563
  "max_input_tokens": 131072,
19520
19564
  "max_output_tokens": 131072,
19521
- "max_tokens": 131072,
19522
- "mode": "chat",
19565
+ "input_cost_per_token": 5.5e-08,
19523
19566
  "output_cost_per_token": 5.5e-08,
19567
+ "litellm_provider": "deepinfra",
19568
+ "mode": "chat",
19524
19569
  "supports_tool_choice": false
19525
19570
  },
19526
19571
  "deepinfra/mistralai/Mistral-Small-24B-Instruct-2501": {
19527
- "input_cost_per_token": 5e-08,
19528
- "litellm_provider": "deepinfra",
19572
+ "max_tokens": 32768,
19529
19573
  "max_input_tokens": 32768,
19530
19574
  "max_output_tokens": 32768,
19531
- "max_tokens": 32768,
19532
- "mode": "chat",
19575
+ "input_cost_per_token": 5e-08,
19533
19576
  "output_cost_per_token": 8e-08,
19577
+ "litellm_provider": "deepinfra",
19578
+ "mode": "chat",
19534
19579
  "supports_tool_choice": true
19535
19580
  },
19536
19581
  "deepinfra/anthropic/claude-4-opus": {
19537
- "input_cost_per_token": 1.65e-05,
19538
- "litellm_provider": "deepinfra",
19582
+ "max_tokens": 200000,
19539
19583
  "max_input_tokens": 200000,
19540
19584
  "max_output_tokens": 200000,
19541
- "max_tokens": 200000,
19542
- "mode": "chat",
19585
+ "input_cost_per_token": 1.65e-05,
19543
19586
  "output_cost_per_token": 8.25e-05,
19587
+ "litellm_provider": "deepinfra",
19588
+ "mode": "chat",
19544
19589
  "supports_tool_choice": true
19545
19590
  },
19546
19591
  "deepinfra/openchat/openchat-3.6-8b": {
@@ -19554,13 +19599,13 @@
19554
19599
  "supports_tool_choice": false
19555
19600
  },
19556
19601
  "deepinfra/google/gemma-3-27b-it": {
19557
- "input_cost_per_token": 9e-08,
19558
- "litellm_provider": "deepinfra",
19602
+ "max_tokens": 131072,
19559
19603
  "max_input_tokens": 131072,
19560
19604
  "max_output_tokens": 131072,
19561
- "max_tokens": 131072,
19605
+ "input_cost_per_token": 9e-08,
19606
+ "output_cost_per_token": 1.6e-07,
19607
+ "litellm_provider": "deepinfra",
19562
19608
  "mode": "chat",
19563
- "output_cost_per_token": 1.7e-07,
19564
19609
  "supports_tool_choice": true
19565
19610
  },
19566
19611
  "deepinfra/Austism/chronos-hermes-13b-v2": {
@@ -19574,13 +19619,13 @@
19574
19619
  "supports_tool_choice": true
19575
19620
  },
19576
19621
  "deepinfra/Sao10K/L3.1-70B-Euryale-v2.2": {
19577
- "input_cost_per_token": 6.5e-07,
19578
- "litellm_provider": "deepinfra",
19622
+ "max_tokens": 131072,
19579
19623
  "max_input_tokens": 131072,
19580
19624
  "max_output_tokens": 131072,
19581
- "max_tokens": 131072,
19582
- "mode": "chat",
19625
+ "input_cost_per_token": 6.5e-07,
19583
19626
  "output_cost_per_token": 7.5e-07,
19627
+ "litellm_provider": "deepinfra",
19628
+ "mode": "chat",
19584
19629
  "supports_tool_choice": false
19585
19630
  },
19586
19631
  "deepinfra/Qwen/QwQ-32B-Preview": {
@@ -19594,13 +19639,13 @@
19594
19639
  "supports_tool_choice": false
19595
19640
  },
19596
19641
  "deepinfra/anthropic/claude-4-sonnet": {
19597
- "input_cost_per_token": 3.3e-06,
19598
- "litellm_provider": "deepinfra",
19642
+ "max_tokens": 200000,
19599
19643
  "max_input_tokens": 200000,
19600
19644
  "max_output_tokens": 200000,
19601
- "max_tokens": 200000,
19602
- "mode": "chat",
19645
+ "input_cost_per_token": 3.3e-06,
19603
19646
  "output_cost_per_token": 1.65e-05,
19647
+ "litellm_provider": "deepinfra",
19648
+ "mode": "chat",
19604
19649
  "supports_tool_choice": true
19605
19650
  },
19606
19651
  "deepinfra/microsoft/Phi-3-medium-4k-instruct": {
@@ -19624,36 +19669,36 @@
19624
19669
  "supports_tool_choice": false
19625
19670
  },
19626
19671
  "deepinfra/Sao10K/L3.3-70B-Euryale-v2.3": {
19627
- "input_cost_per_token": 6.5e-07,
19628
- "litellm_provider": "deepinfra",
19672
+ "max_tokens": 131072,
19629
19673
  "max_input_tokens": 131072,
19630
19674
  "max_output_tokens": 131072,
19631
- "max_tokens": 131072,
19632
- "mode": "chat",
19675
+ "input_cost_per_token": 6.5e-07,
19633
19676
  "output_cost_per_token": 7.5e-07,
19677
+ "litellm_provider": "deepinfra",
19678
+ "mode": "chat",
19634
19679
  "supports_tool_choice": false
19635
19680
  },
19636
19681
  "deepinfra/meta-llama/Meta-Llama-3.1-70B-Instruct": {
19637
- "input_cost_per_token": 2.3e-07,
19638
- "litellm_provider": "deepinfra",
19682
+ "max_tokens": 131072,
19639
19683
  "max_input_tokens": 131072,
19640
19684
  "max_output_tokens": 131072,
19641
- "max_tokens": 131072,
19642
- "mode": "chat",
19685
+ "input_cost_per_token": 4e-07,
19643
19686
  "output_cost_per_token": 4e-07,
19687
+ "litellm_provider": "deepinfra",
19688
+ "mode": "chat",
19644
19689
  "supports_tool_choice": true
19645
19690
  },
19646
19691
  "deepinfra/deepseek-ai/DeepSeek-V3.1": {
19647
- "cache_read_input_token_cost": 2.16e-07,
19648
- "input_cost_per_token": 2.7e-07,
19649
- "litellm_provider": "deepinfra",
19692
+ "max_tokens": 163840,
19650
19693
  "max_input_tokens": 163840,
19651
19694
  "max_output_tokens": 163840,
19652
- "max_tokens": 163840,
19653
- "mode": "chat",
19695
+ "input_cost_per_token": 2.7e-07,
19654
19696
  "output_cost_per_token": 1e-06,
19655
- "supports_reasoning": true,
19656
- "supports_tool_choice": true
19697
+ "cache_read_input_token_cost": 2.16e-07,
19698
+ "litellm_provider": "deepinfra",
19699
+ "mode": "chat",
19700
+ "supports_tool_choice": true,
19701
+ "supports_reasoning": true
19657
19702
  },
19658
19703
  "deepinfra/Qwen/Qwen2.5-Coder-7B": {
19659
19704
  "max_tokens": 32768,
@@ -19677,23 +19722,23 @@
19677
19722
  "supports_reasoning": true
19678
19723
  },
19679
19724
  "deepinfra/zai-org/GLM-4.5": {
19680
- "input_cost_per_token": 5.5e-07,
19681
- "litellm_provider": "deepinfra",
19725
+ "max_tokens": 131072,
19682
19726
  "max_input_tokens": 131072,
19683
19727
  "max_output_tokens": 131072,
19684
- "max_tokens": 131072,
19728
+ "input_cost_per_token": 4e-07,
19729
+ "output_cost_per_token": 1.6e-06,
19730
+ "litellm_provider": "deepinfra",
19685
19731
  "mode": "chat",
19686
- "output_cost_per_token": 2e-06,
19687
19732
  "supports_tool_choice": true
19688
19733
  },
19689
19734
  "deepinfra/meta-llama/Llama-3.2-3B-Instruct": {
19690
- "input_cost_per_token": 1.2e-08,
19691
- "litellm_provider": "deepinfra",
19735
+ "max_tokens": 131072,
19692
19736
  "max_input_tokens": 131072,
19693
19737
  "max_output_tokens": 131072,
19694
- "max_tokens": 131072,
19738
+ "input_cost_per_token": 2e-08,
19739
+ "output_cost_per_token": 2e-08,
19740
+ "litellm_provider": "deepinfra",
19695
19741
  "mode": "chat",
19696
- "output_cost_per_token": 2.4e-08,
19697
19742
  "supports_tool_choice": true
19698
19743
  },
19699
19744
  "deepinfra/google/gemini-1.5-flash": {
@@ -19717,43 +19762,43 @@
19717
19762
  "supports_tool_choice": true
19718
19763
  },
19719
19764
  "deepinfra/google/gemini-2.5-pro": {
19720
- "input_cost_per_token": 8.75e-07,
19721
- "litellm_provider": "deepinfra",
19765
+ "max_tokens": 1000000,
19722
19766
  "max_input_tokens": 1000000,
19723
19767
  "max_output_tokens": 1000000,
19724
- "max_tokens": 1000000,
19768
+ "input_cost_per_token": 1.25e-06,
19769
+ "output_cost_per_token": 1e-05,
19770
+ "litellm_provider": "deepinfra",
19725
19771
  "mode": "chat",
19726
- "output_cost_per_token": 7e-06,
19727
19772
  "supports_tool_choice": true
19728
19773
  },
19729
19774
  "deepinfra/Qwen/Qwen3-30B-A3B": {
19730
- "input_cost_per_token": 8e-08,
19731
- "litellm_provider": "deepinfra",
19775
+ "max_tokens": 40960,
19732
19776
  "max_input_tokens": 40960,
19733
19777
  "max_output_tokens": 40960,
19734
- "max_tokens": 40960,
19735
- "mode": "chat",
19778
+ "input_cost_per_token": 8e-08,
19736
19779
  "output_cost_per_token": 2.9e-07,
19780
+ "litellm_provider": "deepinfra",
19781
+ "mode": "chat",
19737
19782
  "supports_tool_choice": true
19738
19783
  },
19739
19784
  "deepinfra/Qwen/QwQ-32B": {
19740
- "input_cost_per_token": 1.5e-07,
19741
- "litellm_provider": "deepinfra",
19785
+ "max_tokens": 131072,
19742
19786
  "max_input_tokens": 131072,
19743
19787
  "max_output_tokens": 131072,
19744
- "max_tokens": 131072,
19745
- "mode": "chat",
19788
+ "input_cost_per_token": 1.5e-07,
19746
19789
  "output_cost_per_token": 4e-07,
19790
+ "litellm_provider": "deepinfra",
19791
+ "mode": "chat",
19747
19792
  "supports_tool_choice": true
19748
19793
  },
19749
19794
  "deepinfra/moonshotai/Kimi-K2-Instruct": {
19750
- "input_cost_per_token": 5e-07,
19751
- "litellm_provider": "deepinfra",
19795
+ "max_tokens": 131072,
19752
19796
  "max_input_tokens": 131072,
19753
19797
  "max_output_tokens": 131072,
19754
- "max_tokens": 131072,
19755
- "mode": "chat",
19798
+ "input_cost_per_token": 5e-07,
19756
19799
  "output_cost_per_token": 2e-06,
19800
+ "litellm_provider": "deepinfra",
19801
+ "mode": "chat",
19757
19802
  "supports_tool_choice": true
19758
19803
  },
19759
19804
  "deepinfra/Sao10K/L3-70B-Euryale-v2.1": {
@@ -19777,33 +19822,33 @@
19777
19822
  "supports_tool_choice": false
19778
19823
  },
19779
19824
  "deepinfra/google/gemma-3-12b-it": {
19780
- "input_cost_per_token": 5e-08,
19781
- "litellm_provider": "deepinfra",
19825
+ "max_tokens": 131072,
19782
19826
  "max_input_tokens": 131072,
19783
19827
  "max_output_tokens": 131072,
19784
- "max_tokens": 131072,
19785
- "mode": "chat",
19828
+ "input_cost_per_token": 5e-08,
19786
19829
  "output_cost_per_token": 1e-07,
19830
+ "litellm_provider": "deepinfra",
19831
+ "mode": "chat",
19787
19832
  "supports_tool_choice": true
19788
19833
  },
19789
19834
  "deepinfra/google/gemini-2.5-flash": {
19790
- "input_cost_per_token": 2.1e-07,
19791
- "litellm_provider": "deepinfra",
19835
+ "max_tokens": 1000000,
19792
19836
  "max_input_tokens": 1000000,
19793
19837
  "max_output_tokens": 1000000,
19794
- "max_tokens": 1000000,
19838
+ "input_cost_per_token": 3e-07,
19839
+ "output_cost_per_token": 2.5e-06,
19840
+ "litellm_provider": "deepinfra",
19795
19841
  "mode": "chat",
19796
- "output_cost_per_token": 1.75e-06,
19797
19842
  "supports_tool_choice": true
19798
19843
  },
19799
19844
  "deepinfra/deepseek-ai/DeepSeek-R1": {
19800
- "input_cost_per_token": 7e-07,
19801
- "litellm_provider": "deepinfra",
19845
+ "max_tokens": 163840,
19802
19846
  "max_input_tokens": 163840,
19803
19847
  "max_output_tokens": 163840,
19804
- "max_tokens": 163840,
19805
- "mode": "chat",
19848
+ "input_cost_per_token": 7e-07,
19806
19849
  "output_cost_per_token": 2.4e-06,
19850
+ "litellm_provider": "deepinfra",
19851
+ "mode": "chat",
19807
19852
  "supports_tool_choice": true
19808
19853
  },
19809
19854
  "deepinfra/mistralai/Mistral-7B-Instruct-v0.3": {
@@ -19817,63 +19862,63 @@
19817
19862
  "supports_tool_choice": true
19818
19863
  },
19819
19864
  "deepinfra/Qwen/Qwen2.5-72B-Instruct": {
19820
- "input_cost_per_token": 1.2e-07,
19821
- "litellm_provider": "deepinfra",
19865
+ "max_tokens": 32768,
19822
19866
  "max_input_tokens": 32768,
19823
19867
  "max_output_tokens": 32768,
19824
- "max_tokens": 32768,
19825
- "mode": "chat",
19868
+ "input_cost_per_token": 1.2e-07,
19826
19869
  "output_cost_per_token": 3.9e-07,
19870
+ "litellm_provider": "deepinfra",
19871
+ "mode": "chat",
19827
19872
  "supports_tool_choice": true
19828
19873
  },
19829
19874
  "deepinfra/Qwen/Qwen3-14B": {
19830
- "input_cost_per_token": 6e-08,
19831
- "litellm_provider": "deepinfra",
19875
+ "max_tokens": 40960,
19832
19876
  "max_input_tokens": 40960,
19833
19877
  "max_output_tokens": 40960,
19834
- "max_tokens": 40960,
19835
- "mode": "chat",
19878
+ "input_cost_per_token": 6e-08,
19836
19879
  "output_cost_per_token": 2.4e-07,
19880
+ "litellm_provider": "deepinfra",
19881
+ "mode": "chat",
19837
19882
  "supports_tool_choice": true
19838
19883
  },
19839
19884
  "deepinfra/allenai/olmOCR-7B-0725-FP8": {
19840
- "input_cost_per_token": 2.7e-07,
19841
- "litellm_provider": "deepinfra",
19885
+ "max_tokens": 16384,
19842
19886
  "max_input_tokens": 16384,
19843
19887
  "max_output_tokens": 16384,
19844
- "max_tokens": 16384,
19845
- "mode": "chat",
19888
+ "input_cost_per_token": 2.7e-07,
19846
19889
  "output_cost_per_token": 1.5e-06,
19890
+ "litellm_provider": "deepinfra",
19891
+ "mode": "chat",
19847
19892
  "supports_tool_choice": false
19848
19893
  },
19849
19894
  "deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct": {
19850
- "input_cost_per_token": 4e-07,
19851
- "litellm_provider": "deepinfra",
19895
+ "max_tokens": 262144,
19852
19896
  "max_input_tokens": 262144,
19853
19897
  "max_output_tokens": 262144,
19854
- "max_tokens": 262144,
19855
- "mode": "chat",
19898
+ "input_cost_per_token": 4e-07,
19856
19899
  "output_cost_per_token": 1.6e-06,
19900
+ "litellm_provider": "deepinfra",
19901
+ "mode": "chat",
19857
19902
  "supports_tool_choice": true
19858
19903
  },
19859
19904
  "deepinfra/microsoft/phi-4": {
19860
- "input_cost_per_token": 7e-08,
19861
- "litellm_provider": "deepinfra",
19905
+ "max_tokens": 16384,
19862
19906
  "max_input_tokens": 16384,
19863
19907
  "max_output_tokens": 16384,
19864
- "max_tokens": 16384,
19865
- "mode": "chat",
19908
+ "input_cost_per_token": 7e-08,
19866
19909
  "output_cost_per_token": 1.4e-07,
19910
+ "litellm_provider": "deepinfra",
19911
+ "mode": "chat",
19867
19912
  "supports_tool_choice": true
19868
19913
  },
19869
19914
  "deepinfra/NousResearch/Hermes-3-Llama-3.1-405B": {
19870
- "input_cost_per_token": 7e-07,
19871
- "litellm_provider": "deepinfra",
19915
+ "max_tokens": 131072,
19872
19916
  "max_input_tokens": 131072,
19873
19917
  "max_output_tokens": 131072,
19874
- "max_tokens": 131072,
19918
+ "input_cost_per_token": 1e-06,
19919
+ "output_cost_per_token": 1e-06,
19920
+ "litellm_provider": "deepinfra",
19875
19921
  "mode": "chat",
19876
- "output_cost_per_token": 8e-07,
19877
19922
  "supports_tool_choice": true
19878
19923
  },
19879
19924
  "deepinfra/zai-org/GLM-4.5-Air": {
@@ -19887,23 +19932,23 @@
19887
19932
  "supports_tool_choice": true
19888
19933
  },
19889
19934
  "deepinfra/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B": {
19890
- "input_cost_per_token": 7.5e-08,
19891
- "litellm_provider": "deepinfra",
19935
+ "max_tokens": 131072,
19892
19936
  "max_input_tokens": 131072,
19893
19937
  "max_output_tokens": 131072,
19894
- "max_tokens": 131072,
19938
+ "input_cost_per_token": 2.7e-07,
19939
+ "output_cost_per_token": 2.7e-07,
19940
+ "litellm_provider": "deepinfra",
19895
19941
  "mode": "chat",
19896
- "output_cost_per_token": 1.5e-07,
19897
19942
  "supports_tool_choice": true
19898
19943
  },
19899
19944
  "deepinfra/openai/gpt-oss-120b": {
19900
- "input_cost_per_token": 9e-08,
19901
- "litellm_provider": "deepinfra",
19945
+ "max_tokens": 131072,
19902
19946
  "max_input_tokens": 131072,
19903
19947
  "max_output_tokens": 131072,
19904
- "max_tokens": 131072,
19905
- "mode": "chat",
19948
+ "input_cost_per_token": 5e-08,
19906
19949
  "output_cost_per_token": 4.5e-07,
19950
+ "litellm_provider": "deepinfra",
19951
+ "mode": "chat",
19907
19952
  "supports_tool_choice": true
19908
19953
  },
19909
19954
  "deepinfra/google/codegemma-7b-it": {
@@ -19917,24 +19962,23 @@
19917
19962
  "supports_tool_choice": false
19918
19963
  },
19919
19964
  "deepinfra/Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo": {
19920
- "cache_read_input_token_cost": 2.4e-07,
19921
- "input_cost_per_token": 3e-07,
19922
- "litellm_provider": "deepinfra",
19965
+ "max_tokens": 262144,
19923
19966
  "max_input_tokens": 262144,
19924
19967
  "max_output_tokens": 262144,
19925
- "max_tokens": 262144,
19926
- "mode": "chat",
19968
+ "input_cost_per_token": 2.9e-07,
19927
19969
  "output_cost_per_token": 1.2e-06,
19970
+ "litellm_provider": "deepinfra",
19971
+ "mode": "chat",
19928
19972
  "supports_tool_choice": true
19929
19973
  },
19930
19974
  "deepinfra/mistralai/Mistral-Nemo-Instruct-2407": {
19931
- "input_cost_per_token": 2e-08,
19932
- "litellm_provider": "deepinfra",
19975
+ "max_tokens": 131072,
19933
19976
  "max_input_tokens": 131072,
19934
19977
  "max_output_tokens": 131072,
19935
- "max_tokens": 131072,
19936
- "mode": "chat",
19978
+ "input_cost_per_token": 2e-08,
19937
19979
  "output_cost_per_token": 4e-08,
19980
+ "litellm_provider": "deepinfra",
19981
+ "mode": "chat",
19938
19982
  "supports_tool_choice": true
19939
19983
  },
19940
19984
  "deepinfra/openbmb/MiniCPM-Llama3-V-2_5": {
@@ -19958,33 +20002,33 @@
19958
20002
  "supports_tool_choice": false
19959
20003
  },
19960
20004
  "deepinfra/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
19961
- "input_cost_per_token": 1.5e-07,
19962
- "litellm_provider": "deepinfra",
20005
+ "max_tokens": 1048576,
19963
20006
  "max_input_tokens": 1048576,
19964
20007
  "max_output_tokens": 1048576,
19965
- "max_tokens": 1048576,
19966
- "mode": "chat",
20008
+ "input_cost_per_token": 1.5e-07,
19967
20009
  "output_cost_per_token": 6e-07,
20010
+ "litellm_provider": "deepinfra",
20011
+ "mode": "chat",
19968
20012
  "supports_tool_choice": true
19969
20013
  },
19970
20014
  "deepinfra/meta-llama/Llama-4-Scout-17B-16E-Instruct": {
19971
- "input_cost_per_token": 8e-08,
19972
- "litellm_provider": "deepinfra",
20015
+ "max_tokens": 327680,
19973
20016
  "max_input_tokens": 327680,
19974
20017
  "max_output_tokens": 327680,
19975
- "max_tokens": 327680,
19976
- "mode": "chat",
20018
+ "input_cost_per_token": 8e-08,
19977
20019
  "output_cost_per_token": 3e-07,
20020
+ "litellm_provider": "deepinfra",
20021
+ "mode": "chat",
19978
20022
  "supports_tool_choice": true
19979
20023
  },
19980
20024
  "deepinfra/google/gemini-2.0-flash-001": {
19981
- "input_cost_per_token": 1e-07,
19982
- "litellm_provider": "deepinfra",
20025
+ "max_tokens": 1000000,
19983
20026
  "max_input_tokens": 1000000,
19984
20027
  "max_output_tokens": 1000000,
19985
- "max_tokens": 1000000,
19986
- "mode": "chat",
20028
+ "input_cost_per_token": 1e-07,
19987
20029
  "output_cost_per_token": 4e-07,
20030
+ "litellm_provider": "deepinfra",
20031
+ "mode": "chat",
19988
20032
  "supports_tool_choice": true
19989
20033
  },
19990
20034
  "deepinfra/Gryphe/MythoMax-L2-13b-turbo": {
@@ -20008,43 +20052,43 @@
20008
20052
  "supports_tool_choice": true
20009
20053
  },
20010
20054
  "deepinfra/meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo": {
20011
- "input_cost_per_token": 1e-07,
20012
- "litellm_provider": "deepinfra",
20055
+ "max_tokens": 131072,
20013
20056
  "max_input_tokens": 131072,
20014
20057
  "max_output_tokens": 131072,
20015
- "max_tokens": 131072,
20016
- "mode": "chat",
20058
+ "input_cost_per_token": 1e-07,
20017
20059
  "output_cost_per_token": 2.8e-07,
20060
+ "litellm_provider": "deepinfra",
20061
+ "mode": "chat",
20018
20062
  "supports_tool_choice": true
20019
20063
  },
20020
20064
  "deepinfra/meta-llama/Meta-Llama-3.1-8B-Instruct": {
20021
- "input_cost_per_token": 3e-08,
20022
- "litellm_provider": "deepinfra",
20065
+ "max_tokens": 131072,
20023
20066
  "max_input_tokens": 131072,
20024
20067
  "max_output_tokens": 131072,
20025
- "max_tokens": 131072,
20026
- "mode": "chat",
20068
+ "input_cost_per_token": 3e-08,
20027
20069
  "output_cost_per_token": 5e-08,
20070
+ "litellm_provider": "deepinfra",
20071
+ "mode": "chat",
20028
20072
  "supports_tool_choice": true
20029
20073
  },
20030
20074
  "deepinfra/Qwen/Qwen3-32B": {
20031
- "input_cost_per_token": 1e-07,
20032
- "litellm_provider": "deepinfra",
20075
+ "max_tokens": 40960,
20033
20076
  "max_input_tokens": 40960,
20034
20077
  "max_output_tokens": 40960,
20035
- "max_tokens": 40960,
20078
+ "input_cost_per_token": 1e-07,
20079
+ "output_cost_per_token": 2.8e-07,
20080
+ "litellm_provider": "deepinfra",
20036
20081
  "mode": "chat",
20037
- "output_cost_per_token": 3e-07,
20038
20082
  "supports_tool_choice": true
20039
20083
  },
20040
20084
  "deepinfra/Qwen/Qwen3-235B-A22B-Thinking-2507": {
20041
- "input_cost_per_token": 1.3e-07,
20042
- "litellm_provider": "deepinfra",
20085
+ "max_tokens": 262144,
20043
20086
  "max_input_tokens": 262144,
20044
20087
  "max_output_tokens": 262144,
20045
- "max_tokens": 262144,
20088
+ "input_cost_per_token": 3e-07,
20089
+ "output_cost_per_token": 2.9e-06,
20090
+ "litellm_provider": "deepinfra",
20046
20091
  "mode": "chat",
20047
- "output_cost_per_token": 6e-07,
20048
20092
  "supports_tool_choice": true
20049
20093
  },
20050
20094
  "deepinfra/nvidia/Nemotron-4-340B-Instruct": {
@@ -20058,24 +20102,24 @@
20058
20102
  "supports_tool_choice": true
20059
20103
  },
20060
20104
  "deepinfra/deepseek-ai/DeepSeek-R1-0528": {
20061
- "cache_read_input_token_cost": 4e-07,
20062
- "input_cost_per_token": 5e-07,
20063
- "litellm_provider": "deepinfra",
20105
+ "max_tokens": 163840,
20064
20106
  "max_input_tokens": 163840,
20065
20107
  "max_output_tokens": 163840,
20066
- "max_tokens": 163840,
20067
- "mode": "chat",
20108
+ "input_cost_per_token": 5e-07,
20068
20109
  "output_cost_per_token": 2.15e-06,
20110
+ "cache_read_input_token_cost": 4e-07,
20111
+ "litellm_provider": "deepinfra",
20112
+ "mode": "chat",
20069
20113
  "supports_tool_choice": true
20070
20114
  },
20071
20115
  "deepinfra/deepseek-ai/DeepSeek-R1-Turbo": {
20072
- "input_cost_per_token": 1e-06,
20073
- "litellm_provider": "deepinfra",
20116
+ "max_tokens": 40960,
20074
20117
  "max_input_tokens": 40960,
20075
20118
  "max_output_tokens": 40960,
20076
- "max_tokens": 40960,
20077
- "mode": "chat",
20119
+ "input_cost_per_token": 1e-06,
20078
20120
  "output_cost_per_token": 3e-06,
20121
+ "litellm_provider": "deepinfra",
20122
+ "mode": "chat",
20079
20123
  "supports_tool_choice": true
20080
20124
  },
20081
20125
  "deepinfra/NovaSky-AI/Sky-T1-32B-Preview": {
@@ -20089,13 +20133,13 @@
20089
20133
  "supports_tool_choice": false
20090
20134
  },
20091
20135
  "deepinfra/Qwen/Qwen3-235B-A22B-Instruct-2507": {
20092
- "input_cost_per_token": 1.3e-07,
20093
- "litellm_provider": "deepinfra",
20136
+ "max_tokens": 262144,
20094
20137
  "max_input_tokens": 262144,
20095
20138
  "max_output_tokens": 262144,
20096
- "max_tokens": 262144,
20097
- "mode": "chat",
20139
+ "input_cost_per_token": 9e-08,
20098
20140
  "output_cost_per_token": 6e-07,
20141
+ "litellm_provider": "deepinfra",
20142
+ "mode": "chat",
20099
20143
  "supports_tool_choice": true
20100
20144
  },
20101
20145
  "deepinfra/mistralai/Mistral-Small-3.1-24B-Instruct-2503": {
@@ -20139,13 +20183,13 @@
20139
20183
  "supports_tool_choice": false
20140
20184
  },
20141
20185
  "deepinfra/google/gemma-3-4b-it": {
20142
- "input_cost_per_token": 4e-08,
20143
- "litellm_provider": "deepinfra",
20186
+ "max_tokens": 131072,
20144
20187
  "max_input_tokens": 131072,
20145
20188
  "max_output_tokens": 131072,
20146
- "max_tokens": 131072,
20147
- "mode": "chat",
20189
+ "input_cost_per_token": 4e-08,
20148
20190
  "output_cost_per_token": 8e-08,
20191
+ "litellm_provider": "deepinfra",
20192
+ "mode": "chat",
20149
20193
  "supports_tool_choice": true
20150
20194
  },
20151
20195
  "deepinfra/mistralai/Mistral-7B-Instruct-v0.2": {
@@ -20159,13 +20203,13 @@
20159
20203
  "supports_tool_choice": false
20160
20204
  },
20161
20205
  "deepinfra/meta-llama/Llama-3.3-70B-Instruct-Turbo": {
20162
- "input_cost_per_token": 3.8e-08,
20163
- "litellm_provider": "deepinfra",
20206
+ "max_tokens": 131072,
20164
20207
  "max_input_tokens": 131072,
20165
20208
  "max_output_tokens": 131072,
20166
- "max_tokens": 131072,
20209
+ "input_cost_per_token": 1.3e-07,
20210
+ "output_cost_per_token": 3.9e-07,
20211
+ "litellm_provider": "deepinfra",
20167
20212
  "mode": "chat",
20168
- "output_cost_per_token": 1.2e-07,
20169
20213
  "supports_tool_choice": true
20170
20214
  },
20171
20215
  "deepinfra/mistralai/Devstral-Small-2505": {
@@ -20179,23 +20223,23 @@
20179
20223
  "supports_tool_choice": true
20180
20224
  },
20181
20225
  "deepinfra/meta-llama/Llama-3.2-11B-Vision-Instruct": {
20182
- "input_cost_per_token": 4.9e-08,
20183
- "litellm_provider": "deepinfra",
20226
+ "max_tokens": 131072,
20184
20227
  "max_input_tokens": 131072,
20185
20228
  "max_output_tokens": 131072,
20186
- "max_tokens": 131072,
20187
- "mode": "chat",
20229
+ "input_cost_per_token": 4.9e-08,
20188
20230
  "output_cost_per_token": 4.9e-08,
20231
+ "litellm_provider": "deepinfra",
20232
+ "mode": "chat",
20189
20233
  "supports_tool_choice": false
20190
20234
  },
20191
20235
  "deepinfra/openai/gpt-oss-20b": {
20192
- "input_cost_per_token": 4e-08,
20193
- "litellm_provider": "deepinfra",
20236
+ "max_tokens": 131072,
20194
20237
  "max_input_tokens": 131072,
20195
20238
  "max_output_tokens": 131072,
20196
- "max_tokens": 131072,
20239
+ "input_cost_per_token": 4e-08,
20240
+ "output_cost_per_token": 1.5e-07,
20241
+ "litellm_provider": "deepinfra",
20197
20242
  "mode": "chat",
20198
- "output_cost_per_token": 1.6e-07,
20199
20243
  "supports_tool_choice": true
20200
20244
  },
20201
20245
  "voyage/voyage-context-3": {
@@ -20431,6 +20475,8 @@
20431
20475
  },
20432
20476
  "openrouter/anthropic/claude-opus-4": {
20433
20477
  "input_cost_per_image": 0.0048,
20478
+ "cache_creation_input_token_cost": 1.875e-05,
20479
+ "cache_read_input_token_cost": 1.5e-06,
20434
20480
  "input_cost_per_token": 1.5e-05,
20435
20481
  "litellm_provider": "openrouter",
20436
20482
  "max_input_tokens": 200000,
@@ -20441,6 +20487,7 @@
20441
20487
  "supports_assistant_prefill": true,
20442
20488
  "supports_computer_use": true,
20443
20489
  "supports_function_calling": true,
20490
+ "supports_prompt_caching": true,
20444
20491
  "supports_reasoning": true,
20445
20492
  "supports_tool_choice": true,
20446
20493
  "supports_vision": true,
@@ -20448,6 +20495,9 @@
20448
20495
  },
20449
20496
  "openrouter/anthropic/claude-opus-4.1": {
20450
20497
  "input_cost_per_image": 0.0048,
20498
+ "cache_creation_input_token_cost": 1.875e-05,
20499
+ "cache_creation_input_token_cost_above_1hr": 3e-05,
20500
+ "cache_read_input_token_cost": 1.5e-06,
20451
20501
  "input_cost_per_token": 1.5e-05,
20452
20502
  "litellm_provider": "openrouter",
20453
20503
  "max_input_tokens": 200000,
@@ -20458,6 +20508,7 @@
20458
20508
  "supports_assistant_prefill": true,
20459
20509
  "supports_computer_use": true,
20460
20510
  "supports_function_calling": true,
20511
+ "supports_prompt_caching": true,
20461
20512
  "supports_reasoning": true,
20462
20513
  "supports_tool_choice": true,
20463
20514
  "supports_vision": true,
@@ -22719,11 +22770,12 @@
22719
22770
  "max_tokens": 2000000.0,
22720
22771
  "mode": "chat",
22721
22772
  "input_cost_per_token": 2e-07,
22773
+ "input_cost_per_token_above_128k_tokens": 4e-07,
22722
22774
  "output_cost_per_token": 5e-07,
22775
+ "output_cost_per_token_above_128k_tokens": 1e-06,
22723
22776
  "cache_read_input_token_cost": 5e-08,
22724
22777
  "source": "https://docs.x.ai/docs/models",
22725
22778
  "supports_function_calling": true,
22726
- "supports_reasoning": true,
22727
22779
  "supports_tool_choice": true,
22728
22780
  "supports_web_search": true
22729
22781
  },
@@ -22735,7 +22787,9 @@
22735
22787
  "max_tokens": 2000000.0,
22736
22788
  "mode": "chat",
22737
22789
  "input_cost_per_token": 2e-07,
22790
+ "input_cost_per_token_above_128k_tokens": 4e-07,
22738
22791
  "output_cost_per_token": 5e-07,
22792
+ "output_cost_per_token_above_128k_tokens": 1e-06,
22739
22793
  "source": "https://docs.x.ai/docs/models",
22740
22794
  "supports_function_calling": true,
22741
22795
  "supports_tool_choice": true,
@@ -22825,11 +22879,11 @@
22825
22879
  "text"
22826
22880
  ],
22827
22881
  "supports_function_calling": true,
22828
- "supports_native_streaming": false,
22882
+ "supports_native_streaming": true,
22829
22883
  "supports_parallel_function_calling": true,
22830
22884
  "supports_pdf_input": true,
22831
22885
  "supports_prompt_caching": true,
22832
- "supports_reasoning": false,
22886
+ "supports_reasoning": true,
22833
22887
  "supports_response_schema": true,
22834
22888
  "supports_system_messages": false,
22835
22889
  "supports_tool_choice": true,
@@ -23363,10 +23417,14 @@
23363
23417
  "cache_creation_input_token_cost": 3.75e-06,
23364
23418
  "cache_read_input_token_cost": 3e-07,
23365
23419
  "input_cost_per_token": 3e-06,
23420
+ "input_cost_per_token_above_200k_tokens": 6e-06,
23421
+ "output_cost_per_token_above_200k_tokens": 2.25e-05,
23422
+ "cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
23423
+ "cache_read_input_token_cost_above_200k_tokens": 6e-07,
23366
23424
  "litellm_provider": "anthropic",
23367
23425
  "max_input_tokens": 200000,
23368
23426
  "max_output_tokens": 64000,
23369
- "max_tokens": 200000,
23427
+ "max_tokens": 64000,
23370
23428
  "mode": "chat",
23371
23429
  "output_cost_per_token": 1.5e-05,
23372
23430
  "search_context_cost_per_query": {
@@ -23383,6 +23441,7 @@
23383
23441
  "supports_response_schema": true,
23384
23442
  "supports_tool_choice": true,
23385
23443
  "supports_vision": true,
23444
+ "supports_web_search": true,
23386
23445
  "tool_use_system_prompt_tokens": 346
23387
23446
  },
23388
23447
  "ollama/deepseek-v3.1:671b-cloud": {
@@ -23426,15 +23485,19 @@
23426
23485
  "supports_function_calling": true
23427
23486
  },
23428
23487
  "us.anthropic.claude-sonnet-4-5-20250929-v1:0": {
23429
- "cache_creation_input_token_cost": 3.75e-06,
23430
- "cache_read_input_token_cost": 3e-07,
23431
- "input_cost_per_token": 3e-06,
23488
+ "cache_creation_input_token_cost": 4.125e-06,
23489
+ "cache_read_input_token_cost": 3.3e-07,
23490
+ "input_cost_per_token": 3.3e-06,
23491
+ "input_cost_per_token_above_200k_tokens": 6.6e-06,
23492
+ "output_cost_per_token_above_200k_tokens": 2.475e-05,
23493
+ "cache_creation_input_token_cost_above_200k_tokens": 8.25e-06,
23494
+ "cache_read_input_token_cost_above_200k_tokens": 6.6e-07,
23432
23495
  "litellm_provider": "bedrock_converse",
23433
23496
  "max_input_tokens": 200000,
23434
23497
  "max_output_tokens": 64000,
23435
- "max_tokens": 200000,
23498
+ "max_tokens": 64000,
23436
23499
  "mode": "chat",
23437
- "output_cost_per_token": 1.5e-05,
23500
+ "output_cost_per_token": 1.65e-05,
23438
23501
  "search_context_cost_per_query": {
23439
23502
  "search_context_size_high": 0.01,
23440
23503
  "search_context_size_low": 0.01,
@@ -23455,11 +23518,15 @@
23455
23518
  "cache_creation_input_token_cost": 3.75e-06,
23456
23519
  "cache_read_input_token_cost": 3e-07,
23457
23520
  "input_cost_per_token": 3e-06,
23521
+ "input_cost_per_token_above_200k_tokens": 6e-06,
23522
+ "output_cost_per_token_above_200k_tokens": 2.25e-05,
23523
+ "cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
23524
+ "cache_read_input_token_cost_above_200k_tokens": 6e-07,
23458
23525
  "input_cost_per_token_batches": 1.5e-06,
23459
23526
  "litellm_provider": "vertex_ai-anthropic_models",
23460
23527
  "max_input_tokens": 200000,
23461
23528
  "max_output_tokens": 64000,
23462
- "max_tokens": 200000,
23529
+ "max_tokens": 64000,
23463
23530
  "mode": "chat",
23464
23531
  "output_cost_per_token": 1.5e-05,
23465
23532
  "output_cost_per_token_batches": 7.5e-06,
@@ -23477,11 +23544,15 @@
23477
23544
  "cache_creation_input_token_cost": 3.75e-06,
23478
23545
  "cache_read_input_token_cost": 3e-07,
23479
23546
  "input_cost_per_token": 3e-06,
23547
+ "input_cost_per_token_above_200k_tokens": 6e-06,
23548
+ "output_cost_per_token_above_200k_tokens": 2.25e-05,
23549
+ "cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
23550
+ "cache_read_input_token_cost_above_200k_tokens": 6e-07,
23480
23551
  "input_cost_per_token_batches": 1.5e-06,
23481
23552
  "litellm_provider": "vertex_ai-anthropic_models",
23482
23553
  "max_input_tokens": 200000,
23483
23554
  "max_output_tokens": 64000,
23484
- "max_tokens": 200000,
23555
+ "max_tokens": 64000,
23485
23556
  "mode": "chat",
23486
23557
  "output_cost_per_token": 1.5e-05,
23487
23558
  "output_cost_per_token_batches": 7.5e-06,
@@ -23499,10 +23570,14 @@
23499
23570
  "cache_creation_input_token_cost": 3.75e-06,
23500
23571
  "cache_read_input_token_cost": 3e-07,
23501
23572
  "input_cost_per_token": 3e-06,
23573
+ "input_cost_per_token_above_200k_tokens": 6e-06,
23574
+ "output_cost_per_token_above_200k_tokens": 2.25e-05,
23575
+ "cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
23576
+ "cache_read_input_token_cost_above_200k_tokens": 6e-07,
23502
23577
  "litellm_provider": "anthropic",
23503
23578
  "max_input_tokens": 200000,
23504
23579
  "max_output_tokens": 64000,
23505
- "max_tokens": 200000,
23580
+ "max_tokens": 64000,
23506
23581
  "mode": "chat",
23507
23582
  "output_cost_per_token": 1.5e-05,
23508
23583
  "search_context_cost_per_query": {
@@ -23524,8 +23599,8 @@
23524
23599
  "lemonade/Qwen3-Coder-30B-A3B-Instruct-GGUF": {
23525
23600
  "input_cost_per_token": 0,
23526
23601
  "litellm_provider": "lemonade",
23527
- "max_tokens": 32768,
23528
- "max_input_tokens": 32768,
23602
+ "max_tokens": 262144,
23603
+ "max_input_tokens": 262144,
23529
23604
  "max_output_tokens": 32768,
23530
23605
  "mode": "chat",
23531
23606
  "output_cost_per_token": 0,
@@ -23545,5 +23620,2094 @@
23545
23620
  "supports_function_calling": true,
23546
23621
  "supports_response_schema": true,
23547
23622
  "supports_tool_choice": true
23623
+ },
23624
+ "azure_ai/grok-4": {
23625
+ "input_cost_per_token": 5.5e-06,
23626
+ "litellm_provider": "azure_ai",
23627
+ "max_input_tokens": 131072,
23628
+ "max_output_tokens": 131072,
23629
+ "max_tokens": 131072,
23630
+ "mode": "chat",
23631
+ "output_cost_per_token": 2.75e-05,
23632
+ "source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/",
23633
+ "supports_function_calling": true,
23634
+ "supports_response_schema": true,
23635
+ "supports_tool_choice": true,
23636
+ "supports_web_search": true
23637
+ },
23638
+ "azure_ai/grok-4-fast-non-reasoning": {
23639
+ "input_cost_per_token": 4.3e-07,
23640
+ "output_cost_per_token": 1.73e-06,
23641
+ "litellm_provider": "azure_ai",
23642
+ "max_input_tokens": 131072,
23643
+ "max_output_tokens": 131072,
23644
+ "max_tokens": 131072,
23645
+ "mode": "chat",
23646
+ "supports_function_calling": true,
23647
+ "supports_response_schema": true,
23648
+ "supports_tool_choice": true,
23649
+ "supports_web_search": true
23650
+ },
23651
+ "azure_ai/grok-4-fast-reasoning": {
23652
+ "input_cost_per_token": 4.3e-07,
23653
+ "output_cost_per_token": 1.73e-06,
23654
+ "litellm_provider": "azure_ai",
23655
+ "max_input_tokens": 131072,
23656
+ "max_output_tokens": 131072,
23657
+ "max_tokens": 131072,
23658
+ "mode": "chat",
23659
+ "source": "https://techcommunity.microsoft.com/blog/azure-ai-foundry-blog/announcing-the-grok-4-fast-models-from-xai-now-available-in-azure-ai-foundry/4456701",
23660
+ "supports_function_calling": true,
23661
+ "supports_response_schema": true,
23662
+ "supports_tool_choice": true,
23663
+ "supports_web_search": true
23664
+ },
23665
+ "azure_ai/grok-code-fast-1": {
23666
+ "input_cost_per_token": 3.5e-06,
23667
+ "litellm_provider": "azure_ai",
23668
+ "max_input_tokens": 131072,
23669
+ "max_output_tokens": 131072,
23670
+ "max_tokens": 131072,
23671
+ "mode": "chat",
23672
+ "output_cost_per_token": 1.75e-05,
23673
+ "source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/",
23674
+ "supports_function_calling": true,
23675
+ "supports_response_schema": true,
23676
+ "supports_tool_choice": true,
23677
+ "supports_web_search": true
23678
+ },
23679
+ "eu.anthropic.claude-sonnet-4-5-20250929-v1:0": {
23680
+ "cache_creation_input_token_cost": 4.125e-06,
23681
+ "cache_read_input_token_cost": 3.3e-07,
23682
+ "input_cost_per_token": 3.3e-06,
23683
+ "input_cost_per_token_above_200k_tokens": 6.6e-06,
23684
+ "output_cost_per_token_above_200k_tokens": 2.475e-05,
23685
+ "cache_creation_input_token_cost_above_200k_tokens": 8.25e-06,
23686
+ "cache_read_input_token_cost_above_200k_tokens": 6.6e-07,
23687
+ "litellm_provider": "bedrock_converse",
23688
+ "max_input_tokens": 200000,
23689
+ "max_output_tokens": 64000,
23690
+ "max_tokens": 64000,
23691
+ "mode": "chat",
23692
+ "output_cost_per_token": 1.65e-05,
23693
+ "search_context_cost_per_query": {
23694
+ "search_context_size_high": 0.01,
23695
+ "search_context_size_low": 0.01,
23696
+ "search_context_size_medium": 0.01
23697
+ },
23698
+ "supports_assistant_prefill": true,
23699
+ "supports_computer_use": true,
23700
+ "supports_function_calling": true,
23701
+ "supports_pdf_input": true,
23702
+ "supports_prompt_caching": true,
23703
+ "supports_reasoning": true,
23704
+ "supports_response_schema": true,
23705
+ "supports_tool_choice": true,
23706
+ "supports_vision": true,
23707
+ "tool_use_system_prompt_tokens": 346
23708
+ },
23709
+ "nvidia_nim/nvidia/nv-rerankqa-mistral-4b-v3": {
23710
+ "input_cost_per_query": 0.0,
23711
+ "input_cost_per_token": 0.0,
23712
+ "litellm_provider": "nvidia_nim",
23713
+ "mode": "rerank",
23714
+ "output_cost_per_token": 0.0
23715
+ },
23716
+ "nvidia_nim/nvidia/llama-3_2-nv-rerankqa-1b-v2": {
23717
+ "input_cost_per_query": 0.0,
23718
+ "input_cost_per_token": 0.0,
23719
+ "litellm_provider": "nvidia_nim",
23720
+ "mode": "rerank",
23721
+ "output_cost_per_token": 0.0
23722
+ },
23723
+ "deepinfra/Qwen/Qwen3-Next-80B-A3B-Instruct": {
23724
+ "max_tokens": 262144,
23725
+ "max_input_tokens": 262144,
23726
+ "max_output_tokens": 262144,
23727
+ "input_cost_per_token": 1.4e-07,
23728
+ "output_cost_per_token": 1.4e-06,
23729
+ "litellm_provider": "deepinfra",
23730
+ "mode": "chat",
23731
+ "supports_tool_choice": true
23732
+ },
23733
+ "deepinfra/Qwen/Qwen3-Next-80B-A3B-Thinking": {
23734
+ "max_tokens": 262144,
23735
+ "max_input_tokens": 262144,
23736
+ "max_output_tokens": 262144,
23737
+ "input_cost_per_token": 1.4e-07,
23738
+ "output_cost_per_token": 1.4e-06,
23739
+ "litellm_provider": "deepinfra",
23740
+ "mode": "chat",
23741
+ "supports_tool_choice": true
23742
+ },
23743
+ "deepinfra/deepseek-ai/DeepSeek-V3.1-Terminus": {
23744
+ "max_tokens": 163840,
23745
+ "max_input_tokens": 163840,
23746
+ "max_output_tokens": 163840,
23747
+ "input_cost_per_token": 2.7e-07,
23748
+ "output_cost_per_token": 1e-06,
23749
+ "cache_read_input_token_cost": 2.16e-07,
23750
+ "litellm_provider": "deepinfra",
23751
+ "mode": "chat",
23752
+ "supports_tool_choice": true
23753
+ },
23754
+ "deepinfra/moonshotai/Kimi-K2-Instruct-0905": {
23755
+ "max_tokens": 262144,
23756
+ "max_input_tokens": 262144,
23757
+ "max_output_tokens": 262144,
23758
+ "input_cost_per_token": 5e-07,
23759
+ "output_cost_per_token": 2e-06,
23760
+ "cache_read_input_token_cost": 4e-07,
23761
+ "litellm_provider": "deepinfra",
23762
+ "mode": "chat",
23763
+ "supports_tool_choice": true
23764
+ },
23765
+ "deepinfra/nvidia/Llama-3.3-Nemotron-Super-49B-v1.5": {
23766
+ "max_tokens": 131072,
23767
+ "max_input_tokens": 131072,
23768
+ "max_output_tokens": 131072,
23769
+ "input_cost_per_token": 1e-07,
23770
+ "output_cost_per_token": 4e-07,
23771
+ "litellm_provider": "deepinfra",
23772
+ "mode": "chat",
23773
+ "supports_tool_choice": true
23774
+ },
23775
+ "deepinfra/nvidia/NVIDIA-Nemotron-Nano-9B-v2": {
23776
+ "max_tokens": 131072,
23777
+ "max_input_tokens": 131072,
23778
+ "max_output_tokens": 131072,
23779
+ "input_cost_per_token": 4e-08,
23780
+ "output_cost_per_token": 1.6e-07,
23781
+ "litellm_provider": "deepinfra",
23782
+ "mode": "chat",
23783
+ "supports_tool_choice": true
23784
+ },
23785
+ "jp.anthropic.claude-sonnet-4-5-20250929-v1:0": {
23786
+ "cache_creation_input_token_cost": 4.125e-06,
23787
+ "cache_read_input_token_cost": 3.3e-07,
23788
+ "input_cost_per_token": 3.3e-06,
23789
+ "input_cost_per_token_above_200k_tokens": 6.6e-06,
23790
+ "output_cost_per_token_above_200k_tokens": 2.475e-05,
23791
+ "cache_creation_input_token_cost_above_200k_tokens": 8.25e-06,
23792
+ "cache_read_input_token_cost_above_200k_tokens": 6.6e-07,
23793
+ "litellm_provider": "bedrock_converse",
23794
+ "max_input_tokens": 200000,
23795
+ "max_output_tokens": 64000,
23796
+ "max_tokens": 64000,
23797
+ "mode": "chat",
23798
+ "output_cost_per_token": 1.65e-05,
23799
+ "search_context_cost_per_query": {
23800
+ "search_context_size_high": 0.01,
23801
+ "search_context_size_low": 0.01,
23802
+ "search_context_size_medium": 0.01
23803
+ },
23804
+ "supports_assistant_prefill": true,
23805
+ "supports_computer_use": true,
23806
+ "supports_function_calling": true,
23807
+ "supports_pdf_input": true,
23808
+ "supports_prompt_caching": true,
23809
+ "supports_reasoning": true,
23810
+ "supports_response_schema": true,
23811
+ "supports_tool_choice": true,
23812
+ "supports_vision": true,
23813
+ "tool_use_system_prompt_tokens": 346
23814
+ },
23815
+ "watsonx/bigscience/mt0-xxl-13b": {
23816
+ "max_tokens": 8192,
23817
+ "max_input_tokens": 8192,
23818
+ "max_output_tokens": 8192,
23819
+ "input_cost_per_token": 0.0005,
23820
+ "output_cost_per_token": 0.002,
23821
+ "litellm_provider": "watsonx",
23822
+ "mode": "chat",
23823
+ "supports_function_calling": false,
23824
+ "supports_parallel_function_calling": false,
23825
+ "supports_vision": false
23826
+ },
23827
+ "watsonx/core42/jais-13b-chat": {
23828
+ "max_tokens": 8192,
23829
+ "max_input_tokens": 8192,
23830
+ "max_output_tokens": 8192,
23831
+ "input_cost_per_token": 0.0005,
23832
+ "output_cost_per_token": 0.002,
23833
+ "litellm_provider": "watsonx",
23834
+ "mode": "chat",
23835
+ "supports_function_calling": false,
23836
+ "supports_parallel_function_calling": false,
23837
+ "supports_vision": false
23838
+ },
23839
+ "watsonx/google/flan-t5-xl-3b": {
23840
+ "max_tokens": 8192,
23841
+ "max_input_tokens": 8192,
23842
+ "max_output_tokens": 8192,
23843
+ "input_cost_per_token": 6e-07,
23844
+ "output_cost_per_token": 6e-07,
23845
+ "litellm_provider": "watsonx",
23846
+ "mode": "chat",
23847
+ "supports_function_calling": false,
23848
+ "supports_parallel_function_calling": false,
23849
+ "supports_vision": false
23850
+ },
23851
+ "watsonx/ibm/granite-13b-chat-v2": {
23852
+ "max_tokens": 8192,
23853
+ "max_input_tokens": 8192,
23854
+ "max_output_tokens": 8192,
23855
+ "input_cost_per_token": 6e-07,
23856
+ "output_cost_per_token": 6e-07,
23857
+ "litellm_provider": "watsonx",
23858
+ "mode": "chat",
23859
+ "supports_function_calling": false,
23860
+ "supports_parallel_function_calling": false,
23861
+ "supports_vision": false
23862
+ },
23863
+ "watsonx/ibm/granite-13b-instruct-v2": {
23864
+ "max_tokens": 8192,
23865
+ "max_input_tokens": 8192,
23866
+ "max_output_tokens": 8192,
23867
+ "input_cost_per_token": 6e-07,
23868
+ "output_cost_per_token": 6e-07,
23869
+ "litellm_provider": "watsonx",
23870
+ "mode": "chat",
23871
+ "supports_function_calling": false,
23872
+ "supports_parallel_function_calling": false,
23873
+ "supports_vision": false
23874
+ },
23875
+ "watsonx/ibm/granite-3-3-8b-instruct": {
23876
+ "max_tokens": 8192,
23877
+ "max_input_tokens": 8192,
23878
+ "max_output_tokens": 8192,
23879
+ "input_cost_per_token": 2e-07,
23880
+ "output_cost_per_token": 2e-07,
23881
+ "litellm_provider": "watsonx",
23882
+ "mode": "chat",
23883
+ "supports_function_calling": true,
23884
+ "supports_parallel_function_calling": true,
23885
+ "supports_vision": false
23886
+ },
23887
+ "watsonx/ibm/granite-4-h-small": {
23888
+ "max_tokens": 20480,
23889
+ "max_input_tokens": 20480,
23890
+ "max_output_tokens": 20480,
23891
+ "input_cost_per_token": 6e-08,
23892
+ "output_cost_per_token": 2.5e-07,
23893
+ "litellm_provider": "watsonx",
23894
+ "mode": "chat",
23895
+ "supports_function_calling": true,
23896
+ "supports_parallel_function_calling": true,
23897
+ "supports_vision": false
23898
+ },
23899
+ "watsonx/ibm/granite-guardian-3-2-2b": {
23900
+ "max_tokens": 8192,
23901
+ "max_input_tokens": 8192,
23902
+ "max_output_tokens": 8192,
23903
+ "input_cost_per_token": 1e-07,
23904
+ "output_cost_per_token": 1e-07,
23905
+ "litellm_provider": "watsonx",
23906
+ "mode": "chat",
23907
+ "supports_function_calling": false,
23908
+ "supports_parallel_function_calling": false,
23909
+ "supports_vision": false
23910
+ },
23911
+ "watsonx/ibm/granite-guardian-3-3-8b": {
23912
+ "max_tokens": 8192,
23913
+ "max_input_tokens": 8192,
23914
+ "max_output_tokens": 8192,
23915
+ "input_cost_per_token": 2e-07,
23916
+ "output_cost_per_token": 2e-07,
23917
+ "litellm_provider": "watsonx",
23918
+ "mode": "chat",
23919
+ "supports_function_calling": false,
23920
+ "supports_parallel_function_calling": false,
23921
+ "supports_vision": false
23922
+ },
23923
+ "watsonx/ibm/granite-ttm-1024-96-r2": {
23924
+ "max_tokens": 512,
23925
+ "max_input_tokens": 512,
23926
+ "max_output_tokens": 512,
23927
+ "input_cost_per_token": 3.8e-07,
23928
+ "output_cost_per_token": 3.8e-07,
23929
+ "litellm_provider": "watsonx",
23930
+ "mode": "chat",
23931
+ "supports_function_calling": false,
23932
+ "supports_parallel_function_calling": false,
23933
+ "supports_vision": false
23934
+ },
23935
+ "watsonx/ibm/granite-ttm-1536-96-r2": {
23936
+ "max_tokens": 512,
23937
+ "max_input_tokens": 512,
23938
+ "max_output_tokens": 512,
23939
+ "input_cost_per_token": 3.8e-07,
23940
+ "output_cost_per_token": 3.8e-07,
23941
+ "litellm_provider": "watsonx",
23942
+ "mode": "chat",
23943
+ "supports_function_calling": false,
23944
+ "supports_parallel_function_calling": false,
23945
+ "supports_vision": false
23946
+ },
23947
+ "watsonx/ibm/granite-ttm-512-96-r2": {
23948
+ "max_tokens": 512,
23949
+ "max_input_tokens": 512,
23950
+ "max_output_tokens": 512,
23951
+ "input_cost_per_token": 3.8e-07,
23952
+ "output_cost_per_token": 3.8e-07,
23953
+ "litellm_provider": "watsonx",
23954
+ "mode": "chat",
23955
+ "supports_function_calling": false,
23956
+ "supports_parallel_function_calling": false,
23957
+ "supports_vision": false
23958
+ },
23959
+ "watsonx/ibm/granite-vision-3-2-2b": {
23960
+ "max_tokens": 8192,
23961
+ "max_input_tokens": 8192,
23962
+ "max_output_tokens": 8192,
23963
+ "input_cost_per_token": 1e-07,
23964
+ "output_cost_per_token": 1e-07,
23965
+ "litellm_provider": "watsonx",
23966
+ "mode": "chat",
23967
+ "supports_function_calling": false,
23968
+ "supports_parallel_function_calling": false,
23969
+ "supports_vision": true
23970
+ },
23971
+ "watsonx/meta-llama/llama-3-2-11b-vision-instruct": {
23972
+ "max_tokens": 128000,
23973
+ "max_input_tokens": 128000,
23974
+ "max_output_tokens": 128000,
23975
+ "input_cost_per_token": 3.5e-07,
23976
+ "output_cost_per_token": 3.5e-07,
23977
+ "litellm_provider": "watsonx",
23978
+ "mode": "chat",
23979
+ "supports_function_calling": true,
23980
+ "supports_parallel_function_calling": true,
23981
+ "supports_vision": true
23982
+ },
23983
+ "watsonx/meta-llama/llama-3-2-1b-instruct": {
23984
+ "max_tokens": 128000,
23985
+ "max_input_tokens": 128000,
23986
+ "max_output_tokens": 128000,
23987
+ "input_cost_per_token": 1e-07,
23988
+ "output_cost_per_token": 1e-07,
23989
+ "litellm_provider": "watsonx",
23990
+ "mode": "chat",
23991
+ "supports_function_calling": true,
23992
+ "supports_parallel_function_calling": true,
23993
+ "supports_vision": false
23994
+ },
23995
+ "watsonx/meta-llama/llama-3-2-3b-instruct": {
23996
+ "max_tokens": 128000,
23997
+ "max_input_tokens": 128000,
23998
+ "max_output_tokens": 128000,
23999
+ "input_cost_per_token": 1.5e-07,
24000
+ "output_cost_per_token": 1.5e-07,
24001
+ "litellm_provider": "watsonx",
24002
+ "mode": "chat",
24003
+ "supports_function_calling": true,
24004
+ "supports_parallel_function_calling": true,
24005
+ "supports_vision": false
24006
+ },
24007
+ "watsonx/meta-llama/llama-3-2-90b-vision-instruct": {
24008
+ "max_tokens": 128000,
24009
+ "max_input_tokens": 128000,
24010
+ "max_output_tokens": 128000,
24011
+ "input_cost_per_token": 2e-06,
24012
+ "output_cost_per_token": 2e-06,
24013
+ "litellm_provider": "watsonx",
24014
+ "mode": "chat",
24015
+ "supports_function_calling": true,
24016
+ "supports_parallel_function_calling": true,
24017
+ "supports_vision": true
24018
+ },
24019
+ "watsonx/meta-llama/llama-3-3-70b-instruct": {
24020
+ "max_tokens": 128000,
24021
+ "max_input_tokens": 128000,
24022
+ "max_output_tokens": 128000,
24023
+ "input_cost_per_token": 7.1e-07,
24024
+ "output_cost_per_token": 7.1e-07,
24025
+ "litellm_provider": "watsonx",
24026
+ "mode": "chat",
24027
+ "supports_function_calling": true,
24028
+ "supports_parallel_function_calling": true,
24029
+ "supports_vision": false
24030
+ },
24031
+ "watsonx/meta-llama/llama-4-maverick-17b": {
24032
+ "max_tokens": 128000,
24033
+ "max_input_tokens": 128000,
24034
+ "max_output_tokens": 128000,
24035
+ "input_cost_per_token": 3.5e-07,
24036
+ "output_cost_per_token": 1.4e-06,
24037
+ "litellm_provider": "watsonx",
24038
+ "mode": "chat",
24039
+ "supports_function_calling": true,
24040
+ "supports_parallel_function_calling": true,
24041
+ "supports_vision": false
24042
+ },
24043
+ "watsonx/meta-llama/llama-guard-3-11b-vision": {
24044
+ "max_tokens": 128000,
24045
+ "max_input_tokens": 128000,
24046
+ "max_output_tokens": 128000,
24047
+ "input_cost_per_token": 3.5e-07,
24048
+ "output_cost_per_token": 3.5e-07,
24049
+ "litellm_provider": "watsonx",
24050
+ "mode": "chat",
24051
+ "supports_function_calling": false,
24052
+ "supports_parallel_function_calling": false,
24053
+ "supports_vision": true
24054
+ },
24055
+ "watsonx/mistralai/mistral-medium-2505": {
24056
+ "max_tokens": 128000,
24057
+ "max_input_tokens": 128000,
24058
+ "max_output_tokens": 128000,
24059
+ "input_cost_per_token": 3e-06,
24060
+ "output_cost_per_token": 1e-05,
24061
+ "litellm_provider": "watsonx",
24062
+ "mode": "chat",
24063
+ "supports_function_calling": true,
24064
+ "supports_parallel_function_calling": true,
24065
+ "supports_vision": false
24066
+ },
24067
+ "watsonx/mistralai/mistral-small-2503": {
24068
+ "max_tokens": 32000,
24069
+ "max_input_tokens": 32000,
24070
+ "max_output_tokens": 32000,
24071
+ "input_cost_per_token": 1e-07,
24072
+ "output_cost_per_token": 3e-07,
24073
+ "litellm_provider": "watsonx",
24074
+ "mode": "chat",
24075
+ "supports_function_calling": true,
24076
+ "supports_parallel_function_calling": true,
24077
+ "supports_vision": false
24078
+ },
24079
+ "watsonx/mistralai/pixtral-12b-2409": {
24080
+ "max_tokens": 128000,
24081
+ "max_input_tokens": 128000,
24082
+ "max_output_tokens": 128000,
24083
+ "input_cost_per_token": 3.5e-07,
24084
+ "output_cost_per_token": 3.5e-07,
24085
+ "litellm_provider": "watsonx",
24086
+ "mode": "chat",
24087
+ "supports_function_calling": false,
24088
+ "supports_parallel_function_calling": false,
24089
+ "supports_vision": true
24090
+ },
24091
+ "watsonx/openai/gpt-oss-120b": {
24092
+ "max_tokens": 8192,
24093
+ "max_input_tokens": 8192,
24094
+ "max_output_tokens": 8192,
24095
+ "input_cost_per_token": 1.5e-07,
24096
+ "output_cost_per_token": 6e-07,
24097
+ "litellm_provider": "watsonx",
24098
+ "mode": "chat",
24099
+ "supports_function_calling": false,
24100
+ "supports_parallel_function_calling": false,
24101
+ "supports_vision": false
24102
+ },
24103
+ "watsonx/sdaia/allam-1-13b-instruct": {
24104
+ "max_tokens": 8192,
24105
+ "max_input_tokens": 8192,
24106
+ "max_output_tokens": 8192,
24107
+ "input_cost_per_token": 1.8e-06,
24108
+ "output_cost_per_token": 1.8e-06,
24109
+ "litellm_provider": "watsonx",
24110
+ "mode": "chat",
24111
+ "supports_function_calling": false,
24112
+ "supports_parallel_function_calling": false,
24113
+ "supports_vision": false
24114
+ },
24115
+ "global.anthropic.claude-sonnet-4-5-20250929-v1:0": {
24116
+ "cache_creation_input_token_cost": 3.75e-06,
24117
+ "cache_read_input_token_cost": 3e-07,
24118
+ "input_cost_per_token": 3e-06,
24119
+ "input_cost_per_token_above_200k_tokens": 6e-06,
24120
+ "output_cost_per_token_above_200k_tokens": 2.25e-05,
24121
+ "cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
24122
+ "cache_read_input_token_cost_above_200k_tokens": 6e-07,
24123
+ "litellm_provider": "bedrock_converse",
24124
+ "max_input_tokens": 200000,
24125
+ "max_output_tokens": 64000,
24126
+ "max_tokens": 64000,
24127
+ "mode": "chat",
24128
+ "output_cost_per_token": 1.5e-05,
24129
+ "search_context_cost_per_query": {
24130
+ "search_context_size_high": 0.01,
24131
+ "search_context_size_low": 0.01,
24132
+ "search_context_size_medium": 0.01
24133
+ },
24134
+ "supports_assistant_prefill": true,
24135
+ "supports_computer_use": true,
24136
+ "supports_function_calling": true,
24137
+ "supports_pdf_input": true,
24138
+ "supports_prompt_caching": true,
24139
+ "supports_reasoning": true,
24140
+ "supports_response_schema": true,
24141
+ "supports_tool_choice": true,
24142
+ "supports_vision": true,
24143
+ "tool_use_system_prompt_tokens": 346
24144
+ },
24145
+ "global.anthropic.claude-sonnet-4-20250514-v1:0": {
24146
+ "cache_creation_input_token_cost": 3.75e-06,
24147
+ "cache_read_input_token_cost": 3e-07,
24148
+ "input_cost_per_token": 3e-06,
24149
+ "input_cost_per_token_above_200k_tokens": 6e-06,
24150
+ "output_cost_per_token_above_200k_tokens": 2.25e-05,
24151
+ "cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
24152
+ "cache_read_input_token_cost_above_200k_tokens": 6e-07,
24153
+ "litellm_provider": "bedrock_converse",
24154
+ "max_input_tokens": 1000000,
24155
+ "max_output_tokens": 64000,
24156
+ "max_tokens": 64000,
24157
+ "mode": "chat",
24158
+ "output_cost_per_token": 1.5e-05,
24159
+ "search_context_cost_per_query": {
24160
+ "search_context_size_high": 0.01,
24161
+ "search_context_size_low": 0.01,
24162
+ "search_context_size_medium": 0.01
24163
+ },
24164
+ "supports_assistant_prefill": true,
24165
+ "supports_computer_use": true,
24166
+ "supports_function_calling": true,
24167
+ "supports_pdf_input": true,
24168
+ "supports_prompt_caching": true,
24169
+ "supports_reasoning": true,
24170
+ "supports_response_schema": true,
24171
+ "supports_tool_choice": true,
24172
+ "supports_vision": true,
24173
+ "tool_use_system_prompt_tokens": 159
24174
+ },
24175
+ "gpt-5-pro": {
24176
+ "input_cost_per_token": 1.5e-05,
24177
+ "input_cost_per_token_batches": 7.5e-06,
24178
+ "litellm_provider": "openai",
24179
+ "max_input_tokens": 400000,
24180
+ "max_output_tokens": 272000,
24181
+ "max_tokens": 272000,
24182
+ "mode": "responses",
24183
+ "output_cost_per_token": 0.00012,
24184
+ "output_cost_per_token_batches": 6e-05,
24185
+ "supported_endpoints": [
24186
+ "/v1/batch",
24187
+ "/v1/responses"
24188
+ ],
24189
+ "supported_modalities": [
24190
+ "text",
24191
+ "image"
24192
+ ],
24193
+ "supported_output_modalities": [
24194
+ "text"
24195
+ ],
24196
+ "supports_function_calling": true,
24197
+ "supports_native_streaming": false,
24198
+ "supports_parallel_function_calling": true,
24199
+ "supports_pdf_input": true,
24200
+ "supports_prompt_caching": true,
24201
+ "supports_reasoning": true,
24202
+ "supports_response_schema": true,
24203
+ "supports_system_messages": true,
24204
+ "supports_tool_choice": true,
24205
+ "supports_vision": true,
24206
+ "supports_web_search": true
24207
+ },
24208
+ "gpt-image-1-mini": {
24209
+ "cache_read_input_image_token_cost": 2.5e-07,
24210
+ "cache_read_input_token_cost": 2e-07,
24211
+ "input_cost_per_image_token": 2.5e-06,
24212
+ "input_cost_per_token": 2e-06,
24213
+ "litellm_provider": "openai",
24214
+ "mode": "chat",
24215
+ "output_cost_per_image_token": 8e-06,
24216
+ "supported_endpoints": [
24217
+ "/v1/images/generations",
24218
+ "/v1/images/edits"
24219
+ ]
24220
+ },
24221
+ "gpt-realtime-mini": {
24222
+ "cache_creation_input_audio_token_cost": 3e-07,
24223
+ "cache_read_input_audio_token_cost": 3e-07,
24224
+ "input_cost_per_audio_token": 1e-05,
24225
+ "input_cost_per_token": 6e-07,
24226
+ "litellm_provider": "openai",
24227
+ "max_input_tokens": 128000,
24228
+ "max_output_tokens": 4096,
24229
+ "max_tokens": 4096,
24230
+ "mode": "chat",
24231
+ "output_cost_per_audio_token": 2e-05,
24232
+ "output_cost_per_token": 2.4e-06,
24233
+ "supported_endpoints": [
24234
+ "/v1/realtime"
24235
+ ],
24236
+ "supported_modalities": [
24237
+ "text",
24238
+ "image",
24239
+ "audio"
24240
+ ],
24241
+ "supported_output_modalities": [
24242
+ "text",
24243
+ "audio"
24244
+ ],
24245
+ "supports_audio_input": true,
24246
+ "supports_audio_output": true,
24247
+ "supports_function_calling": true,
24248
+ "supports_parallel_function_calling": true,
24249
+ "supports_system_messages": true,
24250
+ "supports_tool_choice": true
24251
+ },
24252
+ "low/1024-x-1024/gpt-image-1-mini": {
24253
+ "input_cost_per_image": 0.005,
24254
+ "litellm_provider": "openai",
24255
+ "mode": "image_generation",
24256
+ "supported_endpoints": [
24257
+ "/v1/images/generations"
24258
+ ]
24259
+ },
24260
+ "low/1024-x-1536/gpt-image-1-mini": {
24261
+ "input_cost_per_image": 0.006,
24262
+ "litellm_provider": "openai",
24263
+ "mode": "image_generation",
24264
+ "supported_endpoints": [
24265
+ "/v1/images/generations"
24266
+ ]
24267
+ },
24268
+ "low/1536-x-1024/gpt-image-1-mini": {
24269
+ "input_cost_per_image": 0.006,
24270
+ "litellm_provider": "openai",
24271
+ "mode": "image_generation",
24272
+ "supported_endpoints": [
24273
+ "/v1/images/generations"
24274
+ ]
24275
+ },
24276
+ "medium/1024-x-1024/gpt-image-1-mini": {
24277
+ "input_cost_per_image": 0.011,
24278
+ "litellm_provider": "openai",
24279
+ "mode": "image_generation",
24280
+ "supported_endpoints": [
24281
+ "/v1/images/generations"
24282
+ ]
24283
+ },
24284
+ "medium/1024-x-1536/gpt-image-1-mini": {
24285
+ "input_cost_per_image": 0.015,
24286
+ "litellm_provider": "openai",
24287
+ "mode": "image_generation",
24288
+ "supported_endpoints": [
24289
+ "/v1/images/generations"
24290
+ ]
24291
+ },
24292
+ "medium/1536-x-1024/gpt-image-1-mini": {
24293
+ "input_cost_per_image": 0.015,
24294
+ "litellm_provider": "openai",
24295
+ "mode": "image_generation",
24296
+ "supported_endpoints": [
24297
+ "/v1/images/generations"
24298
+ ]
24299
+ },
24300
+ "cohere.embed-v4:0": {
24301
+ "input_cost_per_token": 1.2e-07,
24302
+ "litellm_provider": "bedrock",
24303
+ "max_input_tokens": 128000,
24304
+ "max_tokens": 128000,
24305
+ "mode": "embedding",
24306
+ "output_cost_per_token": 0.0,
24307
+ "output_vector_size": 1536,
24308
+ "supports_embedding_image_input": true
24309
+ },
24310
+ "gpt-5-pro-2025-10-06": {
24311
+ "input_cost_per_token": 1.5e-05,
24312
+ "input_cost_per_token_batches": 7.5e-06,
24313
+ "litellm_provider": "openai",
24314
+ "max_input_tokens": 400000,
24315
+ "max_output_tokens": 272000,
24316
+ "max_tokens": 272000,
24317
+ "mode": "responses",
24318
+ "output_cost_per_token": 0.00012,
24319
+ "output_cost_per_token_batches": 6e-05,
24320
+ "supported_endpoints": [
24321
+ "/v1/batch",
24322
+ "/v1/responses"
24323
+ ],
24324
+ "supported_modalities": [
24325
+ "text",
24326
+ "image"
24327
+ ],
24328
+ "supported_output_modalities": [
24329
+ "text"
24330
+ ],
24331
+ "supports_function_calling": true,
24332
+ "supports_native_streaming": false,
24333
+ "supports_parallel_function_calling": true,
24334
+ "supports_pdf_input": true,
24335
+ "supports_prompt_caching": true,
24336
+ "supports_reasoning": true,
24337
+ "supports_response_schema": true,
24338
+ "supports_system_messages": true,
24339
+ "supports_tool_choice": true,
24340
+ "supports_vision": true,
24341
+ "supports_web_search": true
24342
+ },
24343
+ "azure_ai/Phi-4-mini-reasoning": {
24344
+ "input_cost_per_token": 8e-08,
24345
+ "litellm_provider": "azure_ai",
24346
+ "max_input_tokens": 131072,
24347
+ "max_output_tokens": 4096,
24348
+ "max_tokens": 4096,
24349
+ "mode": "chat",
24350
+ "output_cost_per_token": 3.2e-07,
24351
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/microsoft/",
24352
+ "supports_function_calling": true
24353
+ },
24354
+ "azure_ai/Phi-4-reasoning": {
24355
+ "input_cost_per_token": 1.25e-07,
24356
+ "litellm_provider": "azure_ai",
24357
+ "max_input_tokens": 32768,
24358
+ "max_output_tokens": 4096,
24359
+ "max_tokens": 4096,
24360
+ "mode": "chat",
24361
+ "output_cost_per_token": 5e-07,
24362
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/microsoft/",
24363
+ "supports_function_calling": true,
24364
+ "supports_tool_choice": true,
24365
+ "supports_reasoning": true
24366
+ },
24367
+ "azure_ai/MAI-DS-R1": {
24368
+ "input_cost_per_token": 1.35e-06,
24369
+ "litellm_provider": "azure_ai",
24370
+ "max_input_tokens": 128000,
24371
+ "max_output_tokens": 8192,
24372
+ "max_tokens": 8192,
24373
+ "mode": "chat",
24374
+ "output_cost_per_token": 5.4e-06,
24375
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/microsoft/",
24376
+ "supports_reasoning": true,
24377
+ "supports_tool_choice": true
24378
+ },
24379
+ "au.anthropic.claude-sonnet-4-5-20250929-v1:0": {
24380
+ "cache_creation_input_token_cost": 4.125e-06,
24381
+ "cache_read_input_token_cost": 3.3e-07,
24382
+ "input_cost_per_token": 3.3e-06,
24383
+ "input_cost_per_token_above_200k_tokens": 6.6e-06,
24384
+ "output_cost_per_token_above_200k_tokens": 2.475e-05,
24385
+ "cache_creation_input_token_cost_above_200k_tokens": 8.25e-06,
24386
+ "cache_read_input_token_cost_above_200k_tokens": 6.6e-07,
24387
+ "litellm_provider": "bedrock_converse",
24388
+ "max_input_tokens": 200000,
24389
+ "max_output_tokens": 64000,
24390
+ "max_tokens": 64000,
24391
+ "mode": "chat",
24392
+ "output_cost_per_token": 1.65e-05,
24393
+ "search_context_cost_per_query": {
24394
+ "search_context_size_high": 0.01,
24395
+ "search_context_size_low": 0.01,
24396
+ "search_context_size_medium": 0.01
24397
+ },
24398
+ "supports_assistant_prefill": true,
24399
+ "supports_computer_use": true,
24400
+ "supports_function_calling": true,
24401
+ "supports_pdf_input": true,
24402
+ "supports_prompt_caching": true,
24403
+ "supports_reasoning": true,
24404
+ "supports_response_schema": true,
24405
+ "supports_tool_choice": true,
24406
+ "supports_vision": true,
24407
+ "tool_use_system_prompt_tokens": 346
24408
+ },
24409
+ "oci/cohere.command-latest": {
24410
+ "input_cost_per_token": 1.56e-06,
24411
+ "litellm_provider": "oci",
24412
+ "max_input_tokens": 128000,
24413
+ "max_output_tokens": 4000,
24414
+ "max_tokens": 128000,
24415
+ "mode": "chat",
24416
+ "output_cost_per_token": 1.56e-06,
24417
+ "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/",
24418
+ "supports_function_calling": true,
24419
+ "supports_response_schema": false
24420
+ },
24421
+ "oci/cohere.command-a-03-2025": {
24422
+ "input_cost_per_token": 1.56e-06,
24423
+ "litellm_provider": "oci",
24424
+ "max_input_tokens": 256000,
24425
+ "max_output_tokens": 4000,
24426
+ "max_tokens": 256000,
24427
+ "mode": "chat",
24428
+ "output_cost_per_token": 1.56e-06,
24429
+ "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/",
24430
+ "supports_function_calling": true,
24431
+ "supports_response_schema": false
24432
+ },
24433
+ "oci/cohere.command-plus-latest": {
24434
+ "input_cost_per_token": 1.56e-06,
24435
+ "litellm_provider": "oci",
24436
+ "max_input_tokens": 128000,
24437
+ "max_output_tokens": 4000,
24438
+ "max_tokens": 128000,
24439
+ "mode": "chat",
24440
+ "output_cost_per_token": 1.56e-06,
24441
+ "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/",
24442
+ "supports_function_calling": true,
24443
+ "supports_response_schema": false
24444
+ },
24445
+ "together_ai/moonshotai/Kimi-K2-Instruct-0905": {
24446
+ "input_cost_per_token": 1e-06,
24447
+ "litellm_provider": "together_ai",
24448
+ "max_input_tokens": 262144,
24449
+ "mode": "chat",
24450
+ "output_cost_per_token": 3e-06,
24451
+ "source": "https://www.together.ai/models/kimi-k2-0905",
24452
+ "supports_function_calling": true,
24453
+ "supports_parallel_function_calling": true,
24454
+ "supports_tool_choice": true
24455
+ },
24456
+ "together_ai/Qwen/Qwen3-Next-80B-A3B-Instruct": {
24457
+ "input_cost_per_token": 1.5e-07,
24458
+ "litellm_provider": "together_ai",
24459
+ "max_input_tokens": 262144,
24460
+ "mode": "chat",
24461
+ "output_cost_per_token": 1.5e-06,
24462
+ "source": "https://www.together.ai/models/qwen3-next-80b-a3b-instruct",
24463
+ "supports_function_calling": true,
24464
+ "supports_parallel_function_calling": true,
24465
+ "supports_tool_choice": true
24466
+ },
24467
+ "together_ai/Qwen/Qwen3-Next-80B-A3B-Thinking": {
24468
+ "input_cost_per_token": 1.5e-07,
24469
+ "litellm_provider": "together_ai",
24470
+ "max_input_tokens": 262144,
24471
+ "mode": "chat",
24472
+ "output_cost_per_token": 1.5e-06,
24473
+ "source": "https://www.together.ai/models/qwen3-next-80b-a3b-thinking",
24474
+ "supports_function_calling": true,
24475
+ "supports_parallel_function_calling": true,
24476
+ "supports_tool_choice": true
24477
+ },
24478
+ "together_ai/baai/bge-base-en-v1.5": {
24479
+ "input_cost_per_token": 8e-09,
24480
+ "litellm_provider": "together_ai",
24481
+ "max_input_tokens": 512,
24482
+ "mode": "embedding",
24483
+ "output_cost_per_token": 0.0,
24484
+ "output_vector_size": 768
24485
+ },
24486
+ "together_ai/BAAI/bge-base-en-v1.5": {
24487
+ "input_cost_per_token": 8e-09,
24488
+ "litellm_provider": "together_ai",
24489
+ "max_input_tokens": 512,
24490
+ "mode": "embedding",
24491
+ "output_cost_per_token": 0.0,
24492
+ "output_vector_size": 768
24493
+ },
24494
+ "openrouter/anthropic/claude-sonnet-4.5": {
24495
+ "input_cost_per_image": 0.0048,
24496
+ "cache_creation_input_token_cost": 3.75e-06,
24497
+ "cache_read_input_token_cost": 3e-07,
24498
+ "input_cost_per_token": 3e-06,
24499
+ "input_cost_per_token_above_200k_tokens": 6e-06,
24500
+ "output_cost_per_token_above_200k_tokens": 2.25e-05,
24501
+ "cache_creation_input_token_cost_above_200k_tokens": 7.5e-06,
24502
+ "cache_read_input_token_cost_above_200k_tokens": 6e-07,
24503
+ "litellm_provider": "openrouter",
24504
+ "max_input_tokens": 1000000,
24505
+ "max_output_tokens": 1000000,
24506
+ "max_tokens": 1000000,
24507
+ "mode": "chat",
24508
+ "output_cost_per_token": 1.5e-05,
24509
+ "supports_assistant_prefill": true,
24510
+ "supports_computer_use": true,
24511
+ "supports_function_calling": true,
24512
+ "supports_prompt_caching": true,
24513
+ "supports_reasoning": true,
24514
+ "supports_tool_choice": true,
24515
+ "supports_vision": true,
24516
+ "tool_use_system_prompt_tokens": 159
24517
+ },
24518
+ "anthropic.claude-haiku-4-5-20251001-v1:0": {
24519
+ "cache_creation_input_token_cost": 1.25e-06,
24520
+ "cache_read_input_token_cost": 1e-07,
24521
+ "input_cost_per_token": 1e-06,
24522
+ "litellm_provider": "bedrock",
24523
+ "max_input_tokens": 200000,
24524
+ "max_output_tokens": 8192,
24525
+ "max_tokens": 8192,
24526
+ "mode": "chat",
24527
+ "output_cost_per_token": 5e-06,
24528
+ "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
24529
+ "supports_assistant_prefill": true,
24530
+ "supports_function_calling": true,
24531
+ "supports_pdf_input": true,
24532
+ "supports_prompt_caching": true,
24533
+ "supports_reasoning": true,
24534
+ "supports_response_schema": true,
24535
+ "supports_tool_choice": true
24536
+ },
24537
+ "anthropic.claude-haiku-4-5@20251001": {
24538
+ "cache_creation_input_token_cost": 1.25e-06,
24539
+ "cache_read_input_token_cost": 1e-07,
24540
+ "input_cost_per_token": 1e-06,
24541
+ "litellm_provider": "bedrock",
24542
+ "max_input_tokens": 200000,
24543
+ "max_output_tokens": 8192,
24544
+ "max_tokens": 8192,
24545
+ "mode": "chat",
24546
+ "output_cost_per_token": 5e-06,
24547
+ "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
24548
+ "supports_assistant_prefill": true,
24549
+ "supports_function_calling": true,
24550
+ "supports_pdf_input": true,
24551
+ "supports_prompt_caching": true,
24552
+ "supports_reasoning": true,
24553
+ "supports_response_schema": true,
24554
+ "supports_tool_choice": true
24555
+ },
24556
+ "apac.anthropic.claude-haiku-4-5-20251001-v1:0": {
24557
+ "cache_creation_input_token_cost": 1.375e-06,
24558
+ "cache_read_input_token_cost": 1.1e-07,
24559
+ "input_cost_per_token": 1.1e-06,
24560
+ "litellm_provider": "bedrock",
24561
+ "max_input_tokens": 200000,
24562
+ "max_output_tokens": 8192,
24563
+ "max_tokens": 8192,
24564
+ "mode": "chat",
24565
+ "output_cost_per_token": 5.5e-06,
24566
+ "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
24567
+ "supports_assistant_prefill": true,
24568
+ "supports_function_calling": true,
24569
+ "supports_pdf_input": true,
24570
+ "supports_prompt_caching": true,
24571
+ "supports_reasoning": true,
24572
+ "supports_response_schema": true,
24573
+ "supports_tool_choice": true
24574
+ },
24575
+ "claude-haiku-4-5-20251001": {
24576
+ "cache_creation_input_token_cost": 1.25e-06,
24577
+ "cache_creation_input_token_cost_above_1hr": 2e-06,
24578
+ "cache_read_input_token_cost": 1e-07,
24579
+ "input_cost_per_token": 1e-06,
24580
+ "litellm_provider": "anthropic",
24581
+ "max_input_tokens": 200000,
24582
+ "max_output_tokens": 64000,
24583
+ "max_tokens": 64000,
24584
+ "mode": "chat",
24585
+ "output_cost_per_token": 5e-06,
24586
+ "supports_assistant_prefill": true,
24587
+ "supports_function_calling": true,
24588
+ "supports_computer_use": true,
24589
+ "supports_pdf_input": true,
24590
+ "supports_prompt_caching": true,
24591
+ "supports_reasoning": true,
24592
+ "supports_response_schema": true,
24593
+ "supports_tool_choice": true,
24594
+ "supports_vision": true
24595
+ },
24596
+ "claude-haiku-4-5": {
24597
+ "cache_creation_input_token_cost": 1.25e-06,
24598
+ "cache_creation_input_token_cost_above_1hr": 2e-06,
24599
+ "cache_read_input_token_cost": 1e-07,
24600
+ "input_cost_per_token": 1e-06,
24601
+ "litellm_provider": "anthropic",
24602
+ "max_input_tokens": 200000,
24603
+ "max_output_tokens": 64000,
24604
+ "max_tokens": 64000,
24605
+ "mode": "chat",
24606
+ "output_cost_per_token": 5e-06,
24607
+ "supports_assistant_prefill": true,
24608
+ "supports_function_calling": true,
24609
+ "supports_computer_use": true,
24610
+ "supports_pdf_input": true,
24611
+ "supports_prompt_caching": true,
24612
+ "supports_reasoning": true,
24613
+ "supports_response_schema": true,
24614
+ "supports_tool_choice": true,
24615
+ "supports_vision": true
24616
+ },
24617
+ "eu.anthropic.claude-haiku-4-5-20251001-v1:0": {
24618
+ "cache_creation_input_token_cost": 1.375e-06,
24619
+ "cache_read_input_token_cost": 1.1e-07,
24620
+ "input_cost_per_token": 1.1e-06,
24621
+ "deprecation_date": "2026-10-15",
24622
+ "litellm_provider": "bedrock",
24623
+ "max_input_tokens": 200000,
24624
+ "max_output_tokens": 8192,
24625
+ "max_tokens": 8192,
24626
+ "mode": "chat",
24627
+ "output_cost_per_token": 5.5e-06,
24628
+ "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
24629
+ "supports_assistant_prefill": true,
24630
+ "supports_function_calling": true,
24631
+ "supports_pdf_input": true,
24632
+ "supports_prompt_caching": true,
24633
+ "supports_reasoning": true,
24634
+ "supports_response_schema": true,
24635
+ "supports_tool_choice": true
24636
+ },
24637
+ "gemini-2.5-flash-image": {
24638
+ "cache_read_input_token_cost": 3e-08,
24639
+ "input_cost_per_audio_token": 1e-06,
24640
+ "input_cost_per_token": 3e-07,
24641
+ "litellm_provider": "vertex_ai-language-models",
24642
+ "max_audio_length_hours": 8.4,
24643
+ "max_audio_per_prompt": 1,
24644
+ "max_images_per_prompt": 3000,
24645
+ "max_input_tokens": 32768,
24646
+ "max_output_tokens": 32768,
24647
+ "max_tokens": 32768,
24648
+ "max_pdf_size_mb": 30,
24649
+ "max_video_length": 1,
24650
+ "max_videos_per_prompt": 10,
24651
+ "mode": "image_generation",
24652
+ "output_cost_per_image": 0.039,
24653
+ "output_cost_per_reasoning_token": 2.5e-06,
24654
+ "output_cost_per_token": 2.5e-06,
24655
+ "rpm": 100000,
24656
+ "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-flash-image",
24657
+ "supported_endpoints": [
24658
+ "/v1/chat/completions",
24659
+ "/v1/completions",
24660
+ "/v1/batch"
24661
+ ],
24662
+ "supported_modalities": [
24663
+ "text",
24664
+ "image",
24665
+ "audio",
24666
+ "video"
24667
+ ],
24668
+ "supported_output_modalities": [
24669
+ "text",
24670
+ "image"
24671
+ ],
24672
+ "supports_audio_output": false,
24673
+ "supports_function_calling": true,
24674
+ "supports_parallel_function_calling": true,
24675
+ "supports_pdf_input": true,
24676
+ "supports_prompt_caching": true,
24677
+ "supports_response_schema": true,
24678
+ "supports_system_messages": true,
24679
+ "supports_tool_choice": true,
24680
+ "supports_url_context": true,
24681
+ "supports_vision": true,
24682
+ "supports_web_search": false,
24683
+ "tpm": 8000000
24684
+ },
24685
+ "gemini/gemini-2.5-flash-image": {
24686
+ "cache_read_input_token_cost": 3e-08,
24687
+ "input_cost_per_audio_token": 1e-06,
24688
+ "input_cost_per_token": 3e-07,
24689
+ "litellm_provider": "vertex_ai-language-models",
24690
+ "max_audio_length_hours": 8.4,
24691
+ "max_audio_per_prompt": 1,
24692
+ "max_images_per_prompt": 3000,
24693
+ "max_input_tokens": 32768,
24694
+ "max_output_tokens": 32768,
24695
+ "max_tokens": 32768,
24696
+ "max_pdf_size_mb": 30,
24697
+ "max_video_length": 1,
24698
+ "max_videos_per_prompt": 10,
24699
+ "mode": "image_generation",
24700
+ "output_cost_per_image": 0.039,
24701
+ "output_cost_per_reasoning_token": 2.5e-06,
24702
+ "output_cost_per_token": 2.5e-06,
24703
+ "rpm": 100000,
24704
+ "source": "https://ai.google.dev/gemini-api/docs/pricing#gemini-2.5-flash-image",
24705
+ "supported_endpoints": [
24706
+ "/v1/chat/completions",
24707
+ "/v1/completions",
24708
+ "/v1/batch"
24709
+ ],
24710
+ "supported_modalities": [
24711
+ "text",
24712
+ "image",
24713
+ "audio",
24714
+ "video"
24715
+ ],
24716
+ "supported_output_modalities": [
24717
+ "text",
24718
+ "image"
24719
+ ],
24720
+ "supports_audio_output": false,
24721
+ "supports_function_calling": true,
24722
+ "supports_parallel_function_calling": true,
24723
+ "supports_pdf_input": true,
24724
+ "supports_prompt_caching": true,
24725
+ "supports_response_schema": true,
24726
+ "supports_system_messages": true,
24727
+ "supports_tool_choice": true,
24728
+ "supports_url_context": true,
24729
+ "supports_vision": true,
24730
+ "supports_web_search": true,
24731
+ "tpm": 8000000
24732
+ },
24733
+ "lemonade/gpt-oss-20b-mxfp4-GGUF": {
24734
+ "input_cost_per_token": 0,
24735
+ "litellm_provider": "lemonade",
24736
+ "max_tokens": 131072,
24737
+ "max_input_tokens": 131072,
24738
+ "max_output_tokens": 32768,
24739
+ "mode": "chat",
24740
+ "output_cost_per_token": 0,
24741
+ "supports_function_calling": true,
24742
+ "supports_response_schema": true,
24743
+ "supports_tool_choice": true
24744
+ },
24745
+ "lemonade/gpt-oss-120b-mxfp-GGUF": {
24746
+ "input_cost_per_token": 0,
24747
+ "litellm_provider": "lemonade",
24748
+ "max_tokens": 131072,
24749
+ "max_input_tokens": 131072,
24750
+ "max_output_tokens": 32768,
24751
+ "mode": "chat",
24752
+ "output_cost_per_token": 0,
24753
+ "supports_function_calling": true,
24754
+ "supports_response_schema": true,
24755
+ "supports_tool_choice": true
24756
+ },
24757
+ "lemonade/Gemma-3-4b-it-GGUF": {
24758
+ "input_cost_per_token": 0,
24759
+ "litellm_provider": "lemonade",
24760
+ "max_tokens": 128000,
24761
+ "max_input_tokens": 128000,
24762
+ "max_output_tokens": 8192,
24763
+ "mode": "chat",
24764
+ "output_cost_per_token": 0,
24765
+ "supports_function_calling": true,
24766
+ "supports_response_schema": true,
24767
+ "supports_tool_choice": true
24768
+ },
24769
+ "lemonade/Qwen3-4B-Instruct-2507-GGUF": {
24770
+ "input_cost_per_token": 0,
24771
+ "litellm_provider": "lemonade",
24772
+ "max_tokens": 262144,
24773
+ "max_input_tokens": 262144,
24774
+ "max_output_tokens": 32768,
24775
+ "mode": "chat",
24776
+ "output_cost_per_token": 0,
24777
+ "supports_function_calling": true,
24778
+ "supports_response_schema": true,
24779
+ "supports_tool_choice": true
24780
+ },
24781
+ "jp.anthropic.claude-haiku-4-5-20251001-v1:0": {
24782
+ "cache_creation_input_token_cost": 1.375e-06,
24783
+ "cache_read_input_token_cost": 1.1e-07,
24784
+ "input_cost_per_token": 1.1e-06,
24785
+ "litellm_provider": "bedrock",
24786
+ "max_input_tokens": 200000,
24787
+ "max_output_tokens": 8192,
24788
+ "max_tokens": 8192,
24789
+ "mode": "chat",
24790
+ "output_cost_per_token": 5.5e-06,
24791
+ "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
24792
+ "supports_assistant_prefill": true,
24793
+ "supports_function_calling": true,
24794
+ "supports_pdf_input": true,
24795
+ "supports_prompt_caching": true,
24796
+ "supports_reasoning": true,
24797
+ "supports_response_schema": true,
24798
+ "supports_tool_choice": true
24799
+ },
24800
+ "us.anthropic.claude-haiku-4-5-20251001-v1:0": {
24801
+ "cache_creation_input_token_cost": 1.375e-06,
24802
+ "cache_read_input_token_cost": 1.1e-07,
24803
+ "input_cost_per_token": 1.1e-06,
24804
+ "litellm_provider": "bedrock",
24805
+ "max_input_tokens": 200000,
24806
+ "max_output_tokens": 8192,
24807
+ "max_tokens": 8192,
24808
+ "mode": "chat",
24809
+ "output_cost_per_token": 5.5e-06,
24810
+ "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
24811
+ "supports_assistant_prefill": true,
24812
+ "supports_function_calling": true,
24813
+ "supports_pdf_input": true,
24814
+ "supports_prompt_caching": true,
24815
+ "supports_reasoning": true,
24816
+ "supports_response_schema": true,
24817
+ "supports_tool_choice": true
24818
+ },
24819
+ "vertex_ai/claude-haiku-4-5@20251001": {
24820
+ "cache_creation_input_token_cost": 1.25e-06,
24821
+ "cache_read_input_token_cost": 1e-07,
24822
+ "input_cost_per_token": 1e-06,
24823
+ "litellm_provider": "vertex_ai-anthropic_models",
24824
+ "max_input_tokens": 200000,
24825
+ "max_output_tokens": 8192,
24826
+ "max_tokens": 8192,
24827
+ "mode": "chat",
24828
+ "output_cost_per_token": 5e-06,
24829
+ "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/partner-models/claude/haiku-4-5",
24830
+ "supports_assistant_prefill": true,
24831
+ "supports_function_calling": true,
24832
+ "supports_pdf_input": true,
24833
+ "supports_prompt_caching": true,
24834
+ "supports_reasoning": true,
24835
+ "supports_response_schema": true,
24836
+ "supports_tool_choice": true
24837
+ },
24838
+ "azure_ai/mistral-document-ai-2505": {
24839
+ "litellm_provider": "azure_ai",
24840
+ "ocr_cost_per_page": 0.003,
24841
+ "mode": "ocr",
24842
+ "supported_endpoints": [
24843
+ "/v1/ocr"
24844
+ ],
24845
+ "source": "https://devblogs.microsoft.com/foundry/whats-new-in-azure-ai-foundry-august-2025/#mistral-document-ai-(ocr)-%E2%80%94-serverless-in-foundry"
24846
+ },
24847
+ "mistral/mistral-ocr-latest": {
24848
+ "litellm_provider": "mistral",
24849
+ "ocr_cost_per_page": 0.001,
24850
+ "annotation_cost_per_page": 0.003,
24851
+ "mode": "ocr",
24852
+ "supported_endpoints": [
24853
+ "/v1/ocr"
24854
+ ],
24855
+ "source": "https://mistral.ai/pricing#api-pricing"
24856
+ },
24857
+ "mistral/mistral-ocr-2505-completion": {
24858
+ "litellm_provider": "mistral",
24859
+ "ocr_cost_per_page": 0.001,
24860
+ "annotation_cost_per_page": 0.003,
24861
+ "mode": "ocr",
24862
+ "supported_endpoints": [
24863
+ "/v1/ocr"
24864
+ ],
24865
+ "source": "https://mistral.ai/pricing#api-pricing"
24866
+ },
24867
+ "watsonx/mistralai/mistral-small-3-1-24b-instruct-2503": {
24868
+ "max_tokens": 32000,
24869
+ "max_input_tokens": 32000,
24870
+ "max_output_tokens": 32000,
24871
+ "input_cost_per_token": 1e-07,
24872
+ "output_cost_per_token": 3e-07,
24873
+ "litellm_provider": "watsonx",
24874
+ "mode": "chat",
24875
+ "supports_function_calling": true,
24876
+ "supports_parallel_function_calling": true,
24877
+ "supports_vision": false
24878
+ },
24879
+ "global.anthropic.claude-haiku-4-5-20251001-v1:0": {
24880
+ "cache_creation_input_token_cost": 1.375e-06,
24881
+ "cache_read_input_token_cost": 1.1e-07,
24882
+ "input_cost_per_token": 1.1e-06,
24883
+ "litellm_provider": "bedrock_converse",
24884
+ "max_input_tokens": 200000,
24885
+ "max_output_tokens": 8192,
24886
+ "max_tokens": 8192,
24887
+ "mode": "chat",
24888
+ "output_cost_per_token": 5.5e-06,
24889
+ "source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
24890
+ "supports_assistant_prefill": true,
24891
+ "supports_function_calling": true,
24892
+ "supports_pdf_input": true,
24893
+ "supports_prompt_caching": true,
24894
+ "supports_reasoning": true,
24895
+ "supports_response_schema": true,
24896
+ "supports_tool_choice": true,
24897
+ "supports_vision": true,
24898
+ "tool_use_system_prompt_tokens": 346
24899
+ },
24900
+ "au.anthropic.claude-haiku-4-5-20251001-v1:0": {
24901
+ "cache_creation_input_token_cost": 1.375e-06,
24902
+ "cache_read_input_token_cost": 1.1e-07,
24903
+ "input_cost_per_token": 1.1e-06,
24904
+ "litellm_provider": "bedrock_converse",
24905
+ "max_input_tokens": 200000,
24906
+ "max_output_tokens": 8192,
24907
+ "max_tokens": 8192,
24908
+ "mode": "chat",
24909
+ "output_cost_per_token": 5.5e-06,
24910
+ "supports_assistant_prefill": true,
24911
+ "supports_function_calling": true,
24912
+ "supports_pdf_input": true,
24913
+ "supports_prompt_caching": true,
24914
+ "supports_response_schema": true,
24915
+ "supports_tool_choice": true,
24916
+ "supports_vision": true,
24917
+ "tool_use_system_prompt_tokens": 346
24918
+ },
24919
+ "vercel_ai_gateway/glm-4.6": {
24920
+ "litellm_provider": "vercel_ai_gateway",
24921
+ "cache_read_input_token_cost": 1.1e-07,
24922
+ "input_cost_per_token": 6e-07,
24923
+ "max_input_tokens": 200000,
24924
+ "max_output_tokens": 200000,
24925
+ "max_tokens": 200000,
24926
+ "mode": "chat",
24927
+ "output_cost_per_token": 2.2e-06
24928
+ },
24929
+ "azure/speech/azure-tts": {
24930
+ "input_cost_per_character": 1.5e-05,
24931
+ "litellm_provider": "azure",
24932
+ "mode": "audio_speech",
24933
+ "source": "https://azure.microsoft.com/en-us/pricing/calculator/"
24934
+ },
24935
+ "azure/speech/azure-tts-hd": {
24936
+ "input_cost_per_character": 3e-05,
24937
+ "litellm_provider": "azure",
24938
+ "mode": "audio_speech",
24939
+ "source": "https://azure.microsoft.com/en-us/pricing/calculator/"
24940
+ },
24941
+ "bedrock/us-gov-west-1/anthropic.claude-3-7-sonnet-20240620-v1:0": {
24942
+ "cache_creation_input_token_cost": 4.5e-06,
24943
+ "cache_read_input_token_cost": 3.6e-07,
24944
+ "input_cost_per_token": 3.6e-06,
24945
+ "litellm_provider": "bedrock",
24946
+ "max_input_tokens": 200000,
24947
+ "max_output_tokens": 8192,
24948
+ "max_tokens": 8192,
24949
+ "mode": "chat",
24950
+ "output_cost_per_token": 1.8e-05,
24951
+ "supports_assistant_prefill": true,
24952
+ "supports_computer_use": true,
24953
+ "supports_function_calling": true,
24954
+ "supports_pdf_input": true,
24955
+ "supports_prompt_caching": true,
24956
+ "supports_reasoning": true,
24957
+ "supports_response_schema": true,
24958
+ "supports_tool_choice": true,
24959
+ "supports_vision": true
24960
+ },
24961
+ "anthropic.claude-3-7-sonnet-20240620-v1:0": {
24962
+ "cache_creation_input_token_cost": 4.5e-06,
24963
+ "cache_read_input_token_cost": 3.6e-07,
24964
+ "input_cost_per_token": 3.6e-06,
24965
+ "litellm_provider": "bedrock",
24966
+ "max_input_tokens": 200000,
24967
+ "max_output_tokens": 8192,
24968
+ "max_tokens": 8192,
24969
+ "mode": "chat",
24970
+ "output_cost_per_token": 1.8e-05,
24971
+ "supports_assistant_prefill": true,
24972
+ "supports_computer_use": true,
24973
+ "supports_function_calling": true,
24974
+ "supports_pdf_input": true,
24975
+ "supports_prompt_caching": true,
24976
+ "supports_reasoning": true,
24977
+ "supports_response_schema": true,
24978
+ "supports_tool_choice": true,
24979
+ "supports_vision": true
24980
+ },
24981
+ "bedrock/us-gov-west-1/anthropic.claude-3-7-sonnet-20250219-v1:0": {
24982
+ "cache_creation_input_token_cost": 4.5e-06,
24983
+ "cache_read_input_token_cost": 3.6e-07,
24984
+ "input_cost_per_token": 3.6e-06,
24985
+ "litellm_provider": "bedrock",
24986
+ "max_input_tokens": 200000,
24987
+ "max_output_tokens": 8192,
24988
+ "max_tokens": 8192,
24989
+ "mode": "chat",
24990
+ "output_cost_per_token": 1.8e-05,
24991
+ "supports_assistant_prefill": true,
24992
+ "supports_computer_use": true,
24993
+ "supports_function_calling": true,
24994
+ "supports_pdf_input": true,
24995
+ "supports_prompt_caching": true,
24996
+ "supports_reasoning": true,
24997
+ "supports_response_schema": true,
24998
+ "supports_tool_choice": true,
24999
+ "supports_vision": true
25000
+ },
25001
+ "dataforseo/search": {
25002
+ "input_cost_per_query": 0.003,
25003
+ "litellm_provider": "dataforseo",
25004
+ "mode": "search"
25005
+ },
25006
+ "exa_ai/search": {
25007
+ "litellm_provider": "exa_ai",
25008
+ "mode": "search",
25009
+ "tiered_pricing": [
25010
+ {
25011
+ "input_cost_per_query": 0.005,
25012
+ "max_results_range": [
25013
+ 0,
25014
+ 25
25015
+ ]
25016
+ },
25017
+ {
25018
+ "input_cost_per_query": 0.025,
25019
+ "max_results_range": [
25020
+ 26,
25021
+ 100
25022
+ ]
25023
+ }
25024
+ ]
25025
+ },
25026
+ "perplexity/search": {
25027
+ "input_cost_per_query": 0.005,
25028
+ "litellm_provider": "perplexity",
25029
+ "mode": "search"
25030
+ },
25031
+ "google_pse/search": {
25032
+ "input_cost_per_query": 0.005,
25033
+ "litellm_provider": "google_pse",
25034
+ "mode": "search"
25035
+ },
25036
+ "parallel_ai/search": {
25037
+ "input_cost_per_query": 0.004,
25038
+ "litellm_provider": "parallel_ai",
25039
+ "mode": "search"
25040
+ },
25041
+ "parallel_ai/search-pro": {
25042
+ "input_cost_per_query": 0.009,
25043
+ "litellm_provider": "parallel_ai",
25044
+ "mode": "search"
25045
+ },
25046
+ "tavily/search": {
25047
+ "input_cost_per_query": 0.008,
25048
+ "litellm_provider": "tavily",
25049
+ "mode": "search"
25050
+ },
25051
+ "tavily/search-advanced": {
25052
+ "input_cost_per_query": 0.016,
25053
+ "litellm_provider": "tavily",
25054
+ "mode": "search"
25055
+ },
25056
+ "vertex_ai/mistralai/codestral-2@001": {
25057
+ "input_cost_per_token": 3e-07,
25058
+ "litellm_provider": "vertex_ai-mistral_models",
25059
+ "max_input_tokens": 128000,
25060
+ "max_output_tokens": 128000,
25061
+ "max_tokens": 128000,
25062
+ "mode": "chat",
25063
+ "output_cost_per_token": 9e-07,
25064
+ "supports_function_calling": true,
25065
+ "supports_tool_choice": true
25066
+ },
25067
+ "vertex_ai/codestral-2": {
25068
+ "input_cost_per_token": 3e-07,
25069
+ "litellm_provider": "vertex_ai-mistral_models",
25070
+ "max_input_tokens": 128000,
25071
+ "max_output_tokens": 128000,
25072
+ "max_tokens": 128000,
25073
+ "mode": "chat",
25074
+ "output_cost_per_token": 9e-07,
25075
+ "supports_function_calling": true,
25076
+ "supports_tool_choice": true
25077
+ },
25078
+ "vertex_ai/codestral-2@001": {
25079
+ "input_cost_per_token": 3e-07,
25080
+ "litellm_provider": "vertex_ai-mistral_models",
25081
+ "max_input_tokens": 128000,
25082
+ "max_output_tokens": 128000,
25083
+ "max_tokens": 128000,
25084
+ "mode": "chat",
25085
+ "output_cost_per_token": 9e-07,
25086
+ "supports_function_calling": true,
25087
+ "supports_tool_choice": true
25088
+ },
25089
+ "vertex_ai/mistralai/codestral-2": {
25090
+ "input_cost_per_token": 3e-07,
25091
+ "litellm_provider": "vertex_ai-mistral_models",
25092
+ "max_input_tokens": 128000,
25093
+ "max_output_tokens": 128000,
25094
+ "max_tokens": 128000,
25095
+ "mode": "chat",
25096
+ "output_cost_per_token": 9e-07,
25097
+ "supports_function_calling": true,
25098
+ "supports_tool_choice": true
25099
+ },
25100
+ "vertex_ai/mistral-medium-3": {
25101
+ "input_cost_per_token": 4e-07,
25102
+ "litellm_provider": "vertex_ai-mistral_models",
25103
+ "max_input_tokens": 128000,
25104
+ "max_output_tokens": 8191,
25105
+ "max_tokens": 8191,
25106
+ "mode": "chat",
25107
+ "output_cost_per_token": 2e-06,
25108
+ "supports_function_calling": true,
25109
+ "supports_tool_choice": true
25110
+ },
25111
+ "vertex_ai/mistral-medium-3@001": {
25112
+ "input_cost_per_token": 4e-07,
25113
+ "litellm_provider": "vertex_ai-mistral_models",
25114
+ "max_input_tokens": 128000,
25115
+ "max_output_tokens": 8191,
25116
+ "max_tokens": 8191,
25117
+ "mode": "chat",
25118
+ "output_cost_per_token": 2e-06,
25119
+ "supports_function_calling": true,
25120
+ "supports_tool_choice": true
25121
+ },
25122
+ "vertex_ai/mistralai/mistral-medium-3": {
25123
+ "input_cost_per_token": 4e-07,
25124
+ "litellm_provider": "vertex_ai-mistral_models",
25125
+ "max_input_tokens": 128000,
25126
+ "max_output_tokens": 8191,
25127
+ "max_tokens": 8191,
25128
+ "mode": "chat",
25129
+ "output_cost_per_token": 2e-06,
25130
+ "supports_function_calling": true,
25131
+ "supports_tool_choice": true
25132
+ },
25133
+ "vertex_ai/mistralai/mistral-medium-3@001": {
25134
+ "input_cost_per_token": 4e-07,
25135
+ "litellm_provider": "vertex_ai-mistral_models",
25136
+ "max_input_tokens": 128000,
25137
+ "max_output_tokens": 8191,
25138
+ "max_tokens": 8191,
25139
+ "mode": "chat",
25140
+ "output_cost_per_token": 2e-06,
25141
+ "supports_function_calling": true,
25142
+ "supports_tool_choice": true
25143
+ },
25144
+ "openai/sora-2": {
25145
+ "litellm_provider": "openai",
25146
+ "mode": "video_generation",
25147
+ "output_cost_per_video_per_second": 0.1,
25148
+ "source": "https://platform.openai.com/docs/api-reference/videos",
25149
+ "supported_modalities": [
25150
+ "text",
25151
+ "image"
25152
+ ],
25153
+ "supported_output_modalities": [
25154
+ "video"
25155
+ ],
25156
+ "supported_resolutions": [
25157
+ "720x1280",
25158
+ "1280x720"
25159
+ ]
25160
+ },
25161
+ "azure/sora-2": {
25162
+ "litellm_provider": "azure",
25163
+ "mode": "video_generation",
25164
+ "output_cost_per_video_per_second": 0.1,
25165
+ "source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation",
25166
+ "supported_modalities": [
25167
+ "text"
25168
+ ],
25169
+ "supported_output_modalities": [
25170
+ "video"
25171
+ ],
25172
+ "supported_resolutions": [
25173
+ "720x1280",
25174
+ "1280x720"
25175
+ ]
25176
+ },
25177
+ "azure/sora-2-pro": {
25178
+ "litellm_provider": "azure",
25179
+ "mode": "video_generation",
25180
+ "output_cost_per_video_per_second": 0.3,
25181
+ "source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation",
25182
+ "supported_modalities": [
25183
+ "text"
25184
+ ],
25185
+ "supported_output_modalities": [
25186
+ "video"
25187
+ ],
25188
+ "supported_resolutions": [
25189
+ "720x1280",
25190
+ "1280x720"
25191
+ ]
25192
+ },
25193
+ "azure/sora-2-pro-high-res": {
25194
+ "litellm_provider": "azure",
25195
+ "mode": "video_generation",
25196
+ "output_cost_per_video_per_second": 0.5,
25197
+ "source": "https://azure.microsoft.com/en-us/products/ai-services/video-generation",
25198
+ "supported_modalities": [
25199
+ "text"
25200
+ ],
25201
+ "supported_output_modalities": [
25202
+ "video"
25203
+ ],
25204
+ "supported_resolutions": [
25205
+ "1024x1792",
25206
+ "1792x1024"
25207
+ ]
25208
+ },
25209
+ "amazon.titan-image-generator-v1": {
25210
+ "input_cost_per_image": 0.0,
25211
+ "output_cost_per_image": 0.008,
25212
+ "output_cost_per_image_premium_image": 0.01,
25213
+ "output_cost_per_image_above_512_and_512_pixels": 0.01,
25214
+ "output_cost_per_image_above_512_and_512_pixels_and_premium_image": 0.012,
25215
+ "litellm_provider": "bedrock",
25216
+ "mode": "image_generation"
25217
+ },
25218
+ "amazon.titan-image-generator-v2": {
25219
+ "input_cost_per_image": 0.0,
25220
+ "output_cost_per_image": 0.008,
25221
+ "output_cost_per_image_premium_image": 0.01,
25222
+ "output_cost_per_image_above_1024_and_1024_pixels": 0.01,
25223
+ "output_cost_per_image_above_1024_and_1024_pixels_and_premium_image": 0.012,
25224
+ "litellm_provider": "bedrock",
25225
+ "mode": "image_generation"
25226
+ },
25227
+ "vertex_ai/search_api": {
25228
+ "input_cost_per_query": 0.0015,
25229
+ "litellm_provider": "vertex_ai",
25230
+ "mode": "vector_store"
25231
+ },
25232
+ "openai/sora-2-pro": {
25233
+ "litellm_provider": "openai",
25234
+ "mode": "video_generation",
25235
+ "output_cost_per_video_per_second": 0.3,
25236
+ "source": "https://platform.openai.com/docs/api-reference/videos",
25237
+ "supported_modalities": [
25238
+ "text",
25239
+ "image"
25240
+ ],
25241
+ "supported_output_modalities": [
25242
+ "video"
25243
+ ],
25244
+ "supported_resolutions": [
25245
+ "720x1280",
25246
+ "1280x720"
25247
+ ]
25248
+ },
25249
+ "openrouter/anthropic/claude-haiku-4.5": {
25250
+ "cache_creation_input_token_cost": 1.25e-06,
25251
+ "cache_read_input_token_cost": 1e-07,
25252
+ "input_cost_per_token": 1e-06,
25253
+ "litellm_provider": "openrouter",
25254
+ "max_input_tokens": 200000,
25255
+ "max_output_tokens": 200000,
25256
+ "max_tokens": 200000,
25257
+ "mode": "chat",
25258
+ "output_cost_per_token": 5e-06,
25259
+ "supports_assistant_prefill": true,
25260
+ "supports_computer_use": true,
25261
+ "supports_function_calling": true,
25262
+ "supports_prompt_caching": true,
25263
+ "supports_reasoning": true,
25264
+ "supports_tool_choice": true,
25265
+ "supports_vision": true,
25266
+ "tool_use_system_prompt_tokens": 346
25267
+ },
25268
+ "fal_ai/bria/text-to-image/3.2": {
25269
+ "litellm_provider": "fal_ai",
25270
+ "mode": "image_generation",
25271
+ "output_cost_per_image": 0.0398,
25272
+ "supported_endpoints": [
25273
+ "/v1/images/generations"
25274
+ ]
25275
+ },
25276
+ "fal_ai/fal-ai/flux-pro/v1.1-ultra": {
25277
+ "litellm_provider": "fal_ai",
25278
+ "mode": "image_generation",
25279
+ "output_cost_per_image": 0.0398,
25280
+ "supported_endpoints": [
25281
+ "/v1/images/generations"
25282
+ ]
25283
+ },
25284
+ "fal_ai/fal-ai/imagen4/preview": {
25285
+ "litellm_provider": "fal_ai",
25286
+ "mode": "image_generation",
25287
+ "output_cost_per_image": 0.0398,
25288
+ "supported_endpoints": [
25289
+ "/v1/images/generations"
25290
+ ]
25291
+ },
25292
+ "fal_ai/fal-ai/recraft/v3/text-to-image": {
25293
+ "litellm_provider": "fal_ai",
25294
+ "mode": "image_generation",
25295
+ "output_cost_per_image": 0.0398,
25296
+ "supported_endpoints": [
25297
+ "/v1/images/generations"
25298
+ ]
25299
+ },
25300
+ "fal_ai/fal-ai/stable-diffusion-v35-medium": {
25301
+ "litellm_provider": "fal_ai",
25302
+ "mode": "image_generation",
25303
+ "output_cost_per_image": 0.0398,
25304
+ "supported_endpoints": [
25305
+ "/v1/images/generations"
25306
+ ]
25307
+ },
25308
+ "mistral/codestral-embed": {
25309
+ "input_cost_per_token": 1.5e-07,
25310
+ "litellm_provider": "mistral",
25311
+ "max_input_tokens": 8192,
25312
+ "max_tokens": 8192,
25313
+ "mode": "embedding"
25314
+ },
25315
+ "mistral/codestral-embed-2505": {
25316
+ "input_cost_per_token": 1.5e-07,
25317
+ "litellm_provider": "mistral",
25318
+ "max_input_tokens": 8192,
25319
+ "max_tokens": 8192,
25320
+ "mode": "embedding"
25321
+ },
25322
+ "gemini/gemini-embedding-001": {
25323
+ "input_cost_per_token": 1.5e-07,
25324
+ "litellm_provider": "gemini",
25325
+ "max_input_tokens": 2048,
25326
+ "max_tokens": 2048,
25327
+ "mode": "embedding",
25328
+ "output_cost_per_token": 0,
25329
+ "output_vector_size": 3072,
25330
+ "rpm": 10000,
25331
+ "source": "https://ai.google.dev/gemini-api/docs/embeddings#model-versions",
25332
+ "tpm": 10000000
25333
+ },
25334
+ "vercel_ai_gateway/zai/glm-4.6": {
25335
+ "litellm_provider": "vercel_ai_gateway",
25336
+ "cache_read_input_token_cost": 1.1e-07,
25337
+ "input_cost_per_token": 4.5e-07,
25338
+ "max_input_tokens": 200000,
25339
+ "max_output_tokens": 200000,
25340
+ "max_tokens": 200000,
25341
+ "mode": "chat",
25342
+ "output_cost_per_token": 1.8e-06,
25343
+ "source": "https://vercel.com/ai-gateway/models/glm-4.6",
25344
+ "supports_function_calling": true,
25345
+ "supports_parallel_function_calling": true,
25346
+ "supports_tool_choice": true
25347
+ },
25348
+ "openrouter/nvidia/nemotron-nano-9b-v2:free": {
25349
+ "input_cost_per_token": 0,
25350
+ "litellm_provider": "openrouter",
25351
+ "max_tokens": 128000,
25352
+ "mode": "chat",
25353
+ "output_cost_per_token": 0,
25354
+ "source": "https://openrouter.ai/nvidia/nemotron-nano-9b-v2:free",
25355
+ "supports_tool_choice": true
25356
+ },
25357
+ "openai/container": {
25358
+ "code_interpreter_cost_per_session": 0.03,
25359
+ "litellm_provider": "openai",
25360
+ "mode": "chat"
25361
+ },
25362
+ "azure/container": {
25363
+ "code_interpreter_cost_per_session": 0.03,
25364
+ "litellm_provider": "azure",
25365
+ "mode": "chat"
25366
+ },
25367
+ "vertex_ai/mistral-ocr-2505": {
25368
+ "litellm_provider": "vertex_ai",
25369
+ "mode": "ocr",
25370
+ "ocr_cost_per_page": 0.0005,
25371
+ "supported_endpoints": [
25372
+ "/v1/ocr"
25373
+ ],
25374
+ "source": "https://cloud.google.com/generative-ai-app-builder/pricing"
25375
+ },
25376
+ "azure_ai/doc-intelligence/prebuilt-read": {
25377
+ "litellm_provider": "azure_ai",
25378
+ "ocr_cost_per_page": 0.0015,
25379
+ "mode": "ocr",
25380
+ "supported_endpoints": [
25381
+ "/v1/ocr"
25382
+ ],
25383
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/"
25384
+ },
25385
+ "azure_ai/doc-intelligence/prebuilt-layout": {
25386
+ "litellm_provider": "azure_ai",
25387
+ "ocr_cost_per_page": 0.01,
25388
+ "mode": "ocr",
25389
+ "supported_endpoints": [
25390
+ "/v1/ocr"
25391
+ ],
25392
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/"
25393
+ },
25394
+ "azure_ai/doc-intelligence/prebuilt-document": {
25395
+ "litellm_provider": "azure_ai",
25396
+ "ocr_cost_per_page": 0.01,
25397
+ "mode": "ocr",
25398
+ "supported_endpoints": [
25399
+ "/v1/ocr"
25400
+ ],
25401
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-document-intelligence/"
25402
+ },
25403
+ "gemini-live-2.5-flash-preview-native-audio-09-2025": {
25404
+ "cache_read_input_token_cost": 7.5e-08,
25405
+ "input_cost_per_audio_token": 3e-06,
25406
+ "input_cost_per_token": 3e-07,
25407
+ "litellm_provider": "vertex_ai-language-models",
25408
+ "max_audio_length_hours": 8.4,
25409
+ "max_audio_per_prompt": 1,
25410
+ "max_images_per_prompt": 3000,
25411
+ "max_input_tokens": 1048576,
25412
+ "max_output_tokens": 65535,
25413
+ "max_pdf_size_mb": 30,
25414
+ "max_tokens": 65535,
25415
+ "max_video_length": 1,
25416
+ "max_videos_per_prompt": 10,
25417
+ "mode": "chat",
25418
+ "output_cost_per_audio_token": 1.2e-05,
25419
+ "output_cost_per_token": 2e-06,
25420
+ "source": "https://ai.google.dev/gemini-api/docs/pricing",
25421
+ "supported_endpoints": [
25422
+ "/v1/chat/completions",
25423
+ "/v1/completions"
25424
+ ],
25425
+ "supported_modalities": [
25426
+ "text",
25427
+ "image",
25428
+ "audio",
25429
+ "video"
25430
+ ],
25431
+ "supported_output_modalities": [
25432
+ "text",
25433
+ "audio"
25434
+ ],
25435
+ "supports_audio_input": true,
25436
+ "supports_audio_output": true,
25437
+ "supports_function_calling": true,
25438
+ "supports_parallel_function_calling": true,
25439
+ "supports_pdf_input": true,
25440
+ "supports_prompt_caching": true,
25441
+ "supports_response_schema": true,
25442
+ "supports_system_messages": true,
25443
+ "supports_tool_choice": true,
25444
+ "supports_url_context": true,
25445
+ "supports_vision": true,
25446
+ "supports_web_search": true
25447
+ },
25448
+ "gemini/gemini-live-2.5-flash-preview-native-audio-09-2025": {
25449
+ "cache_read_input_token_cost": 7.5e-08,
25450
+ "input_cost_per_audio_token": 3e-06,
25451
+ "input_cost_per_token": 3e-07,
25452
+ "litellm_provider": "gemini",
25453
+ "max_audio_length_hours": 8.4,
25454
+ "max_audio_per_prompt": 1,
25455
+ "max_images_per_prompt": 3000,
25456
+ "max_input_tokens": 1048576,
25457
+ "max_output_tokens": 65535,
25458
+ "max_pdf_size_mb": 30,
25459
+ "max_tokens": 65535,
25460
+ "max_video_length": 1,
25461
+ "max_videos_per_prompt": 10,
25462
+ "mode": "chat",
25463
+ "output_cost_per_audio_token": 1.2e-05,
25464
+ "output_cost_per_token": 2e-06,
25465
+ "rpm": 100000,
25466
+ "source": "https://ai.google.dev/gemini-api/docs/pricing",
25467
+ "supported_endpoints": [
25468
+ "/v1/chat/completions",
25469
+ "/v1/completions"
25470
+ ],
25471
+ "supported_modalities": [
25472
+ "text",
25473
+ "image",
25474
+ "audio",
25475
+ "video"
25476
+ ],
25477
+ "supported_output_modalities": [
25478
+ "text",
25479
+ "audio"
25480
+ ],
25481
+ "supports_audio_input": true,
25482
+ "supports_audio_output": true,
25483
+ "supports_function_calling": true,
25484
+ "supports_parallel_function_calling": true,
25485
+ "supports_pdf_input": true,
25486
+ "supports_prompt_caching": true,
25487
+ "supports_response_schema": true,
25488
+ "supports_system_messages": true,
25489
+ "supports_tool_choice": true,
25490
+ "supports_url_context": true,
25491
+ "supports_vision": true,
25492
+ "supports_web_search": true,
25493
+ "tpm": 8000000
25494
+ },
25495
+ "azure/gpt-image-1-mini": {
25496
+ "input_cost_per_pixel": 8.0566406e-09,
25497
+ "litellm_provider": "azure",
25498
+ "mode": "image_generation",
25499
+ "output_cost_per_pixel": 0.0,
25500
+ "supported_endpoints": [
25501
+ "/v1/images/generations"
25502
+ ]
25503
+ },
25504
+ "azure/low/1024-x-1024/gpt-image-1-mini": {
25505
+ "input_cost_per_pixel": 2.0751953125e-09,
25506
+ "litellm_provider": "azure",
25507
+ "mode": "image_generation",
25508
+ "output_cost_per_pixel": 0.0,
25509
+ "supported_endpoints": [
25510
+ "/v1/images/generations"
25511
+ ]
25512
+ },
25513
+ "azure/low/1024-x-1536/gpt-image-1-mini": {
25514
+ "input_cost_per_pixel": 2.0751953125e-09,
25515
+ "litellm_provider": "azure",
25516
+ "mode": "image_generation",
25517
+ "output_cost_per_pixel": 0.0,
25518
+ "supported_endpoints": [
25519
+ "/v1/images/generations"
25520
+ ]
25521
+ },
25522
+ "azure/low/1536-x-1024/gpt-image-1-mini": {
25523
+ "input_cost_per_pixel": 2.0345052083e-09,
25524
+ "litellm_provider": "azure",
25525
+ "mode": "image_generation",
25526
+ "output_cost_per_pixel": 0.0,
25527
+ "supported_endpoints": [
25528
+ "/v1/images/generations"
25529
+ ]
25530
+ },
25531
+ "azure/medium/1024-x-1024/gpt-image-1-mini": {
25532
+ "input_cost_per_pixel": 8.056640625e-09,
25533
+ "litellm_provider": "azure",
25534
+ "mode": "image_generation",
25535
+ "output_cost_per_pixel": 0.0,
25536
+ "supported_endpoints": [
25537
+ "/v1/images/generations"
25538
+ ]
25539
+ },
25540
+ "azure/medium/1024-x-1536/gpt-image-1-mini": {
25541
+ "input_cost_per_pixel": 8.056640625e-09,
25542
+ "litellm_provider": "azure",
25543
+ "mode": "image_generation",
25544
+ "output_cost_per_pixel": 0.0,
25545
+ "supported_endpoints": [
25546
+ "/v1/images/generations"
25547
+ ]
25548
+ },
25549
+ "azure/medium/1536-x-1024/gpt-image-1-mini": {
25550
+ "input_cost_per_pixel": 7.9752604167e-09,
25551
+ "litellm_provider": "azure",
25552
+ "mode": "image_generation",
25553
+ "output_cost_per_pixel": 0.0,
25554
+ "supported_endpoints": [
25555
+ "/v1/images/generations"
25556
+ ]
25557
+ },
25558
+ "azure/high/1024-x-1024/gpt-image-1-mini": {
25559
+ "input_cost_per_pixel": 3.173828125e-08,
25560
+ "litellm_provider": "azure",
25561
+ "mode": "image_generation",
25562
+ "output_cost_per_pixel": 0.0,
25563
+ "supported_endpoints": [
25564
+ "/v1/images/generations"
25565
+ ]
25566
+ },
25567
+ "azure/high/1024-x-1536/gpt-image-1-mini": {
25568
+ "input_cost_per_pixel": 3.173828125e-08,
25569
+ "litellm_provider": "azure",
25570
+ "mode": "image_generation",
25571
+ "output_cost_per_pixel": 0.0,
25572
+ "supported_endpoints": [
25573
+ "/v1/images/generations"
25574
+ ]
25575
+ },
25576
+ "azure/high/1536-x-1024/gpt-image-1-mini": {
25577
+ "input_cost_per_pixel": 3.1575520833e-08,
25578
+ "litellm_provider": "azure",
25579
+ "mode": "image_generation",
25580
+ "output_cost_per_pixel": 0.0,
25581
+ "supported_endpoints": [
25582
+ "/v1/images/generations"
25583
+ ]
25584
+ },
25585
+ "firecrawl/search": {
25586
+ "litellm_provider": "firecrawl",
25587
+ "mode": "search",
25588
+ "tiered_pricing": [
25589
+ {
25590
+ "input_cost_per_query": 0.00166,
25591
+ "max_results_range": [
25592
+ 1,
25593
+ 10
25594
+ ]
25595
+ },
25596
+ {
25597
+ "input_cost_per_query": 0.00332,
25598
+ "max_results_range": [
25599
+ 11,
25600
+ 20
25601
+ ]
25602
+ },
25603
+ {
25604
+ "input_cost_per_query": 0.00498,
25605
+ "max_results_range": [
25606
+ 21,
25607
+ 30
25608
+ ]
25609
+ },
25610
+ {
25611
+ "input_cost_per_query": 0.00664,
25612
+ "max_results_range": [
25613
+ 31,
25614
+ 40
25615
+ ]
25616
+ },
25617
+ {
25618
+ "input_cost_per_query": 0.0083,
25619
+ "max_results_range": [
25620
+ 41,
25621
+ 50
25622
+ ]
25623
+ },
25624
+ {
25625
+ "input_cost_per_query": 0.00996,
25626
+ "max_results_range": [
25627
+ 51,
25628
+ 60
25629
+ ]
25630
+ },
25631
+ {
25632
+ "input_cost_per_query": 0.01162,
25633
+ "max_results_range": [
25634
+ 61,
25635
+ 70
25636
+ ]
25637
+ },
25638
+ {
25639
+ "input_cost_per_query": 0.01328,
25640
+ "max_results_range": [
25641
+ 71,
25642
+ 80
25643
+ ]
25644
+ },
25645
+ {
25646
+ "input_cost_per_query": 0.01494,
25647
+ "max_results_range": [
25648
+ 81,
25649
+ 90
25650
+ ]
25651
+ },
25652
+ {
25653
+ "input_cost_per_query": 0.0166,
25654
+ "max_results_range": [
25655
+ 91,
25656
+ 100
25657
+ ]
25658
+ }
25659
+ ],
25660
+ "metadata": {
25661
+ "notes": "Firecrawl search pricing: $83 for 100,000 credits, 2 credits per 10 results. Cost = ceiling(limit/10) * 2 * $0.00083"
25662
+ }
25663
+ },
25664
+ "searxng/search": {
25665
+ "litellm_provider": "searxng",
25666
+ "mode": "search",
25667
+ "input_cost_per_query": 0.0,
25668
+ "metadata": {
25669
+ "notes": "SearXNG is an open-source metasearch engine. Free to use when self-hosted or using public instances."
25670
+ }
25671
+ },
25672
+ "azure/gpt-5-pro": {
25673
+ "input_cost_per_token": 1.5e-05,
25674
+ "litellm_provider": "azure",
25675
+ "max_input_tokens": 272000,
25676
+ "max_output_tokens": 128000,
25677
+ "max_tokens": 400000,
25678
+ "mode": "responses",
25679
+ "output_cost_per_token": 0.00012,
25680
+ "source": "https://learn.microsoft.com/en-us/azure/ai-foundry/foundry-models/concepts/models-sold-directly-by-azure?pivots=azure-openai&tabs=global-standard-aoai%2Cstandard-chat-completions%2Cglobal-standard#gpt-5",
25681
+ "supported_endpoints": [
25682
+ "/v1/responses"
25683
+ ],
25684
+ "supported_modalities": [
25685
+ "text",
25686
+ "image"
25687
+ ],
25688
+ "supported_output_modalities": [
25689
+ "text"
25690
+ ],
25691
+ "supports_function_calling": true,
25692
+ "supports_parallel_function_calling": true,
25693
+ "supports_pdf_input": true,
25694
+ "supports_prompt_caching": true,
25695
+ "supports_reasoning": true,
25696
+ "supports_response_schema": true,
25697
+ "supports_system_messages": true,
25698
+ "supports_tool_choice": true,
25699
+ "supports_vision": true
25700
+ },
25701
+ "vertex_ai/minimaxai/minimax-m2-maas": {
25702
+ "input_cost_per_token": 3e-07,
25703
+ "litellm_provider": "vertex_ai-minimax_models",
25704
+ "max_input_tokens": 196608,
25705
+ "max_output_tokens": 196608,
25706
+ "max_tokens": 196608,
25707
+ "mode": "chat",
25708
+ "output_cost_per_token": 1.2e-06,
25709
+ "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
25710
+ "supports_function_calling": true,
25711
+ "supports_tool_choice": true
23548
25712
  }
23549
25713
  }