tokencostauto 0.1.447__py3-none-any.whl → 0.1.449__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- tokencostauto/model_prices.json +329 -253
- {tokencostauto-0.1.447.dist-info → tokencostauto-0.1.449.dist-info}/METADATA +1 -1
- tokencostauto-0.1.449.dist-info/RECORD +9 -0
- tokencostauto-0.1.447.dist-info/RECORD +0 -9
- {tokencostauto-0.1.447.dist-info → tokencostauto-0.1.449.dist-info}/WHEEL +0 -0
- {tokencostauto-0.1.447.dist-info → tokencostauto-0.1.449.dist-info}/licenses/LICENSE +0 -0
- {tokencostauto-0.1.447.dist-info → tokencostauto-0.1.449.dist-info}/top_level.txt +0 -0
tokencostauto/model_prices.json
CHANGED
|
@@ -1656,14 +1656,14 @@
|
|
|
1656
1656
|
"mode": "embedding"
|
|
1657
1657
|
},
|
|
1658
1658
|
"deepseek-chat": {
|
|
1659
|
-
"cache_read_input_token_cost":
|
|
1660
|
-
"input_cost_per_token":
|
|
1659
|
+
"cache_read_input_token_cost": 2.8e-08,
|
|
1660
|
+
"input_cost_per_token": 2.8e-07,
|
|
1661
1661
|
"litellm_provider": "deepseek",
|
|
1662
1662
|
"max_input_tokens": 131072,
|
|
1663
1663
|
"max_output_tokens": 8192,
|
|
1664
1664
|
"max_tokens": 8192,
|
|
1665
1665
|
"mode": "chat",
|
|
1666
|
-
"output_cost_per_token":
|
|
1666
|
+
"output_cost_per_token": 4.2e-07,
|
|
1667
1667
|
"source": "https://api-docs.deepseek.com/quick_start/pricing",
|
|
1668
1668
|
"supported_endpoints": [
|
|
1669
1669
|
"/v1/chat/completions"
|
|
@@ -8316,15 +8316,15 @@
|
|
|
8316
8316
|
},
|
|
8317
8317
|
"deepseek/deepseek-chat": {
|
|
8318
8318
|
"cache_creation_input_token_cost": 0.0,
|
|
8319
|
-
"cache_read_input_token_cost":
|
|
8320
|
-
"input_cost_per_token": 2.
|
|
8321
|
-
"input_cost_per_token_cache_hit":
|
|
8319
|
+
"cache_read_input_token_cost": 2.8e-08,
|
|
8320
|
+
"input_cost_per_token": 2.8e-07,
|
|
8321
|
+
"input_cost_per_token_cache_hit": 2.8e-08,
|
|
8322
8322
|
"litellm_provider": "deepseek",
|
|
8323
|
-
"max_input_tokens":
|
|
8323
|
+
"max_input_tokens": 128000,
|
|
8324
8324
|
"max_output_tokens": 8192,
|
|
8325
8325
|
"max_tokens": 8192,
|
|
8326
8326
|
"mode": "chat",
|
|
8327
|
-
"output_cost_per_token":
|
|
8327
|
+
"output_cost_per_token": 4.2e-07,
|
|
8328
8328
|
"supports_assistant_prefill": true,
|
|
8329
8329
|
"supports_function_calling": true,
|
|
8330
8330
|
"supports_prompt_caching": true,
|
|
@@ -8995,14 +8995,15 @@
|
|
|
8995
8995
|
"supports_tool_choice": true
|
|
8996
8996
|
},
|
|
8997
8997
|
"deepseek/deepseek-reasoner": {
|
|
8998
|
-
"
|
|
8999
|
-
"
|
|
8998
|
+
"cache_read_input_token_cost": 2.8e-08,
|
|
8999
|
+
"input_cost_per_token": 2.8e-07,
|
|
9000
|
+
"input_cost_per_token_cache_hit": 2.8e-08,
|
|
9000
9001
|
"litellm_provider": "deepseek",
|
|
9001
|
-
"max_input_tokens":
|
|
9002
|
+
"max_input_tokens": 128000,
|
|
9002
9003
|
"max_output_tokens": 8192,
|
|
9003
9004
|
"max_tokens": 8192,
|
|
9004
9005
|
"mode": "chat",
|
|
9005
|
-
"output_cost_per_token":
|
|
9006
|
+
"output_cost_per_token": 4.2e-07,
|
|
9006
9007
|
"supports_assistant_prefill": true,
|
|
9007
9008
|
"supports_function_calling": true,
|
|
9008
9009
|
"supports_prompt_caching": true,
|
|
@@ -12284,7 +12285,8 @@
|
|
|
12284
12285
|
"supports_tool_choice": true
|
|
12285
12286
|
},
|
|
12286
12287
|
"gemini-2.5-pro-exp-03-25": {
|
|
12287
|
-
"cache_read_input_token_cost":
|
|
12288
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
12289
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
12288
12290
|
"input_cost_per_token": 1.25e-06,
|
|
12289
12291
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
12290
12292
|
"litellm_provider": "vertex_ai-language-models",
|
|
@@ -12570,7 +12572,8 @@
|
|
|
12570
12572
|
},
|
|
12571
12573
|
"gemini-2.5-pro-preview-05-06": {
|
|
12572
12574
|
"deprecation_date": "2025-12-02",
|
|
12573
|
-
"cache_read_input_token_cost":
|
|
12575
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
12576
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
12574
12577
|
"input_cost_per_audio_token": 1.25e-06,
|
|
12575
12578
|
"input_cost_per_token": 1.25e-06,
|
|
12576
12579
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -12619,7 +12622,8 @@
|
|
|
12619
12622
|
},
|
|
12620
12623
|
"gemini-2.5-pro-preview-03-25": {
|
|
12621
12624
|
"deprecation_date": "2025-12-02",
|
|
12622
|
-
"cache_read_input_token_cost":
|
|
12625
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
12626
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
12623
12627
|
"input_cost_per_audio_token": 1.25e-06,
|
|
12624
12628
|
"input_cost_per_token": 1.25e-06,
|
|
12625
12629
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -12701,7 +12705,8 @@
|
|
|
12701
12705
|
},
|
|
12702
12706
|
"gemini/gemini-2.5-pro-preview-05-06": {
|
|
12703
12707
|
"deprecation_date": "2025-12-02",
|
|
12704
|
-
"cache_read_input_token_cost":
|
|
12708
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
12709
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
12705
12710
|
"input_cost_per_audio_token": 7e-07,
|
|
12706
12711
|
"input_cost_per_token": 1.25e-06,
|
|
12707
12712
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -12743,7 +12748,8 @@
|
|
|
12743
12748
|
},
|
|
12744
12749
|
"gemini/gemini-2.5-pro-preview-03-25": {
|
|
12745
12750
|
"deprecation_date": "2025-12-02",
|
|
12746
|
-
"cache_read_input_token_cost":
|
|
12751
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
12752
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
12747
12753
|
"input_cost_per_audio_token": 7e-07,
|
|
12748
12754
|
"input_cost_per_token": 1.25e-06,
|
|
12749
12755
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -13921,7 +13927,8 @@
|
|
|
13921
13927
|
"supports_web_search": true
|
|
13922
13928
|
},
|
|
13923
13929
|
"gemini-2.5-pro-preview-tts": {
|
|
13924
|
-
"cache_read_input_token_cost":
|
|
13930
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
13931
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
13925
13932
|
"input_cost_per_audio_token": 7e-07,
|
|
13926
13933
|
"input_cost_per_token": 1.25e-06,
|
|
13927
13934
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -13996,7 +14003,8 @@
|
|
|
13996
14003
|
"tpm": 10000000
|
|
13997
14004
|
},
|
|
13998
14005
|
"gemini/gemini-2.5-pro-preview-tts": {
|
|
13999
|
-
"cache_read_input_token_cost":
|
|
14006
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
14007
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
14000
14008
|
"input_cost_per_audio_token": 7e-07,
|
|
14001
14009
|
"input_cost_per_token": 1.25e-06,
|
|
14002
14010
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -14563,7 +14571,8 @@
|
|
|
14563
14571
|
]
|
|
14564
14572
|
},
|
|
14565
14573
|
"gemini-2.5-pro-preview-06-05": {
|
|
14566
|
-
"cache_read_input_token_cost":
|
|
14574
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
14575
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
14567
14576
|
"input_cost_per_audio_token": 1.25e-06,
|
|
14568
14577
|
"input_cost_per_token": 1.25e-06,
|
|
14569
14578
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -14608,7 +14617,8 @@
|
|
|
14608
14617
|
"supports_web_search": true
|
|
14609
14618
|
},
|
|
14610
14619
|
"gemini/gemini-2.5-pro-preview-06-05": {
|
|
14611
|
-
"cache_read_input_token_cost":
|
|
14620
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
14621
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
14612
14622
|
"input_cost_per_audio_token": 7e-07,
|
|
14613
14623
|
"input_cost_per_token": 1.25e-06,
|
|
14614
14624
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -15441,7 +15451,8 @@
|
|
|
15441
15451
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing"
|
|
15442
15452
|
},
|
|
15443
15453
|
"gemini/gemini-2.5-pro": {
|
|
15444
|
-
"cache_read_input_token_cost":
|
|
15454
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
15455
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
15445
15456
|
"input_cost_per_token": 1.25e-06,
|
|
15446
15457
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
15447
15458
|
"litellm_provider": "gemini",
|
|
@@ -15756,6 +15767,7 @@
|
|
|
15756
15767
|
},
|
|
15757
15768
|
"gemini-2.5-pro": {
|
|
15758
15769
|
"cache_read_input_token_cost": 1.25e-07,
|
|
15770
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
15759
15771
|
"cache_creation_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
15760
15772
|
"input_cost_per_token": 1.25e-06,
|
|
15761
15773
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -18167,7 +18179,7 @@
|
|
|
18167
18179
|
"input_cost_per_token_flex": 6.25e-07,
|
|
18168
18180
|
"input_cost_per_token_priority": 2.5e-06,
|
|
18169
18181
|
"litellm_provider": "openai",
|
|
18170
|
-
"max_input_tokens":
|
|
18182
|
+
"max_input_tokens": 400000,
|
|
18171
18183
|
"max_output_tokens": 128000,
|
|
18172
18184
|
"max_tokens": 128000,
|
|
18173
18185
|
"mode": "chat",
|
|
@@ -18206,7 +18218,7 @@
|
|
|
18206
18218
|
"input_cost_per_token_flex": 1.25e-07,
|
|
18207
18219
|
"input_cost_per_token_priority": 4.5e-07,
|
|
18208
18220
|
"litellm_provider": "openai",
|
|
18209
|
-
"max_input_tokens":
|
|
18221
|
+
"max_input_tokens": 400000,
|
|
18210
18222
|
"max_output_tokens": 128000,
|
|
18211
18223
|
"max_tokens": 128000,
|
|
18212
18224
|
"mode": "chat",
|
|
@@ -18244,7 +18256,7 @@
|
|
|
18244
18256
|
"input_cost_per_token_flex": 2.5e-08,
|
|
18245
18257
|
"input_cost_per_token_priority": 2.5e-06,
|
|
18246
18258
|
"litellm_provider": "openai",
|
|
18247
|
-
"max_input_tokens":
|
|
18259
|
+
"max_input_tokens": 400000,
|
|
18248
18260
|
"max_output_tokens": 128000,
|
|
18249
18261
|
"max_tokens": 128000,
|
|
18250
18262
|
"mode": "chat",
|
|
@@ -18277,9 +18289,9 @@
|
|
|
18277
18289
|
"cache_read_input_token_cost": 1.25e-07,
|
|
18278
18290
|
"input_cost_per_token": 1.25e-06,
|
|
18279
18291
|
"litellm_provider": "openai",
|
|
18280
|
-
"max_input_tokens":
|
|
18281
|
-
"max_output_tokens":
|
|
18282
|
-
"max_tokens":
|
|
18292
|
+
"max_input_tokens": 128000,
|
|
18293
|
+
"max_output_tokens": 16384,
|
|
18294
|
+
"max_tokens": 16384,
|
|
18283
18295
|
"mode": "chat",
|
|
18284
18296
|
"output_cost_per_token": 1e-05,
|
|
18285
18297
|
"supported_endpoints": [
|
|
@@ -18345,7 +18357,7 @@
|
|
|
18345
18357
|
"input_cost_per_token_flex": 6.25e-07,
|
|
18346
18358
|
"input_cost_per_token_priority": 2.5e-06,
|
|
18347
18359
|
"litellm_provider": "openai",
|
|
18348
|
-
"max_input_tokens":
|
|
18360
|
+
"max_input_tokens": 400000,
|
|
18349
18361
|
"max_output_tokens": 128000,
|
|
18350
18362
|
"max_tokens": 128000,
|
|
18351
18363
|
"mode": "chat",
|
|
@@ -18384,7 +18396,7 @@
|
|
|
18384
18396
|
"input_cost_per_token_flex": 1.25e-07,
|
|
18385
18397
|
"input_cost_per_token_priority": 4.5e-07,
|
|
18386
18398
|
"litellm_provider": "openai",
|
|
18387
|
-
"max_input_tokens":
|
|
18399
|
+
"max_input_tokens": 400000,
|
|
18388
18400
|
"max_output_tokens": 128000,
|
|
18389
18401
|
"max_tokens": 128000,
|
|
18390
18402
|
"mode": "chat",
|
|
@@ -18421,7 +18433,7 @@
|
|
|
18421
18433
|
"input_cost_per_token": 5e-08,
|
|
18422
18434
|
"input_cost_per_token_flex": 2.5e-08,
|
|
18423
18435
|
"litellm_provider": "openai",
|
|
18424
|
-
"max_input_tokens":
|
|
18436
|
+
"max_input_tokens": 400000,
|
|
18425
18437
|
"max_output_tokens": 128000,
|
|
18426
18438
|
"max_tokens": 128000,
|
|
18427
18439
|
"mode": "chat",
|
|
@@ -18646,9 +18658,9 @@
|
|
|
18646
18658
|
"cache_read_input_token_cost": 1.25e-07,
|
|
18647
18659
|
"input_cost_per_token": 1.25e-06,
|
|
18648
18660
|
"litellm_provider": "azure",
|
|
18649
|
-
"max_input_tokens":
|
|
18650
|
-
"max_output_tokens":
|
|
18651
|
-
"max_tokens":
|
|
18661
|
+
"max_input_tokens": 128000,
|
|
18662
|
+
"max_output_tokens": 16384,
|
|
18663
|
+
"max_tokens": 16384,
|
|
18652
18664
|
"mode": "chat",
|
|
18653
18665
|
"output_cost_per_token": 1e-05,
|
|
18654
18666
|
"source": "https://azure.microsoft.com/en-us/blog/gpt-5-in-azure-ai-foundry-the-future-of-ai-apps-and-agents-starts-here/",
|
|
@@ -20611,7 +20623,7 @@
|
|
|
20611
20623
|
"cache_read_input_token_cost": 2.5e-08,
|
|
20612
20624
|
"input_cost_per_token": 2.5e-07,
|
|
20613
20625
|
"litellm_provider": "openrouter",
|
|
20614
|
-
"max_input_tokens":
|
|
20626
|
+
"max_input_tokens": 400000,
|
|
20615
20627
|
"max_output_tokens": 128000,
|
|
20616
20628
|
"max_tokens": 128000,
|
|
20617
20629
|
"mode": "chat",
|
|
@@ -20630,7 +20642,7 @@
|
|
|
20630
20642
|
"cache_read_input_token_cost": 5e-09,
|
|
20631
20643
|
"input_cost_per_token": 5e-08,
|
|
20632
20644
|
"litellm_provider": "openrouter",
|
|
20633
|
-
"max_input_tokens":
|
|
20645
|
+
"max_input_tokens": 400000,
|
|
20634
20646
|
"max_output_tokens": 128000,
|
|
20635
20647
|
"max_tokens": 128000,
|
|
20636
20648
|
"mode": "chat",
|
|
@@ -20649,9 +20661,9 @@
|
|
|
20649
20661
|
"cache_read_input_token_cost": 1.25e-07,
|
|
20650
20662
|
"input_cost_per_token": 1.25e-06,
|
|
20651
20663
|
"litellm_provider": "openrouter",
|
|
20652
|
-
"max_input_tokens":
|
|
20653
|
-
"max_output_tokens":
|
|
20654
|
-
"max_tokens":
|
|
20664
|
+
"max_input_tokens": 128000,
|
|
20665
|
+
"max_output_tokens": 16384,
|
|
20666
|
+
"max_tokens": 16384,
|
|
20655
20667
|
"mode": "chat",
|
|
20656
20668
|
"output_cost_per_token": 1e-05,
|
|
20657
20669
|
"supported_modalities": [
|
|
@@ -22912,14 +22924,14 @@
|
|
|
22912
22924
|
"supports_vision": true
|
|
22913
22925
|
},
|
|
22914
22926
|
"deepseek-reasoner": {
|
|
22915
|
-
"cache_read_input_token_cost":
|
|
22916
|
-
"input_cost_per_token":
|
|
22927
|
+
"cache_read_input_token_cost": 2.8e-08,
|
|
22928
|
+
"input_cost_per_token": 2.8e-07,
|
|
22917
22929
|
"litellm_provider": "deepseek",
|
|
22918
22930
|
"max_input_tokens": 131072,
|
|
22919
22931
|
"max_output_tokens": 65536,
|
|
22920
22932
|
"max_tokens": 65536,
|
|
22921
22933
|
"mode": "chat",
|
|
22922
|
-
"output_cost_per_token":
|
|
22934
|
+
"output_cost_per_token": 4.2e-07,
|
|
22923
22935
|
"source": "https://api-docs.deepseek.com/quick_start/pricing",
|
|
22924
22936
|
"supported_endpoints": [
|
|
22925
22937
|
"/v1/chat/completions"
|
|
@@ -22949,7 +22961,7 @@
|
|
|
22949
22961
|
"cache_read_input_token_cost": 1.25e-07,
|
|
22950
22962
|
"input_cost_per_token": 1.25e-06,
|
|
22951
22963
|
"litellm_provider": "openai",
|
|
22952
|
-
"max_input_tokens":
|
|
22964
|
+
"max_input_tokens": 400000,
|
|
22953
22965
|
"max_output_tokens": 128000,
|
|
22954
22966
|
"max_tokens": 128000,
|
|
22955
22967
|
"mode": "responses",
|
|
@@ -23439,7 +23451,7 @@
|
|
|
23439
23451
|
"cache_read_input_token_cost": 1.25e-07,
|
|
23440
23452
|
"input_cost_per_token": 1.25e-06,
|
|
23441
23453
|
"litellm_provider": "openrouter",
|
|
23442
|
-
"max_input_tokens":
|
|
23454
|
+
"max_input_tokens": 400000,
|
|
23443
23455
|
"max_output_tokens": 128000,
|
|
23444
23456
|
"max_tokens": 128000,
|
|
23445
23457
|
"mode": "chat",
|
|
@@ -23458,7 +23470,7 @@
|
|
|
23458
23470
|
"cache_read_input_token_cost": 1.25e-07,
|
|
23459
23471
|
"input_cost_per_token": 1.25e-06,
|
|
23460
23472
|
"litellm_provider": "openrouter",
|
|
23461
|
-
"max_input_tokens":
|
|
23473
|
+
"max_input_tokens": 400000,
|
|
23462
23474
|
"max_output_tokens": 128000,
|
|
23463
23475
|
"max_tokens": 128000,
|
|
23464
23476
|
"mode": "chat",
|
|
@@ -26083,7 +26095,7 @@
|
|
|
26083
26095
|
"input_cost_per_token": 1.25e-06,
|
|
26084
26096
|
"input_cost_per_token_priority": 2.5e-06,
|
|
26085
26097
|
"litellm_provider": "openai",
|
|
26086
|
-
"max_input_tokens":
|
|
26098
|
+
"max_input_tokens": 400000,
|
|
26087
26099
|
"max_output_tokens": 128000,
|
|
26088
26100
|
"max_tokens": 128000,
|
|
26089
26101
|
"mode": "chat",
|
|
@@ -26119,7 +26131,7 @@
|
|
|
26119
26131
|
"input_cost_per_token": 1.25e-06,
|
|
26120
26132
|
"input_cost_per_token_priority": 2.5e-06,
|
|
26121
26133
|
"litellm_provider": "openai",
|
|
26122
|
-
"max_input_tokens":
|
|
26134
|
+
"max_input_tokens": 400000,
|
|
26123
26135
|
"max_output_tokens": 128000,
|
|
26124
26136
|
"max_tokens": 128000,
|
|
26125
26137
|
"mode": "chat",
|
|
@@ -26190,7 +26202,7 @@
|
|
|
26190
26202
|
"input_cost_per_token": 1.25e-06,
|
|
26191
26203
|
"input_cost_per_token_priority": 2.5e-06,
|
|
26192
26204
|
"litellm_provider": "openai",
|
|
26193
|
-
"max_input_tokens":
|
|
26205
|
+
"max_input_tokens": 400000,
|
|
26194
26206
|
"max_output_tokens": 128000,
|
|
26195
26207
|
"max_tokens": 128000,
|
|
26196
26208
|
"mode": "responses",
|
|
@@ -26223,7 +26235,7 @@
|
|
|
26223
26235
|
"input_cost_per_token": 2.5e-07,
|
|
26224
26236
|
"input_cost_per_token_priority": 4.5e-07,
|
|
26225
26237
|
"litellm_provider": "openai",
|
|
26226
|
-
"max_input_tokens":
|
|
26238
|
+
"max_input_tokens": 400000,
|
|
26227
26239
|
"max_output_tokens": 128000,
|
|
26228
26240
|
"max_tokens": 128000,
|
|
26229
26241
|
"mode": "responses",
|
|
@@ -26980,7 +26992,7 @@
|
|
|
26980
26992
|
"input_cost_per_token": 1.25e-06,
|
|
26981
26993
|
"input_cost_per_token_priority": 2.5e-06,
|
|
26982
26994
|
"litellm_provider": "azure",
|
|
26983
|
-
"max_input_tokens":
|
|
26995
|
+
"max_input_tokens": 400000,
|
|
26984
26996
|
"max_output_tokens": 128000,
|
|
26985
26997
|
"max_tokens": 128000,
|
|
26986
26998
|
"mode": "chat",
|
|
@@ -27051,7 +27063,7 @@
|
|
|
27051
27063
|
"input_cost_per_token": 1.25e-06,
|
|
27052
27064
|
"input_cost_per_token_priority": 2.5e-06,
|
|
27053
27065
|
"litellm_provider": "azure",
|
|
27054
|
-
"max_input_tokens":
|
|
27066
|
+
"max_input_tokens": 400000,
|
|
27055
27067
|
"max_output_tokens": 128000,
|
|
27056
27068
|
"max_tokens": 128000,
|
|
27057
27069
|
"mode": "responses",
|
|
@@ -27084,7 +27096,7 @@
|
|
|
27084
27096
|
"input_cost_per_token": 2.5e-07,
|
|
27085
27097
|
"input_cost_per_token_priority": 4.5e-07,
|
|
27086
27098
|
"litellm_provider": "azure",
|
|
27087
|
-
"max_input_tokens":
|
|
27099
|
+
"max_input_tokens": 400000,
|
|
27088
27100
|
"max_output_tokens": 128000,
|
|
27089
27101
|
"max_tokens": 128000,
|
|
27090
27102
|
"mode": "responses",
|
|
@@ -27148,7 +27160,7 @@
|
|
|
27148
27160
|
"cache_read_input_token_cost": 1.4e-07,
|
|
27149
27161
|
"input_cost_per_token": 1.38e-06,
|
|
27150
27162
|
"litellm_provider": "azure",
|
|
27151
|
-
"max_input_tokens":
|
|
27163
|
+
"max_input_tokens": 128000,
|
|
27152
27164
|
"max_output_tokens": 128000,
|
|
27153
27165
|
"max_tokens": 128000,
|
|
27154
27166
|
"mode": "chat",
|
|
@@ -27274,7 +27286,7 @@
|
|
|
27274
27286
|
"cache_read_input_token_cost": 1.25e-07,
|
|
27275
27287
|
"input_cost_per_token": 1.25e-06,
|
|
27276
27288
|
"litellm_provider": "azure",
|
|
27277
|
-
"max_input_tokens":
|
|
27289
|
+
"max_input_tokens": 128000,
|
|
27278
27290
|
"max_output_tokens": 128000,
|
|
27279
27291
|
"max_tokens": 128000,
|
|
27280
27292
|
"mode": "chat",
|
|
@@ -27400,7 +27412,7 @@
|
|
|
27400
27412
|
"cache_read_input_token_cost": 1.25e-07,
|
|
27401
27413
|
"input_cost_per_token": 1.25e-06,
|
|
27402
27414
|
"litellm_provider": "azure",
|
|
27403
|
-
"max_input_tokens":
|
|
27415
|
+
"max_input_tokens": 128000,
|
|
27404
27416
|
"max_output_tokens": 128000,
|
|
27405
27417
|
"max_tokens": 128000,
|
|
27406
27418
|
"mode": "chat",
|
|
@@ -27526,7 +27538,7 @@
|
|
|
27526
27538
|
"cache_read_input_token_cost": 1.4e-07,
|
|
27527
27539
|
"input_cost_per_token": 1.38e-06,
|
|
27528
27540
|
"litellm_provider": "azure",
|
|
27529
|
-
"max_input_tokens":
|
|
27541
|
+
"max_input_tokens": 128000,
|
|
27530
27542
|
"max_output_tokens": 128000,
|
|
27531
27543
|
"max_tokens": 128000,
|
|
27532
27544
|
"mode": "chat",
|
|
@@ -29110,7 +29122,7 @@
|
|
|
29110
29122
|
"cache_read_input_token_cost": 1.25e-07,
|
|
29111
29123
|
"input_cost_per_token": 1.25e-06,
|
|
29112
29124
|
"litellm_provider": "azure",
|
|
29113
|
-
"max_input_tokens":
|
|
29125
|
+
"max_input_tokens": 272000,
|
|
29114
29126
|
"max_output_tokens": 128000,
|
|
29115
29127
|
"max_tokens": 128000,
|
|
29116
29128
|
"mode": "responses",
|
|
@@ -31913,7 +31925,7 @@
|
|
|
31913
31925
|
"cache_read_input_token_cost": 1.75e-07,
|
|
31914
31926
|
"input_cost_per_token": 1.75e-06,
|
|
31915
31927
|
"litellm_provider": "azure",
|
|
31916
|
-
"max_input_tokens":
|
|
31928
|
+
"max_input_tokens": 272000,
|
|
31917
31929
|
"max_output_tokens": 128000,
|
|
31918
31930
|
"max_tokens": 128000,
|
|
31919
31931
|
"mode": "chat",
|
|
@@ -32016,7 +32028,7 @@
|
|
|
32016
32028
|
"azure/gpt-5.2-pro": {
|
|
32017
32029
|
"input_cost_per_token": 2.1e-05,
|
|
32018
32030
|
"litellm_provider": "azure",
|
|
32019
|
-
"max_input_tokens":
|
|
32031
|
+
"max_input_tokens": 272000,
|
|
32020
32032
|
"max_output_tokens": 128000,
|
|
32021
32033
|
"max_tokens": 128000,
|
|
32022
32034
|
"mode": "responses",
|
|
@@ -32047,7 +32059,7 @@
|
|
|
32047
32059
|
"azure/gpt-5.2-pro-2025-12-11": {
|
|
32048
32060
|
"input_cost_per_token": 2.1e-05,
|
|
32049
32061
|
"litellm_provider": "azure",
|
|
32050
|
-
"max_input_tokens":
|
|
32062
|
+
"max_input_tokens": 272000,
|
|
32051
32063
|
"max_output_tokens": 128000,
|
|
32052
32064
|
"max_tokens": 128000,
|
|
32053
32065
|
"mode": "responses",
|
|
@@ -33985,8 +33997,8 @@
|
|
|
33985
33997
|
"novita/deepseek/deepseek-v3.2": {
|
|
33986
33998
|
"litellm_provider": "novita",
|
|
33987
33999
|
"mode": "chat",
|
|
33988
|
-
"input_cost_per_token":
|
|
33989
|
-
"output_cost_per_token":
|
|
34000
|
+
"input_cost_per_token": 2.69e-07,
|
|
34001
|
+
"output_cost_per_token": 4e-07,
|
|
33990
34002
|
"max_input_tokens": 163840,
|
|
33991
34003
|
"max_output_tokens": 65536,
|
|
33992
34004
|
"max_tokens": 65536,
|
|
@@ -33995,15 +34007,15 @@
|
|
|
33995
34007
|
"supports_tool_choice": true,
|
|
33996
34008
|
"supports_system_messages": true,
|
|
33997
34009
|
"supports_response_schema": true,
|
|
33998
|
-
"cache_read_input_token_cost":
|
|
33999
|
-
"input_cost_per_token_cache_hit":
|
|
34010
|
+
"cache_read_input_token_cost": 1.345e-07,
|
|
34011
|
+
"input_cost_per_token_cache_hit": 1.345e-07,
|
|
34000
34012
|
"supports_reasoning": true
|
|
34001
34013
|
},
|
|
34002
34014
|
"novita/minimax/minimax-m2.1": {
|
|
34003
34015
|
"litellm_provider": "novita",
|
|
34004
34016
|
"mode": "chat",
|
|
34005
|
-
"input_cost_per_token":
|
|
34006
|
-
"output_cost_per_token":
|
|
34017
|
+
"input_cost_per_token": 3e-07,
|
|
34018
|
+
"output_cost_per_token": 1.2e-06,
|
|
34007
34019
|
"max_input_tokens": 204800,
|
|
34008
34020
|
"max_output_tokens": 131072,
|
|
34009
34021
|
"max_tokens": 131072,
|
|
@@ -34012,15 +34024,14 @@
|
|
|
34012
34024
|
"supports_tool_choice": true,
|
|
34013
34025
|
"supports_system_messages": true,
|
|
34014
34026
|
"supports_response_schema": true,
|
|
34015
|
-
"cache_read_input_token_cost":
|
|
34016
|
-
"input_cost_per_token_cache_hit":
|
|
34017
|
-
"supports_reasoning": true
|
|
34027
|
+
"cache_read_input_token_cost": 3e-08,
|
|
34028
|
+
"input_cost_per_token_cache_hit": 3e-08
|
|
34018
34029
|
},
|
|
34019
34030
|
"novita/zai-org/glm-4.7": {
|
|
34020
34031
|
"litellm_provider": "novita",
|
|
34021
34032
|
"mode": "chat",
|
|
34022
|
-
"input_cost_per_token":
|
|
34023
|
-
"output_cost_per_token":
|
|
34033
|
+
"input_cost_per_token": 6e-07,
|
|
34034
|
+
"output_cost_per_token": 2.2e-06,
|
|
34024
34035
|
"max_input_tokens": 204800,
|
|
34025
34036
|
"max_output_tokens": 131072,
|
|
34026
34037
|
"max_tokens": 131072,
|
|
@@ -34029,15 +34040,15 @@
|
|
|
34029
34040
|
"supports_tool_choice": true,
|
|
34030
34041
|
"supports_system_messages": true,
|
|
34031
34042
|
"supports_response_schema": true,
|
|
34032
|
-
"cache_read_input_token_cost":
|
|
34033
|
-
"input_cost_per_token_cache_hit":
|
|
34043
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
34044
|
+
"input_cost_per_token_cache_hit": 1.1e-07,
|
|
34034
34045
|
"supports_reasoning": true
|
|
34035
34046
|
},
|
|
34036
34047
|
"novita/xiaomimimo/mimo-v2-flash": {
|
|
34037
34048
|
"litellm_provider": "novita",
|
|
34038
34049
|
"mode": "chat",
|
|
34039
|
-
"input_cost_per_token":
|
|
34040
|
-
"output_cost_per_token":
|
|
34050
|
+
"input_cost_per_token": 1e-07,
|
|
34051
|
+
"output_cost_per_token": 3e-07,
|
|
34041
34052
|
"max_input_tokens": 262144,
|
|
34042
34053
|
"max_output_tokens": 32000,
|
|
34043
34054
|
"max_tokens": 32000,
|
|
@@ -34046,15 +34057,15 @@
|
|
|
34046
34057
|
"supports_tool_choice": true,
|
|
34047
34058
|
"supports_system_messages": true,
|
|
34048
34059
|
"supports_response_schema": true,
|
|
34049
|
-
"cache_read_input_token_cost":
|
|
34050
|
-
"input_cost_per_token_cache_hit":
|
|
34060
|
+
"cache_read_input_token_cost": 2e-08,
|
|
34061
|
+
"input_cost_per_token_cache_hit": 2e-08,
|
|
34051
34062
|
"supports_reasoning": true
|
|
34052
34063
|
},
|
|
34053
34064
|
"novita/zai-org/autoglm-phone-9b-multilingual": {
|
|
34054
34065
|
"litellm_provider": "novita",
|
|
34055
34066
|
"mode": "chat",
|
|
34056
|
-
"input_cost_per_token":
|
|
34057
|
-
"output_cost_per_token":
|
|
34067
|
+
"input_cost_per_token": 3.5e-08,
|
|
34068
|
+
"output_cost_per_token": 1.38e-07,
|
|
34058
34069
|
"max_input_tokens": 65536,
|
|
34059
34070
|
"max_output_tokens": 65536,
|
|
34060
34071
|
"max_tokens": 65536,
|
|
@@ -34064,8 +34075,8 @@
|
|
|
34064
34075
|
"novita/moonshotai/kimi-k2-thinking": {
|
|
34065
34076
|
"litellm_provider": "novita",
|
|
34066
34077
|
"mode": "chat",
|
|
34067
|
-
"input_cost_per_token":
|
|
34068
|
-
"output_cost_per_token":
|
|
34078
|
+
"input_cost_per_token": 6e-07,
|
|
34079
|
+
"output_cost_per_token": 2.5e-06,
|
|
34069
34080
|
"max_input_tokens": 262144,
|
|
34070
34081
|
"max_output_tokens": 262144,
|
|
34071
34082
|
"max_tokens": 262144,
|
|
@@ -34079,8 +34090,8 @@
|
|
|
34079
34090
|
"novita/minimax/minimax-m2": {
|
|
34080
34091
|
"litellm_provider": "novita",
|
|
34081
34092
|
"mode": "chat",
|
|
34082
|
-
"input_cost_per_token":
|
|
34083
|
-
"output_cost_per_token":
|
|
34093
|
+
"input_cost_per_token": 3e-07,
|
|
34094
|
+
"output_cost_per_token": 1.2e-06,
|
|
34084
34095
|
"max_input_tokens": 204800,
|
|
34085
34096
|
"max_output_tokens": 131072,
|
|
34086
34097
|
"max_tokens": 131072,
|
|
@@ -34088,15 +34099,15 @@
|
|
|
34088
34099
|
"supports_parallel_function_calling": true,
|
|
34089
34100
|
"supports_tool_choice": true,
|
|
34090
34101
|
"supports_system_messages": true,
|
|
34091
|
-
"cache_read_input_token_cost":
|
|
34092
|
-
"input_cost_per_token_cache_hit":
|
|
34102
|
+
"cache_read_input_token_cost": 3e-08,
|
|
34103
|
+
"input_cost_per_token_cache_hit": 3e-08,
|
|
34093
34104
|
"supports_reasoning": true
|
|
34094
34105
|
},
|
|
34095
34106
|
"novita/paddlepaddle/paddleocr-vl": {
|
|
34096
34107
|
"litellm_provider": "novita",
|
|
34097
34108
|
"mode": "chat",
|
|
34098
|
-
"input_cost_per_token":
|
|
34099
|
-
"output_cost_per_token":
|
|
34109
|
+
"input_cost_per_token": 2e-08,
|
|
34110
|
+
"output_cost_per_token": 2e-08,
|
|
34100
34111
|
"max_input_tokens": 16384,
|
|
34101
34112
|
"max_output_tokens": 16384,
|
|
34102
34113
|
"max_tokens": 16384,
|
|
@@ -34106,8 +34117,8 @@
|
|
|
34106
34117
|
"novita/deepseek/deepseek-v3.2-exp": {
|
|
34107
34118
|
"litellm_provider": "novita",
|
|
34108
34119
|
"mode": "chat",
|
|
34109
|
-
"input_cost_per_token":
|
|
34110
|
-
"output_cost_per_token":
|
|
34120
|
+
"input_cost_per_token": 2.7e-07,
|
|
34121
|
+
"output_cost_per_token": 4.1e-07,
|
|
34111
34122
|
"max_input_tokens": 163840,
|
|
34112
34123
|
"max_output_tokens": 65536,
|
|
34113
34124
|
"max_tokens": 65536,
|
|
@@ -34121,8 +34132,8 @@
|
|
|
34121
34132
|
"novita/qwen/qwen3-vl-235b-a22b-thinking": {
|
|
34122
34133
|
"litellm_provider": "novita",
|
|
34123
34134
|
"mode": "chat",
|
|
34124
|
-
"input_cost_per_token":
|
|
34125
|
-
"output_cost_per_token":
|
|
34135
|
+
"input_cost_per_token": 9.8e-07,
|
|
34136
|
+
"output_cost_per_token": 3.95e-06,
|
|
34126
34137
|
"max_input_tokens": 131072,
|
|
34127
34138
|
"max_output_tokens": 32768,
|
|
34128
34139
|
"max_tokens": 32768,
|
|
@@ -34133,8 +34144,8 @@
|
|
|
34133
34144
|
"novita/zai-org/glm-4.6v": {
|
|
34134
34145
|
"litellm_provider": "novita",
|
|
34135
34146
|
"mode": "chat",
|
|
34136
|
-
"input_cost_per_token":
|
|
34137
|
-
"output_cost_per_token":
|
|
34147
|
+
"input_cost_per_token": 3e-07,
|
|
34148
|
+
"output_cost_per_token": 9e-07,
|
|
34138
34149
|
"max_input_tokens": 131072,
|
|
34139
34150
|
"max_output_tokens": 32768,
|
|
34140
34151
|
"max_tokens": 32768,
|
|
@@ -34144,15 +34155,15 @@
|
|
|
34144
34155
|
"supports_vision": true,
|
|
34145
34156
|
"supports_system_messages": true,
|
|
34146
34157
|
"supports_response_schema": true,
|
|
34147
|
-
"cache_read_input_token_cost":
|
|
34148
|
-
"input_cost_per_token_cache_hit":
|
|
34158
|
+
"cache_read_input_token_cost": 5.5e-08,
|
|
34159
|
+
"input_cost_per_token_cache_hit": 5.5e-08,
|
|
34149
34160
|
"supports_reasoning": true
|
|
34150
34161
|
},
|
|
34151
34162
|
"novita/zai-org/glm-4.6": {
|
|
34152
34163
|
"litellm_provider": "novita",
|
|
34153
34164
|
"mode": "chat",
|
|
34154
|
-
"input_cost_per_token":
|
|
34155
|
-
"output_cost_per_token":
|
|
34165
|
+
"input_cost_per_token": 5.5e-07,
|
|
34166
|
+
"output_cost_per_token": 2.2e-06,
|
|
34156
34167
|
"max_input_tokens": 204800,
|
|
34157
34168
|
"max_output_tokens": 131072,
|
|
34158
34169
|
"max_tokens": 131072,
|
|
@@ -34161,15 +34172,15 @@
|
|
|
34161
34172
|
"supports_tool_choice": true,
|
|
34162
34173
|
"supports_system_messages": true,
|
|
34163
34174
|
"supports_response_schema": true,
|
|
34164
|
-
"cache_read_input_token_cost":
|
|
34165
|
-
"input_cost_per_token_cache_hit":
|
|
34175
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
34176
|
+
"input_cost_per_token_cache_hit": 1.1e-07,
|
|
34166
34177
|
"supports_reasoning": true
|
|
34167
34178
|
},
|
|
34168
34179
|
"novita/qwen/qwen3-next-80b-a3b-instruct": {
|
|
34169
34180
|
"litellm_provider": "novita",
|
|
34170
34181
|
"mode": "chat",
|
|
34171
|
-
"input_cost_per_token":
|
|
34172
|
-
"output_cost_per_token":
|
|
34182
|
+
"input_cost_per_token": 1.5e-07,
|
|
34183
|
+
"output_cost_per_token": 1.5e-06,
|
|
34173
34184
|
"max_input_tokens": 131072,
|
|
34174
34185
|
"max_output_tokens": 32768,
|
|
34175
34186
|
"max_tokens": 32768,
|
|
@@ -34182,8 +34193,8 @@
|
|
|
34182
34193
|
"novita/qwen/qwen3-next-80b-a3b-thinking": {
|
|
34183
34194
|
"litellm_provider": "novita",
|
|
34184
34195
|
"mode": "chat",
|
|
34185
|
-
"input_cost_per_token":
|
|
34186
|
-
"output_cost_per_token":
|
|
34196
|
+
"input_cost_per_token": 1.5e-07,
|
|
34197
|
+
"output_cost_per_token": 1.5e-06,
|
|
34187
34198
|
"max_input_tokens": 131072,
|
|
34188
34199
|
"max_output_tokens": 32768,
|
|
34189
34200
|
"max_tokens": 32768,
|
|
@@ -34197,8 +34208,8 @@
|
|
|
34197
34208
|
"novita/deepseek/deepseek-ocr": {
|
|
34198
34209
|
"litellm_provider": "novita",
|
|
34199
34210
|
"mode": "chat",
|
|
34200
|
-
"input_cost_per_token":
|
|
34201
|
-
"output_cost_per_token":
|
|
34211
|
+
"input_cost_per_token": 3e-08,
|
|
34212
|
+
"output_cost_per_token": 3e-08,
|
|
34202
34213
|
"max_input_tokens": 8192,
|
|
34203
34214
|
"max_output_tokens": 8192,
|
|
34204
34215
|
"max_tokens": 8192,
|
|
@@ -34210,8 +34221,8 @@
|
|
|
34210
34221
|
"novita/deepseek/deepseek-v3.1-terminus": {
|
|
34211
34222
|
"litellm_provider": "novita",
|
|
34212
34223
|
"mode": "chat",
|
|
34213
|
-
"input_cost_per_token":
|
|
34214
|
-
"output_cost_per_token":
|
|
34224
|
+
"input_cost_per_token": 2.7e-07,
|
|
34225
|
+
"output_cost_per_token": 1e-06,
|
|
34215
34226
|
"max_input_tokens": 131072,
|
|
34216
34227
|
"max_output_tokens": 32768,
|
|
34217
34228
|
"max_tokens": 32768,
|
|
@@ -34220,15 +34231,15 @@
|
|
|
34220
34231
|
"supports_tool_choice": true,
|
|
34221
34232
|
"supports_system_messages": true,
|
|
34222
34233
|
"supports_response_schema": true,
|
|
34223
|
-
"cache_read_input_token_cost":
|
|
34224
|
-
"input_cost_per_token_cache_hit":
|
|
34234
|
+
"cache_read_input_token_cost": 1.35e-07,
|
|
34235
|
+
"input_cost_per_token_cache_hit": 1.35e-07,
|
|
34225
34236
|
"supports_reasoning": true
|
|
34226
34237
|
},
|
|
34227
34238
|
"novita/qwen/qwen3-vl-235b-a22b-instruct": {
|
|
34228
34239
|
"litellm_provider": "novita",
|
|
34229
34240
|
"mode": "chat",
|
|
34230
|
-
"input_cost_per_token":
|
|
34231
|
-
"output_cost_per_token":
|
|
34241
|
+
"input_cost_per_token": 3e-07,
|
|
34242
|
+
"output_cost_per_token": 1.5e-06,
|
|
34232
34243
|
"max_input_tokens": 131072,
|
|
34233
34244
|
"max_output_tokens": 32768,
|
|
34234
34245
|
"max_tokens": 32768,
|
|
@@ -34242,8 +34253,8 @@
|
|
|
34242
34253
|
"novita/qwen/qwen3-max": {
|
|
34243
34254
|
"litellm_provider": "novita",
|
|
34244
34255
|
"mode": "chat",
|
|
34245
|
-
"input_cost_per_token":
|
|
34246
|
-
"output_cost_per_token":
|
|
34256
|
+
"input_cost_per_token": 2.11e-06,
|
|
34257
|
+
"output_cost_per_token": 8.45e-06,
|
|
34247
34258
|
"max_input_tokens": 262144,
|
|
34248
34259
|
"max_output_tokens": 65536,
|
|
34249
34260
|
"max_tokens": 65536,
|
|
@@ -34256,8 +34267,8 @@
|
|
|
34256
34267
|
"novita/skywork/r1v4-lite": {
|
|
34257
34268
|
"litellm_provider": "novita",
|
|
34258
34269
|
"mode": "chat",
|
|
34259
|
-
"input_cost_per_token":
|
|
34260
|
-
"output_cost_per_token":
|
|
34270
|
+
"input_cost_per_token": 2e-07,
|
|
34271
|
+
"output_cost_per_token": 6e-07,
|
|
34261
34272
|
"max_input_tokens": 262144,
|
|
34262
34273
|
"max_output_tokens": 65536,
|
|
34263
34274
|
"max_tokens": 65536,
|
|
@@ -34269,8 +34280,8 @@
|
|
|
34269
34280
|
"novita/deepseek/deepseek-v3.1": {
|
|
34270
34281
|
"litellm_provider": "novita",
|
|
34271
34282
|
"mode": "chat",
|
|
34272
|
-
"input_cost_per_token":
|
|
34273
|
-
"output_cost_per_token":
|
|
34283
|
+
"input_cost_per_token": 2.7e-07,
|
|
34284
|
+
"output_cost_per_token": 1e-06,
|
|
34274
34285
|
"max_input_tokens": 131072,
|
|
34275
34286
|
"max_output_tokens": 32768,
|
|
34276
34287
|
"max_tokens": 32768,
|
|
@@ -34279,15 +34290,15 @@
|
|
|
34279
34290
|
"supports_tool_choice": true,
|
|
34280
34291
|
"supports_system_messages": true,
|
|
34281
34292
|
"supports_response_schema": true,
|
|
34282
|
-
"cache_read_input_token_cost":
|
|
34283
|
-
"input_cost_per_token_cache_hit":
|
|
34293
|
+
"cache_read_input_token_cost": 1.35e-07,
|
|
34294
|
+
"input_cost_per_token_cache_hit": 1.35e-07,
|
|
34284
34295
|
"supports_reasoning": true
|
|
34285
34296
|
},
|
|
34286
34297
|
"novita/moonshotai/kimi-k2-0905": {
|
|
34287
34298
|
"litellm_provider": "novita",
|
|
34288
34299
|
"mode": "chat",
|
|
34289
|
-
"input_cost_per_token":
|
|
34290
|
-
"output_cost_per_token":
|
|
34300
|
+
"input_cost_per_token": 6e-07,
|
|
34301
|
+
"output_cost_per_token": 2.5e-06,
|
|
34291
34302
|
"max_input_tokens": 262144,
|
|
34292
34303
|
"max_output_tokens": 262144,
|
|
34293
34304
|
"max_tokens": 262144,
|
|
@@ -34300,8 +34311,8 @@
|
|
|
34300
34311
|
"novita/qwen/qwen3-coder-480b-a35b-instruct": {
|
|
34301
34312
|
"litellm_provider": "novita",
|
|
34302
34313
|
"mode": "chat",
|
|
34303
|
-
"input_cost_per_token":
|
|
34304
|
-
"output_cost_per_token":
|
|
34314
|
+
"input_cost_per_token": 3e-07,
|
|
34315
|
+
"output_cost_per_token": 1.3e-06,
|
|
34305
34316
|
"max_input_tokens": 262144,
|
|
34306
34317
|
"max_output_tokens": 65536,
|
|
34307
34318
|
"max_tokens": 65536,
|
|
@@ -34314,8 +34325,8 @@
|
|
|
34314
34325
|
"novita/qwen/qwen3-coder-30b-a3b-instruct": {
|
|
34315
34326
|
"litellm_provider": "novita",
|
|
34316
34327
|
"mode": "chat",
|
|
34317
|
-
"input_cost_per_token":
|
|
34318
|
-
"output_cost_per_token":
|
|
34328
|
+
"input_cost_per_token": 7e-08,
|
|
34329
|
+
"output_cost_per_token": 2.7e-07,
|
|
34319
34330
|
"max_input_tokens": 160000,
|
|
34320
34331
|
"max_output_tokens": 32768,
|
|
34321
34332
|
"max_tokens": 32768,
|
|
@@ -34328,8 +34339,8 @@
|
|
|
34328
34339
|
"novita/openai/gpt-oss-120b": {
|
|
34329
34340
|
"litellm_provider": "novita",
|
|
34330
34341
|
"mode": "chat",
|
|
34331
|
-
"input_cost_per_token":
|
|
34332
|
-
"output_cost_per_token":
|
|
34342
|
+
"input_cost_per_token": 5e-08,
|
|
34343
|
+
"output_cost_per_token": 2.5e-07,
|
|
34333
34344
|
"max_input_tokens": 131072,
|
|
34334
34345
|
"max_output_tokens": 32768,
|
|
34335
34346
|
"max_tokens": 32768,
|
|
@@ -34344,8 +34355,8 @@
|
|
|
34344
34355
|
"novita/moonshotai/kimi-k2-instruct": {
|
|
34345
34356
|
"litellm_provider": "novita",
|
|
34346
34357
|
"mode": "chat",
|
|
34347
|
-
"input_cost_per_token":
|
|
34348
|
-
"output_cost_per_token":
|
|
34358
|
+
"input_cost_per_token": 5.7e-07,
|
|
34359
|
+
"output_cost_per_token": 2.3e-06,
|
|
34349
34360
|
"max_input_tokens": 131072,
|
|
34350
34361
|
"max_output_tokens": 131072,
|
|
34351
34362
|
"max_tokens": 131072,
|
|
@@ -34358,8 +34369,8 @@
|
|
|
34358
34369
|
"novita/deepseek/deepseek-v3-0324": {
|
|
34359
34370
|
"litellm_provider": "novita",
|
|
34360
34371
|
"mode": "chat",
|
|
34361
|
-
"input_cost_per_token":
|
|
34362
|
-
"output_cost_per_token":
|
|
34372
|
+
"input_cost_per_token": 2.7e-07,
|
|
34373
|
+
"output_cost_per_token": 1.12e-06,
|
|
34363
34374
|
"max_input_tokens": 163840,
|
|
34364
34375
|
"max_output_tokens": 163840,
|
|
34365
34376
|
"max_tokens": 163840,
|
|
@@ -34368,14 +34379,14 @@
|
|
|
34368
34379
|
"supports_tool_choice": true,
|
|
34369
34380
|
"supports_system_messages": true,
|
|
34370
34381
|
"supports_response_schema": true,
|
|
34371
|
-
"cache_read_input_token_cost":
|
|
34372
|
-
"input_cost_per_token_cache_hit":
|
|
34382
|
+
"cache_read_input_token_cost": 1.35e-07,
|
|
34383
|
+
"input_cost_per_token_cache_hit": 1.35e-07
|
|
34373
34384
|
},
|
|
34374
34385
|
"novita/zai-org/glm-4.5": {
|
|
34375
34386
|
"litellm_provider": "novita",
|
|
34376
34387
|
"mode": "chat",
|
|
34377
|
-
"input_cost_per_token":
|
|
34378
|
-
"output_cost_per_token":
|
|
34388
|
+
"input_cost_per_token": 6e-07,
|
|
34389
|
+
"output_cost_per_token": 2.2e-06,
|
|
34379
34390
|
"max_input_tokens": 131072,
|
|
34380
34391
|
"max_output_tokens": 98304,
|
|
34381
34392
|
"max_tokens": 98304,
|
|
@@ -34383,15 +34394,15 @@
|
|
|
34383
34394
|
"supports_parallel_function_calling": true,
|
|
34384
34395
|
"supports_tool_choice": true,
|
|
34385
34396
|
"supports_system_messages": true,
|
|
34386
|
-
"cache_read_input_token_cost":
|
|
34387
|
-
"input_cost_per_token_cache_hit":
|
|
34397
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
34398
|
+
"input_cost_per_token_cache_hit": 1.1e-07,
|
|
34388
34399
|
"supports_reasoning": true
|
|
34389
34400
|
},
|
|
34390
34401
|
"novita/qwen/qwen3-235b-a22b-thinking-2507": {
|
|
34391
34402
|
"litellm_provider": "novita",
|
|
34392
34403
|
"mode": "chat",
|
|
34393
|
-
"input_cost_per_token":
|
|
34394
|
-
"output_cost_per_token":
|
|
34404
|
+
"input_cost_per_token": 3e-07,
|
|
34405
|
+
"output_cost_per_token": 3e-06,
|
|
34395
34406
|
"max_input_tokens": 131072,
|
|
34396
34407
|
"max_output_tokens": 32768,
|
|
34397
34408
|
"max_tokens": 32768,
|
|
@@ -34404,8 +34415,8 @@
|
|
|
34404
34415
|
"novita/meta-llama/llama-3.1-8b-instruct": {
|
|
34405
34416
|
"litellm_provider": "novita",
|
|
34406
34417
|
"mode": "chat",
|
|
34407
|
-
"input_cost_per_token":
|
|
34408
|
-
"output_cost_per_token":
|
|
34418
|
+
"input_cost_per_token": 2e-08,
|
|
34419
|
+
"output_cost_per_token": 5e-08,
|
|
34409
34420
|
"max_input_tokens": 16384,
|
|
34410
34421
|
"max_output_tokens": 16384,
|
|
34411
34422
|
"max_tokens": 16384,
|
|
@@ -34414,8 +34425,8 @@
|
|
|
34414
34425
|
"novita/google/gemma-3-12b-it": {
|
|
34415
34426
|
"litellm_provider": "novita",
|
|
34416
34427
|
"mode": "chat",
|
|
34417
|
-
"input_cost_per_token":
|
|
34418
|
-
"output_cost_per_token":
|
|
34428
|
+
"input_cost_per_token": 5e-08,
|
|
34429
|
+
"output_cost_per_token": 1e-07,
|
|
34419
34430
|
"max_input_tokens": 131072,
|
|
34420
34431
|
"max_output_tokens": 8192,
|
|
34421
34432
|
"max_tokens": 8192,
|
|
@@ -34427,8 +34438,8 @@
|
|
|
34427
34438
|
"novita/zai-org/glm-4.5v": {
|
|
34428
34439
|
"litellm_provider": "novita",
|
|
34429
34440
|
"mode": "chat",
|
|
34430
|
-
"input_cost_per_token":
|
|
34431
|
-
"output_cost_per_token":
|
|
34441
|
+
"input_cost_per_token": 6e-07,
|
|
34442
|
+
"output_cost_per_token": 1.8e-06,
|
|
34432
34443
|
"max_input_tokens": 65536,
|
|
34433
34444
|
"max_output_tokens": 16384,
|
|
34434
34445
|
"max_tokens": 16384,
|
|
@@ -34438,15 +34449,15 @@
|
|
|
34438
34449
|
"supports_vision": true,
|
|
34439
34450
|
"supports_system_messages": true,
|
|
34440
34451
|
"supports_response_schema": true,
|
|
34441
|
-
"cache_read_input_token_cost":
|
|
34442
|
-
"input_cost_per_token_cache_hit":
|
|
34452
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
34453
|
+
"input_cost_per_token_cache_hit": 1.1e-07,
|
|
34443
34454
|
"supports_reasoning": true
|
|
34444
34455
|
},
|
|
34445
34456
|
"novita/openai/gpt-oss-20b": {
|
|
34446
34457
|
"litellm_provider": "novita",
|
|
34447
34458
|
"mode": "chat",
|
|
34448
|
-
"input_cost_per_token":
|
|
34449
|
-
"output_cost_per_token":
|
|
34459
|
+
"input_cost_per_token": 4e-08,
|
|
34460
|
+
"output_cost_per_token": 1.5e-07,
|
|
34450
34461
|
"max_input_tokens": 131072,
|
|
34451
34462
|
"max_output_tokens": 32768,
|
|
34452
34463
|
"max_tokens": 32768,
|
|
@@ -34459,8 +34470,8 @@
|
|
|
34459
34470
|
"novita/qwen/qwen3-235b-a22b-instruct-2507": {
|
|
34460
34471
|
"litellm_provider": "novita",
|
|
34461
34472
|
"mode": "chat",
|
|
34462
|
-
"input_cost_per_token":
|
|
34463
|
-
"output_cost_per_token":
|
|
34473
|
+
"input_cost_per_token": 9e-08,
|
|
34474
|
+
"output_cost_per_token": 5.8e-07,
|
|
34464
34475
|
"max_input_tokens": 131072,
|
|
34465
34476
|
"max_output_tokens": 16384,
|
|
34466
34477
|
"max_tokens": 16384,
|
|
@@ -34473,8 +34484,8 @@
|
|
|
34473
34484
|
"novita/deepseek/deepseek-r1-distill-qwen-14b": {
|
|
34474
34485
|
"litellm_provider": "novita",
|
|
34475
34486
|
"mode": "chat",
|
|
34476
|
-
"input_cost_per_token":
|
|
34477
|
-
"output_cost_per_token":
|
|
34487
|
+
"input_cost_per_token": 1.5e-07,
|
|
34488
|
+
"output_cost_per_token": 1.5e-07,
|
|
34478
34489
|
"max_input_tokens": 32768,
|
|
34479
34490
|
"max_output_tokens": 16384,
|
|
34480
34491
|
"max_tokens": 16384,
|
|
@@ -34486,8 +34497,8 @@
|
|
|
34486
34497
|
"novita/meta-llama/llama-3.3-70b-instruct": {
|
|
34487
34498
|
"litellm_provider": "novita",
|
|
34488
34499
|
"mode": "chat",
|
|
34489
|
-
"input_cost_per_token":
|
|
34490
|
-
"output_cost_per_token":
|
|
34500
|
+
"input_cost_per_token": 1.35e-07,
|
|
34501
|
+
"output_cost_per_token": 4e-07,
|
|
34491
34502
|
"max_input_tokens": 131072,
|
|
34492
34503
|
"max_output_tokens": 120000,
|
|
34493
34504
|
"max_tokens": 120000,
|
|
@@ -34499,8 +34510,8 @@
|
|
|
34499
34510
|
"novita/qwen/qwen-2.5-72b-instruct": {
|
|
34500
34511
|
"litellm_provider": "novita",
|
|
34501
34512
|
"mode": "chat",
|
|
34502
|
-
"input_cost_per_token":
|
|
34503
|
-
"output_cost_per_token":
|
|
34513
|
+
"input_cost_per_token": 3.8e-07,
|
|
34514
|
+
"output_cost_per_token": 4e-07,
|
|
34504
34515
|
"max_input_tokens": 32000,
|
|
34505
34516
|
"max_output_tokens": 8192,
|
|
34506
34517
|
"max_tokens": 8192,
|
|
@@ -34513,8 +34524,8 @@
|
|
|
34513
34524
|
"novita/mistralai/mistral-nemo": {
|
|
34514
34525
|
"litellm_provider": "novita",
|
|
34515
34526
|
"mode": "chat",
|
|
34516
|
-
"input_cost_per_token":
|
|
34517
|
-
"output_cost_per_token":
|
|
34527
|
+
"input_cost_per_token": 4e-08,
|
|
34528
|
+
"output_cost_per_token": 1.7e-07,
|
|
34518
34529
|
"max_input_tokens": 60288,
|
|
34519
34530
|
"max_output_tokens": 16000,
|
|
34520
34531
|
"max_tokens": 16000,
|
|
@@ -34525,8 +34536,8 @@
|
|
|
34525
34536
|
"novita/minimaxai/minimax-m1-80k": {
|
|
34526
34537
|
"litellm_provider": "novita",
|
|
34527
34538
|
"mode": "chat",
|
|
34528
|
-
"input_cost_per_token":
|
|
34529
|
-
"output_cost_per_token":
|
|
34539
|
+
"input_cost_per_token": 5.5e-07,
|
|
34540
|
+
"output_cost_per_token": 2.2e-06,
|
|
34530
34541
|
"max_input_tokens": 1000000,
|
|
34531
34542
|
"max_output_tokens": 40000,
|
|
34532
34543
|
"max_tokens": 40000,
|
|
@@ -34539,8 +34550,8 @@
|
|
|
34539
34550
|
"novita/deepseek/deepseek-r1-0528": {
|
|
34540
34551
|
"litellm_provider": "novita",
|
|
34541
34552
|
"mode": "chat",
|
|
34542
|
-
"input_cost_per_token":
|
|
34543
|
-
"output_cost_per_token":
|
|
34553
|
+
"input_cost_per_token": 7e-07,
|
|
34554
|
+
"output_cost_per_token": 2.5e-06,
|
|
34544
34555
|
"max_input_tokens": 163840,
|
|
34545
34556
|
"max_output_tokens": 32768,
|
|
34546
34557
|
"max_tokens": 32768,
|
|
@@ -34549,15 +34560,15 @@
|
|
|
34549
34560
|
"supports_tool_choice": true,
|
|
34550
34561
|
"supports_system_messages": true,
|
|
34551
34562
|
"supports_response_schema": true,
|
|
34552
|
-
"cache_read_input_token_cost":
|
|
34553
|
-
"input_cost_per_token_cache_hit":
|
|
34563
|
+
"cache_read_input_token_cost": 3.5e-07,
|
|
34564
|
+
"input_cost_per_token_cache_hit": 3.5e-07,
|
|
34554
34565
|
"supports_reasoning": true
|
|
34555
34566
|
},
|
|
34556
34567
|
"novita/deepseek/deepseek-r1-distill-qwen-32b": {
|
|
34557
34568
|
"litellm_provider": "novita",
|
|
34558
34569
|
"mode": "chat",
|
|
34559
|
-
"input_cost_per_token":
|
|
34560
|
-
"output_cost_per_token":
|
|
34570
|
+
"input_cost_per_token": 3e-07,
|
|
34571
|
+
"output_cost_per_token": 3e-07,
|
|
34561
34572
|
"max_input_tokens": 64000,
|
|
34562
34573
|
"max_output_tokens": 32000,
|
|
34563
34574
|
"max_tokens": 32000,
|
|
@@ -34569,8 +34580,8 @@
|
|
|
34569
34580
|
"novita/meta-llama/llama-3-8b-instruct": {
|
|
34570
34581
|
"litellm_provider": "novita",
|
|
34571
34582
|
"mode": "chat",
|
|
34572
|
-
"input_cost_per_token":
|
|
34573
|
-
"output_cost_per_token":
|
|
34583
|
+
"input_cost_per_token": 4e-08,
|
|
34584
|
+
"output_cost_per_token": 4e-08,
|
|
34574
34585
|
"max_input_tokens": 8192,
|
|
34575
34586
|
"max_output_tokens": 8192,
|
|
34576
34587
|
"max_tokens": 8192,
|
|
@@ -34579,8 +34590,8 @@
|
|
|
34579
34590
|
"novita/microsoft/wizardlm-2-8x22b": {
|
|
34580
34591
|
"litellm_provider": "novita",
|
|
34581
34592
|
"mode": "chat",
|
|
34582
|
-
"input_cost_per_token":
|
|
34583
|
-
"output_cost_per_token":
|
|
34593
|
+
"input_cost_per_token": 6.2e-07,
|
|
34594
|
+
"output_cost_per_token": 6.2e-07,
|
|
34584
34595
|
"max_input_tokens": 65535,
|
|
34585
34596
|
"max_output_tokens": 8000,
|
|
34586
34597
|
"max_tokens": 8000,
|
|
@@ -34589,8 +34600,8 @@
|
|
|
34589
34600
|
"novita/deepseek/deepseek-r1-0528-qwen3-8b": {
|
|
34590
34601
|
"litellm_provider": "novita",
|
|
34591
34602
|
"mode": "chat",
|
|
34592
|
-
"input_cost_per_token":
|
|
34593
|
-
"output_cost_per_token":
|
|
34603
|
+
"input_cost_per_token": 6e-08,
|
|
34604
|
+
"output_cost_per_token": 9e-08,
|
|
34594
34605
|
"max_input_tokens": 128000,
|
|
34595
34606
|
"max_output_tokens": 32000,
|
|
34596
34607
|
"max_tokens": 32000,
|
|
@@ -34600,8 +34611,8 @@
|
|
|
34600
34611
|
"novita/deepseek/deepseek-r1-distill-llama-70b": {
|
|
34601
34612
|
"litellm_provider": "novita",
|
|
34602
34613
|
"mode": "chat",
|
|
34603
|
-
"input_cost_per_token":
|
|
34604
|
-
"output_cost_per_token":
|
|
34614
|
+
"input_cost_per_token": 8e-07,
|
|
34615
|
+
"output_cost_per_token": 8e-07,
|
|
34605
34616
|
"max_input_tokens": 8192,
|
|
34606
34617
|
"max_output_tokens": 8192,
|
|
34607
34618
|
"max_tokens": 8192,
|
|
@@ -34613,8 +34624,8 @@
|
|
|
34613
34624
|
"novita/meta-llama/llama-3-70b-instruct": {
|
|
34614
34625
|
"litellm_provider": "novita",
|
|
34615
34626
|
"mode": "chat",
|
|
34616
|
-
"input_cost_per_token":
|
|
34617
|
-
"output_cost_per_token":
|
|
34627
|
+
"input_cost_per_token": 5.1e-07,
|
|
34628
|
+
"output_cost_per_token": 7.4e-07,
|
|
34618
34629
|
"max_input_tokens": 8192,
|
|
34619
34630
|
"max_output_tokens": 8000,
|
|
34620
34631
|
"max_tokens": 8000,
|
|
@@ -34625,8 +34636,8 @@
|
|
|
34625
34636
|
"novita/qwen/qwen3-235b-a22b-fp8": {
|
|
34626
34637
|
"litellm_provider": "novita",
|
|
34627
34638
|
"mode": "chat",
|
|
34628
|
-
"input_cost_per_token":
|
|
34629
|
-
"output_cost_per_token":
|
|
34639
|
+
"input_cost_per_token": 2e-07,
|
|
34640
|
+
"output_cost_per_token": 8e-07,
|
|
34630
34641
|
"max_input_tokens": 40960,
|
|
34631
34642
|
"max_output_tokens": 20000,
|
|
34632
34643
|
"max_tokens": 20000,
|
|
@@ -34636,8 +34647,8 @@
|
|
|
34636
34647
|
"novita/meta-llama/llama-4-maverick-17b-128e-instruct-fp8": {
|
|
34637
34648
|
"litellm_provider": "novita",
|
|
34638
34649
|
"mode": "chat",
|
|
34639
|
-
"input_cost_per_token":
|
|
34640
|
-
"output_cost_per_token":
|
|
34650
|
+
"input_cost_per_token": 2.7e-07,
|
|
34651
|
+
"output_cost_per_token": 8.5e-07,
|
|
34641
34652
|
"max_input_tokens": 1048576,
|
|
34642
34653
|
"max_output_tokens": 8192,
|
|
34643
34654
|
"max_tokens": 8192,
|
|
@@ -34647,8 +34658,8 @@
|
|
|
34647
34658
|
"novita/meta-llama/llama-4-scout-17b-16e-instruct": {
|
|
34648
34659
|
"litellm_provider": "novita",
|
|
34649
34660
|
"mode": "chat",
|
|
34650
|
-
"input_cost_per_token":
|
|
34651
|
-
"output_cost_per_token":
|
|
34661
|
+
"input_cost_per_token": 1.8e-07,
|
|
34662
|
+
"output_cost_per_token": 5.9e-07,
|
|
34652
34663
|
"max_input_tokens": 131072,
|
|
34653
34664
|
"max_output_tokens": 131072,
|
|
34654
34665
|
"max_tokens": 131072,
|
|
@@ -34658,8 +34669,8 @@
|
|
|
34658
34669
|
"novita/nousresearch/hermes-2-pro-llama-3-8b": {
|
|
34659
34670
|
"litellm_provider": "novita",
|
|
34660
34671
|
"mode": "chat",
|
|
34661
|
-
"input_cost_per_token":
|
|
34662
|
-
"output_cost_per_token":
|
|
34672
|
+
"input_cost_per_token": 1.4e-07,
|
|
34673
|
+
"output_cost_per_token": 1.4e-07,
|
|
34663
34674
|
"max_input_tokens": 8192,
|
|
34664
34675
|
"max_output_tokens": 8192,
|
|
34665
34676
|
"max_tokens": 8192,
|
|
@@ -34670,8 +34681,8 @@
|
|
|
34670
34681
|
"novita/qwen/qwen2.5-vl-72b-instruct": {
|
|
34671
34682
|
"litellm_provider": "novita",
|
|
34672
34683
|
"mode": "chat",
|
|
34673
|
-
"input_cost_per_token":
|
|
34674
|
-
"output_cost_per_token":
|
|
34684
|
+
"input_cost_per_token": 8e-07,
|
|
34685
|
+
"output_cost_per_token": 8e-07,
|
|
34675
34686
|
"max_input_tokens": 32768,
|
|
34676
34687
|
"max_output_tokens": 32768,
|
|
34677
34688
|
"max_tokens": 32768,
|
|
@@ -34681,8 +34692,8 @@
|
|
|
34681
34692
|
"novita/sao10k/l3-70b-euryale-v2.1": {
|
|
34682
34693
|
"litellm_provider": "novita",
|
|
34683
34694
|
"mode": "chat",
|
|
34684
|
-
"input_cost_per_token":
|
|
34685
|
-
"output_cost_per_token":
|
|
34695
|
+
"input_cost_per_token": 1.48e-06,
|
|
34696
|
+
"output_cost_per_token": 1.48e-06,
|
|
34686
34697
|
"max_input_tokens": 8192,
|
|
34687
34698
|
"max_output_tokens": 8192,
|
|
34688
34699
|
"max_tokens": 8192,
|
|
@@ -34694,8 +34705,8 @@
|
|
|
34694
34705
|
"novita/baidu/ernie-4.5-21B-a3b-thinking": {
|
|
34695
34706
|
"litellm_provider": "novita",
|
|
34696
34707
|
"mode": "chat",
|
|
34697
|
-
"input_cost_per_token":
|
|
34698
|
-
"output_cost_per_token":
|
|
34708
|
+
"input_cost_per_token": 7e-08,
|
|
34709
|
+
"output_cost_per_token": 2.8e-07,
|
|
34699
34710
|
"max_input_tokens": 131072,
|
|
34700
34711
|
"max_output_tokens": 65536,
|
|
34701
34712
|
"max_tokens": 65536,
|
|
@@ -34705,8 +34716,8 @@
|
|
|
34705
34716
|
"novita/sao10k/l3-8b-lunaris": {
|
|
34706
34717
|
"litellm_provider": "novita",
|
|
34707
34718
|
"mode": "chat",
|
|
34708
|
-
"input_cost_per_token":
|
|
34709
|
-
"output_cost_per_token":
|
|
34719
|
+
"input_cost_per_token": 5e-08,
|
|
34720
|
+
"output_cost_per_token": 5e-08,
|
|
34710
34721
|
"max_input_tokens": 8192,
|
|
34711
34722
|
"max_output_tokens": 8192,
|
|
34712
34723
|
"max_tokens": 8192,
|
|
@@ -34717,15 +34728,12 @@
|
|
|
34717
34728
|
"novita/baichuan/baichuan-m2-32b": {
|
|
34718
34729
|
"litellm_provider": "novita",
|
|
34719
34730
|
"mode": "chat",
|
|
34720
|
-
"input_cost_per_token":
|
|
34721
|
-
"output_cost_per_token":
|
|
34731
|
+
"input_cost_per_token": 7e-08,
|
|
34732
|
+
"output_cost_per_token": 7e-08,
|
|
34722
34733
|
"max_input_tokens": 131072,
|
|
34723
34734
|
"max_output_tokens": 131072,
|
|
34724
34735
|
"max_tokens": 131072,
|
|
34725
|
-
"
|
|
34726
|
-
"supports_system_messages": true,
|
|
34727
|
-
"supports_response_schema": true,
|
|
34728
|
-
"supports_reasoning": true
|
|
34736
|
+
"supports_system_messages": true
|
|
34729
34737
|
},
|
|
34730
34738
|
"novita/thudm/glm-4.1v-9b-thinking": {
|
|
34731
34739
|
"litellm_provider": "novita",
|
|
@@ -34742,8 +34750,8 @@
|
|
|
34742
34750
|
"novita/baidu/ernie-4.5-vl-424b-a47b": {
|
|
34743
34751
|
"litellm_provider": "novita",
|
|
34744
34752
|
"mode": "chat",
|
|
34745
|
-
"input_cost_per_token":
|
|
34746
|
-
"output_cost_per_token":
|
|
34753
|
+
"input_cost_per_token": 4.2e-07,
|
|
34754
|
+
"output_cost_per_token": 1.25e-06,
|
|
34747
34755
|
"max_input_tokens": 123000,
|
|
34748
34756
|
"max_output_tokens": 16000,
|
|
34749
34757
|
"max_tokens": 16000,
|
|
@@ -34754,8 +34762,8 @@
|
|
|
34754
34762
|
"novita/baidu/ernie-4.5-300b-a47b-paddle": {
|
|
34755
34763
|
"litellm_provider": "novita",
|
|
34756
34764
|
"mode": "chat",
|
|
34757
|
-
"input_cost_per_token":
|
|
34758
|
-
"output_cost_per_token":
|
|
34765
|
+
"input_cost_per_token": 2.8e-07,
|
|
34766
|
+
"output_cost_per_token": 1.1e-06,
|
|
34759
34767
|
"max_input_tokens": 123000,
|
|
34760
34768
|
"max_output_tokens": 12000,
|
|
34761
34769
|
"max_tokens": 12000,
|
|
@@ -34766,8 +34774,8 @@
|
|
|
34766
34774
|
"novita/deepseek/deepseek-prover-v2-671b": {
|
|
34767
34775
|
"litellm_provider": "novita",
|
|
34768
34776
|
"mode": "chat",
|
|
34769
|
-
"input_cost_per_token":
|
|
34770
|
-
"output_cost_per_token":
|
|
34777
|
+
"input_cost_per_token": 7e-07,
|
|
34778
|
+
"output_cost_per_token": 2.5e-06,
|
|
34771
34779
|
"max_input_tokens": 160000,
|
|
34772
34780
|
"max_output_tokens": 160000,
|
|
34773
34781
|
"max_tokens": 160000,
|
|
@@ -34776,8 +34784,8 @@
|
|
|
34776
34784
|
"novita/qwen/qwen3-32b-fp8": {
|
|
34777
34785
|
"litellm_provider": "novita",
|
|
34778
34786
|
"mode": "chat",
|
|
34779
|
-
"input_cost_per_token":
|
|
34780
|
-
"output_cost_per_token":
|
|
34787
|
+
"input_cost_per_token": 1e-07,
|
|
34788
|
+
"output_cost_per_token": 4.5e-07,
|
|
34781
34789
|
"max_input_tokens": 40960,
|
|
34782
34790
|
"max_output_tokens": 20000,
|
|
34783
34791
|
"max_tokens": 20000,
|
|
@@ -34787,8 +34795,8 @@
|
|
|
34787
34795
|
"novita/qwen/qwen3-30b-a3b-fp8": {
|
|
34788
34796
|
"litellm_provider": "novita",
|
|
34789
34797
|
"mode": "chat",
|
|
34790
|
-
"input_cost_per_token":
|
|
34791
|
-
"output_cost_per_token":
|
|
34798
|
+
"input_cost_per_token": 9e-08,
|
|
34799
|
+
"output_cost_per_token": 4.5e-07,
|
|
34792
34800
|
"max_input_tokens": 40960,
|
|
34793
34801
|
"max_output_tokens": 20000,
|
|
34794
34802
|
"max_tokens": 20000,
|
|
@@ -34798,8 +34806,8 @@
|
|
|
34798
34806
|
"novita/google/gemma-3-27b-it": {
|
|
34799
34807
|
"litellm_provider": "novita",
|
|
34800
34808
|
"mode": "chat",
|
|
34801
|
-
"input_cost_per_token":
|
|
34802
|
-
"output_cost_per_token":
|
|
34809
|
+
"input_cost_per_token": 1.19e-07,
|
|
34810
|
+
"output_cost_per_token": 2e-07,
|
|
34803
34811
|
"max_input_tokens": 98304,
|
|
34804
34812
|
"max_output_tokens": 16384,
|
|
34805
34813
|
"max_tokens": 16384,
|
|
@@ -34809,8 +34817,8 @@
|
|
|
34809
34817
|
"novita/deepseek/deepseek-v3-turbo": {
|
|
34810
34818
|
"litellm_provider": "novita",
|
|
34811
34819
|
"mode": "chat",
|
|
34812
|
-
"input_cost_per_token":
|
|
34813
|
-
"output_cost_per_token":
|
|
34820
|
+
"input_cost_per_token": 4e-07,
|
|
34821
|
+
"output_cost_per_token": 1.3e-06,
|
|
34814
34822
|
"max_input_tokens": 64000,
|
|
34815
34823
|
"max_output_tokens": 16000,
|
|
34816
34824
|
"max_tokens": 16000,
|
|
@@ -34822,8 +34830,8 @@
|
|
|
34822
34830
|
"novita/deepseek/deepseek-r1-turbo": {
|
|
34823
34831
|
"litellm_provider": "novita",
|
|
34824
34832
|
"mode": "chat",
|
|
34825
|
-
"input_cost_per_token":
|
|
34826
|
-
"output_cost_per_token":
|
|
34833
|
+
"input_cost_per_token": 7e-07,
|
|
34834
|
+
"output_cost_per_token": 2.5e-06,
|
|
34827
34835
|
"max_input_tokens": 64000,
|
|
34828
34836
|
"max_output_tokens": 16000,
|
|
34829
34837
|
"max_tokens": 16000,
|
|
@@ -34836,8 +34844,8 @@
|
|
|
34836
34844
|
"novita/Sao10K/L3-8B-Stheno-v3.2": {
|
|
34837
34845
|
"litellm_provider": "novita",
|
|
34838
34846
|
"mode": "chat",
|
|
34839
|
-
"input_cost_per_token":
|
|
34840
|
-
"output_cost_per_token":
|
|
34847
|
+
"input_cost_per_token": 5e-08,
|
|
34848
|
+
"output_cost_per_token": 5e-08,
|
|
34841
34849
|
"max_input_tokens": 8192,
|
|
34842
34850
|
"max_output_tokens": 32000,
|
|
34843
34851
|
"max_tokens": 32000,
|
|
@@ -34849,8 +34857,8 @@
|
|
|
34849
34857
|
"novita/gryphe/mythomax-l2-13b": {
|
|
34850
34858
|
"litellm_provider": "novita",
|
|
34851
34859
|
"mode": "chat",
|
|
34852
|
-
"input_cost_per_token":
|
|
34853
|
-
"output_cost_per_token":
|
|
34860
|
+
"input_cost_per_token": 9e-08,
|
|
34861
|
+
"output_cost_per_token": 9e-08,
|
|
34854
34862
|
"max_input_tokens": 4096,
|
|
34855
34863
|
"max_output_tokens": 3200,
|
|
34856
34864
|
"max_tokens": 3200,
|
|
@@ -34859,8 +34867,8 @@
|
|
|
34859
34867
|
"novita/baidu/ernie-4.5-vl-28b-a3b-thinking": {
|
|
34860
34868
|
"litellm_provider": "novita",
|
|
34861
34869
|
"mode": "chat",
|
|
34862
|
-
"input_cost_per_token":
|
|
34863
|
-
"output_cost_per_token":
|
|
34870
|
+
"input_cost_per_token": 3.9e-07,
|
|
34871
|
+
"output_cost_per_token": 3.9e-07,
|
|
34864
34872
|
"max_input_tokens": 131072,
|
|
34865
34873
|
"max_output_tokens": 65536,
|
|
34866
34874
|
"max_tokens": 65536,
|
|
@@ -34875,8 +34883,8 @@
|
|
|
34875
34883
|
"novita/qwen/qwen3-vl-8b-instruct": {
|
|
34876
34884
|
"litellm_provider": "novita",
|
|
34877
34885
|
"mode": "chat",
|
|
34878
|
-
"input_cost_per_token":
|
|
34879
|
-
"output_cost_per_token":
|
|
34886
|
+
"input_cost_per_token": 8e-08,
|
|
34887
|
+
"output_cost_per_token": 5e-07,
|
|
34880
34888
|
"max_input_tokens": 131072,
|
|
34881
34889
|
"max_output_tokens": 32768,
|
|
34882
34890
|
"max_tokens": 32768,
|
|
@@ -34890,8 +34898,8 @@
|
|
|
34890
34898
|
"novita/zai-org/glm-4.5-air": {
|
|
34891
34899
|
"litellm_provider": "novita",
|
|
34892
34900
|
"mode": "chat",
|
|
34893
|
-
"input_cost_per_token":
|
|
34894
|
-
"output_cost_per_token":
|
|
34901
|
+
"input_cost_per_token": 1.3e-07,
|
|
34902
|
+
"output_cost_per_token": 8.5e-07,
|
|
34895
34903
|
"max_input_tokens": 131072,
|
|
34896
34904
|
"max_output_tokens": 98304,
|
|
34897
34905
|
"max_tokens": 98304,
|
|
@@ -34904,8 +34912,8 @@
|
|
|
34904
34912
|
"novita/qwen/qwen3-vl-30b-a3b-instruct": {
|
|
34905
34913
|
"litellm_provider": "novita",
|
|
34906
34914
|
"mode": "chat",
|
|
34907
|
-
"input_cost_per_token":
|
|
34908
|
-
"output_cost_per_token":
|
|
34915
|
+
"input_cost_per_token": 2e-07,
|
|
34916
|
+
"output_cost_per_token": 7e-07,
|
|
34909
34917
|
"max_input_tokens": 131072,
|
|
34910
34918
|
"max_output_tokens": 32768,
|
|
34911
34919
|
"max_tokens": 32768,
|
|
@@ -34919,8 +34927,8 @@
|
|
|
34919
34927
|
"novita/qwen/qwen3-vl-30b-a3b-thinking": {
|
|
34920
34928
|
"litellm_provider": "novita",
|
|
34921
34929
|
"mode": "chat",
|
|
34922
|
-
"input_cost_per_token":
|
|
34923
|
-
"output_cost_per_token":
|
|
34930
|
+
"input_cost_per_token": 2e-07,
|
|
34931
|
+
"output_cost_per_token": 1e-06,
|
|
34924
34932
|
"max_input_tokens": 131072,
|
|
34925
34933
|
"max_output_tokens": 32768,
|
|
34926
34934
|
"max_tokens": 32768,
|
|
@@ -34934,8 +34942,8 @@
|
|
|
34934
34942
|
"novita/qwen/qwen-mt-plus": {
|
|
34935
34943
|
"litellm_provider": "novita",
|
|
34936
34944
|
"mode": "chat",
|
|
34937
|
-
"input_cost_per_token":
|
|
34938
|
-
"output_cost_per_token":
|
|
34945
|
+
"input_cost_per_token": 2.5e-07,
|
|
34946
|
+
"output_cost_per_token": 7.5e-07,
|
|
34939
34947
|
"max_input_tokens": 16384,
|
|
34940
34948
|
"max_output_tokens": 8192,
|
|
34941
34949
|
"max_tokens": 8192,
|
|
@@ -34944,8 +34952,8 @@
|
|
|
34944
34952
|
"novita/baidu/ernie-4.5-vl-28b-a3b": {
|
|
34945
34953
|
"litellm_provider": "novita",
|
|
34946
34954
|
"mode": "chat",
|
|
34947
|
-
"input_cost_per_token":
|
|
34948
|
-
"output_cost_per_token":
|
|
34955
|
+
"input_cost_per_token": 1.4e-07,
|
|
34956
|
+
"output_cost_per_token": 5.6e-07,
|
|
34949
34957
|
"max_input_tokens": 30000,
|
|
34950
34958
|
"max_output_tokens": 8000,
|
|
34951
34959
|
"max_tokens": 8000,
|
|
@@ -34959,8 +34967,8 @@
|
|
|
34959
34967
|
"novita/baidu/ernie-4.5-21B-a3b": {
|
|
34960
34968
|
"litellm_provider": "novita",
|
|
34961
34969
|
"mode": "chat",
|
|
34962
|
-
"input_cost_per_token":
|
|
34963
|
-
"output_cost_per_token":
|
|
34970
|
+
"input_cost_per_token": 7e-08,
|
|
34971
|
+
"output_cost_per_token": 2.8e-07,
|
|
34964
34972
|
"max_input_tokens": 120000,
|
|
34965
34973
|
"max_output_tokens": 8000,
|
|
34966
34974
|
"max_tokens": 8000,
|
|
@@ -34972,8 +34980,8 @@
|
|
|
34972
34980
|
"novita/qwen/qwen3-8b-fp8": {
|
|
34973
34981
|
"litellm_provider": "novita",
|
|
34974
34982
|
"mode": "chat",
|
|
34975
|
-
"input_cost_per_token":
|
|
34976
|
-
"output_cost_per_token":
|
|
34983
|
+
"input_cost_per_token": 3.5e-08,
|
|
34984
|
+
"output_cost_per_token": 1.38e-07,
|
|
34977
34985
|
"max_input_tokens": 128000,
|
|
34978
34986
|
"max_output_tokens": 20000,
|
|
34979
34987
|
"max_tokens": 20000,
|
|
@@ -34983,8 +34991,8 @@
|
|
|
34983
34991
|
"novita/qwen/qwen3-4b-fp8": {
|
|
34984
34992
|
"litellm_provider": "novita",
|
|
34985
34993
|
"mode": "chat",
|
|
34986
|
-
"input_cost_per_token":
|
|
34987
|
-
"output_cost_per_token":
|
|
34994
|
+
"input_cost_per_token": 3e-08,
|
|
34995
|
+
"output_cost_per_token": 3e-08,
|
|
34988
34996
|
"max_input_tokens": 128000,
|
|
34989
34997
|
"max_output_tokens": 20000,
|
|
34990
34998
|
"max_tokens": 20000,
|
|
@@ -34994,8 +35002,8 @@
|
|
|
34994
35002
|
"novita/qwen/qwen2.5-7b-instruct": {
|
|
34995
35003
|
"litellm_provider": "novita",
|
|
34996
35004
|
"mode": "chat",
|
|
34997
|
-
"input_cost_per_token":
|
|
34998
|
-
"output_cost_per_token":
|
|
35005
|
+
"input_cost_per_token": 7e-08,
|
|
35006
|
+
"output_cost_per_token": 7e-08,
|
|
34999
35007
|
"max_input_tokens": 32000,
|
|
35000
35008
|
"max_output_tokens": 32000,
|
|
35001
35009
|
"max_tokens": 32000,
|
|
@@ -35008,8 +35016,8 @@
|
|
|
35008
35016
|
"novita/meta-llama/llama-3.2-3b-instruct": {
|
|
35009
35017
|
"litellm_provider": "novita",
|
|
35010
35018
|
"mode": "chat",
|
|
35011
|
-
"input_cost_per_token":
|
|
35012
|
-
"output_cost_per_token":
|
|
35019
|
+
"input_cost_per_token": 3e-08,
|
|
35020
|
+
"output_cost_per_token": 5e-08,
|
|
35013
35021
|
"max_input_tokens": 32768,
|
|
35014
35022
|
"max_output_tokens": 32000,
|
|
35015
35023
|
"max_tokens": 32000,
|
|
@@ -35021,8 +35029,8 @@
|
|
|
35021
35029
|
"novita/sao10k/l31-70b-euryale-v2.2": {
|
|
35022
35030
|
"litellm_provider": "novita",
|
|
35023
35031
|
"mode": "chat",
|
|
35024
|
-
"input_cost_per_token":
|
|
35025
|
-
"output_cost_per_token":
|
|
35032
|
+
"input_cost_per_token": 1.48e-06,
|
|
35033
|
+
"output_cost_per_token": 1.48e-06,
|
|
35026
35034
|
"max_input_tokens": 8192,
|
|
35027
35035
|
"max_output_tokens": 8192,
|
|
35028
35036
|
"max_tokens": 8192,
|
|
@@ -35034,7 +35042,7 @@
|
|
|
35034
35042
|
"novita/qwen/qwen3-embedding-0.6b": {
|
|
35035
35043
|
"litellm_provider": "novita",
|
|
35036
35044
|
"mode": "embedding",
|
|
35037
|
-
"input_cost_per_token":
|
|
35045
|
+
"input_cost_per_token": 7e-08,
|
|
35038
35046
|
"output_cost_per_token": 0,
|
|
35039
35047
|
"max_input_tokens": 32768,
|
|
35040
35048
|
"max_output_tokens": 32768,
|
|
@@ -35043,7 +35051,7 @@
|
|
|
35043
35051
|
"novita/qwen/qwen3-embedding-8b": {
|
|
35044
35052
|
"litellm_provider": "novita",
|
|
35045
35053
|
"mode": "embedding",
|
|
35046
|
-
"input_cost_per_token":
|
|
35054
|
+
"input_cost_per_token": 7e-08,
|
|
35047
35055
|
"output_cost_per_token": 0,
|
|
35048
35056
|
"max_input_tokens": 32768,
|
|
35049
35057
|
"max_output_tokens": 4096,
|
|
@@ -35052,8 +35060,8 @@
|
|
|
35052
35060
|
"novita/baai/bge-m3": {
|
|
35053
35061
|
"litellm_provider": "novita",
|
|
35054
35062
|
"mode": "embedding",
|
|
35055
|
-
"input_cost_per_token":
|
|
35056
|
-
"output_cost_per_token":
|
|
35063
|
+
"input_cost_per_token": 1e-08,
|
|
35064
|
+
"output_cost_per_token": 1e-08,
|
|
35057
35065
|
"max_input_tokens": 8192,
|
|
35058
35066
|
"max_output_tokens": 96000,
|
|
35059
35067
|
"max_tokens": 96000
|
|
@@ -35061,8 +35069,8 @@
|
|
|
35061
35069
|
"novita/qwen/qwen3-reranker-8b": {
|
|
35062
35070
|
"litellm_provider": "novita",
|
|
35063
35071
|
"mode": "rerank",
|
|
35064
|
-
"input_cost_per_token":
|
|
35065
|
-
"output_cost_per_token":
|
|
35072
|
+
"input_cost_per_token": 5e-08,
|
|
35073
|
+
"output_cost_per_token": 5e-08,
|
|
35066
35074
|
"max_input_tokens": 32768,
|
|
35067
35075
|
"max_output_tokens": 4096,
|
|
35068
35076
|
"max_tokens": 4096
|
|
@@ -35070,8 +35078,8 @@
|
|
|
35070
35078
|
"novita/baai/bge-reranker-v2-m3": {
|
|
35071
35079
|
"litellm_provider": "novita",
|
|
35072
35080
|
"mode": "rerank",
|
|
35073
|
-
"input_cost_per_token":
|
|
35074
|
-
"output_cost_per_token":
|
|
35081
|
+
"input_cost_per_token": 1e-08,
|
|
35082
|
+
"output_cost_per_token": 1e-08,
|
|
35075
35083
|
"max_input_tokens": 8000,
|
|
35076
35084
|
"max_output_tokens": 8000,
|
|
35077
35085
|
"max_tokens": 8000
|
|
@@ -35369,5 +35377,73 @@
|
|
|
35369
35377
|
"max_tokens": 8192,
|
|
35370
35378
|
"supports_reasoning": true,
|
|
35371
35379
|
"supports_system_messages": true
|
|
35380
|
+
},
|
|
35381
|
+
"azure_ai/claude-opus-4-5": {
|
|
35382
|
+
"input_cost_per_token": 5e-06,
|
|
35383
|
+
"litellm_provider": "azure_ai",
|
|
35384
|
+
"max_input_tokens": 200000,
|
|
35385
|
+
"max_output_tokens": 64000,
|
|
35386
|
+
"max_tokens": 64000,
|
|
35387
|
+
"mode": "chat",
|
|
35388
|
+
"output_cost_per_token": 2.5e-05,
|
|
35389
|
+
"supports_assistant_prefill": true,
|
|
35390
|
+
"supports_computer_use": true,
|
|
35391
|
+
"supports_function_calling": true,
|
|
35392
|
+
"supports_pdf_input": true,
|
|
35393
|
+
"supports_prompt_caching": true,
|
|
35394
|
+
"supports_reasoning": true,
|
|
35395
|
+
"supports_response_schema": true,
|
|
35396
|
+
"supports_tool_choice": true,
|
|
35397
|
+
"supports_vision": true
|
|
35398
|
+
},
|
|
35399
|
+
"novita/kwaipilot/kat-coder-pro": {
|
|
35400
|
+
"litellm_provider": "novita",
|
|
35401
|
+
"mode": "chat",
|
|
35402
|
+
"input_cost_per_token": 3e-07,
|
|
35403
|
+
"output_cost_per_token": 1.2e-06,
|
|
35404
|
+
"max_input_tokens": 256000,
|
|
35405
|
+
"max_output_tokens": 128000,
|
|
35406
|
+
"max_tokens": 128000,
|
|
35407
|
+
"supports_function_calling": true,
|
|
35408
|
+
"supports_parallel_function_calling": true,
|
|
35409
|
+
"supports_tool_choice": true,
|
|
35410
|
+
"supports_system_messages": true,
|
|
35411
|
+
"supports_response_schema": true,
|
|
35412
|
+
"cache_read_input_token_cost": 6e-08,
|
|
35413
|
+
"input_cost_per_token_cache_hit": 6e-08
|
|
35414
|
+
},
|
|
35415
|
+
"novita/qwen/qwen3-omni-30b-a3b-thinking": {
|
|
35416
|
+
"litellm_provider": "novita",
|
|
35417
|
+
"mode": "chat",
|
|
35418
|
+
"input_cost_per_token": 2.5e-07,
|
|
35419
|
+
"output_cost_per_token": 9.7e-07,
|
|
35420
|
+
"max_input_tokens": 65536,
|
|
35421
|
+
"max_output_tokens": 16384,
|
|
35422
|
+
"max_tokens": 16384,
|
|
35423
|
+
"supports_function_calling": true,
|
|
35424
|
+
"supports_parallel_function_calling": true,
|
|
35425
|
+
"supports_tool_choice": true,
|
|
35426
|
+
"supports_vision": true,
|
|
35427
|
+
"supports_system_messages": true,
|
|
35428
|
+
"supports_response_schema": true,
|
|
35429
|
+
"supports_reasoning": true,
|
|
35430
|
+
"supports_audio_input": true
|
|
35431
|
+
},
|
|
35432
|
+
"novita/qwen/qwen3-omni-30b-a3b-instruct": {
|
|
35433
|
+
"litellm_provider": "novita",
|
|
35434
|
+
"mode": "chat",
|
|
35435
|
+
"input_cost_per_token": 2.5e-07,
|
|
35436
|
+
"output_cost_per_token": 9.7e-07,
|
|
35437
|
+
"max_input_tokens": 65536,
|
|
35438
|
+
"max_output_tokens": 16384,
|
|
35439
|
+
"max_tokens": 16384,
|
|
35440
|
+
"supports_function_calling": true,
|
|
35441
|
+
"supports_parallel_function_calling": true,
|
|
35442
|
+
"supports_tool_choice": true,
|
|
35443
|
+
"supports_vision": true,
|
|
35444
|
+
"supports_system_messages": true,
|
|
35445
|
+
"supports_response_schema": true,
|
|
35446
|
+
"supports_audio_input": true,
|
|
35447
|
+
"supports_audio_output": true
|
|
35372
35448
|
}
|
|
35373
35449
|
}
|