tokencostauto 0.1.447__py3-none-any.whl → 0.1.452__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- tokencostauto/model_prices.json +408 -266
- {tokencostauto-0.1.447.dist-info → tokencostauto-0.1.452.dist-info}/METADATA +1 -1
- tokencostauto-0.1.452.dist-info/RECORD +9 -0
- tokencostauto-0.1.447.dist-info/RECORD +0 -9
- {tokencostauto-0.1.447.dist-info → tokencostauto-0.1.452.dist-info}/WHEEL +0 -0
- {tokencostauto-0.1.447.dist-info → tokencostauto-0.1.452.dist-info}/licenses/LICENSE +0 -0
- {tokencostauto-0.1.447.dist-info → tokencostauto-0.1.452.dist-info}/top_level.txt +0 -0
tokencostauto/model_prices.json
CHANGED
|
@@ -1656,14 +1656,14 @@
|
|
|
1656
1656
|
"mode": "embedding"
|
|
1657
1657
|
},
|
|
1658
1658
|
"deepseek-chat": {
|
|
1659
|
-
"cache_read_input_token_cost":
|
|
1660
|
-
"input_cost_per_token":
|
|
1659
|
+
"cache_read_input_token_cost": 2.8e-08,
|
|
1660
|
+
"input_cost_per_token": 2.8e-07,
|
|
1661
1661
|
"litellm_provider": "deepseek",
|
|
1662
1662
|
"max_input_tokens": 131072,
|
|
1663
1663
|
"max_output_tokens": 8192,
|
|
1664
1664
|
"max_tokens": 8192,
|
|
1665
1665
|
"mode": "chat",
|
|
1666
|
-
"output_cost_per_token":
|
|
1666
|
+
"output_cost_per_token": 4.2e-07,
|
|
1667
1667
|
"source": "https://api-docs.deepseek.com/quick_start/pricing",
|
|
1668
1668
|
"supported_endpoints": [
|
|
1669
1669
|
"/v1/chat/completions"
|
|
@@ -8316,15 +8316,15 @@
|
|
|
8316
8316
|
},
|
|
8317
8317
|
"deepseek/deepseek-chat": {
|
|
8318
8318
|
"cache_creation_input_token_cost": 0.0,
|
|
8319
|
-
"cache_read_input_token_cost":
|
|
8320
|
-
"input_cost_per_token": 2.
|
|
8321
|
-
"input_cost_per_token_cache_hit":
|
|
8319
|
+
"cache_read_input_token_cost": 2.8e-08,
|
|
8320
|
+
"input_cost_per_token": 2.8e-07,
|
|
8321
|
+
"input_cost_per_token_cache_hit": 2.8e-08,
|
|
8322
8322
|
"litellm_provider": "deepseek",
|
|
8323
|
-
"max_input_tokens":
|
|
8323
|
+
"max_input_tokens": 128000,
|
|
8324
8324
|
"max_output_tokens": 8192,
|
|
8325
8325
|
"max_tokens": 8192,
|
|
8326
8326
|
"mode": "chat",
|
|
8327
|
-
"output_cost_per_token":
|
|
8327
|
+
"output_cost_per_token": 4.2e-07,
|
|
8328
8328
|
"supports_assistant_prefill": true,
|
|
8329
8329
|
"supports_function_calling": true,
|
|
8330
8330
|
"supports_prompt_caching": true,
|
|
@@ -8995,14 +8995,15 @@
|
|
|
8995
8995
|
"supports_tool_choice": true
|
|
8996
8996
|
},
|
|
8997
8997
|
"deepseek/deepseek-reasoner": {
|
|
8998
|
-
"
|
|
8999
|
-
"
|
|
8998
|
+
"cache_read_input_token_cost": 2.8e-08,
|
|
8999
|
+
"input_cost_per_token": 2.8e-07,
|
|
9000
|
+
"input_cost_per_token_cache_hit": 2.8e-08,
|
|
9000
9001
|
"litellm_provider": "deepseek",
|
|
9001
|
-
"max_input_tokens":
|
|
9002
|
+
"max_input_tokens": 128000,
|
|
9002
9003
|
"max_output_tokens": 8192,
|
|
9003
9004
|
"max_tokens": 8192,
|
|
9004
9005
|
"mode": "chat",
|
|
9005
|
-
"output_cost_per_token":
|
|
9006
|
+
"output_cost_per_token": 4.2e-07,
|
|
9006
9007
|
"supports_assistant_prefill": true,
|
|
9007
9008
|
"supports_function_calling": true,
|
|
9008
9009
|
"supports_prompt_caching": true,
|
|
@@ -12284,7 +12285,8 @@
|
|
|
12284
12285
|
"supports_tool_choice": true
|
|
12285
12286
|
},
|
|
12286
12287
|
"gemini-2.5-pro-exp-03-25": {
|
|
12287
|
-
"cache_read_input_token_cost":
|
|
12288
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
12289
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
12288
12290
|
"input_cost_per_token": 1.25e-06,
|
|
12289
12291
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
12290
12292
|
"litellm_provider": "vertex_ai-language-models",
|
|
@@ -12570,7 +12572,8 @@
|
|
|
12570
12572
|
},
|
|
12571
12573
|
"gemini-2.5-pro-preview-05-06": {
|
|
12572
12574
|
"deprecation_date": "2025-12-02",
|
|
12573
|
-
"cache_read_input_token_cost":
|
|
12575
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
12576
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
12574
12577
|
"input_cost_per_audio_token": 1.25e-06,
|
|
12575
12578
|
"input_cost_per_token": 1.25e-06,
|
|
12576
12579
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -12619,7 +12622,8 @@
|
|
|
12619
12622
|
},
|
|
12620
12623
|
"gemini-2.5-pro-preview-03-25": {
|
|
12621
12624
|
"deprecation_date": "2025-12-02",
|
|
12622
|
-
"cache_read_input_token_cost":
|
|
12625
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
12626
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
12623
12627
|
"input_cost_per_audio_token": 1.25e-06,
|
|
12624
12628
|
"input_cost_per_token": 1.25e-06,
|
|
12625
12629
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -12701,7 +12705,8 @@
|
|
|
12701
12705
|
},
|
|
12702
12706
|
"gemini/gemini-2.5-pro-preview-05-06": {
|
|
12703
12707
|
"deprecation_date": "2025-12-02",
|
|
12704
|
-
"cache_read_input_token_cost":
|
|
12708
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
12709
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
12705
12710
|
"input_cost_per_audio_token": 7e-07,
|
|
12706
12711
|
"input_cost_per_token": 1.25e-06,
|
|
12707
12712
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -12743,7 +12748,8 @@
|
|
|
12743
12748
|
},
|
|
12744
12749
|
"gemini/gemini-2.5-pro-preview-03-25": {
|
|
12745
12750
|
"deprecation_date": "2025-12-02",
|
|
12746
|
-
"cache_read_input_token_cost":
|
|
12751
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
12752
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
12747
12753
|
"input_cost_per_audio_token": 7e-07,
|
|
12748
12754
|
"input_cost_per_token": 1.25e-06,
|
|
12749
12755
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -13921,7 +13927,8 @@
|
|
|
13921
13927
|
"supports_web_search": true
|
|
13922
13928
|
},
|
|
13923
13929
|
"gemini-2.5-pro-preview-tts": {
|
|
13924
|
-
"cache_read_input_token_cost":
|
|
13930
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
13931
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
13925
13932
|
"input_cost_per_audio_token": 7e-07,
|
|
13926
13933
|
"input_cost_per_token": 1.25e-06,
|
|
13927
13934
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -13996,7 +14003,8 @@
|
|
|
13996
14003
|
"tpm": 10000000
|
|
13997
14004
|
},
|
|
13998
14005
|
"gemini/gemini-2.5-pro-preview-tts": {
|
|
13999
|
-
"cache_read_input_token_cost":
|
|
14006
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
14007
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
14000
14008
|
"input_cost_per_audio_token": 7e-07,
|
|
14001
14009
|
"input_cost_per_token": 1.25e-06,
|
|
14002
14010
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -14563,7 +14571,8 @@
|
|
|
14563
14571
|
]
|
|
14564
14572
|
},
|
|
14565
14573
|
"gemini-2.5-pro-preview-06-05": {
|
|
14566
|
-
"cache_read_input_token_cost":
|
|
14574
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
14575
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
14567
14576
|
"input_cost_per_audio_token": 1.25e-06,
|
|
14568
14577
|
"input_cost_per_token": 1.25e-06,
|
|
14569
14578
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -14608,7 +14617,8 @@
|
|
|
14608
14617
|
"supports_web_search": true
|
|
14609
14618
|
},
|
|
14610
14619
|
"gemini/gemini-2.5-pro-preview-06-05": {
|
|
14611
|
-
"cache_read_input_token_cost":
|
|
14620
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
14621
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
14612
14622
|
"input_cost_per_audio_token": 7e-07,
|
|
14613
14623
|
"input_cost_per_token": 1.25e-06,
|
|
14614
14624
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -15441,7 +15451,8 @@
|
|
|
15441
15451
|
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing"
|
|
15442
15452
|
},
|
|
15443
15453
|
"gemini/gemini-2.5-pro": {
|
|
15444
|
-
"cache_read_input_token_cost":
|
|
15454
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
15455
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
15445
15456
|
"input_cost_per_token": 1.25e-06,
|
|
15446
15457
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
15447
15458
|
"litellm_provider": "gemini",
|
|
@@ -15756,6 +15767,7 @@
|
|
|
15756
15767
|
},
|
|
15757
15768
|
"gemini-2.5-pro": {
|
|
15758
15769
|
"cache_read_input_token_cost": 1.25e-07,
|
|
15770
|
+
"cache_read_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
15759
15771
|
"cache_creation_input_token_cost_above_200k_tokens": 2.5e-07,
|
|
15760
15772
|
"input_cost_per_token": 1.25e-06,
|
|
15761
15773
|
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
@@ -16605,14 +16617,14 @@
|
|
|
16605
16617
|
"supports_tool_choice": true
|
|
16606
16618
|
},
|
|
16607
16619
|
"azure_ai/grok-3": {
|
|
16608
|
-
"input_cost_per_token":
|
|
16620
|
+
"input_cost_per_token": 3e-06,
|
|
16609
16621
|
"litellm_provider": "azure_ai",
|
|
16610
16622
|
"max_input_tokens": 131072,
|
|
16611
16623
|
"max_output_tokens": 131072,
|
|
16612
16624
|
"max_tokens": 131072,
|
|
16613
16625
|
"mode": "chat",
|
|
16614
|
-
"output_cost_per_token": 1.
|
|
16615
|
-
"source": "https://
|
|
16626
|
+
"output_cost_per_token": 1.5e-05,
|
|
16627
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
|
|
16616
16628
|
"supports_function_calling": true,
|
|
16617
16629
|
"supports_response_schema": false,
|
|
16618
16630
|
"supports_tool_choice": true,
|
|
@@ -16648,14 +16660,14 @@
|
|
|
16648
16660
|
"supports_web_search": true
|
|
16649
16661
|
},
|
|
16650
16662
|
"azure_ai/grok-3-mini": {
|
|
16651
|
-
"input_cost_per_token": 2.
|
|
16663
|
+
"input_cost_per_token": 2.5e-07,
|
|
16652
16664
|
"litellm_provider": "azure_ai",
|
|
16653
16665
|
"max_input_tokens": 131072,
|
|
16654
16666
|
"max_output_tokens": 131072,
|
|
16655
16667
|
"max_tokens": 131072,
|
|
16656
16668
|
"mode": "chat",
|
|
16657
|
-
"output_cost_per_token": 1.
|
|
16658
|
-
"source": "https://
|
|
16669
|
+
"output_cost_per_token": 1.27e-06,
|
|
16670
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
|
|
16659
16671
|
"supports_function_calling": true,
|
|
16660
16672
|
"supports_reasoning": true,
|
|
16661
16673
|
"supports_response_schema": false,
|
|
@@ -18277,9 +18289,9 @@
|
|
|
18277
18289
|
"cache_read_input_token_cost": 1.25e-07,
|
|
18278
18290
|
"input_cost_per_token": 1.25e-06,
|
|
18279
18291
|
"litellm_provider": "openai",
|
|
18280
|
-
"max_input_tokens":
|
|
18281
|
-
"max_output_tokens":
|
|
18282
|
-
"max_tokens":
|
|
18292
|
+
"max_input_tokens": 128000,
|
|
18293
|
+
"max_output_tokens": 16384,
|
|
18294
|
+
"max_tokens": 16384,
|
|
18283
18295
|
"mode": "chat",
|
|
18284
18296
|
"output_cost_per_token": 1e-05,
|
|
18285
18297
|
"supported_endpoints": [
|
|
@@ -18646,9 +18658,9 @@
|
|
|
18646
18658
|
"cache_read_input_token_cost": 1.25e-07,
|
|
18647
18659
|
"input_cost_per_token": 1.25e-06,
|
|
18648
18660
|
"litellm_provider": "azure",
|
|
18649
|
-
"max_input_tokens":
|
|
18650
|
-
"max_output_tokens":
|
|
18651
|
-
"max_tokens":
|
|
18661
|
+
"max_input_tokens": 128000,
|
|
18662
|
+
"max_output_tokens": 16384,
|
|
18663
|
+
"max_tokens": 16384,
|
|
18652
18664
|
"mode": "chat",
|
|
18653
18665
|
"output_cost_per_token": 1e-05,
|
|
18654
18666
|
"source": "https://azure.microsoft.com/en-us/blog/gpt-5-in-azure-ai-foundry-the-future-of-ai-apps-and-agents-starts-here/",
|
|
@@ -20649,9 +20661,9 @@
|
|
|
20649
20661
|
"cache_read_input_token_cost": 1.25e-07,
|
|
20650
20662
|
"input_cost_per_token": 1.25e-06,
|
|
20651
20663
|
"litellm_provider": "openrouter",
|
|
20652
|
-
"max_input_tokens":
|
|
20653
|
-
"max_output_tokens":
|
|
20654
|
-
"max_tokens":
|
|
20664
|
+
"max_input_tokens": 128000,
|
|
20665
|
+
"max_output_tokens": 16384,
|
|
20666
|
+
"max_tokens": 16384,
|
|
20655
20667
|
"mode": "chat",
|
|
20656
20668
|
"output_cost_per_token": 1e-05,
|
|
20657
20669
|
"supported_modalities": [
|
|
@@ -22912,14 +22924,14 @@
|
|
|
22912
22924
|
"supports_vision": true
|
|
22913
22925
|
},
|
|
22914
22926
|
"deepseek-reasoner": {
|
|
22915
|
-
"cache_read_input_token_cost":
|
|
22916
|
-
"input_cost_per_token":
|
|
22927
|
+
"cache_read_input_token_cost": 2.8e-08,
|
|
22928
|
+
"input_cost_per_token": 2.8e-07,
|
|
22917
22929
|
"litellm_provider": "deepseek",
|
|
22918
22930
|
"max_input_tokens": 131072,
|
|
22919
22931
|
"max_output_tokens": 65536,
|
|
22920
22932
|
"max_tokens": 65536,
|
|
22921
22933
|
"mode": "chat",
|
|
22922
|
-
"output_cost_per_token":
|
|
22934
|
+
"output_cost_per_token": 4.2e-07,
|
|
22923
22935
|
"source": "https://api-docs.deepseek.com/quick_start/pricing",
|
|
22924
22936
|
"supported_endpoints": [
|
|
22925
22937
|
"/v1/chat/completions"
|
|
@@ -23708,22 +23720,22 @@
|
|
|
23708
23720
|
"supports_tool_choice": true
|
|
23709
23721
|
},
|
|
23710
23722
|
"azure_ai/grok-4": {
|
|
23711
|
-
"input_cost_per_token":
|
|
23723
|
+
"input_cost_per_token": 3e-06,
|
|
23712
23724
|
"litellm_provider": "azure_ai",
|
|
23713
23725
|
"max_input_tokens": 131072,
|
|
23714
23726
|
"max_output_tokens": 131072,
|
|
23715
23727
|
"max_tokens": 131072,
|
|
23716
23728
|
"mode": "chat",
|
|
23717
|
-
"output_cost_per_token":
|
|
23718
|
-
"source": "https://azure.microsoft.com/en-us/
|
|
23729
|
+
"output_cost_per_token": 1.5e-05,
|
|
23730
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
|
|
23719
23731
|
"supports_function_calling": true,
|
|
23720
23732
|
"supports_response_schema": true,
|
|
23721
23733
|
"supports_tool_choice": true,
|
|
23722
23734
|
"supports_web_search": true
|
|
23723
23735
|
},
|
|
23724
23736
|
"azure_ai/grok-4-fast-non-reasoning": {
|
|
23725
|
-
"input_cost_per_token":
|
|
23726
|
-
"output_cost_per_token":
|
|
23737
|
+
"input_cost_per_token": 2e-07,
|
|
23738
|
+
"output_cost_per_token": 5e-07,
|
|
23727
23739
|
"litellm_provider": "azure_ai",
|
|
23728
23740
|
"max_input_tokens": 131072,
|
|
23729
23741
|
"max_output_tokens": 131072,
|
|
@@ -23735,28 +23747,28 @@
|
|
|
23735
23747
|
"supports_web_search": true
|
|
23736
23748
|
},
|
|
23737
23749
|
"azure_ai/grok-4-fast-reasoning": {
|
|
23738
|
-
"input_cost_per_token":
|
|
23739
|
-
"output_cost_per_token":
|
|
23750
|
+
"input_cost_per_token": 2e-07,
|
|
23751
|
+
"output_cost_per_token": 5e-07,
|
|
23740
23752
|
"litellm_provider": "azure_ai",
|
|
23741
23753
|
"max_input_tokens": 131072,
|
|
23742
23754
|
"max_output_tokens": 131072,
|
|
23743
23755
|
"max_tokens": 131072,
|
|
23744
23756
|
"mode": "chat",
|
|
23745
|
-
"source": "https://
|
|
23757
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
|
|
23746
23758
|
"supports_function_calling": true,
|
|
23747
23759
|
"supports_response_schema": true,
|
|
23748
23760
|
"supports_tool_choice": true,
|
|
23749
23761
|
"supports_web_search": true
|
|
23750
23762
|
},
|
|
23751
23763
|
"azure_ai/grok-code-fast-1": {
|
|
23752
|
-
"input_cost_per_token":
|
|
23764
|
+
"input_cost_per_token": 2e-07,
|
|
23753
23765
|
"litellm_provider": "azure_ai",
|
|
23754
23766
|
"max_input_tokens": 131072,
|
|
23755
23767
|
"max_output_tokens": 131072,
|
|
23756
23768
|
"max_tokens": 131072,
|
|
23757
23769
|
"mode": "chat",
|
|
23758
|
-
"output_cost_per_token": 1.
|
|
23759
|
-
"source": "https://azure.microsoft.com/en-us/
|
|
23770
|
+
"output_cost_per_token": 1.5e-06,
|
|
23771
|
+
"source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
|
|
23760
23772
|
"supports_function_calling": true,
|
|
23761
23773
|
"supports_response_schema": true,
|
|
23762
23774
|
"supports_tool_choice": true,
|
|
@@ -24262,7 +24274,7 @@
|
|
|
24262
24274
|
"input_cost_per_token": 1.5e-05,
|
|
24263
24275
|
"input_cost_per_token_batches": 7.5e-06,
|
|
24264
24276
|
"litellm_provider": "openai",
|
|
24265
|
-
"max_input_tokens":
|
|
24277
|
+
"max_input_tokens": 128000,
|
|
24266
24278
|
"max_output_tokens": 272000,
|
|
24267
24279
|
"max_tokens": 272000,
|
|
24268
24280
|
"mode": "responses",
|
|
@@ -24397,7 +24409,7 @@
|
|
|
24397
24409
|
"input_cost_per_token": 1.5e-05,
|
|
24398
24410
|
"input_cost_per_token_batches": 7.5e-06,
|
|
24399
24411
|
"litellm_provider": "openai",
|
|
24400
|
-
"max_input_tokens":
|
|
24412
|
+
"max_input_tokens": 128000,
|
|
24401
24413
|
"max_output_tokens": 272000,
|
|
24402
24414
|
"max_tokens": 272000,
|
|
24403
24415
|
"mode": "responses",
|
|
@@ -27148,7 +27160,7 @@
|
|
|
27148
27160
|
"cache_read_input_token_cost": 1.4e-07,
|
|
27149
27161
|
"input_cost_per_token": 1.38e-06,
|
|
27150
27162
|
"litellm_provider": "azure",
|
|
27151
|
-
"max_input_tokens":
|
|
27163
|
+
"max_input_tokens": 128000,
|
|
27152
27164
|
"max_output_tokens": 128000,
|
|
27153
27165
|
"max_tokens": 128000,
|
|
27154
27166
|
"mode": "chat",
|
|
@@ -27274,7 +27286,7 @@
|
|
|
27274
27286
|
"cache_read_input_token_cost": 1.25e-07,
|
|
27275
27287
|
"input_cost_per_token": 1.25e-06,
|
|
27276
27288
|
"litellm_provider": "azure",
|
|
27277
|
-
"max_input_tokens":
|
|
27289
|
+
"max_input_tokens": 128000,
|
|
27278
27290
|
"max_output_tokens": 128000,
|
|
27279
27291
|
"max_tokens": 128000,
|
|
27280
27292
|
"mode": "chat",
|
|
@@ -27400,7 +27412,7 @@
|
|
|
27400
27412
|
"cache_read_input_token_cost": 1.25e-07,
|
|
27401
27413
|
"input_cost_per_token": 1.25e-06,
|
|
27402
27414
|
"litellm_provider": "azure",
|
|
27403
|
-
"max_input_tokens":
|
|
27415
|
+
"max_input_tokens": 128000,
|
|
27404
27416
|
"max_output_tokens": 128000,
|
|
27405
27417
|
"max_tokens": 128000,
|
|
27406
27418
|
"mode": "chat",
|
|
@@ -27526,7 +27538,7 @@
|
|
|
27526
27538
|
"cache_read_input_token_cost": 1.4e-07,
|
|
27527
27539
|
"input_cost_per_token": 1.38e-06,
|
|
27528
27540
|
"litellm_provider": "azure",
|
|
27529
|
-
"max_input_tokens":
|
|
27541
|
+
"max_input_tokens": 128000,
|
|
27530
27542
|
"max_output_tokens": 128000,
|
|
27531
27543
|
"max_tokens": 128000,
|
|
27532
27544
|
"mode": "chat",
|
|
@@ -28646,7 +28658,7 @@
|
|
|
28646
28658
|
"input_cost_per_token": 1.24999e-06,
|
|
28647
28659
|
"input_dbu_cost_per_token": 1.7857e-05,
|
|
28648
28660
|
"litellm_provider": "databricks",
|
|
28649
|
-
"max_input_tokens":
|
|
28661
|
+
"max_input_tokens": 272000,
|
|
28650
28662
|
"max_output_tokens": 128000,
|
|
28651
28663
|
"max_tokens": 128000,
|
|
28652
28664
|
"metadata": {
|
|
@@ -28661,7 +28673,7 @@
|
|
|
28661
28673
|
"input_cost_per_token": 1.24999e-06,
|
|
28662
28674
|
"input_dbu_cost_per_token": 1.7857e-05,
|
|
28663
28675
|
"litellm_provider": "databricks",
|
|
28664
|
-
"max_input_tokens":
|
|
28676
|
+
"max_input_tokens": 272000,
|
|
28665
28677
|
"max_output_tokens": 128000,
|
|
28666
28678
|
"max_tokens": 128000,
|
|
28667
28679
|
"metadata": {
|
|
@@ -28676,7 +28688,7 @@
|
|
|
28676
28688
|
"input_cost_per_token": 2.4997000000000006e-07,
|
|
28677
28689
|
"input_dbu_cost_per_token": 3.571e-06,
|
|
28678
28690
|
"litellm_provider": "databricks",
|
|
28679
|
-
"max_input_tokens":
|
|
28691
|
+
"max_input_tokens": 272000,
|
|
28680
28692
|
"max_output_tokens": 128000,
|
|
28681
28693
|
"max_tokens": 128000,
|
|
28682
28694
|
"metadata": {
|
|
@@ -28691,7 +28703,7 @@
|
|
|
28691
28703
|
"input_cost_per_token": 4.998e-08,
|
|
28692
28704
|
"input_dbu_cost_per_token": 7.14e-07,
|
|
28693
28705
|
"litellm_provider": "databricks",
|
|
28694
|
-
"max_input_tokens":
|
|
28706
|
+
"max_input_tokens": 272000,
|
|
28695
28707
|
"max_output_tokens": 128000,
|
|
28696
28708
|
"max_tokens": 128000,
|
|
28697
28709
|
"metadata": {
|
|
@@ -29110,7 +29122,7 @@
|
|
|
29110
29122
|
"cache_read_input_token_cost": 1.25e-07,
|
|
29111
29123
|
"input_cost_per_token": 1.25e-06,
|
|
29112
29124
|
"litellm_provider": "azure",
|
|
29113
|
-
"max_input_tokens":
|
|
29125
|
+
"max_input_tokens": 272000,
|
|
29114
29126
|
"max_output_tokens": 128000,
|
|
29115
29127
|
"max_tokens": 128000,
|
|
29116
29128
|
"mode": "responses",
|
|
@@ -29153,7 +29165,7 @@
|
|
|
29153
29165
|
"cache_read_input_token_cost": 1.25e-07,
|
|
29154
29166
|
"input_cost_per_token": 1.25e-06,
|
|
29155
29167
|
"litellm_provider": "openai",
|
|
29156
|
-
"max_input_tokens":
|
|
29168
|
+
"max_input_tokens": 272000,
|
|
29157
29169
|
"max_output_tokens": 128000,
|
|
29158
29170
|
"max_tokens": 128000,
|
|
29159
29171
|
"mode": "responses",
|
|
@@ -29513,6 +29525,7 @@
|
|
|
29513
29525
|
"supports_pdf_input": true
|
|
29514
29526
|
},
|
|
29515
29527
|
"cerebras/zai-glm-4.6": {
|
|
29528
|
+
"deprecation_date": "2026-01-20",
|
|
29516
29529
|
"input_cost_per_token": 2.25e-06,
|
|
29517
29530
|
"litellm_provider": "cerebras",
|
|
29518
29531
|
"max_input_tokens": 128000,
|
|
@@ -31703,7 +31716,7 @@
|
|
|
31703
31716
|
"input_cost_per_token": 1.75e-06,
|
|
31704
31717
|
"input_cost_per_token_priority": 3.5e-06,
|
|
31705
31718
|
"litellm_provider": "openai",
|
|
31706
|
-
"max_input_tokens":
|
|
31719
|
+
"max_input_tokens": 272000,
|
|
31707
31720
|
"max_output_tokens": 128000,
|
|
31708
31721
|
"max_tokens": 128000,
|
|
31709
31722
|
"mode": "chat",
|
|
@@ -31740,7 +31753,7 @@
|
|
|
31740
31753
|
"input_cost_per_token": 1.75e-06,
|
|
31741
31754
|
"input_cost_per_token_priority": 3.5e-06,
|
|
31742
31755
|
"litellm_provider": "openai",
|
|
31743
|
-
"max_input_tokens":
|
|
31756
|
+
"max_input_tokens": 272000,
|
|
31744
31757
|
"max_output_tokens": 128000,
|
|
31745
31758
|
"max_tokens": 128000,
|
|
31746
31759
|
"mode": "chat",
|
|
@@ -31808,7 +31821,7 @@
|
|
|
31808
31821
|
"gpt-5.2-pro": {
|
|
31809
31822
|
"input_cost_per_token": 2.1e-05,
|
|
31810
31823
|
"litellm_provider": "openai",
|
|
31811
|
-
"max_input_tokens":
|
|
31824
|
+
"max_input_tokens": 272000,
|
|
31812
31825
|
"max_output_tokens": 128000,
|
|
31813
31826
|
"max_tokens": 128000,
|
|
31814
31827
|
"mode": "responses",
|
|
@@ -31839,7 +31852,7 @@
|
|
|
31839
31852
|
"gpt-5.2-pro-2025-12-11": {
|
|
31840
31853
|
"input_cost_per_token": 2.1e-05,
|
|
31841
31854
|
"litellm_provider": "openai",
|
|
31842
|
-
"max_input_tokens":
|
|
31855
|
+
"max_input_tokens": 272000,
|
|
31843
31856
|
"max_output_tokens": 128000,
|
|
31844
31857
|
"max_tokens": 128000,
|
|
31845
31858
|
"mode": "responses",
|
|
@@ -31913,7 +31926,7 @@
|
|
|
31913
31926
|
"cache_read_input_token_cost": 1.75e-07,
|
|
31914
31927
|
"input_cost_per_token": 1.75e-06,
|
|
31915
31928
|
"litellm_provider": "azure",
|
|
31916
|
-
"max_input_tokens":
|
|
31929
|
+
"max_input_tokens": 272000,
|
|
31917
31930
|
"max_output_tokens": 128000,
|
|
31918
31931
|
"max_tokens": 128000,
|
|
31919
31932
|
"mode": "chat",
|
|
@@ -31948,7 +31961,7 @@
|
|
|
31948
31961
|
"input_cost_per_token": 1.75e-06,
|
|
31949
31962
|
"input_cost_per_token_priority": 3.5e-06,
|
|
31950
31963
|
"litellm_provider": "azure",
|
|
31951
|
-
"max_input_tokens":
|
|
31964
|
+
"max_input_tokens": 272000,
|
|
31952
31965
|
"max_output_tokens": 128000,
|
|
31953
31966
|
"max_tokens": 128000,
|
|
31954
31967
|
"mode": "chat",
|
|
@@ -32016,7 +32029,7 @@
|
|
|
32016
32029
|
"azure/gpt-5.2-pro": {
|
|
32017
32030
|
"input_cost_per_token": 2.1e-05,
|
|
32018
32031
|
"litellm_provider": "azure",
|
|
32019
|
-
"max_input_tokens":
|
|
32032
|
+
"max_input_tokens": 272000,
|
|
32020
32033
|
"max_output_tokens": 128000,
|
|
32021
32034
|
"max_tokens": 128000,
|
|
32022
32035
|
"mode": "responses",
|
|
@@ -32047,7 +32060,7 @@
|
|
|
32047
32060
|
"azure/gpt-5.2-pro-2025-12-11": {
|
|
32048
32061
|
"input_cost_per_token": 2.1e-05,
|
|
32049
32062
|
"litellm_provider": "azure",
|
|
32050
|
-
"max_input_tokens":
|
|
32063
|
+
"max_input_tokens": 272000,
|
|
32051
32064
|
"max_output_tokens": 128000,
|
|
32052
32065
|
"max_tokens": 128000,
|
|
32053
32066
|
"mode": "responses",
|
|
@@ -32284,7 +32297,7 @@
|
|
|
32284
32297
|
"cache_read_input_token_cost": 1.75e-07,
|
|
32285
32298
|
"input_cost_per_token": 1.75e-06,
|
|
32286
32299
|
"litellm_provider": "openrouter",
|
|
32287
|
-
"max_input_tokens":
|
|
32300
|
+
"max_input_tokens": 272000,
|
|
32288
32301
|
"max_output_tokens": 128000,
|
|
32289
32302
|
"max_tokens": 128000,
|
|
32290
32303
|
"mode": "chat",
|
|
@@ -32314,7 +32327,7 @@
|
|
|
32314
32327
|
"input_cost_per_image": 0,
|
|
32315
32328
|
"input_cost_per_token": 2.1e-05,
|
|
32316
32329
|
"litellm_provider": "openrouter",
|
|
32317
|
-
"max_input_tokens":
|
|
32330
|
+
"max_input_tokens": 272000,
|
|
32318
32331
|
"max_output_tokens": 128000,
|
|
32319
32332
|
"max_tokens": 128000,
|
|
32320
32333
|
"mode": "chat",
|
|
@@ -33985,8 +33998,8 @@
|
|
|
33985
33998
|
"novita/deepseek/deepseek-v3.2": {
|
|
33986
33999
|
"litellm_provider": "novita",
|
|
33987
34000
|
"mode": "chat",
|
|
33988
|
-
"input_cost_per_token":
|
|
33989
|
-
"output_cost_per_token":
|
|
34001
|
+
"input_cost_per_token": 2.69e-07,
|
|
34002
|
+
"output_cost_per_token": 4e-07,
|
|
33990
34003
|
"max_input_tokens": 163840,
|
|
33991
34004
|
"max_output_tokens": 65536,
|
|
33992
34005
|
"max_tokens": 65536,
|
|
@@ -33995,15 +34008,15 @@
|
|
|
33995
34008
|
"supports_tool_choice": true,
|
|
33996
34009
|
"supports_system_messages": true,
|
|
33997
34010
|
"supports_response_schema": true,
|
|
33998
|
-
"cache_read_input_token_cost":
|
|
33999
|
-
"input_cost_per_token_cache_hit":
|
|
34011
|
+
"cache_read_input_token_cost": 1.345e-07,
|
|
34012
|
+
"input_cost_per_token_cache_hit": 1.345e-07,
|
|
34000
34013
|
"supports_reasoning": true
|
|
34001
34014
|
},
|
|
34002
34015
|
"novita/minimax/minimax-m2.1": {
|
|
34003
34016
|
"litellm_provider": "novita",
|
|
34004
34017
|
"mode": "chat",
|
|
34005
|
-
"input_cost_per_token":
|
|
34006
|
-
"output_cost_per_token":
|
|
34018
|
+
"input_cost_per_token": 3e-07,
|
|
34019
|
+
"output_cost_per_token": 1.2e-06,
|
|
34007
34020
|
"max_input_tokens": 204800,
|
|
34008
34021
|
"max_output_tokens": 131072,
|
|
34009
34022
|
"max_tokens": 131072,
|
|
@@ -34012,15 +34025,14 @@
|
|
|
34012
34025
|
"supports_tool_choice": true,
|
|
34013
34026
|
"supports_system_messages": true,
|
|
34014
34027
|
"supports_response_schema": true,
|
|
34015
|
-
"cache_read_input_token_cost":
|
|
34016
|
-
"input_cost_per_token_cache_hit":
|
|
34017
|
-
"supports_reasoning": true
|
|
34028
|
+
"cache_read_input_token_cost": 3e-08,
|
|
34029
|
+
"input_cost_per_token_cache_hit": 3e-08
|
|
34018
34030
|
},
|
|
34019
34031
|
"novita/zai-org/glm-4.7": {
|
|
34020
34032
|
"litellm_provider": "novita",
|
|
34021
34033
|
"mode": "chat",
|
|
34022
|
-
"input_cost_per_token":
|
|
34023
|
-
"output_cost_per_token":
|
|
34034
|
+
"input_cost_per_token": 6e-07,
|
|
34035
|
+
"output_cost_per_token": 2.2e-06,
|
|
34024
34036
|
"max_input_tokens": 204800,
|
|
34025
34037
|
"max_output_tokens": 131072,
|
|
34026
34038
|
"max_tokens": 131072,
|
|
@@ -34029,15 +34041,15 @@
|
|
|
34029
34041
|
"supports_tool_choice": true,
|
|
34030
34042
|
"supports_system_messages": true,
|
|
34031
34043
|
"supports_response_schema": true,
|
|
34032
|
-
"cache_read_input_token_cost":
|
|
34033
|
-
"input_cost_per_token_cache_hit":
|
|
34044
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
34045
|
+
"input_cost_per_token_cache_hit": 1.1e-07,
|
|
34034
34046
|
"supports_reasoning": true
|
|
34035
34047
|
},
|
|
34036
34048
|
"novita/xiaomimimo/mimo-v2-flash": {
|
|
34037
34049
|
"litellm_provider": "novita",
|
|
34038
34050
|
"mode": "chat",
|
|
34039
|
-
"input_cost_per_token":
|
|
34040
|
-
"output_cost_per_token":
|
|
34051
|
+
"input_cost_per_token": 1e-07,
|
|
34052
|
+
"output_cost_per_token": 3e-07,
|
|
34041
34053
|
"max_input_tokens": 262144,
|
|
34042
34054
|
"max_output_tokens": 32000,
|
|
34043
34055
|
"max_tokens": 32000,
|
|
@@ -34046,15 +34058,15 @@
|
|
|
34046
34058
|
"supports_tool_choice": true,
|
|
34047
34059
|
"supports_system_messages": true,
|
|
34048
34060
|
"supports_response_schema": true,
|
|
34049
|
-
"cache_read_input_token_cost":
|
|
34050
|
-
"input_cost_per_token_cache_hit":
|
|
34061
|
+
"cache_read_input_token_cost": 2e-08,
|
|
34062
|
+
"input_cost_per_token_cache_hit": 2e-08,
|
|
34051
34063
|
"supports_reasoning": true
|
|
34052
34064
|
},
|
|
34053
34065
|
"novita/zai-org/autoglm-phone-9b-multilingual": {
|
|
34054
34066
|
"litellm_provider": "novita",
|
|
34055
34067
|
"mode": "chat",
|
|
34056
|
-
"input_cost_per_token":
|
|
34057
|
-
"output_cost_per_token":
|
|
34068
|
+
"input_cost_per_token": 3.5e-08,
|
|
34069
|
+
"output_cost_per_token": 1.38e-07,
|
|
34058
34070
|
"max_input_tokens": 65536,
|
|
34059
34071
|
"max_output_tokens": 65536,
|
|
34060
34072
|
"max_tokens": 65536,
|
|
@@ -34064,8 +34076,8 @@
|
|
|
34064
34076
|
"novita/moonshotai/kimi-k2-thinking": {
|
|
34065
34077
|
"litellm_provider": "novita",
|
|
34066
34078
|
"mode": "chat",
|
|
34067
|
-
"input_cost_per_token":
|
|
34068
|
-
"output_cost_per_token":
|
|
34079
|
+
"input_cost_per_token": 6e-07,
|
|
34080
|
+
"output_cost_per_token": 2.5e-06,
|
|
34069
34081
|
"max_input_tokens": 262144,
|
|
34070
34082
|
"max_output_tokens": 262144,
|
|
34071
34083
|
"max_tokens": 262144,
|
|
@@ -34079,8 +34091,8 @@
|
|
|
34079
34091
|
"novita/minimax/minimax-m2": {
|
|
34080
34092
|
"litellm_provider": "novita",
|
|
34081
34093
|
"mode": "chat",
|
|
34082
|
-
"input_cost_per_token":
|
|
34083
|
-
"output_cost_per_token":
|
|
34094
|
+
"input_cost_per_token": 3e-07,
|
|
34095
|
+
"output_cost_per_token": 1.2e-06,
|
|
34084
34096
|
"max_input_tokens": 204800,
|
|
34085
34097
|
"max_output_tokens": 131072,
|
|
34086
34098
|
"max_tokens": 131072,
|
|
@@ -34088,15 +34100,15 @@
|
|
|
34088
34100
|
"supports_parallel_function_calling": true,
|
|
34089
34101
|
"supports_tool_choice": true,
|
|
34090
34102
|
"supports_system_messages": true,
|
|
34091
|
-
"cache_read_input_token_cost":
|
|
34092
|
-
"input_cost_per_token_cache_hit":
|
|
34103
|
+
"cache_read_input_token_cost": 3e-08,
|
|
34104
|
+
"input_cost_per_token_cache_hit": 3e-08,
|
|
34093
34105
|
"supports_reasoning": true
|
|
34094
34106
|
},
|
|
34095
34107
|
"novita/paddlepaddle/paddleocr-vl": {
|
|
34096
34108
|
"litellm_provider": "novita",
|
|
34097
34109
|
"mode": "chat",
|
|
34098
|
-
"input_cost_per_token":
|
|
34099
|
-
"output_cost_per_token":
|
|
34110
|
+
"input_cost_per_token": 2e-08,
|
|
34111
|
+
"output_cost_per_token": 2e-08,
|
|
34100
34112
|
"max_input_tokens": 16384,
|
|
34101
34113
|
"max_output_tokens": 16384,
|
|
34102
34114
|
"max_tokens": 16384,
|
|
@@ -34106,8 +34118,8 @@
|
|
|
34106
34118
|
"novita/deepseek/deepseek-v3.2-exp": {
|
|
34107
34119
|
"litellm_provider": "novita",
|
|
34108
34120
|
"mode": "chat",
|
|
34109
|
-
"input_cost_per_token":
|
|
34110
|
-
"output_cost_per_token":
|
|
34121
|
+
"input_cost_per_token": 2.7e-07,
|
|
34122
|
+
"output_cost_per_token": 4.1e-07,
|
|
34111
34123
|
"max_input_tokens": 163840,
|
|
34112
34124
|
"max_output_tokens": 65536,
|
|
34113
34125
|
"max_tokens": 65536,
|
|
@@ -34121,8 +34133,8 @@
|
|
|
34121
34133
|
"novita/qwen/qwen3-vl-235b-a22b-thinking": {
|
|
34122
34134
|
"litellm_provider": "novita",
|
|
34123
34135
|
"mode": "chat",
|
|
34124
|
-
"input_cost_per_token":
|
|
34125
|
-
"output_cost_per_token":
|
|
34136
|
+
"input_cost_per_token": 9.8e-07,
|
|
34137
|
+
"output_cost_per_token": 3.95e-06,
|
|
34126
34138
|
"max_input_tokens": 131072,
|
|
34127
34139
|
"max_output_tokens": 32768,
|
|
34128
34140
|
"max_tokens": 32768,
|
|
@@ -34133,8 +34145,8 @@
|
|
|
34133
34145
|
"novita/zai-org/glm-4.6v": {
|
|
34134
34146
|
"litellm_provider": "novita",
|
|
34135
34147
|
"mode": "chat",
|
|
34136
|
-
"input_cost_per_token":
|
|
34137
|
-
"output_cost_per_token":
|
|
34148
|
+
"input_cost_per_token": 3e-07,
|
|
34149
|
+
"output_cost_per_token": 9e-07,
|
|
34138
34150
|
"max_input_tokens": 131072,
|
|
34139
34151
|
"max_output_tokens": 32768,
|
|
34140
34152
|
"max_tokens": 32768,
|
|
@@ -34144,15 +34156,15 @@
|
|
|
34144
34156
|
"supports_vision": true,
|
|
34145
34157
|
"supports_system_messages": true,
|
|
34146
34158
|
"supports_response_schema": true,
|
|
34147
|
-
"cache_read_input_token_cost":
|
|
34148
|
-
"input_cost_per_token_cache_hit":
|
|
34159
|
+
"cache_read_input_token_cost": 5.5e-08,
|
|
34160
|
+
"input_cost_per_token_cache_hit": 5.5e-08,
|
|
34149
34161
|
"supports_reasoning": true
|
|
34150
34162
|
},
|
|
34151
34163
|
"novita/zai-org/glm-4.6": {
|
|
34152
34164
|
"litellm_provider": "novita",
|
|
34153
34165
|
"mode": "chat",
|
|
34154
|
-
"input_cost_per_token":
|
|
34155
|
-
"output_cost_per_token":
|
|
34166
|
+
"input_cost_per_token": 5.5e-07,
|
|
34167
|
+
"output_cost_per_token": 2.2e-06,
|
|
34156
34168
|
"max_input_tokens": 204800,
|
|
34157
34169
|
"max_output_tokens": 131072,
|
|
34158
34170
|
"max_tokens": 131072,
|
|
@@ -34161,15 +34173,15 @@
|
|
|
34161
34173
|
"supports_tool_choice": true,
|
|
34162
34174
|
"supports_system_messages": true,
|
|
34163
34175
|
"supports_response_schema": true,
|
|
34164
|
-
"cache_read_input_token_cost":
|
|
34165
|
-
"input_cost_per_token_cache_hit":
|
|
34176
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
34177
|
+
"input_cost_per_token_cache_hit": 1.1e-07,
|
|
34166
34178
|
"supports_reasoning": true
|
|
34167
34179
|
},
|
|
34168
34180
|
"novita/qwen/qwen3-next-80b-a3b-instruct": {
|
|
34169
34181
|
"litellm_provider": "novita",
|
|
34170
34182
|
"mode": "chat",
|
|
34171
|
-
"input_cost_per_token":
|
|
34172
|
-
"output_cost_per_token":
|
|
34183
|
+
"input_cost_per_token": 1.5e-07,
|
|
34184
|
+
"output_cost_per_token": 1.5e-06,
|
|
34173
34185
|
"max_input_tokens": 131072,
|
|
34174
34186
|
"max_output_tokens": 32768,
|
|
34175
34187
|
"max_tokens": 32768,
|
|
@@ -34182,8 +34194,8 @@
|
|
|
34182
34194
|
"novita/qwen/qwen3-next-80b-a3b-thinking": {
|
|
34183
34195
|
"litellm_provider": "novita",
|
|
34184
34196
|
"mode": "chat",
|
|
34185
|
-
"input_cost_per_token":
|
|
34186
|
-
"output_cost_per_token":
|
|
34197
|
+
"input_cost_per_token": 1.5e-07,
|
|
34198
|
+
"output_cost_per_token": 1.5e-06,
|
|
34187
34199
|
"max_input_tokens": 131072,
|
|
34188
34200
|
"max_output_tokens": 32768,
|
|
34189
34201
|
"max_tokens": 32768,
|
|
@@ -34197,8 +34209,8 @@
|
|
|
34197
34209
|
"novita/deepseek/deepseek-ocr": {
|
|
34198
34210
|
"litellm_provider": "novita",
|
|
34199
34211
|
"mode": "chat",
|
|
34200
|
-
"input_cost_per_token":
|
|
34201
|
-
"output_cost_per_token":
|
|
34212
|
+
"input_cost_per_token": 3e-08,
|
|
34213
|
+
"output_cost_per_token": 3e-08,
|
|
34202
34214
|
"max_input_tokens": 8192,
|
|
34203
34215
|
"max_output_tokens": 8192,
|
|
34204
34216
|
"max_tokens": 8192,
|
|
@@ -34210,8 +34222,8 @@
|
|
|
34210
34222
|
"novita/deepseek/deepseek-v3.1-terminus": {
|
|
34211
34223
|
"litellm_provider": "novita",
|
|
34212
34224
|
"mode": "chat",
|
|
34213
|
-
"input_cost_per_token":
|
|
34214
|
-
"output_cost_per_token":
|
|
34225
|
+
"input_cost_per_token": 2.7e-07,
|
|
34226
|
+
"output_cost_per_token": 1e-06,
|
|
34215
34227
|
"max_input_tokens": 131072,
|
|
34216
34228
|
"max_output_tokens": 32768,
|
|
34217
34229
|
"max_tokens": 32768,
|
|
@@ -34220,15 +34232,15 @@
|
|
|
34220
34232
|
"supports_tool_choice": true,
|
|
34221
34233
|
"supports_system_messages": true,
|
|
34222
34234
|
"supports_response_schema": true,
|
|
34223
|
-
"cache_read_input_token_cost":
|
|
34224
|
-
"input_cost_per_token_cache_hit":
|
|
34235
|
+
"cache_read_input_token_cost": 1.35e-07,
|
|
34236
|
+
"input_cost_per_token_cache_hit": 1.35e-07,
|
|
34225
34237
|
"supports_reasoning": true
|
|
34226
34238
|
},
|
|
34227
34239
|
"novita/qwen/qwen3-vl-235b-a22b-instruct": {
|
|
34228
34240
|
"litellm_provider": "novita",
|
|
34229
34241
|
"mode": "chat",
|
|
34230
|
-
"input_cost_per_token":
|
|
34231
|
-
"output_cost_per_token":
|
|
34242
|
+
"input_cost_per_token": 3e-07,
|
|
34243
|
+
"output_cost_per_token": 1.5e-06,
|
|
34232
34244
|
"max_input_tokens": 131072,
|
|
34233
34245
|
"max_output_tokens": 32768,
|
|
34234
34246
|
"max_tokens": 32768,
|
|
@@ -34242,8 +34254,8 @@
|
|
|
34242
34254
|
"novita/qwen/qwen3-max": {
|
|
34243
34255
|
"litellm_provider": "novita",
|
|
34244
34256
|
"mode": "chat",
|
|
34245
|
-
"input_cost_per_token":
|
|
34246
|
-
"output_cost_per_token":
|
|
34257
|
+
"input_cost_per_token": 2.11e-06,
|
|
34258
|
+
"output_cost_per_token": 8.45e-06,
|
|
34247
34259
|
"max_input_tokens": 262144,
|
|
34248
34260
|
"max_output_tokens": 65536,
|
|
34249
34261
|
"max_tokens": 65536,
|
|
@@ -34256,8 +34268,8 @@
|
|
|
34256
34268
|
"novita/skywork/r1v4-lite": {
|
|
34257
34269
|
"litellm_provider": "novita",
|
|
34258
34270
|
"mode": "chat",
|
|
34259
|
-
"input_cost_per_token":
|
|
34260
|
-
"output_cost_per_token":
|
|
34271
|
+
"input_cost_per_token": 2e-07,
|
|
34272
|
+
"output_cost_per_token": 6e-07,
|
|
34261
34273
|
"max_input_tokens": 262144,
|
|
34262
34274
|
"max_output_tokens": 65536,
|
|
34263
34275
|
"max_tokens": 65536,
|
|
@@ -34269,8 +34281,8 @@
|
|
|
34269
34281
|
"novita/deepseek/deepseek-v3.1": {
|
|
34270
34282
|
"litellm_provider": "novita",
|
|
34271
34283
|
"mode": "chat",
|
|
34272
|
-
"input_cost_per_token":
|
|
34273
|
-
"output_cost_per_token":
|
|
34284
|
+
"input_cost_per_token": 2.7e-07,
|
|
34285
|
+
"output_cost_per_token": 1e-06,
|
|
34274
34286
|
"max_input_tokens": 131072,
|
|
34275
34287
|
"max_output_tokens": 32768,
|
|
34276
34288
|
"max_tokens": 32768,
|
|
@@ -34279,15 +34291,15 @@
|
|
|
34279
34291
|
"supports_tool_choice": true,
|
|
34280
34292
|
"supports_system_messages": true,
|
|
34281
34293
|
"supports_response_schema": true,
|
|
34282
|
-
"cache_read_input_token_cost":
|
|
34283
|
-
"input_cost_per_token_cache_hit":
|
|
34294
|
+
"cache_read_input_token_cost": 1.35e-07,
|
|
34295
|
+
"input_cost_per_token_cache_hit": 1.35e-07,
|
|
34284
34296
|
"supports_reasoning": true
|
|
34285
34297
|
},
|
|
34286
34298
|
"novita/moonshotai/kimi-k2-0905": {
|
|
34287
34299
|
"litellm_provider": "novita",
|
|
34288
34300
|
"mode": "chat",
|
|
34289
|
-
"input_cost_per_token":
|
|
34290
|
-
"output_cost_per_token":
|
|
34301
|
+
"input_cost_per_token": 6e-07,
|
|
34302
|
+
"output_cost_per_token": 2.5e-06,
|
|
34291
34303
|
"max_input_tokens": 262144,
|
|
34292
34304
|
"max_output_tokens": 262144,
|
|
34293
34305
|
"max_tokens": 262144,
|
|
@@ -34300,8 +34312,8 @@
|
|
|
34300
34312
|
"novita/qwen/qwen3-coder-480b-a35b-instruct": {
|
|
34301
34313
|
"litellm_provider": "novita",
|
|
34302
34314
|
"mode": "chat",
|
|
34303
|
-
"input_cost_per_token":
|
|
34304
|
-
"output_cost_per_token":
|
|
34315
|
+
"input_cost_per_token": 3e-07,
|
|
34316
|
+
"output_cost_per_token": 1.3e-06,
|
|
34305
34317
|
"max_input_tokens": 262144,
|
|
34306
34318
|
"max_output_tokens": 65536,
|
|
34307
34319
|
"max_tokens": 65536,
|
|
@@ -34314,8 +34326,8 @@
|
|
|
34314
34326
|
"novita/qwen/qwen3-coder-30b-a3b-instruct": {
|
|
34315
34327
|
"litellm_provider": "novita",
|
|
34316
34328
|
"mode": "chat",
|
|
34317
|
-
"input_cost_per_token":
|
|
34318
|
-
"output_cost_per_token":
|
|
34329
|
+
"input_cost_per_token": 7e-08,
|
|
34330
|
+
"output_cost_per_token": 2.7e-07,
|
|
34319
34331
|
"max_input_tokens": 160000,
|
|
34320
34332
|
"max_output_tokens": 32768,
|
|
34321
34333
|
"max_tokens": 32768,
|
|
@@ -34328,8 +34340,8 @@
|
|
|
34328
34340
|
"novita/openai/gpt-oss-120b": {
|
|
34329
34341
|
"litellm_provider": "novita",
|
|
34330
34342
|
"mode": "chat",
|
|
34331
|
-
"input_cost_per_token":
|
|
34332
|
-
"output_cost_per_token":
|
|
34343
|
+
"input_cost_per_token": 5e-08,
|
|
34344
|
+
"output_cost_per_token": 2.5e-07,
|
|
34333
34345
|
"max_input_tokens": 131072,
|
|
34334
34346
|
"max_output_tokens": 32768,
|
|
34335
34347
|
"max_tokens": 32768,
|
|
@@ -34344,8 +34356,8 @@
|
|
|
34344
34356
|
"novita/moonshotai/kimi-k2-instruct": {
|
|
34345
34357
|
"litellm_provider": "novita",
|
|
34346
34358
|
"mode": "chat",
|
|
34347
|
-
"input_cost_per_token":
|
|
34348
|
-
"output_cost_per_token":
|
|
34359
|
+
"input_cost_per_token": 5.7e-07,
|
|
34360
|
+
"output_cost_per_token": 2.3e-06,
|
|
34349
34361
|
"max_input_tokens": 131072,
|
|
34350
34362
|
"max_output_tokens": 131072,
|
|
34351
34363
|
"max_tokens": 131072,
|
|
@@ -34358,8 +34370,8 @@
|
|
|
34358
34370
|
"novita/deepseek/deepseek-v3-0324": {
|
|
34359
34371
|
"litellm_provider": "novita",
|
|
34360
34372
|
"mode": "chat",
|
|
34361
|
-
"input_cost_per_token":
|
|
34362
|
-
"output_cost_per_token":
|
|
34373
|
+
"input_cost_per_token": 2.7e-07,
|
|
34374
|
+
"output_cost_per_token": 1.12e-06,
|
|
34363
34375
|
"max_input_tokens": 163840,
|
|
34364
34376
|
"max_output_tokens": 163840,
|
|
34365
34377
|
"max_tokens": 163840,
|
|
@@ -34368,14 +34380,14 @@
|
|
|
34368
34380
|
"supports_tool_choice": true,
|
|
34369
34381
|
"supports_system_messages": true,
|
|
34370
34382
|
"supports_response_schema": true,
|
|
34371
|
-
"cache_read_input_token_cost":
|
|
34372
|
-
"input_cost_per_token_cache_hit":
|
|
34383
|
+
"cache_read_input_token_cost": 1.35e-07,
|
|
34384
|
+
"input_cost_per_token_cache_hit": 1.35e-07
|
|
34373
34385
|
},
|
|
34374
34386
|
"novita/zai-org/glm-4.5": {
|
|
34375
34387
|
"litellm_provider": "novita",
|
|
34376
34388
|
"mode": "chat",
|
|
34377
|
-
"input_cost_per_token":
|
|
34378
|
-
"output_cost_per_token":
|
|
34389
|
+
"input_cost_per_token": 6e-07,
|
|
34390
|
+
"output_cost_per_token": 2.2e-06,
|
|
34379
34391
|
"max_input_tokens": 131072,
|
|
34380
34392
|
"max_output_tokens": 98304,
|
|
34381
34393
|
"max_tokens": 98304,
|
|
@@ -34383,15 +34395,15 @@
|
|
|
34383
34395
|
"supports_parallel_function_calling": true,
|
|
34384
34396
|
"supports_tool_choice": true,
|
|
34385
34397
|
"supports_system_messages": true,
|
|
34386
|
-
"cache_read_input_token_cost":
|
|
34387
|
-
"input_cost_per_token_cache_hit":
|
|
34398
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
34399
|
+
"input_cost_per_token_cache_hit": 1.1e-07,
|
|
34388
34400
|
"supports_reasoning": true
|
|
34389
34401
|
},
|
|
34390
34402
|
"novita/qwen/qwen3-235b-a22b-thinking-2507": {
|
|
34391
34403
|
"litellm_provider": "novita",
|
|
34392
34404
|
"mode": "chat",
|
|
34393
|
-
"input_cost_per_token":
|
|
34394
|
-
"output_cost_per_token":
|
|
34405
|
+
"input_cost_per_token": 3e-07,
|
|
34406
|
+
"output_cost_per_token": 3e-06,
|
|
34395
34407
|
"max_input_tokens": 131072,
|
|
34396
34408
|
"max_output_tokens": 32768,
|
|
34397
34409
|
"max_tokens": 32768,
|
|
@@ -34404,8 +34416,8 @@
|
|
|
34404
34416
|
"novita/meta-llama/llama-3.1-8b-instruct": {
|
|
34405
34417
|
"litellm_provider": "novita",
|
|
34406
34418
|
"mode": "chat",
|
|
34407
|
-
"input_cost_per_token":
|
|
34408
|
-
"output_cost_per_token":
|
|
34419
|
+
"input_cost_per_token": 2e-08,
|
|
34420
|
+
"output_cost_per_token": 5e-08,
|
|
34409
34421
|
"max_input_tokens": 16384,
|
|
34410
34422
|
"max_output_tokens": 16384,
|
|
34411
34423
|
"max_tokens": 16384,
|
|
@@ -34414,8 +34426,8 @@
|
|
|
34414
34426
|
"novita/google/gemma-3-12b-it": {
|
|
34415
34427
|
"litellm_provider": "novita",
|
|
34416
34428
|
"mode": "chat",
|
|
34417
|
-
"input_cost_per_token":
|
|
34418
|
-
"output_cost_per_token":
|
|
34429
|
+
"input_cost_per_token": 5e-08,
|
|
34430
|
+
"output_cost_per_token": 1e-07,
|
|
34419
34431
|
"max_input_tokens": 131072,
|
|
34420
34432
|
"max_output_tokens": 8192,
|
|
34421
34433
|
"max_tokens": 8192,
|
|
@@ -34427,8 +34439,8 @@
|
|
|
34427
34439
|
"novita/zai-org/glm-4.5v": {
|
|
34428
34440
|
"litellm_provider": "novita",
|
|
34429
34441
|
"mode": "chat",
|
|
34430
|
-
"input_cost_per_token":
|
|
34431
|
-
"output_cost_per_token":
|
|
34442
|
+
"input_cost_per_token": 6e-07,
|
|
34443
|
+
"output_cost_per_token": 1.8e-06,
|
|
34432
34444
|
"max_input_tokens": 65536,
|
|
34433
34445
|
"max_output_tokens": 16384,
|
|
34434
34446
|
"max_tokens": 16384,
|
|
@@ -34438,15 +34450,15 @@
|
|
|
34438
34450
|
"supports_vision": true,
|
|
34439
34451
|
"supports_system_messages": true,
|
|
34440
34452
|
"supports_response_schema": true,
|
|
34441
|
-
"cache_read_input_token_cost":
|
|
34442
|
-
"input_cost_per_token_cache_hit":
|
|
34453
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
34454
|
+
"input_cost_per_token_cache_hit": 1.1e-07,
|
|
34443
34455
|
"supports_reasoning": true
|
|
34444
34456
|
},
|
|
34445
34457
|
"novita/openai/gpt-oss-20b": {
|
|
34446
34458
|
"litellm_provider": "novita",
|
|
34447
34459
|
"mode": "chat",
|
|
34448
|
-
"input_cost_per_token":
|
|
34449
|
-
"output_cost_per_token":
|
|
34460
|
+
"input_cost_per_token": 4e-08,
|
|
34461
|
+
"output_cost_per_token": 1.5e-07,
|
|
34450
34462
|
"max_input_tokens": 131072,
|
|
34451
34463
|
"max_output_tokens": 32768,
|
|
34452
34464
|
"max_tokens": 32768,
|
|
@@ -34459,8 +34471,8 @@
|
|
|
34459
34471
|
"novita/qwen/qwen3-235b-a22b-instruct-2507": {
|
|
34460
34472
|
"litellm_provider": "novita",
|
|
34461
34473
|
"mode": "chat",
|
|
34462
|
-
"input_cost_per_token":
|
|
34463
|
-
"output_cost_per_token":
|
|
34474
|
+
"input_cost_per_token": 9e-08,
|
|
34475
|
+
"output_cost_per_token": 5.8e-07,
|
|
34464
34476
|
"max_input_tokens": 131072,
|
|
34465
34477
|
"max_output_tokens": 16384,
|
|
34466
34478
|
"max_tokens": 16384,
|
|
@@ -34473,8 +34485,8 @@
|
|
|
34473
34485
|
"novita/deepseek/deepseek-r1-distill-qwen-14b": {
|
|
34474
34486
|
"litellm_provider": "novita",
|
|
34475
34487
|
"mode": "chat",
|
|
34476
|
-
"input_cost_per_token":
|
|
34477
|
-
"output_cost_per_token":
|
|
34488
|
+
"input_cost_per_token": 1.5e-07,
|
|
34489
|
+
"output_cost_per_token": 1.5e-07,
|
|
34478
34490
|
"max_input_tokens": 32768,
|
|
34479
34491
|
"max_output_tokens": 16384,
|
|
34480
34492
|
"max_tokens": 16384,
|
|
@@ -34486,8 +34498,8 @@
|
|
|
34486
34498
|
"novita/meta-llama/llama-3.3-70b-instruct": {
|
|
34487
34499
|
"litellm_provider": "novita",
|
|
34488
34500
|
"mode": "chat",
|
|
34489
|
-
"input_cost_per_token":
|
|
34490
|
-
"output_cost_per_token":
|
|
34501
|
+
"input_cost_per_token": 1.35e-07,
|
|
34502
|
+
"output_cost_per_token": 4e-07,
|
|
34491
34503
|
"max_input_tokens": 131072,
|
|
34492
34504
|
"max_output_tokens": 120000,
|
|
34493
34505
|
"max_tokens": 120000,
|
|
@@ -34499,8 +34511,8 @@
|
|
|
34499
34511
|
"novita/qwen/qwen-2.5-72b-instruct": {
|
|
34500
34512
|
"litellm_provider": "novita",
|
|
34501
34513
|
"mode": "chat",
|
|
34502
|
-
"input_cost_per_token":
|
|
34503
|
-
"output_cost_per_token":
|
|
34514
|
+
"input_cost_per_token": 3.8e-07,
|
|
34515
|
+
"output_cost_per_token": 4e-07,
|
|
34504
34516
|
"max_input_tokens": 32000,
|
|
34505
34517
|
"max_output_tokens": 8192,
|
|
34506
34518
|
"max_tokens": 8192,
|
|
@@ -34513,8 +34525,8 @@
|
|
|
34513
34525
|
"novita/mistralai/mistral-nemo": {
|
|
34514
34526
|
"litellm_provider": "novita",
|
|
34515
34527
|
"mode": "chat",
|
|
34516
|
-
"input_cost_per_token":
|
|
34517
|
-
"output_cost_per_token":
|
|
34528
|
+
"input_cost_per_token": 4e-08,
|
|
34529
|
+
"output_cost_per_token": 1.7e-07,
|
|
34518
34530
|
"max_input_tokens": 60288,
|
|
34519
34531
|
"max_output_tokens": 16000,
|
|
34520
34532
|
"max_tokens": 16000,
|
|
@@ -34525,8 +34537,8 @@
|
|
|
34525
34537
|
"novita/minimaxai/minimax-m1-80k": {
|
|
34526
34538
|
"litellm_provider": "novita",
|
|
34527
34539
|
"mode": "chat",
|
|
34528
|
-
"input_cost_per_token":
|
|
34529
|
-
"output_cost_per_token":
|
|
34540
|
+
"input_cost_per_token": 5.5e-07,
|
|
34541
|
+
"output_cost_per_token": 2.2e-06,
|
|
34530
34542
|
"max_input_tokens": 1000000,
|
|
34531
34543
|
"max_output_tokens": 40000,
|
|
34532
34544
|
"max_tokens": 40000,
|
|
@@ -34539,8 +34551,8 @@
|
|
|
34539
34551
|
"novita/deepseek/deepseek-r1-0528": {
|
|
34540
34552
|
"litellm_provider": "novita",
|
|
34541
34553
|
"mode": "chat",
|
|
34542
|
-
"input_cost_per_token":
|
|
34543
|
-
"output_cost_per_token":
|
|
34554
|
+
"input_cost_per_token": 7e-07,
|
|
34555
|
+
"output_cost_per_token": 2.5e-06,
|
|
34544
34556
|
"max_input_tokens": 163840,
|
|
34545
34557
|
"max_output_tokens": 32768,
|
|
34546
34558
|
"max_tokens": 32768,
|
|
@@ -34549,15 +34561,15 @@
|
|
|
34549
34561
|
"supports_tool_choice": true,
|
|
34550
34562
|
"supports_system_messages": true,
|
|
34551
34563
|
"supports_response_schema": true,
|
|
34552
|
-
"cache_read_input_token_cost":
|
|
34553
|
-
"input_cost_per_token_cache_hit":
|
|
34564
|
+
"cache_read_input_token_cost": 3.5e-07,
|
|
34565
|
+
"input_cost_per_token_cache_hit": 3.5e-07,
|
|
34554
34566
|
"supports_reasoning": true
|
|
34555
34567
|
},
|
|
34556
34568
|
"novita/deepseek/deepseek-r1-distill-qwen-32b": {
|
|
34557
34569
|
"litellm_provider": "novita",
|
|
34558
34570
|
"mode": "chat",
|
|
34559
|
-
"input_cost_per_token":
|
|
34560
|
-
"output_cost_per_token":
|
|
34571
|
+
"input_cost_per_token": 3e-07,
|
|
34572
|
+
"output_cost_per_token": 3e-07,
|
|
34561
34573
|
"max_input_tokens": 64000,
|
|
34562
34574
|
"max_output_tokens": 32000,
|
|
34563
34575
|
"max_tokens": 32000,
|
|
@@ -34569,8 +34581,8 @@
|
|
|
34569
34581
|
"novita/meta-llama/llama-3-8b-instruct": {
|
|
34570
34582
|
"litellm_provider": "novita",
|
|
34571
34583
|
"mode": "chat",
|
|
34572
|
-
"input_cost_per_token":
|
|
34573
|
-
"output_cost_per_token":
|
|
34584
|
+
"input_cost_per_token": 4e-08,
|
|
34585
|
+
"output_cost_per_token": 4e-08,
|
|
34574
34586
|
"max_input_tokens": 8192,
|
|
34575
34587
|
"max_output_tokens": 8192,
|
|
34576
34588
|
"max_tokens": 8192,
|
|
@@ -34579,8 +34591,8 @@
|
|
|
34579
34591
|
"novita/microsoft/wizardlm-2-8x22b": {
|
|
34580
34592
|
"litellm_provider": "novita",
|
|
34581
34593
|
"mode": "chat",
|
|
34582
|
-
"input_cost_per_token":
|
|
34583
|
-
"output_cost_per_token":
|
|
34594
|
+
"input_cost_per_token": 6.2e-07,
|
|
34595
|
+
"output_cost_per_token": 6.2e-07,
|
|
34584
34596
|
"max_input_tokens": 65535,
|
|
34585
34597
|
"max_output_tokens": 8000,
|
|
34586
34598
|
"max_tokens": 8000,
|
|
@@ -34589,8 +34601,8 @@
|
|
|
34589
34601
|
"novita/deepseek/deepseek-r1-0528-qwen3-8b": {
|
|
34590
34602
|
"litellm_provider": "novita",
|
|
34591
34603
|
"mode": "chat",
|
|
34592
|
-
"input_cost_per_token":
|
|
34593
|
-
"output_cost_per_token":
|
|
34604
|
+
"input_cost_per_token": 6e-08,
|
|
34605
|
+
"output_cost_per_token": 9e-08,
|
|
34594
34606
|
"max_input_tokens": 128000,
|
|
34595
34607
|
"max_output_tokens": 32000,
|
|
34596
34608
|
"max_tokens": 32000,
|
|
@@ -34600,8 +34612,8 @@
|
|
|
34600
34612
|
"novita/deepseek/deepseek-r1-distill-llama-70b": {
|
|
34601
34613
|
"litellm_provider": "novita",
|
|
34602
34614
|
"mode": "chat",
|
|
34603
|
-
"input_cost_per_token":
|
|
34604
|
-
"output_cost_per_token":
|
|
34615
|
+
"input_cost_per_token": 8e-07,
|
|
34616
|
+
"output_cost_per_token": 8e-07,
|
|
34605
34617
|
"max_input_tokens": 8192,
|
|
34606
34618
|
"max_output_tokens": 8192,
|
|
34607
34619
|
"max_tokens": 8192,
|
|
@@ -34613,8 +34625,8 @@
|
|
|
34613
34625
|
"novita/meta-llama/llama-3-70b-instruct": {
|
|
34614
34626
|
"litellm_provider": "novita",
|
|
34615
34627
|
"mode": "chat",
|
|
34616
|
-
"input_cost_per_token":
|
|
34617
|
-
"output_cost_per_token":
|
|
34628
|
+
"input_cost_per_token": 5.1e-07,
|
|
34629
|
+
"output_cost_per_token": 7.4e-07,
|
|
34618
34630
|
"max_input_tokens": 8192,
|
|
34619
34631
|
"max_output_tokens": 8000,
|
|
34620
34632
|
"max_tokens": 8000,
|
|
@@ -34625,8 +34637,8 @@
|
|
|
34625
34637
|
"novita/qwen/qwen3-235b-a22b-fp8": {
|
|
34626
34638
|
"litellm_provider": "novita",
|
|
34627
34639
|
"mode": "chat",
|
|
34628
|
-
"input_cost_per_token":
|
|
34629
|
-
"output_cost_per_token":
|
|
34640
|
+
"input_cost_per_token": 2e-07,
|
|
34641
|
+
"output_cost_per_token": 8e-07,
|
|
34630
34642
|
"max_input_tokens": 40960,
|
|
34631
34643
|
"max_output_tokens": 20000,
|
|
34632
34644
|
"max_tokens": 20000,
|
|
@@ -34636,8 +34648,8 @@
|
|
|
34636
34648
|
"novita/meta-llama/llama-4-maverick-17b-128e-instruct-fp8": {
|
|
34637
34649
|
"litellm_provider": "novita",
|
|
34638
34650
|
"mode": "chat",
|
|
34639
|
-
"input_cost_per_token":
|
|
34640
|
-
"output_cost_per_token":
|
|
34651
|
+
"input_cost_per_token": 2.7e-07,
|
|
34652
|
+
"output_cost_per_token": 8.5e-07,
|
|
34641
34653
|
"max_input_tokens": 1048576,
|
|
34642
34654
|
"max_output_tokens": 8192,
|
|
34643
34655
|
"max_tokens": 8192,
|
|
@@ -34647,8 +34659,8 @@
|
|
|
34647
34659
|
"novita/meta-llama/llama-4-scout-17b-16e-instruct": {
|
|
34648
34660
|
"litellm_provider": "novita",
|
|
34649
34661
|
"mode": "chat",
|
|
34650
|
-
"input_cost_per_token":
|
|
34651
|
-
"output_cost_per_token":
|
|
34662
|
+
"input_cost_per_token": 1.8e-07,
|
|
34663
|
+
"output_cost_per_token": 5.9e-07,
|
|
34652
34664
|
"max_input_tokens": 131072,
|
|
34653
34665
|
"max_output_tokens": 131072,
|
|
34654
34666
|
"max_tokens": 131072,
|
|
@@ -34658,8 +34670,8 @@
|
|
|
34658
34670
|
"novita/nousresearch/hermes-2-pro-llama-3-8b": {
|
|
34659
34671
|
"litellm_provider": "novita",
|
|
34660
34672
|
"mode": "chat",
|
|
34661
|
-
"input_cost_per_token":
|
|
34662
|
-
"output_cost_per_token":
|
|
34673
|
+
"input_cost_per_token": 1.4e-07,
|
|
34674
|
+
"output_cost_per_token": 1.4e-07,
|
|
34663
34675
|
"max_input_tokens": 8192,
|
|
34664
34676
|
"max_output_tokens": 8192,
|
|
34665
34677
|
"max_tokens": 8192,
|
|
@@ -34670,8 +34682,8 @@
|
|
|
34670
34682
|
"novita/qwen/qwen2.5-vl-72b-instruct": {
|
|
34671
34683
|
"litellm_provider": "novita",
|
|
34672
34684
|
"mode": "chat",
|
|
34673
|
-
"input_cost_per_token":
|
|
34674
|
-
"output_cost_per_token":
|
|
34685
|
+
"input_cost_per_token": 8e-07,
|
|
34686
|
+
"output_cost_per_token": 8e-07,
|
|
34675
34687
|
"max_input_tokens": 32768,
|
|
34676
34688
|
"max_output_tokens": 32768,
|
|
34677
34689
|
"max_tokens": 32768,
|
|
@@ -34681,8 +34693,8 @@
|
|
|
34681
34693
|
"novita/sao10k/l3-70b-euryale-v2.1": {
|
|
34682
34694
|
"litellm_provider": "novita",
|
|
34683
34695
|
"mode": "chat",
|
|
34684
|
-
"input_cost_per_token":
|
|
34685
|
-
"output_cost_per_token":
|
|
34696
|
+
"input_cost_per_token": 1.48e-06,
|
|
34697
|
+
"output_cost_per_token": 1.48e-06,
|
|
34686
34698
|
"max_input_tokens": 8192,
|
|
34687
34699
|
"max_output_tokens": 8192,
|
|
34688
34700
|
"max_tokens": 8192,
|
|
@@ -34694,8 +34706,8 @@
|
|
|
34694
34706
|
"novita/baidu/ernie-4.5-21B-a3b-thinking": {
|
|
34695
34707
|
"litellm_provider": "novita",
|
|
34696
34708
|
"mode": "chat",
|
|
34697
|
-
"input_cost_per_token":
|
|
34698
|
-
"output_cost_per_token":
|
|
34709
|
+
"input_cost_per_token": 7e-08,
|
|
34710
|
+
"output_cost_per_token": 2.8e-07,
|
|
34699
34711
|
"max_input_tokens": 131072,
|
|
34700
34712
|
"max_output_tokens": 65536,
|
|
34701
34713
|
"max_tokens": 65536,
|
|
@@ -34705,8 +34717,8 @@
|
|
|
34705
34717
|
"novita/sao10k/l3-8b-lunaris": {
|
|
34706
34718
|
"litellm_provider": "novita",
|
|
34707
34719
|
"mode": "chat",
|
|
34708
|
-
"input_cost_per_token":
|
|
34709
|
-
"output_cost_per_token":
|
|
34720
|
+
"input_cost_per_token": 5e-08,
|
|
34721
|
+
"output_cost_per_token": 5e-08,
|
|
34710
34722
|
"max_input_tokens": 8192,
|
|
34711
34723
|
"max_output_tokens": 8192,
|
|
34712
34724
|
"max_tokens": 8192,
|
|
@@ -34717,15 +34729,12 @@
|
|
|
34717
34729
|
"novita/baichuan/baichuan-m2-32b": {
|
|
34718
34730
|
"litellm_provider": "novita",
|
|
34719
34731
|
"mode": "chat",
|
|
34720
|
-
"input_cost_per_token":
|
|
34721
|
-
"output_cost_per_token":
|
|
34732
|
+
"input_cost_per_token": 7e-08,
|
|
34733
|
+
"output_cost_per_token": 7e-08,
|
|
34722
34734
|
"max_input_tokens": 131072,
|
|
34723
34735
|
"max_output_tokens": 131072,
|
|
34724
34736
|
"max_tokens": 131072,
|
|
34725
|
-
"
|
|
34726
|
-
"supports_system_messages": true,
|
|
34727
|
-
"supports_response_schema": true,
|
|
34728
|
-
"supports_reasoning": true
|
|
34737
|
+
"supports_system_messages": true
|
|
34729
34738
|
},
|
|
34730
34739
|
"novita/thudm/glm-4.1v-9b-thinking": {
|
|
34731
34740
|
"litellm_provider": "novita",
|
|
@@ -34742,8 +34751,8 @@
|
|
|
34742
34751
|
"novita/baidu/ernie-4.5-vl-424b-a47b": {
|
|
34743
34752
|
"litellm_provider": "novita",
|
|
34744
34753
|
"mode": "chat",
|
|
34745
|
-
"input_cost_per_token":
|
|
34746
|
-
"output_cost_per_token":
|
|
34754
|
+
"input_cost_per_token": 4.2e-07,
|
|
34755
|
+
"output_cost_per_token": 1.25e-06,
|
|
34747
34756
|
"max_input_tokens": 123000,
|
|
34748
34757
|
"max_output_tokens": 16000,
|
|
34749
34758
|
"max_tokens": 16000,
|
|
@@ -34754,8 +34763,8 @@
|
|
|
34754
34763
|
"novita/baidu/ernie-4.5-300b-a47b-paddle": {
|
|
34755
34764
|
"litellm_provider": "novita",
|
|
34756
34765
|
"mode": "chat",
|
|
34757
|
-
"input_cost_per_token":
|
|
34758
|
-
"output_cost_per_token":
|
|
34766
|
+
"input_cost_per_token": 2.8e-07,
|
|
34767
|
+
"output_cost_per_token": 1.1e-06,
|
|
34759
34768
|
"max_input_tokens": 123000,
|
|
34760
34769
|
"max_output_tokens": 12000,
|
|
34761
34770
|
"max_tokens": 12000,
|
|
@@ -34766,8 +34775,8 @@
|
|
|
34766
34775
|
"novita/deepseek/deepseek-prover-v2-671b": {
|
|
34767
34776
|
"litellm_provider": "novita",
|
|
34768
34777
|
"mode": "chat",
|
|
34769
|
-
"input_cost_per_token":
|
|
34770
|
-
"output_cost_per_token":
|
|
34778
|
+
"input_cost_per_token": 7e-07,
|
|
34779
|
+
"output_cost_per_token": 2.5e-06,
|
|
34771
34780
|
"max_input_tokens": 160000,
|
|
34772
34781
|
"max_output_tokens": 160000,
|
|
34773
34782
|
"max_tokens": 160000,
|
|
@@ -34776,8 +34785,8 @@
|
|
|
34776
34785
|
"novita/qwen/qwen3-32b-fp8": {
|
|
34777
34786
|
"litellm_provider": "novita",
|
|
34778
34787
|
"mode": "chat",
|
|
34779
|
-
"input_cost_per_token":
|
|
34780
|
-
"output_cost_per_token":
|
|
34788
|
+
"input_cost_per_token": 1e-07,
|
|
34789
|
+
"output_cost_per_token": 4.5e-07,
|
|
34781
34790
|
"max_input_tokens": 40960,
|
|
34782
34791
|
"max_output_tokens": 20000,
|
|
34783
34792
|
"max_tokens": 20000,
|
|
@@ -34787,8 +34796,8 @@
|
|
|
34787
34796
|
"novita/qwen/qwen3-30b-a3b-fp8": {
|
|
34788
34797
|
"litellm_provider": "novita",
|
|
34789
34798
|
"mode": "chat",
|
|
34790
|
-
"input_cost_per_token":
|
|
34791
|
-
"output_cost_per_token":
|
|
34799
|
+
"input_cost_per_token": 9e-08,
|
|
34800
|
+
"output_cost_per_token": 4.5e-07,
|
|
34792
34801
|
"max_input_tokens": 40960,
|
|
34793
34802
|
"max_output_tokens": 20000,
|
|
34794
34803
|
"max_tokens": 20000,
|
|
@@ -34798,8 +34807,8 @@
|
|
|
34798
34807
|
"novita/google/gemma-3-27b-it": {
|
|
34799
34808
|
"litellm_provider": "novita",
|
|
34800
34809
|
"mode": "chat",
|
|
34801
|
-
"input_cost_per_token":
|
|
34802
|
-
"output_cost_per_token":
|
|
34810
|
+
"input_cost_per_token": 1.19e-07,
|
|
34811
|
+
"output_cost_per_token": 2e-07,
|
|
34803
34812
|
"max_input_tokens": 98304,
|
|
34804
34813
|
"max_output_tokens": 16384,
|
|
34805
34814
|
"max_tokens": 16384,
|
|
@@ -34809,8 +34818,8 @@
|
|
|
34809
34818
|
"novita/deepseek/deepseek-v3-turbo": {
|
|
34810
34819
|
"litellm_provider": "novita",
|
|
34811
34820
|
"mode": "chat",
|
|
34812
|
-
"input_cost_per_token":
|
|
34813
|
-
"output_cost_per_token":
|
|
34821
|
+
"input_cost_per_token": 4e-07,
|
|
34822
|
+
"output_cost_per_token": 1.3e-06,
|
|
34814
34823
|
"max_input_tokens": 64000,
|
|
34815
34824
|
"max_output_tokens": 16000,
|
|
34816
34825
|
"max_tokens": 16000,
|
|
@@ -34822,8 +34831,8 @@
|
|
|
34822
34831
|
"novita/deepseek/deepseek-r1-turbo": {
|
|
34823
34832
|
"litellm_provider": "novita",
|
|
34824
34833
|
"mode": "chat",
|
|
34825
|
-
"input_cost_per_token":
|
|
34826
|
-
"output_cost_per_token":
|
|
34834
|
+
"input_cost_per_token": 7e-07,
|
|
34835
|
+
"output_cost_per_token": 2.5e-06,
|
|
34827
34836
|
"max_input_tokens": 64000,
|
|
34828
34837
|
"max_output_tokens": 16000,
|
|
34829
34838
|
"max_tokens": 16000,
|
|
@@ -34836,8 +34845,8 @@
|
|
|
34836
34845
|
"novita/Sao10K/L3-8B-Stheno-v3.2": {
|
|
34837
34846
|
"litellm_provider": "novita",
|
|
34838
34847
|
"mode": "chat",
|
|
34839
|
-
"input_cost_per_token":
|
|
34840
|
-
"output_cost_per_token":
|
|
34848
|
+
"input_cost_per_token": 5e-08,
|
|
34849
|
+
"output_cost_per_token": 5e-08,
|
|
34841
34850
|
"max_input_tokens": 8192,
|
|
34842
34851
|
"max_output_tokens": 32000,
|
|
34843
34852
|
"max_tokens": 32000,
|
|
@@ -34849,8 +34858,8 @@
|
|
|
34849
34858
|
"novita/gryphe/mythomax-l2-13b": {
|
|
34850
34859
|
"litellm_provider": "novita",
|
|
34851
34860
|
"mode": "chat",
|
|
34852
|
-
"input_cost_per_token":
|
|
34853
|
-
"output_cost_per_token":
|
|
34861
|
+
"input_cost_per_token": 9e-08,
|
|
34862
|
+
"output_cost_per_token": 9e-08,
|
|
34854
34863
|
"max_input_tokens": 4096,
|
|
34855
34864
|
"max_output_tokens": 3200,
|
|
34856
34865
|
"max_tokens": 3200,
|
|
@@ -34859,8 +34868,8 @@
|
|
|
34859
34868
|
"novita/baidu/ernie-4.5-vl-28b-a3b-thinking": {
|
|
34860
34869
|
"litellm_provider": "novita",
|
|
34861
34870
|
"mode": "chat",
|
|
34862
|
-
"input_cost_per_token":
|
|
34863
|
-
"output_cost_per_token":
|
|
34871
|
+
"input_cost_per_token": 3.9e-07,
|
|
34872
|
+
"output_cost_per_token": 3.9e-07,
|
|
34864
34873
|
"max_input_tokens": 131072,
|
|
34865
34874
|
"max_output_tokens": 65536,
|
|
34866
34875
|
"max_tokens": 65536,
|
|
@@ -34875,8 +34884,8 @@
|
|
|
34875
34884
|
"novita/qwen/qwen3-vl-8b-instruct": {
|
|
34876
34885
|
"litellm_provider": "novita",
|
|
34877
34886
|
"mode": "chat",
|
|
34878
|
-
"input_cost_per_token":
|
|
34879
|
-
"output_cost_per_token":
|
|
34887
|
+
"input_cost_per_token": 8e-08,
|
|
34888
|
+
"output_cost_per_token": 5e-07,
|
|
34880
34889
|
"max_input_tokens": 131072,
|
|
34881
34890
|
"max_output_tokens": 32768,
|
|
34882
34891
|
"max_tokens": 32768,
|
|
@@ -34890,8 +34899,8 @@
|
|
|
34890
34899
|
"novita/zai-org/glm-4.5-air": {
|
|
34891
34900
|
"litellm_provider": "novita",
|
|
34892
34901
|
"mode": "chat",
|
|
34893
|
-
"input_cost_per_token":
|
|
34894
|
-
"output_cost_per_token":
|
|
34902
|
+
"input_cost_per_token": 1.3e-07,
|
|
34903
|
+
"output_cost_per_token": 8.5e-07,
|
|
34895
34904
|
"max_input_tokens": 131072,
|
|
34896
34905
|
"max_output_tokens": 98304,
|
|
34897
34906
|
"max_tokens": 98304,
|
|
@@ -34904,8 +34913,8 @@
|
|
|
34904
34913
|
"novita/qwen/qwen3-vl-30b-a3b-instruct": {
|
|
34905
34914
|
"litellm_provider": "novita",
|
|
34906
34915
|
"mode": "chat",
|
|
34907
|
-
"input_cost_per_token":
|
|
34908
|
-
"output_cost_per_token":
|
|
34916
|
+
"input_cost_per_token": 2e-07,
|
|
34917
|
+
"output_cost_per_token": 7e-07,
|
|
34909
34918
|
"max_input_tokens": 131072,
|
|
34910
34919
|
"max_output_tokens": 32768,
|
|
34911
34920
|
"max_tokens": 32768,
|
|
@@ -34919,8 +34928,8 @@
|
|
|
34919
34928
|
"novita/qwen/qwen3-vl-30b-a3b-thinking": {
|
|
34920
34929
|
"litellm_provider": "novita",
|
|
34921
34930
|
"mode": "chat",
|
|
34922
|
-
"input_cost_per_token":
|
|
34923
|
-
"output_cost_per_token":
|
|
34931
|
+
"input_cost_per_token": 2e-07,
|
|
34932
|
+
"output_cost_per_token": 1e-06,
|
|
34924
34933
|
"max_input_tokens": 131072,
|
|
34925
34934
|
"max_output_tokens": 32768,
|
|
34926
34935
|
"max_tokens": 32768,
|
|
@@ -34934,8 +34943,8 @@
|
|
|
34934
34943
|
"novita/qwen/qwen-mt-plus": {
|
|
34935
34944
|
"litellm_provider": "novita",
|
|
34936
34945
|
"mode": "chat",
|
|
34937
|
-
"input_cost_per_token":
|
|
34938
|
-
"output_cost_per_token":
|
|
34946
|
+
"input_cost_per_token": 2.5e-07,
|
|
34947
|
+
"output_cost_per_token": 7.5e-07,
|
|
34939
34948
|
"max_input_tokens": 16384,
|
|
34940
34949
|
"max_output_tokens": 8192,
|
|
34941
34950
|
"max_tokens": 8192,
|
|
@@ -34944,8 +34953,8 @@
|
|
|
34944
34953
|
"novita/baidu/ernie-4.5-vl-28b-a3b": {
|
|
34945
34954
|
"litellm_provider": "novita",
|
|
34946
34955
|
"mode": "chat",
|
|
34947
|
-
"input_cost_per_token":
|
|
34948
|
-
"output_cost_per_token":
|
|
34956
|
+
"input_cost_per_token": 1.4e-07,
|
|
34957
|
+
"output_cost_per_token": 5.6e-07,
|
|
34949
34958
|
"max_input_tokens": 30000,
|
|
34950
34959
|
"max_output_tokens": 8000,
|
|
34951
34960
|
"max_tokens": 8000,
|
|
@@ -34959,8 +34968,8 @@
|
|
|
34959
34968
|
"novita/baidu/ernie-4.5-21B-a3b": {
|
|
34960
34969
|
"litellm_provider": "novita",
|
|
34961
34970
|
"mode": "chat",
|
|
34962
|
-
"input_cost_per_token":
|
|
34963
|
-
"output_cost_per_token":
|
|
34971
|
+
"input_cost_per_token": 7e-08,
|
|
34972
|
+
"output_cost_per_token": 2.8e-07,
|
|
34964
34973
|
"max_input_tokens": 120000,
|
|
34965
34974
|
"max_output_tokens": 8000,
|
|
34966
34975
|
"max_tokens": 8000,
|
|
@@ -34972,8 +34981,8 @@
|
|
|
34972
34981
|
"novita/qwen/qwen3-8b-fp8": {
|
|
34973
34982
|
"litellm_provider": "novita",
|
|
34974
34983
|
"mode": "chat",
|
|
34975
|
-
"input_cost_per_token":
|
|
34976
|
-
"output_cost_per_token":
|
|
34984
|
+
"input_cost_per_token": 3.5e-08,
|
|
34985
|
+
"output_cost_per_token": 1.38e-07,
|
|
34977
34986
|
"max_input_tokens": 128000,
|
|
34978
34987
|
"max_output_tokens": 20000,
|
|
34979
34988
|
"max_tokens": 20000,
|
|
@@ -34983,8 +34992,8 @@
|
|
|
34983
34992
|
"novita/qwen/qwen3-4b-fp8": {
|
|
34984
34993
|
"litellm_provider": "novita",
|
|
34985
34994
|
"mode": "chat",
|
|
34986
|
-
"input_cost_per_token":
|
|
34987
|
-
"output_cost_per_token":
|
|
34995
|
+
"input_cost_per_token": 3e-08,
|
|
34996
|
+
"output_cost_per_token": 3e-08,
|
|
34988
34997
|
"max_input_tokens": 128000,
|
|
34989
34998
|
"max_output_tokens": 20000,
|
|
34990
34999
|
"max_tokens": 20000,
|
|
@@ -34994,8 +35003,8 @@
|
|
|
34994
35003
|
"novita/qwen/qwen2.5-7b-instruct": {
|
|
34995
35004
|
"litellm_provider": "novita",
|
|
34996
35005
|
"mode": "chat",
|
|
34997
|
-
"input_cost_per_token":
|
|
34998
|
-
"output_cost_per_token":
|
|
35006
|
+
"input_cost_per_token": 7e-08,
|
|
35007
|
+
"output_cost_per_token": 7e-08,
|
|
34999
35008
|
"max_input_tokens": 32000,
|
|
35000
35009
|
"max_output_tokens": 32000,
|
|
35001
35010
|
"max_tokens": 32000,
|
|
@@ -35008,8 +35017,8 @@
|
|
|
35008
35017
|
"novita/meta-llama/llama-3.2-3b-instruct": {
|
|
35009
35018
|
"litellm_provider": "novita",
|
|
35010
35019
|
"mode": "chat",
|
|
35011
|
-
"input_cost_per_token":
|
|
35012
|
-
"output_cost_per_token":
|
|
35020
|
+
"input_cost_per_token": 3e-08,
|
|
35021
|
+
"output_cost_per_token": 5e-08,
|
|
35013
35022
|
"max_input_tokens": 32768,
|
|
35014
35023
|
"max_output_tokens": 32000,
|
|
35015
35024
|
"max_tokens": 32000,
|
|
@@ -35021,8 +35030,8 @@
|
|
|
35021
35030
|
"novita/sao10k/l31-70b-euryale-v2.2": {
|
|
35022
35031
|
"litellm_provider": "novita",
|
|
35023
35032
|
"mode": "chat",
|
|
35024
|
-
"input_cost_per_token":
|
|
35025
|
-
"output_cost_per_token":
|
|
35033
|
+
"input_cost_per_token": 1.48e-06,
|
|
35034
|
+
"output_cost_per_token": 1.48e-06,
|
|
35026
35035
|
"max_input_tokens": 8192,
|
|
35027
35036
|
"max_output_tokens": 8192,
|
|
35028
35037
|
"max_tokens": 8192,
|
|
@@ -35034,7 +35043,7 @@
|
|
|
35034
35043
|
"novita/qwen/qwen3-embedding-0.6b": {
|
|
35035
35044
|
"litellm_provider": "novita",
|
|
35036
35045
|
"mode": "embedding",
|
|
35037
|
-
"input_cost_per_token":
|
|
35046
|
+
"input_cost_per_token": 7e-08,
|
|
35038
35047
|
"output_cost_per_token": 0,
|
|
35039
35048
|
"max_input_tokens": 32768,
|
|
35040
35049
|
"max_output_tokens": 32768,
|
|
@@ -35043,7 +35052,7 @@
|
|
|
35043
35052
|
"novita/qwen/qwen3-embedding-8b": {
|
|
35044
35053
|
"litellm_provider": "novita",
|
|
35045
35054
|
"mode": "embedding",
|
|
35046
|
-
"input_cost_per_token":
|
|
35055
|
+
"input_cost_per_token": 7e-08,
|
|
35047
35056
|
"output_cost_per_token": 0,
|
|
35048
35057
|
"max_input_tokens": 32768,
|
|
35049
35058
|
"max_output_tokens": 4096,
|
|
@@ -35052,8 +35061,8 @@
|
|
|
35052
35061
|
"novita/baai/bge-m3": {
|
|
35053
35062
|
"litellm_provider": "novita",
|
|
35054
35063
|
"mode": "embedding",
|
|
35055
|
-
"input_cost_per_token":
|
|
35056
|
-
"output_cost_per_token":
|
|
35064
|
+
"input_cost_per_token": 1e-08,
|
|
35065
|
+
"output_cost_per_token": 1e-08,
|
|
35057
35066
|
"max_input_tokens": 8192,
|
|
35058
35067
|
"max_output_tokens": 96000,
|
|
35059
35068
|
"max_tokens": 96000
|
|
@@ -35061,8 +35070,8 @@
|
|
|
35061
35070
|
"novita/qwen/qwen3-reranker-8b": {
|
|
35062
35071
|
"litellm_provider": "novita",
|
|
35063
35072
|
"mode": "rerank",
|
|
35064
|
-
"input_cost_per_token":
|
|
35065
|
-
"output_cost_per_token":
|
|
35073
|
+
"input_cost_per_token": 5e-08,
|
|
35074
|
+
"output_cost_per_token": 5e-08,
|
|
35066
35075
|
"max_input_tokens": 32768,
|
|
35067
35076
|
"max_output_tokens": 4096,
|
|
35068
35077
|
"max_tokens": 4096
|
|
@@ -35070,8 +35079,8 @@
|
|
|
35070
35079
|
"novita/baai/bge-reranker-v2-m3": {
|
|
35071
35080
|
"litellm_provider": "novita",
|
|
35072
35081
|
"mode": "rerank",
|
|
35073
|
-
"input_cost_per_token":
|
|
35074
|
-
"output_cost_per_token":
|
|
35082
|
+
"input_cost_per_token": 1e-08,
|
|
35083
|
+
"output_cost_per_token": 1e-08,
|
|
35075
35084
|
"max_input_tokens": 8000,
|
|
35076
35085
|
"max_output_tokens": 8000,
|
|
35077
35086
|
"max_tokens": 8000
|
|
@@ -35369,5 +35378,138 @@
|
|
|
35369
35378
|
"max_tokens": 8192,
|
|
35370
35379
|
"supports_reasoning": true,
|
|
35371
35380
|
"supports_system_messages": true
|
|
35381
|
+
},
|
|
35382
|
+
"azure_ai/claude-opus-4-5": {
|
|
35383
|
+
"input_cost_per_token": 5e-06,
|
|
35384
|
+
"litellm_provider": "azure_ai",
|
|
35385
|
+
"max_input_tokens": 200000,
|
|
35386
|
+
"max_output_tokens": 64000,
|
|
35387
|
+
"max_tokens": 64000,
|
|
35388
|
+
"mode": "chat",
|
|
35389
|
+
"output_cost_per_token": 2.5e-05,
|
|
35390
|
+
"supports_assistant_prefill": true,
|
|
35391
|
+
"supports_computer_use": true,
|
|
35392
|
+
"supports_function_calling": true,
|
|
35393
|
+
"supports_pdf_input": true,
|
|
35394
|
+
"supports_prompt_caching": true,
|
|
35395
|
+
"supports_reasoning": true,
|
|
35396
|
+
"supports_response_schema": true,
|
|
35397
|
+
"supports_tool_choice": true,
|
|
35398
|
+
"supports_vision": true
|
|
35399
|
+
},
|
|
35400
|
+
"novita/kwaipilot/kat-coder-pro": {
|
|
35401
|
+
"litellm_provider": "novita",
|
|
35402
|
+
"mode": "chat",
|
|
35403
|
+
"input_cost_per_token": 3e-07,
|
|
35404
|
+
"output_cost_per_token": 1.2e-06,
|
|
35405
|
+
"max_input_tokens": 256000,
|
|
35406
|
+
"max_output_tokens": 128000,
|
|
35407
|
+
"max_tokens": 128000,
|
|
35408
|
+
"supports_function_calling": true,
|
|
35409
|
+
"supports_parallel_function_calling": true,
|
|
35410
|
+
"supports_tool_choice": true,
|
|
35411
|
+
"supports_system_messages": true,
|
|
35412
|
+
"supports_response_schema": true,
|
|
35413
|
+
"cache_read_input_token_cost": 6e-08,
|
|
35414
|
+
"input_cost_per_token_cache_hit": 6e-08
|
|
35415
|
+
},
|
|
35416
|
+
"novita/qwen/qwen3-omni-30b-a3b-thinking": {
|
|
35417
|
+
"litellm_provider": "novita",
|
|
35418
|
+
"mode": "chat",
|
|
35419
|
+
"input_cost_per_token": 2.5e-07,
|
|
35420
|
+
"output_cost_per_token": 9.7e-07,
|
|
35421
|
+
"max_input_tokens": 65536,
|
|
35422
|
+
"max_output_tokens": 16384,
|
|
35423
|
+
"max_tokens": 16384,
|
|
35424
|
+
"supports_function_calling": true,
|
|
35425
|
+
"supports_parallel_function_calling": true,
|
|
35426
|
+
"supports_tool_choice": true,
|
|
35427
|
+
"supports_vision": true,
|
|
35428
|
+
"supports_system_messages": true,
|
|
35429
|
+
"supports_response_schema": true,
|
|
35430
|
+
"supports_reasoning": true,
|
|
35431
|
+
"supports_audio_input": true
|
|
35432
|
+
},
|
|
35433
|
+
"novita/qwen/qwen3-omni-30b-a3b-instruct": {
|
|
35434
|
+
"litellm_provider": "novita",
|
|
35435
|
+
"mode": "chat",
|
|
35436
|
+
"input_cost_per_token": 2.5e-07,
|
|
35437
|
+
"output_cost_per_token": 9.7e-07,
|
|
35438
|
+
"max_input_tokens": 65536,
|
|
35439
|
+
"max_output_tokens": 16384,
|
|
35440
|
+
"max_tokens": 16384,
|
|
35441
|
+
"supports_function_calling": true,
|
|
35442
|
+
"supports_parallel_function_calling": true,
|
|
35443
|
+
"supports_tool_choice": true,
|
|
35444
|
+
"supports_vision": true,
|
|
35445
|
+
"supports_system_messages": true,
|
|
35446
|
+
"supports_response_schema": true,
|
|
35447
|
+
"supports_audio_input": true,
|
|
35448
|
+
"supports_audio_output": true
|
|
35449
|
+
},
|
|
35450
|
+
"cerebras/zai-glm-4.7": {
|
|
35451
|
+
"input_cost_per_token": 2.25e-06,
|
|
35452
|
+
"litellm_provider": "cerebras",
|
|
35453
|
+
"max_input_tokens": 128000,
|
|
35454
|
+
"max_output_tokens": 128000,
|
|
35455
|
+
"max_tokens": 128000,
|
|
35456
|
+
"mode": "chat",
|
|
35457
|
+
"output_cost_per_token": 2.75e-06,
|
|
35458
|
+
"source": "https://www.cerebras.ai/pricing",
|
|
35459
|
+
"supports_function_calling": true,
|
|
35460
|
+
"supports_reasoning": true,
|
|
35461
|
+
"supports_tool_choice": true
|
|
35462
|
+
},
|
|
35463
|
+
"gpt-5.2-codex": {
|
|
35464
|
+
"cache_read_input_token_cost": 1.75e-07,
|
|
35465
|
+
"cache_read_input_token_cost_priority": 3.5e-07,
|
|
35466
|
+
"input_cost_per_token": 1.75e-06,
|
|
35467
|
+
"input_cost_per_token_priority": 3.5e-06,
|
|
35468
|
+
"litellm_provider": "openai",
|
|
35469
|
+
"max_input_tokens": 400000,
|
|
35470
|
+
"max_output_tokens": 128000,
|
|
35471
|
+
"max_tokens": 128000,
|
|
35472
|
+
"mode": "responses",
|
|
35473
|
+
"output_cost_per_token": 1.4e-05,
|
|
35474
|
+
"output_cost_per_token_priority": 2.8e-05,
|
|
35475
|
+
"supported_endpoints": [
|
|
35476
|
+
"/v1/responses"
|
|
35477
|
+
],
|
|
35478
|
+
"supported_modalities": [
|
|
35479
|
+
"text",
|
|
35480
|
+
"image"
|
|
35481
|
+
],
|
|
35482
|
+
"supported_output_modalities": [
|
|
35483
|
+
"text"
|
|
35484
|
+
],
|
|
35485
|
+
"supports_function_calling": true,
|
|
35486
|
+
"supports_native_streaming": true,
|
|
35487
|
+
"supports_parallel_function_calling": true,
|
|
35488
|
+
"supports_pdf_input": true,
|
|
35489
|
+
"supports_prompt_caching": true,
|
|
35490
|
+
"supports_reasoning": true,
|
|
35491
|
+
"supports_response_schema": true,
|
|
35492
|
+
"supports_system_messages": false,
|
|
35493
|
+
"supports_tool_choice": true,
|
|
35494
|
+
"supports_vision": true
|
|
35495
|
+
},
|
|
35496
|
+
"openrouter/openai/gpt-5.2-codex": {
|
|
35497
|
+
"cache_read_input_token_cost": 1.75e-07,
|
|
35498
|
+
"input_cost_per_token": 1.75e-06,
|
|
35499
|
+
"litellm_provider": "openrouter",
|
|
35500
|
+
"max_input_tokens": 400000,
|
|
35501
|
+
"max_output_tokens": 128000,
|
|
35502
|
+
"max_tokens": 128000,
|
|
35503
|
+
"mode": "chat",
|
|
35504
|
+
"output_cost_per_token": 1.4e-05,
|
|
35505
|
+
"supported_modalities": [
|
|
35506
|
+
"text",
|
|
35507
|
+
"image"
|
|
35508
|
+
],
|
|
35509
|
+
"supported_output_modalities": [
|
|
35510
|
+
"text"
|
|
35511
|
+
],
|
|
35512
|
+
"supports_reasoning": true,
|
|
35513
|
+
"supports_tool_choice": true
|
|
35372
35514
|
}
|
|
35373
35515
|
}
|