tokencostauto 0.1.441__py3-none-any.whl → 0.1.443__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -407,7 +407,7 @@
407
407
  "litellm_provider": "openai",
408
408
  "max_input_tokens": 16385,
409
409
  "max_output_tokens": 4096,
410
- "max_tokens": 4097,
410
+ "max_tokens": 4096,
411
411
  "mode": "chat",
412
412
  "output_cost_per_token": 1.5e-06,
413
413
  "supports_function_calling": true,
@@ -420,7 +420,7 @@
420
420
  "litellm_provider": "openai",
421
421
  "max_input_tokens": 4097,
422
422
  "max_output_tokens": 4096,
423
- "max_tokens": 4097,
423
+ "max_tokens": 4096,
424
424
  "mode": "chat",
425
425
  "output_cost_per_token": 2e-06,
426
426
  "supports_prompt_caching": true,
@@ -432,7 +432,7 @@
432
432
  "litellm_provider": "openai",
433
433
  "max_input_tokens": 4097,
434
434
  "max_output_tokens": 4096,
435
- "max_tokens": 4097,
435
+ "max_tokens": 4096,
436
436
  "mode": "chat",
437
437
  "output_cost_per_token": 2e-06,
438
438
  "supports_function_calling": true,
@@ -446,7 +446,7 @@
446
446
  "litellm_provider": "openai",
447
447
  "max_input_tokens": 16385,
448
448
  "max_output_tokens": 4096,
449
- "max_tokens": 16385,
449
+ "max_tokens": 4096,
450
450
  "mode": "chat",
451
451
  "output_cost_per_token": 2e-06,
452
452
  "supports_function_calling": true,
@@ -460,7 +460,7 @@
460
460
  "litellm_provider": "openai",
461
461
  "max_input_tokens": 16385,
462
462
  "max_output_tokens": 4096,
463
- "max_tokens": 16385,
463
+ "max_tokens": 4096,
464
464
  "mode": "chat",
465
465
  "output_cost_per_token": 1.5e-06,
466
466
  "supports_function_calling": true,
@@ -474,7 +474,7 @@
474
474
  "litellm_provider": "openai",
475
475
  "max_input_tokens": 16385,
476
476
  "max_output_tokens": 4096,
477
- "max_tokens": 16385,
477
+ "max_tokens": 4096,
478
478
  "mode": "chat",
479
479
  "output_cost_per_token": 4e-06,
480
480
  "supports_prompt_caching": true,
@@ -486,7 +486,7 @@
486
486
  "litellm_provider": "openai",
487
487
  "max_input_tokens": 16385,
488
488
  "max_output_tokens": 4096,
489
- "max_tokens": 16385,
489
+ "max_tokens": 4096,
490
490
  "mode": "chat",
491
491
  "output_cost_per_token": 4e-06,
492
492
  "supports_prompt_caching": true,
@@ -597,7 +597,7 @@
597
597
  "litellm_provider": "text-completion-openai",
598
598
  "max_input_tokens": 16384,
599
599
  "max_output_tokens": 4096,
600
- "max_tokens": 16384,
600
+ "max_tokens": 4096,
601
601
  "mode": "completion",
602
602
  "output_cost_per_token": 1.2e-05,
603
603
  "output_cost_per_token_batches": 1e-06
@@ -608,7 +608,7 @@
608
608
  "litellm_provider": "text-completion-openai",
609
609
  "max_input_tokens": 16384,
610
610
  "max_output_tokens": 4096,
611
- "max_tokens": 16384,
611
+ "max_tokens": 4096,
612
612
  "mode": "completion",
613
613
  "output_cost_per_token": 1.6e-06,
614
614
  "output_cost_per_token_batches": 2e-07
@@ -659,7 +659,7 @@
659
659
  "litellm_provider": "openai",
660
660
  "max_input_tokens": 32768,
661
661
  "max_output_tokens": 0,
662
- "max_tokens": 32768,
662
+ "max_tokens": 0,
663
663
  "mode": "moderation",
664
664
  "output_cost_per_token": 0.0
665
665
  },
@@ -668,7 +668,7 @@
668
668
  "litellm_provider": "openai",
669
669
  "max_input_tokens": 32768,
670
670
  "max_output_tokens": 0,
671
- "max_tokens": 32768,
671
+ "max_tokens": 0,
672
672
  "mode": "moderation",
673
673
  "output_cost_per_token": 0.0
674
674
  },
@@ -677,7 +677,7 @@
677
677
  "litellm_provider": "openai",
678
678
  "max_input_tokens": 32768,
679
679
  "max_output_tokens": 0,
680
- "max_tokens": 32768,
680
+ "max_tokens": 0,
681
681
  "mode": "moderation",
682
682
  "output_cost_per_token": 0.0
683
683
  },
@@ -1063,7 +1063,7 @@
1063
1063
  "litellm_provider": "azure",
1064
1064
  "max_input_tokens": 4097,
1065
1065
  "max_output_tokens": 4096,
1066
- "max_tokens": 4097,
1066
+ "max_tokens": 4096,
1067
1067
  "mode": "chat",
1068
1068
  "output_cost_per_token": 2e-06,
1069
1069
  "supports_function_calling": true,
@@ -1076,7 +1076,7 @@
1076
1076
  "litellm_provider": "azure",
1077
1077
  "max_input_tokens": 4097,
1078
1078
  "max_output_tokens": 4096,
1079
- "max_tokens": 4097,
1079
+ "max_tokens": 4096,
1080
1080
  "mode": "chat",
1081
1081
  "output_cost_per_token": 2e-06,
1082
1082
  "supports_function_calling": true,
@@ -1368,7 +1368,7 @@
1368
1368
  "litellm_provider": "text-completion-openai",
1369
1369
  "max_input_tokens": 16384,
1370
1370
  "max_output_tokens": 4096,
1371
- "max_tokens": 16384,
1371
+ "max_tokens": 4096,
1372
1372
  "mode": "completion",
1373
1373
  "output_cost_per_token": 4e-07
1374
1374
  },
@@ -1377,7 +1377,7 @@
1377
1377
  "litellm_provider": "text-completion-openai",
1378
1378
  "max_input_tokens": 16384,
1379
1379
  "max_output_tokens": 4096,
1380
- "max_tokens": 16384,
1380
+ "max_tokens": 4096,
1381
1381
  "mode": "completion",
1382
1382
  "output_cost_per_token": 2e-06
1383
1383
  },
@@ -2496,7 +2496,7 @@
2496
2496
  "litellm_provider": "vertex_ai-language-models",
2497
2497
  "max_input_tokens": 8192,
2498
2498
  "max_output_tokens": 2048,
2499
- "max_tokens": 8192,
2499
+ "max_tokens": 2048,
2500
2500
  "mode": "chat",
2501
2501
  "output_cost_per_character": 3.75e-07,
2502
2502
  "output_cost_per_token": 1.5e-06,
@@ -2513,7 +2513,7 @@
2513
2513
  "litellm_provider": "vertex_ai-language-models",
2514
2514
  "max_input_tokens": 8192,
2515
2515
  "max_output_tokens": 2048,
2516
- "max_tokens": 8192,
2516
+ "max_tokens": 2048,
2517
2517
  "mode": "chat",
2518
2518
  "output_cost_per_character": 3.75e-07,
2519
2519
  "output_cost_per_token": 1.5e-06,
@@ -3098,7 +3098,7 @@
3098
3098
  "litellm_provider": "vertex_ai-llama_models",
3099
3099
  "max_input_tokens": 128000,
3100
3100
  "max_output_tokens": 2048,
3101
- "max_tokens": 128000,
3101
+ "max_tokens": 2048,
3102
3102
  "metadata": {
3103
3103
  "notes": "VertexAI states that The Llama 3.2 API service is at no cost during public preview, and will be priced as per dollar-per-1M-tokens at GA."
3104
3104
  },
@@ -4079,7 +4079,7 @@
4079
4079
  "litellm_provider": "openrouter",
4080
4080
  "max_input_tokens": 66000,
4081
4081
  "max_output_tokens": 4096,
4082
- "max_tokens": 8192,
4082
+ "max_tokens": 4096,
4083
4083
  "mode": "chat",
4084
4084
  "output_cost_per_token": 2.8e-07,
4085
4085
  "supports_prompt_caching": true,
@@ -4410,7 +4410,7 @@
4410
4410
  "input_cost_per_token": 1.63e-06,
4411
4411
  "litellm_provider": "openrouter",
4412
4412
  "max_output_tokens": 8191,
4413
- "max_tokens": 100000,
4413
+ "max_tokens": 8191,
4414
4414
  "mode": "chat",
4415
4415
  "output_cost_per_token": 5.51e-06,
4416
4416
  "supports_tool_choice": true
@@ -4419,7 +4419,7 @@
4419
4419
  "input_cost_per_token": 1.102e-05,
4420
4420
  "litellm_provider": "openrouter",
4421
4421
  "max_output_tokens": 8191,
4422
- "max_tokens": 100000,
4422
+ "max_tokens": 8191,
4423
4423
  "mode": "chat",
4424
4424
  "output_cost_per_token": 3.268e-05,
4425
4425
  "supports_tool_choice": true
@@ -5898,7 +5898,7 @@
5898
5898
  "litellm_provider": "bedrock",
5899
5899
  "max_input_tokens": 128000,
5900
5900
  "max_output_tokens": 2048,
5901
- "max_tokens": 128000,
5901
+ "max_tokens": 2048,
5902
5902
  "mode": "chat",
5903
5903
  "output_cost_per_token": 2.2e-07,
5904
5904
  "supports_function_calling": true,
@@ -5909,7 +5909,7 @@
5909
5909
  "litellm_provider": "bedrock",
5910
5910
  "max_input_tokens": 128000,
5911
5911
  "max_output_tokens": 2048,
5912
- "max_tokens": 128000,
5912
+ "max_tokens": 2048,
5913
5913
  "mode": "chat",
5914
5914
  "output_cost_per_token": 9.9e-07,
5915
5915
  "supports_function_calling": true,
@@ -5920,7 +5920,7 @@
5920
5920
  "litellm_provider": "bedrock",
5921
5921
  "max_input_tokens": 128000,
5922
5922
  "max_output_tokens": 4096,
5923
- "max_tokens": 128000,
5923
+ "max_tokens": 4096,
5924
5924
  "mode": "chat",
5925
5925
  "output_cost_per_token": 1.6e-05,
5926
5926
  "supports_function_calling": true,
@@ -5931,7 +5931,7 @@
5931
5931
  "litellm_provider": "bedrock",
5932
5932
  "max_input_tokens": 128000,
5933
5933
  "max_output_tokens": 4096,
5934
- "max_tokens": 128000,
5934
+ "max_tokens": 4096,
5935
5935
  "mode": "chat",
5936
5936
  "output_cost_per_token": 1e-07,
5937
5937
  "supports_function_calling": true,
@@ -5942,7 +5942,7 @@
5942
5942
  "litellm_provider": "bedrock",
5943
5943
  "max_input_tokens": 128000,
5944
5944
  "max_output_tokens": 4096,
5945
- "max_tokens": 128000,
5945
+ "max_tokens": 4096,
5946
5946
  "mode": "chat",
5947
5947
  "output_cost_per_token": 1e-07,
5948
5948
  "supports_function_calling": true,
@@ -5953,7 +5953,7 @@
5953
5953
  "litellm_provider": "bedrock",
5954
5954
  "max_input_tokens": 128000,
5955
5955
  "max_output_tokens": 4096,
5956
- "max_tokens": 128000,
5956
+ "max_tokens": 4096,
5957
5957
  "mode": "chat",
5958
5958
  "output_cost_per_token": 1.3e-07,
5959
5959
  "supports_function_calling": true,
@@ -5964,7 +5964,7 @@
5964
5964
  "litellm_provider": "bedrock",
5965
5965
  "max_input_tokens": 128000,
5966
5966
  "max_output_tokens": 4096,
5967
- "max_tokens": 128000,
5967
+ "max_tokens": 4096,
5968
5968
  "mode": "chat",
5969
5969
  "output_cost_per_token": 1.5e-07,
5970
5970
  "supports_function_calling": true,
@@ -5975,7 +5975,7 @@
5975
5975
  "litellm_provider": "bedrock",
5976
5976
  "max_input_tokens": 128000,
5977
5977
  "max_output_tokens": 4096,
5978
- "max_tokens": 128000,
5978
+ "max_tokens": 4096,
5979
5979
  "mode": "chat",
5980
5980
  "output_cost_per_token": 1.5e-07,
5981
5981
  "supports_function_calling": true,
@@ -5986,7 +5986,7 @@
5986
5986
  "litellm_provider": "bedrock",
5987
5987
  "max_input_tokens": 128000,
5988
5988
  "max_output_tokens": 4096,
5989
- "max_tokens": 128000,
5989
+ "max_tokens": 4096,
5990
5990
  "mode": "chat",
5991
5991
  "output_cost_per_token": 1.9e-07,
5992
5992
  "supports_function_calling": true,
@@ -5997,7 +5997,7 @@
5997
5997
  "litellm_provider": "bedrock",
5998
5998
  "max_input_tokens": 128000,
5999
5999
  "max_output_tokens": 4096,
6000
- "max_tokens": 128000,
6000
+ "max_tokens": 4096,
6001
6001
  "mode": "chat",
6002
6002
  "output_cost_per_token": 3.5e-07,
6003
6003
  "supports_function_calling": true,
@@ -6009,7 +6009,7 @@
6009
6009
  "litellm_provider": "bedrock",
6010
6010
  "max_input_tokens": 128000,
6011
6011
  "max_output_tokens": 4096,
6012
- "max_tokens": 128000,
6012
+ "max_tokens": 4096,
6013
6013
  "mode": "chat",
6014
6014
  "output_cost_per_token": 3.5e-07,
6015
6015
  "supports_function_calling": true,
@@ -6021,7 +6021,7 @@
6021
6021
  "litellm_provider": "bedrock",
6022
6022
  "max_input_tokens": 128000,
6023
6023
  "max_output_tokens": 4096,
6024
- "max_tokens": 128000,
6024
+ "max_tokens": 4096,
6025
6025
  "mode": "chat",
6026
6026
  "output_cost_per_token": 2e-06,
6027
6027
  "supports_function_calling": true,
@@ -6033,7 +6033,7 @@
6033
6033
  "litellm_provider": "bedrock",
6034
6034
  "max_input_tokens": 128000,
6035
6035
  "max_output_tokens": 4096,
6036
- "max_tokens": 128000,
6036
+ "max_tokens": 4096,
6037
6037
  "mode": "chat",
6038
6038
  "output_cost_per_token": 2e-06,
6039
6039
  "supports_function_calling": true,
@@ -6224,7 +6224,7 @@
6224
6224
  "litellm_provider": "ollama",
6225
6225
  "max_input_tokens": 32768,
6226
6226
  "max_output_tokens": 8192,
6227
- "max_tokens": 32768,
6227
+ "max_tokens": 8192,
6228
6228
  "mode": "chat",
6229
6229
  "output_cost_per_token": 0.0,
6230
6230
  "supports_function_calling": false
@@ -6234,7 +6234,7 @@
6234
6234
  "litellm_provider": "ollama",
6235
6235
  "max_input_tokens": 32768,
6236
6236
  "max_output_tokens": 8192,
6237
- "max_tokens": 32768,
6237
+ "max_tokens": 8192,
6238
6238
  "mode": "chat",
6239
6239
  "output_cost_per_token": 0.0,
6240
6240
  "supports_function_calling": true
@@ -6254,7 +6254,7 @@
6254
6254
  "litellm_provider": "ollama",
6255
6255
  "max_input_tokens": 32768,
6256
6256
  "max_output_tokens": 8192,
6257
- "max_tokens": 32768,
6257
+ "max_tokens": 8192,
6258
6258
  "mode": "chat",
6259
6259
  "output_cost_per_token": 0.0,
6260
6260
  "supports_function_calling": true
@@ -6274,7 +6274,7 @@
6274
6274
  "litellm_provider": "ollama",
6275
6275
  "max_input_tokens": 32768,
6276
6276
  "max_output_tokens": 8192,
6277
- "max_tokens": 32768,
6277
+ "max_tokens": 8192,
6278
6278
  "mode": "chat",
6279
6279
  "output_cost_per_token": 0.0,
6280
6280
  "supports_function_calling": true
@@ -6356,7 +6356,7 @@
6356
6356
  "litellm_provider": "ollama",
6357
6357
  "max_input_tokens": 8192,
6358
6358
  "max_output_tokens": 8192,
6359
- "max_tokens": 32768,
6359
+ "max_tokens": 8192,
6360
6360
  "mode": "chat",
6361
6361
  "output_cost_per_token": 0.0,
6362
6362
  "supports_function_calling": true
@@ -6366,7 +6366,7 @@
6366
6366
  "litellm_provider": "ollama",
6367
6367
  "max_input_tokens": 65536,
6368
6368
  "max_output_tokens": 8192,
6369
- "max_tokens": 65536,
6369
+ "max_tokens": 8192,
6370
6370
  "mode": "chat",
6371
6371
  "output_cost_per_token": 0.0,
6372
6372
  "supports_function_calling": true
@@ -8005,7 +8005,7 @@
8005
8005
  "litellm_provider": "bedrock",
8006
8006
  "max_input_tokens": 128000,
8007
8007
  "max_output_tokens": 2048,
8008
- "max_tokens": 128000,
8008
+ "max_tokens": 2048,
8009
8009
  "mode": "chat",
8010
8010
  "output_cost_per_token": 2.2e-07,
8011
8011
  "supports_function_calling": true,
@@ -8016,7 +8016,7 @@
8016
8016
  "litellm_provider": "bedrock",
8017
8017
  "max_input_tokens": 128000,
8018
8018
  "max_output_tokens": 2048,
8019
- "max_tokens": 128000,
8019
+ "max_tokens": 2048,
8020
8020
  "mode": "chat",
8021
8021
  "output_cost_per_token": 9.9e-07,
8022
8022
  "supports_function_calling": true,
@@ -8027,7 +8027,7 @@
8027
8027
  "litellm_provider": "bedrock",
8028
8028
  "max_input_tokens": 128000,
8029
8029
  "max_output_tokens": 4096,
8030
- "max_tokens": 128000,
8030
+ "max_tokens": 4096,
8031
8031
  "mode": "chat",
8032
8032
  "output_cost_per_token": 1.6e-05,
8033
8033
  "supports_function_calling": true,
@@ -8058,7 +8058,7 @@
8058
8058
  "litellm_provider": "openai",
8059
8059
  "max_input_tokens": 32768,
8060
8060
  "max_output_tokens": 0,
8061
- "max_tokens": 32768,
8061
+ "max_tokens": 0,
8062
8062
  "mode": "moderation",
8063
8063
  "output_cost_per_token": 0.0
8064
8064
  },
@@ -8067,7 +8067,7 @@
8067
8067
  "litellm_provider": "openai",
8068
8068
  "max_input_tokens": 32768,
8069
8069
  "max_output_tokens": 0,
8070
- "max_tokens": 32768,
8070
+ "max_tokens": 0,
8071
8071
  "mode": "moderation",
8072
8072
  "output_cost_per_token": 0.0
8073
8073
  },
@@ -8076,7 +8076,7 @@
8076
8076
  "litellm_provider": "openai",
8077
8077
  "max_input_tokens": 32768,
8078
8078
  "max_output_tokens": 0,
8079
- "max_tokens": 32768,
8079
+ "max_tokens": 0,
8080
8080
  "mode": "moderation",
8081
8081
  "output_cost_per_token": 0.0
8082
8082
  },
@@ -9387,7 +9387,7 @@
9387
9387
  "max_input_tokens": 1048576,
9388
9388
  "max_output_tokens": 65536,
9389
9389
  "max_pdf_size_mb": 30,
9390
- "max_tokens": 8192,
9390
+ "max_tokens": 65536,
9391
9391
  "max_video_length": 1,
9392
9392
  "max_videos_per_prompt": 10,
9393
9393
  "mode": "chat",
@@ -10441,7 +10441,7 @@
10441
10441
  "max_input_tokens": 1048576,
10442
10442
  "max_output_tokens": 65536,
10443
10443
  "max_pdf_size_mb": 30,
10444
- "max_tokens": 8192,
10444
+ "max_tokens": 65536,
10445
10445
  "max_video_length": 1,
10446
10446
  "max_videos_per_prompt": 10,
10447
10447
  "mode": "chat",
@@ -10618,7 +10618,7 @@
10618
10618
  "litellm_provider": "snowflake",
10619
10619
  "max_input_tokens": 32768,
10620
10620
  "max_output_tokens": 8192,
10621
- "max_tokens": 32768,
10621
+ "max_tokens": 8192,
10622
10622
  "mode": "chat",
10623
10623
  "supports_reasoning": true
10624
10624
  },
@@ -10626,14 +10626,14 @@
10626
10626
  "litellm_provider": "snowflake",
10627
10627
  "max_input_tokens": 4096,
10628
10628
  "max_output_tokens": 8192,
10629
- "max_tokens": 4096,
10629
+ "max_tokens": 8192,
10630
10630
  "mode": "chat"
10631
10631
  },
10632
10632
  "snowflake/claude-3-5-sonnet": {
10633
10633
  "litellm_provider": "snowflake",
10634
10634
  "max_input_tokens": 18000,
10635
10635
  "max_output_tokens": 8192,
10636
- "max_tokens": 18000,
10636
+ "max_tokens": 8192,
10637
10637
  "mode": "chat",
10638
10638
  "supports_computer_use": true
10639
10639
  },
@@ -10641,147 +10641,147 @@
10641
10641
  "litellm_provider": "snowflake",
10642
10642
  "max_input_tokens": 32000,
10643
10643
  "max_output_tokens": 8192,
10644
- "max_tokens": 32000,
10644
+ "max_tokens": 8192,
10645
10645
  "mode": "chat"
10646
10646
  },
10647
10647
  "snowflake/mistral-large2": {
10648
10648
  "litellm_provider": "snowflake",
10649
10649
  "max_input_tokens": 128000,
10650
10650
  "max_output_tokens": 8192,
10651
- "max_tokens": 128000,
10651
+ "max_tokens": 8192,
10652
10652
  "mode": "chat"
10653
10653
  },
10654
10654
  "snowflake/reka-flash": {
10655
10655
  "litellm_provider": "snowflake",
10656
10656
  "max_input_tokens": 100000,
10657
10657
  "max_output_tokens": 8192,
10658
- "max_tokens": 100000,
10658
+ "max_tokens": 8192,
10659
10659
  "mode": "chat"
10660
10660
  },
10661
10661
  "snowflake/reka-core": {
10662
10662
  "litellm_provider": "snowflake",
10663
10663
  "max_input_tokens": 32000,
10664
10664
  "max_output_tokens": 8192,
10665
- "max_tokens": 32000,
10665
+ "max_tokens": 8192,
10666
10666
  "mode": "chat"
10667
10667
  },
10668
10668
  "snowflake/jamba-instruct": {
10669
10669
  "litellm_provider": "snowflake",
10670
10670
  "max_input_tokens": 256000,
10671
10671
  "max_output_tokens": 8192,
10672
- "max_tokens": 256000,
10672
+ "max_tokens": 8192,
10673
10673
  "mode": "chat"
10674
10674
  },
10675
10675
  "snowflake/jamba-1.5-mini": {
10676
10676
  "litellm_provider": "snowflake",
10677
10677
  "max_input_tokens": 256000,
10678
10678
  "max_output_tokens": 8192,
10679
- "max_tokens": 256000,
10679
+ "max_tokens": 8192,
10680
10680
  "mode": "chat"
10681
10681
  },
10682
10682
  "snowflake/jamba-1.5-large": {
10683
10683
  "litellm_provider": "snowflake",
10684
10684
  "max_input_tokens": 256000,
10685
10685
  "max_output_tokens": 8192,
10686
- "max_tokens": 256000,
10686
+ "max_tokens": 8192,
10687
10687
  "mode": "chat"
10688
10688
  },
10689
10689
  "snowflake/mixtral-8x7b": {
10690
10690
  "litellm_provider": "snowflake",
10691
10691
  "max_input_tokens": 32000,
10692
10692
  "max_output_tokens": 8192,
10693
- "max_tokens": 32000,
10693
+ "max_tokens": 8192,
10694
10694
  "mode": "chat"
10695
10695
  },
10696
10696
  "snowflake/llama2-70b-chat": {
10697
10697
  "litellm_provider": "snowflake",
10698
10698
  "max_input_tokens": 4096,
10699
10699
  "max_output_tokens": 8192,
10700
- "max_tokens": 4096,
10700
+ "max_tokens": 8192,
10701
10701
  "mode": "chat"
10702
10702
  },
10703
10703
  "snowflake/llama3-8b": {
10704
10704
  "litellm_provider": "snowflake",
10705
10705
  "max_input_tokens": 8000,
10706
10706
  "max_output_tokens": 8192,
10707
- "max_tokens": 8000,
10707
+ "max_tokens": 8192,
10708
10708
  "mode": "chat"
10709
10709
  },
10710
10710
  "snowflake/llama3-70b": {
10711
10711
  "litellm_provider": "snowflake",
10712
10712
  "max_input_tokens": 8000,
10713
10713
  "max_output_tokens": 8192,
10714
- "max_tokens": 8000,
10714
+ "max_tokens": 8192,
10715
10715
  "mode": "chat"
10716
10716
  },
10717
10717
  "snowflake/llama3.1-8b": {
10718
10718
  "litellm_provider": "snowflake",
10719
10719
  "max_input_tokens": 128000,
10720
10720
  "max_output_tokens": 8192,
10721
- "max_tokens": 128000,
10721
+ "max_tokens": 8192,
10722
10722
  "mode": "chat"
10723
10723
  },
10724
10724
  "snowflake/llama3.1-70b": {
10725
10725
  "litellm_provider": "snowflake",
10726
10726
  "max_input_tokens": 128000,
10727
10727
  "max_output_tokens": 8192,
10728
- "max_tokens": 128000,
10728
+ "max_tokens": 8192,
10729
10729
  "mode": "chat"
10730
10730
  },
10731
10731
  "snowflake/llama3.3-70b": {
10732
10732
  "litellm_provider": "snowflake",
10733
10733
  "max_input_tokens": 128000,
10734
10734
  "max_output_tokens": 8192,
10735
- "max_tokens": 128000,
10735
+ "max_tokens": 8192,
10736
10736
  "mode": "chat"
10737
10737
  },
10738
10738
  "snowflake/snowflake-llama-3.3-70b": {
10739
10739
  "litellm_provider": "snowflake",
10740
10740
  "max_input_tokens": 8000,
10741
10741
  "max_output_tokens": 8192,
10742
- "max_tokens": 8000,
10742
+ "max_tokens": 8192,
10743
10743
  "mode": "chat"
10744
10744
  },
10745
10745
  "snowflake/llama3.1-405b": {
10746
10746
  "litellm_provider": "snowflake",
10747
10747
  "max_input_tokens": 128000,
10748
10748
  "max_output_tokens": 8192,
10749
- "max_tokens": 128000,
10749
+ "max_tokens": 8192,
10750
10750
  "mode": "chat"
10751
10751
  },
10752
10752
  "snowflake/snowflake-llama-3.1-405b": {
10753
10753
  "litellm_provider": "snowflake",
10754
10754
  "max_input_tokens": 8000,
10755
10755
  "max_output_tokens": 8192,
10756
- "max_tokens": 8000,
10756
+ "max_tokens": 8192,
10757
10757
  "mode": "chat"
10758
10758
  },
10759
10759
  "snowflake/llama3.2-1b": {
10760
10760
  "litellm_provider": "snowflake",
10761
10761
  "max_input_tokens": 128000,
10762
10762
  "max_output_tokens": 8192,
10763
- "max_tokens": 128000,
10763
+ "max_tokens": 8192,
10764
10764
  "mode": "chat"
10765
10765
  },
10766
10766
  "snowflake/llama3.2-3b": {
10767
10767
  "litellm_provider": "snowflake",
10768
10768
  "max_input_tokens": 128000,
10769
10769
  "max_output_tokens": 8192,
10770
- "max_tokens": 128000,
10770
+ "max_tokens": 8192,
10771
10771
  "mode": "chat"
10772
10772
  },
10773
10773
  "snowflake/mistral-7b": {
10774
10774
  "litellm_provider": "snowflake",
10775
10775
  "max_input_tokens": 32000,
10776
10776
  "max_output_tokens": 8192,
10777
- "max_tokens": 32000,
10777
+ "max_tokens": 8192,
10778
10778
  "mode": "chat"
10779
10779
  },
10780
10780
  "snowflake/gemma-7b": {
10781
10781
  "litellm_provider": "snowflake",
10782
10782
  "max_input_tokens": 8000,
10783
10783
  "max_output_tokens": 8192,
10784
- "max_tokens": 8000,
10784
+ "max_tokens": 8192,
10785
10785
  "mode": "chat"
10786
10786
  },
10787
10787
  "azure/global/gpt-4o-2024-11-20": {
@@ -11234,7 +11234,7 @@
11234
11234
  "litellm_provider": "watsonx",
11235
11235
  "max_input_tokens": 8192,
11236
11236
  "max_output_tokens": 1024,
11237
- "max_tokens": 8192,
11237
+ "max_tokens": 1024,
11238
11238
  "mode": "chat",
11239
11239
  "output_cost_per_token": 2e-07,
11240
11240
  "supports_audio_input": false,
@@ -12221,7 +12221,7 @@
12221
12221
  "litellm_provider": "meta_llama",
12222
12222
  "max_input_tokens": 10000000,
12223
12223
  "max_output_tokens": 4028,
12224
- "max_tokens": 128000,
12224
+ "max_tokens": 4028,
12225
12225
  "mode": "chat",
12226
12226
  "source": "https://llama.developer.meta.com/docs/models",
12227
12227
  "supported_modalities": [
@@ -12238,7 +12238,7 @@
12238
12238
  "litellm_provider": "meta_llama",
12239
12239
  "max_input_tokens": 1000000,
12240
12240
  "max_output_tokens": 4028,
12241
- "max_tokens": 128000,
12241
+ "max_tokens": 4028,
12242
12242
  "mode": "chat",
12243
12243
  "source": "https://llama.developer.meta.com/docs/models",
12244
12244
  "supported_modalities": [
@@ -12255,7 +12255,7 @@
12255
12255
  "litellm_provider": "meta_llama",
12256
12256
  "max_input_tokens": 128000,
12257
12257
  "max_output_tokens": 4028,
12258
- "max_tokens": 128000,
12258
+ "max_tokens": 4028,
12259
12259
  "mode": "chat",
12260
12260
  "source": "https://llama.developer.meta.com/docs/models",
12261
12261
  "supported_modalities": [
@@ -12271,7 +12271,7 @@
12271
12271
  "litellm_provider": "meta_llama",
12272
12272
  "max_input_tokens": 128000,
12273
12273
  "max_output_tokens": 4028,
12274
- "max_tokens": 128000,
12274
+ "max_tokens": 4028,
12275
12275
  "mode": "chat",
12276
12276
  "source": "https://llama.developer.meta.com/docs/models",
12277
12277
  "supported_modalities": [
@@ -13239,7 +13239,7 @@
13239
13239
  "litellm_provider": "databricks",
13240
13240
  "max_input_tokens": 200000,
13241
13241
  "max_output_tokens": 128000,
13242
- "max_tokens": 200000,
13242
+ "max_tokens": 128000,
13243
13243
  "metadata": {
13244
13244
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
13245
13245
  },
@@ -13615,14 +13615,14 @@
13615
13615
  "litellm_provider": "featherless_ai",
13616
13616
  "max_input_tokens": 32768,
13617
13617
  "max_output_tokens": 4096,
13618
- "max_tokens": 32768,
13618
+ "max_tokens": 4096,
13619
13619
  "mode": "chat"
13620
13620
  },
13621
13621
  "featherless_ai/featherless-ai/Qwerky-QwQ-32B": {
13622
13622
  "litellm_provider": "featherless_ai",
13623
13623
  "max_input_tokens": 32768,
13624
13624
  "max_output_tokens": 4096,
13625
- "max_tokens": 32768,
13625
+ "max_tokens": 4096,
13626
13626
  "mode": "chat"
13627
13627
  },
13628
13628
  "sambanova/Llama-4-Maverick-17B-128E-Instruct": {
@@ -14459,7 +14459,7 @@
14459
14459
  "litellm_provider": "anthropic",
14460
14460
  "max_input_tokens": 1000000,
14461
14461
  "max_output_tokens": 64000,
14462
- "max_tokens": 1000000,
14462
+ "max_tokens": 64000,
14463
14463
  "mode": "chat",
14464
14464
  "output_cost_per_token": 1.5e-05,
14465
14465
  "output_cost_per_token_above_200k_tokens": 2.25e-05,
@@ -16243,7 +16243,7 @@
16243
16243
  "litellm_provider": "watsonx",
16244
16244
  "max_input_tokens": 131072,
16245
16245
  "max_output_tokens": 16384,
16246
- "max_tokens": 131072,
16246
+ "max_tokens": 16384,
16247
16247
  "mode": "chat",
16248
16248
  "output_cost_per_token": 1e-05,
16249
16249
  "supports_audio_input": false,
@@ -16364,7 +16364,7 @@
16364
16364
  "litellm_provider": "bedrock_converse",
16365
16365
  "max_input_tokens": 128000,
16366
16366
  "max_output_tokens": 4096,
16367
- "max_tokens": 128000,
16367
+ "max_tokens": 4096,
16368
16368
  "mode": "chat",
16369
16369
  "output_cost_per_token": 6e-06,
16370
16370
  "supports_function_calling": true,
@@ -16375,7 +16375,7 @@
16375
16375
  "litellm_provider": "bedrock_converse",
16376
16376
  "max_input_tokens": 128000,
16377
16377
  "max_output_tokens": 4096,
16378
- "max_tokens": 128000,
16378
+ "max_tokens": 4096,
16379
16379
  "mode": "chat",
16380
16380
  "output_cost_per_token": 6e-06,
16381
16381
  "supports_function_calling": true,
@@ -16386,7 +16386,7 @@
16386
16386
  "litellm_provider": "dashscope",
16387
16387
  "max_input_tokens": 30720,
16388
16388
  "max_output_tokens": 8192,
16389
- "max_tokens": 32768,
16389
+ "max_tokens": 8192,
16390
16390
  "mode": "chat",
16391
16391
  "output_cost_per_token": 6.4e-06,
16392
16392
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
@@ -16398,7 +16398,7 @@
16398
16398
  "litellm_provider": "dashscope",
16399
16399
  "max_input_tokens": 997952,
16400
16400
  "max_output_tokens": 32768,
16401
- "max_tokens": 1000000,
16401
+ "max_tokens": 32768,
16402
16402
  "mode": "chat",
16403
16403
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
16404
16404
  "supports_function_calling": true,
@@ -16430,7 +16430,7 @@
16430
16430
  "litellm_provider": "dashscope",
16431
16431
  "max_input_tokens": 1000000,
16432
16432
  "max_output_tokens": 16384,
16433
- "max_tokens": 1000000,
16433
+ "max_tokens": 16384,
16434
16434
  "mode": "chat",
16435
16435
  "output_cost_per_reasoning_token": 5e-07,
16436
16436
  "output_cost_per_token": 2e-07,
@@ -16443,7 +16443,7 @@
16443
16443
  "litellm_provider": "dashscope",
16444
16444
  "max_input_tokens": 129024,
16445
16445
  "max_output_tokens": 16384,
16446
- "max_tokens": 131072,
16446
+ "max_tokens": 16384,
16447
16447
  "mode": "chat",
16448
16448
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
16449
16449
  "supports_function_calling": true,
@@ -17076,7 +17076,7 @@
17076
17076
  "litellm_provider": "openrouter",
17077
17077
  "max_input_tokens": 8192,
17078
17078
  "max_output_tokens": 2048,
17079
- "max_tokens": 8192,
17079
+ "max_tokens": 2048,
17080
17080
  "mode": "chat",
17081
17081
  "output_cost_per_token": 6.3e-07,
17082
17082
  "supports_tool_choice": true,
@@ -17087,7 +17087,7 @@
17087
17087
  "litellm_provider": "fireworks_ai",
17088
17088
  "max_input_tokens": 131072,
17089
17089
  "max_output_tokens": 16384,
17090
- "max_tokens": 131072,
17090
+ "max_tokens": 16384,
17091
17091
  "mode": "chat",
17092
17092
  "output_cost_per_token": 2.5e-06,
17093
17093
  "source": "https://fireworks.ai/models/fireworks/kimi-k2-instruct",
@@ -17220,7 +17220,7 @@
17220
17220
  "litellm_provider": "lambda_ai",
17221
17221
  "max_input_tokens": 131072,
17222
17222
  "max_output_tokens": 8192,
17223
- "max_tokens": 131072,
17223
+ "max_tokens": 8192,
17224
17224
  "mode": "chat",
17225
17225
  "output_cost_per_token": 1e-07,
17226
17226
  "supports_function_calling": true,
@@ -17233,7 +17233,7 @@
17233
17233
  "litellm_provider": "lambda_ai",
17234
17234
  "max_input_tokens": 16384,
17235
17235
  "max_output_tokens": 8192,
17236
- "max_tokens": 16384,
17236
+ "max_tokens": 8192,
17237
17237
  "mode": "chat",
17238
17238
  "output_cost_per_token": 1e-07,
17239
17239
  "supports_function_calling": true,
@@ -17508,7 +17508,7 @@
17508
17508
  "litellm_provider": "vertex_ai-llama_models",
17509
17509
  "max_input_tokens": 128000,
17510
17510
  "max_output_tokens": 2048,
17511
- "max_tokens": 128000,
17511
+ "max_tokens": 2048,
17512
17512
  "metadata": {
17513
17513
  "notes": "VertexAI states that The Llama 3.1 API service for llama-3.1-70b-instruct-maas and llama-3.1-8b-instruct-maas are in public preview and at no cost."
17514
17514
  },
@@ -17524,7 +17524,7 @@
17524
17524
  "litellm_provider": "vertex_ai-llama_models",
17525
17525
  "max_input_tokens": 128000,
17526
17526
  "max_output_tokens": 2048,
17527
- "max_tokens": 128000,
17527
+ "max_tokens": 2048,
17528
17528
  "mode": "chat",
17529
17529
  "output_cost_per_token": 0.0,
17530
17530
  "source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas",
@@ -17537,7 +17537,7 @@
17537
17537
  "litellm_provider": "vertex_ai-llama_models",
17538
17538
  "max_input_tokens": 128000,
17539
17539
  "max_output_tokens": 2048,
17540
- "max_tokens": 128000,
17540
+ "max_tokens": 2048,
17541
17541
  "mode": "chat",
17542
17542
  "output_cost_per_token": 1.6e-05,
17543
17543
  "source": "https://console.cloud.google.com/vertex-ai/publishers/meta/model-garden/llama-3.2-90b-vision-instruct-maas",
@@ -18769,7 +18769,7 @@
18769
18769
  "litellm_provider": "oci",
18770
18770
  "max_input_tokens": 512000,
18771
18771
  "max_output_tokens": 4000,
18772
- "max_tokens": 512000,
18772
+ "max_tokens": 4000,
18773
18773
  "mode": "chat",
18774
18774
  "output_cost_per_token": 7.2e-07,
18775
18775
  "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing",
@@ -18781,7 +18781,7 @@
18781
18781
  "litellm_provider": "oci",
18782
18782
  "max_input_tokens": 192000,
18783
18783
  "max_output_tokens": 4000,
18784
- "max_tokens": 192000,
18784
+ "max_tokens": 4000,
18785
18785
  "mode": "chat",
18786
18786
  "output_cost_per_token": 7.2e-07,
18787
18787
  "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing",
@@ -18793,7 +18793,7 @@
18793
18793
  "litellm_provider": "oci",
18794
18794
  "max_input_tokens": 128000,
18795
18795
  "max_output_tokens": 4000,
18796
- "max_tokens": 128000,
18796
+ "max_tokens": 4000,
18797
18797
  "mode": "chat",
18798
18798
  "output_cost_per_token": 7.2e-07,
18799
18799
  "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing",
@@ -18805,7 +18805,7 @@
18805
18805
  "litellm_provider": "oci",
18806
18806
  "max_input_tokens": 128000,
18807
18807
  "max_output_tokens": 4000,
18808
- "max_tokens": 128000,
18808
+ "max_tokens": 4000,
18809
18809
  "mode": "chat",
18810
18810
  "output_cost_per_token": 2e-06,
18811
18811
  "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing",
@@ -18817,7 +18817,7 @@
18817
18817
  "litellm_provider": "oci",
18818
18818
  "max_input_tokens": 128000,
18819
18819
  "max_output_tokens": 4000,
18820
- "max_tokens": 128000,
18820
+ "max_tokens": 4000,
18821
18821
  "mode": "chat",
18822
18822
  "output_cost_per_token": 1.068e-05,
18823
18823
  "source": "https://www.oracle.com/artificial-intelligence/generative-ai/generative-ai-service/pricing",
@@ -20331,7 +20331,7 @@
20331
20331
  "litellm_provider": "openrouter",
20332
20332
  "max_input_tokens": 163840,
20333
20333
  "max_output_tokens": 163840,
20334
- "max_tokens": 8192,
20334
+ "max_tokens": 163840,
20335
20335
  "mode": "chat",
20336
20336
  "output_cost_per_token": 8e-07,
20337
20337
  "supports_assistant_prefill": true,
@@ -20774,7 +20774,7 @@
20774
20774
  "litellm_provider": "vercel_ai_gateway",
20775
20775
  "max_input_tokens": 262144,
20776
20776
  "max_output_tokens": 66536,
20777
- "max_tokens": 262144,
20777
+ "max_tokens": 66536,
20778
20778
  "mode": "chat",
20779
20779
  "output_cost_per_token": 1.6e-06
20780
20780
  },
@@ -20792,7 +20792,7 @@
20792
20792
  "litellm_provider": "vercel_ai_gateway",
20793
20793
  "max_input_tokens": 1048576,
20794
20794
  "max_output_tokens": 65536,
20795
- "max_tokens": 1048576,
20795
+ "max_tokens": 65536,
20796
20796
  "mode": "chat",
20797
20797
  "output_cost_per_token": 1e-05
20798
20798
  },
@@ -20801,7 +20801,7 @@
20801
20801
  "litellm_provider": "vercel_ai_gateway",
20802
20802
  "max_input_tokens": 128000,
20803
20803
  "max_output_tokens": 8192,
20804
- "max_tokens": 128000,
20804
+ "max_tokens": 8192,
20805
20805
  "mode": "chat",
20806
20806
  "output_cost_per_token": 9e-07
20807
20807
  },
@@ -20810,7 +20810,7 @@
20810
20810
  "litellm_provider": "vercel_ai_gateway",
20811
20811
  "max_input_tokens": 300000,
20812
20812
  "max_output_tokens": 8192,
20813
- "max_tokens": 300000,
20813
+ "max_tokens": 8192,
20814
20814
  "mode": "chat",
20815
20815
  "output_cost_per_token": 2.4e-07
20816
20816
  },
@@ -20819,7 +20819,7 @@
20819
20819
  "litellm_provider": "vercel_ai_gateway",
20820
20820
  "max_input_tokens": 131072,
20821
20821
  "max_output_tokens": 8192,
20822
- "max_tokens": 131072,
20822
+ "max_tokens": 8192,
20823
20823
  "mode": "chat",
20824
20824
  "output_cost_per_token": 3e-07
20825
20825
  },
@@ -20828,7 +20828,7 @@
20828
20828
  "litellm_provider": "vercel_ai_gateway",
20829
20829
  "max_input_tokens": 128000,
20830
20830
  "max_output_tokens": 8192,
20831
- "max_tokens": 128000,
20831
+ "max_tokens": 8192,
20832
20832
  "mode": "chat",
20833
20833
  "output_cost_per_token": 1e-07
20834
20834
  },
@@ -20837,7 +20837,7 @@
20837
20837
  "litellm_provider": "vercel_ai_gateway",
20838
20838
  "max_input_tokens": 32000,
20839
20839
  "max_output_tokens": 4000,
20840
- "max_tokens": 32000,
20840
+ "max_tokens": 4000,
20841
20841
  "mode": "chat",
20842
20842
  "output_cost_per_token": 3e-07
20843
20843
  },
@@ -20846,7 +20846,7 @@
20846
20846
  "litellm_provider": "vercel_ai_gateway",
20847
20847
  "max_input_tokens": 1000000,
20848
20848
  "max_output_tokens": 65536,
20849
- "max_tokens": 1000000,
20849
+ "max_tokens": 65536,
20850
20850
  "mode": "chat",
20851
20851
  "output_cost_per_token": 2.5e-06
20852
20852
  },
@@ -20855,7 +20855,7 @@
20855
20855
  "litellm_provider": "vercel_ai_gateway",
20856
20856
  "max_input_tokens": 32000,
20857
20857
  "max_output_tokens": 16384,
20858
- "max_tokens": 32000,
20858
+ "max_tokens": 16384,
20859
20859
  "mode": "chat",
20860
20860
  "output_cost_per_token": 1e-06
20861
20861
  },
@@ -20882,7 +20882,7 @@
20882
20882
  "litellm_provider": "vercel_ai_gateway",
20883
20883
  "max_input_tokens": 131072,
20884
20884
  "max_output_tokens": 4000,
20885
- "max_tokens": 131072,
20885
+ "max_tokens": 4000,
20886
20886
  "mode": "chat",
20887
20887
  "output_cost_per_token": 1e-05
20888
20888
  },
@@ -20900,7 +20900,7 @@
20900
20900
  "litellm_provider": "vercel_ai_gateway",
20901
20901
  "max_input_tokens": 128000,
20902
20902
  "max_output_tokens": 8192,
20903
- "max_tokens": 128000,
20903
+ "max_tokens": 8192,
20904
20904
  "mode": "chat",
20905
20905
  "output_cost_per_token": 7.2e-07
20906
20906
  },
@@ -20918,7 +20918,7 @@
20918
20918
  "litellm_provider": "vercel_ai_gateway",
20919
20919
  "max_input_tokens": 40960,
20920
20920
  "max_output_tokens": 16384,
20921
- "max_tokens": 40960,
20921
+ "max_tokens": 16384,
20922
20922
  "mode": "chat",
20923
20923
  "output_cost_per_token": 6e-07
20924
20924
  },
@@ -20936,7 +20936,7 @@
20936
20936
  "litellm_provider": "vercel_ai_gateway",
20937
20937
  "max_input_tokens": 128000,
20938
20938
  "max_output_tokens": 32768,
20939
- "max_tokens": 128000,
20939
+ "max_tokens": 32768,
20940
20940
  "mode": "chat",
20941
20941
  "output_cost_per_token": 1.5e-05
20942
20942
  },
@@ -20947,7 +20947,7 @@
20947
20947
  "litellm_provider": "vercel_ai_gateway",
20948
20948
  "max_input_tokens": 200000,
20949
20949
  "max_output_tokens": 100000,
20950
- "max_tokens": 200000,
20950
+ "max_tokens": 100000,
20951
20951
  "mode": "chat",
20952
20952
  "output_cost_per_token": 4.4e-06
20953
20953
  },
@@ -20956,7 +20956,7 @@
20956
20956
  "litellm_provider": "vercel_ai_gateway",
20957
20957
  "max_input_tokens": 128000,
20958
20958
  "max_output_tokens": 64000,
20959
- "max_tokens": 128000,
20959
+ "max_tokens": 64000,
20960
20960
  "mode": "chat",
20961
20961
  "output_cost_per_token": 5e-06
20962
20962
  },
@@ -20974,7 +20974,7 @@
20974
20974
  "litellm_provider": "vercel_ai_gateway",
20975
20975
  "max_input_tokens": 40960,
20976
20976
  "max_output_tokens": 16384,
20977
- "max_tokens": 40960,
20977
+ "max_tokens": 16384,
20978
20978
  "mode": "chat",
20979
20979
  "output_cost_per_token": 3e-07
20980
20980
  },
@@ -20983,7 +20983,7 @@
20983
20983
  "litellm_provider": "vercel_ai_gateway",
20984
20984
  "max_input_tokens": 128000,
20985
20985
  "max_output_tokens": 96000,
20986
- "max_tokens": 128000,
20986
+ "max_tokens": 96000,
20987
20987
  "mode": "chat",
20988
20988
  "output_cost_per_token": 1.1e-06
20989
20989
  },
@@ -20992,7 +20992,7 @@
20992
20992
  "litellm_provider": "vercel_ai_gateway",
20993
20993
  "max_input_tokens": 128000,
20994
20994
  "max_output_tokens": 4096,
20995
- "max_tokens": 128000,
20995
+ "max_tokens": 4096,
20996
20996
  "mode": "chat",
20997
20997
  "output_cost_per_token": 3e-05
20998
20998
  },
@@ -21001,7 +21001,7 @@
21001
21001
  "litellm_provider": "vercel_ai_gateway",
21002
21002
  "max_input_tokens": 32000,
21003
21003
  "max_output_tokens": 4000,
21004
- "max_tokens": 32000,
21004
+ "max_tokens": 4000,
21005
21005
  "mode": "chat",
21006
21006
  "output_cost_per_token": 6e-06
21007
21007
  },
@@ -21010,7 +21010,7 @@
21010
21010
  "litellm_provider": "vercel_ai_gateway",
21011
21011
  "max_input_tokens": 200000,
21012
21012
  "max_output_tokens": 8000,
21013
- "max_tokens": 200000,
21013
+ "max_tokens": 8000,
21014
21014
  "mode": "chat",
21015
21015
  "output_cost_per_token": 1.5e-05
21016
21016
  },
@@ -21019,7 +21019,7 @@
21019
21019
  "litellm_provider": "vercel_ai_gateway",
21020
21020
  "max_input_tokens": 128000,
21021
21021
  "max_output_tokens": 8192,
21022
- "max_tokens": 128000,
21022
+ "max_tokens": 8192,
21023
21023
  "mode": "chat",
21024
21024
  "output_cost_per_token": 7.2e-07
21025
21025
  },
@@ -21046,7 +21046,7 @@
21046
21046
  "litellm_provider": "vercel_ai_gateway",
21047
21047
  "max_input_tokens": 128000,
21048
21048
  "max_output_tokens": 4000,
21049
- "max_tokens": 128000,
21049
+ "max_tokens": 4000,
21050
21050
  "mode": "chat",
21051
21051
  "output_cost_per_token": 6e-06
21052
21052
  },
@@ -21057,7 +21057,7 @@
21057
21057
  "litellm_provider": "vercel_ai_gateway",
21058
21058
  "max_input_tokens": 200000,
21059
21059
  "max_output_tokens": 8192,
21060
- "max_tokens": 200000,
21060
+ "max_tokens": 8192,
21061
21061
  "mode": "chat",
21062
21062
  "output_cost_per_token": 1.5e-05
21063
21063
  },
@@ -21066,7 +21066,7 @@
21066
21066
  "litellm_provider": "vercel_ai_gateway",
21067
21067
  "max_input_tokens": 128000,
21068
21068
  "max_output_tokens": 8192,
21069
- "max_tokens": 128000,
21069
+ "max_tokens": 8192,
21070
21070
  "mode": "chat",
21071
21071
  "output_cost_per_token": 1.4e-07
21072
21072
  },
@@ -21075,7 +21075,7 @@
21075
21075
  "litellm_provider": "vercel_ai_gateway",
21076
21076
  "max_input_tokens": 128000,
21077
21077
  "max_output_tokens": 4096,
21078
- "max_tokens": 128000,
21078
+ "max_tokens": 4096,
21079
21079
  "mode": "chat",
21080
21080
  "output_cost_per_token": 6e-07
21081
21081
  },
@@ -21084,7 +21084,7 @@
21084
21084
  "litellm_provider": "vercel_ai_gateway",
21085
21085
  "max_input_tokens": 32768,
21086
21086
  "max_output_tokens": 16384,
21087
- "max_tokens": 32768,
21087
+ "max_tokens": 16384,
21088
21088
  "mode": "chat",
21089
21089
  "output_cost_per_token": 1.9e-06
21090
21090
  },
@@ -21093,7 +21093,7 @@
21093
21093
  "litellm_provider": "vercel_ai_gateway",
21094
21094
  "max_input_tokens": 65536,
21095
21095
  "max_output_tokens": 2048,
21096
- "max_tokens": 65536,
21096
+ "max_tokens": 2048,
21097
21097
  "mode": "chat",
21098
21098
  "output_cost_per_token": 1.2e-06
21099
21099
  },
@@ -21111,7 +21111,7 @@
21111
21111
  "litellm_provider": "vercel_ai_gateway",
21112
21112
  "max_input_tokens": 131000,
21113
21113
  "max_output_tokens": 131072,
21114
- "max_tokens": 131000,
21114
+ "max_tokens": 131072,
21115
21115
  "mode": "chat",
21116
21116
  "output_cost_per_token": 8e-08
21117
21117
  },
@@ -21122,7 +21122,7 @@
21122
21122
  "litellm_provider": "vercel_ai_gateway",
21123
21123
  "max_input_tokens": 200000,
21124
21124
  "max_output_tokens": 4096,
21125
- "max_tokens": 200000,
21125
+ "max_tokens": 4096,
21126
21126
  "mode": "chat",
21127
21127
  "output_cost_per_token": 7.5e-05
21128
21128
  },
@@ -21142,7 +21142,7 @@
21142
21142
  "litellm_provider": "vercel_ai_gateway",
21143
21143
  "max_input_tokens": 128000,
21144
21144
  "max_output_tokens": 16384,
21145
- "max_tokens": 128000,
21145
+ "max_tokens": 16384,
21146
21146
  "mode": "chat",
21147
21147
  "output_cost_per_token": 1e-05
21148
21148
  },
@@ -21153,7 +21153,7 @@
21153
21153
  "litellm_provider": "vercel_ai_gateway",
21154
21154
  "max_input_tokens": 200000,
21155
21155
  "max_output_tokens": 100000,
21156
- "max_tokens": 200000,
21156
+ "max_tokens": 100000,
21157
21157
  "mode": "chat",
21158
21158
  "output_cost_per_token": 4.4e-06
21159
21159
  },
@@ -21162,7 +21162,7 @@
21162
21162
  "litellm_provider": "vercel_ai_gateway",
21163
21163
  "max_input_tokens": 128000,
21164
21164
  "max_output_tokens": 4000,
21165
- "max_tokens": 128000,
21165
+ "max_tokens": 4000,
21166
21166
  "mode": "chat",
21167
21167
  "output_cost_per_token": 1e-07
21168
21168
  },
@@ -21173,7 +21173,7 @@
21173
21173
  "litellm_provider": "vercel_ai_gateway",
21174
21174
  "max_input_tokens": 200000,
21175
21175
  "max_output_tokens": 100000,
21176
- "max_tokens": 200000,
21176
+ "max_tokens": 100000,
21177
21177
  "mode": "chat",
21178
21178
  "output_cost_per_token": 8e-06
21179
21179
  },
@@ -21182,7 +21182,7 @@
21182
21182
  "litellm_provider": "vercel_ai_gateway",
21183
21183
  "max_input_tokens": 128000,
21184
21184
  "max_output_tokens": 32000,
21185
- "max_tokens": 128000,
21185
+ "max_tokens": 32000,
21186
21186
  "mode": "chat",
21187
21187
  "output_cost_per_token": 1.5e-05
21188
21188
  },
@@ -21200,7 +21200,7 @@
21200
21200
  "litellm_provider": "vercel_ai_gateway",
21201
21201
  "max_input_tokens": 300000,
21202
21202
  "max_output_tokens": 8192,
21203
- "max_tokens": 300000,
21203
+ "max_tokens": 8192,
21204
21204
  "mode": "chat",
21205
21205
  "output_cost_per_token": 3.2e-06
21206
21206
  },
@@ -21209,7 +21209,7 @@
21209
21209
  "litellm_provider": "vercel_ai_gateway",
21210
21210
  "max_input_tokens": 32768,
21211
21211
  "max_output_tokens": 16384,
21212
- "max_tokens": 32768,
21212
+ "max_tokens": 16384,
21213
21213
  "mode": "chat",
21214
21214
  "output_cost_per_token": 1.2e-06
21215
21215
  },
@@ -21218,7 +21218,7 @@
21218
21218
  "litellm_provider": "vercel_ai_gateway",
21219
21219
  "max_input_tokens": 16385,
21220
21220
  "max_output_tokens": 4096,
21221
- "max_tokens": 16385,
21221
+ "max_tokens": 4096,
21222
21222
  "mode": "chat",
21223
21223
  "output_cost_per_token": 1.5e-06
21224
21224
  },
@@ -21227,7 +21227,7 @@
21227
21227
  "litellm_provider": "vercel_ai_gateway",
21228
21228
  "max_input_tokens": 256000,
21229
21229
  "max_output_tokens": 4000,
21230
- "max_tokens": 256000,
21230
+ "max_tokens": 4000,
21231
21231
  "mode": "chat",
21232
21232
  "output_cost_per_token": 9e-07
21233
21233
  },
@@ -21236,7 +21236,7 @@
21236
21236
  "litellm_provider": "vercel_ai_gateway",
21237
21237
  "max_input_tokens": 128000,
21238
21238
  "max_output_tokens": 8192,
21239
- "max_tokens": 128000,
21239
+ "max_tokens": 8192,
21240
21240
  "mode": "chat",
21241
21241
  "output_cost_per_token": 1.6e-07
21242
21242
  },
@@ -21272,7 +21272,7 @@
21272
21272
  "litellm_provider": "vercel_ai_gateway",
21273
21273
  "max_input_tokens": 1048576,
21274
21274
  "max_output_tokens": 8192,
21275
- "max_tokens": 1048576,
21275
+ "max_tokens": 8192,
21276
21276
  "mode": "chat",
21277
21277
  "output_cost_per_token": 3e-07
21278
21278
  },
@@ -21281,7 +21281,7 @@
21281
21281
  "litellm_provider": "vercel_ai_gateway",
21282
21282
  "max_input_tokens": 128000,
21283
21283
  "max_output_tokens": 4000,
21284
- "max_tokens": 128000,
21284
+ "max_tokens": 4000,
21285
21285
  "mode": "chat",
21286
21286
  "output_cost_per_token": 4e-08
21287
21287
  },
@@ -21290,7 +21290,7 @@
21290
21290
  "litellm_provider": "vercel_ai_gateway",
21291
21291
  "max_input_tokens": 127000,
21292
21292
  "max_output_tokens": 8000,
21293
- "max_tokens": 127000,
21293
+ "max_tokens": 8000,
21294
21294
  "mode": "chat",
21295
21295
  "output_cost_per_token": 8e-06
21296
21296
  },
@@ -21310,7 +21310,7 @@
21310
21310
  "litellm_provider": "vercel_ai_gateway",
21311
21311
  "max_input_tokens": 200000,
21312
21312
  "max_output_tokens": 4096,
21313
- "max_tokens": 200000,
21313
+ "max_tokens": 4096,
21314
21314
  "mode": "chat",
21315
21315
  "output_cost_per_token": 1.25e-06
21316
21316
  },
@@ -21321,7 +21321,7 @@
21321
21321
  "litellm_provider": "vercel_ai_gateway",
21322
21322
  "max_input_tokens": 200000,
21323
21323
  "max_output_tokens": 100000,
21324
- "max_tokens": 200000,
21324
+ "max_tokens": 100000,
21325
21325
  "mode": "chat",
21326
21326
  "output_cost_per_token": 6e-05
21327
21327
  },
@@ -21330,7 +21330,7 @@
21330
21330
  "litellm_provider": "vercel_ai_gateway",
21331
21331
  "max_input_tokens": 128000,
21332
21332
  "max_output_tokens": 8192,
21333
- "max_tokens": 128000,
21333
+ "max_tokens": 8192,
21334
21334
  "mode": "chat",
21335
21335
  "output_cost_per_token": 2.19e-06
21336
21336
  },
@@ -21350,7 +21350,7 @@
21350
21350
  "litellm_provider": "vercel_ai_gateway",
21351
21351
  "max_input_tokens": 1047576,
21352
21352
  "max_output_tokens": 32768,
21353
- "max_tokens": 1047576,
21353
+ "max_tokens": 32768,
21354
21354
  "mode": "chat",
21355
21355
  "output_cost_per_token": 1.6e-06
21356
21356
  },
@@ -21361,7 +21361,7 @@
21361
21361
  "litellm_provider": "vercel_ai_gateway",
21362
21362
  "max_input_tokens": 128000,
21363
21363
  "max_output_tokens": 16384,
21364
- "max_tokens": 128000,
21364
+ "max_tokens": 16384,
21365
21365
  "mode": "chat",
21366
21366
  "output_cost_per_token": 6e-07
21367
21367
  },
@@ -21370,7 +21370,7 @@
21370
21370
  "litellm_provider": "vercel_ai_gateway",
21371
21371
  "max_input_tokens": 40960,
21372
21372
  "max_output_tokens": 16384,
21373
- "max_tokens": 40960,
21373
+ "max_tokens": 16384,
21374
21374
  "mode": "chat",
21375
21375
  "output_cost_per_token": 2.4e-07
21376
21376
  },
@@ -21381,7 +21381,7 @@
21381
21381
  "litellm_provider": "vercel_ai_gateway",
21382
21382
  "max_input_tokens": 200000,
21383
21383
  "max_output_tokens": 32000,
21384
- "max_tokens": 200000,
21384
+ "max_tokens": 32000,
21385
21385
  "mode": "chat",
21386
21386
  "output_cost_per_token": 7.5e-05
21387
21387
  },
@@ -21399,7 +21399,7 @@
21399
21399
  "litellm_provider": "vercel_ai_gateway",
21400
21400
  "max_input_tokens": 127000,
21401
21401
  "max_output_tokens": 8000,
21402
- "max_tokens": 127000,
21402
+ "max_tokens": 8000,
21403
21403
  "mode": "chat",
21404
21404
  "output_cost_per_token": 5e-06
21405
21405
  },
@@ -21410,7 +21410,7 @@
21410
21410
  "litellm_provider": "vercel_ai_gateway",
21411
21411
  "max_input_tokens": 200000,
21412
21412
  "max_output_tokens": 8192,
21413
- "max_tokens": 200000,
21413
+ "max_tokens": 8192,
21414
21414
  "mode": "chat",
21415
21415
  "output_cost_per_token": 4e-06
21416
21416
  },
@@ -21419,7 +21419,7 @@
21419
21419
  "litellm_provider": "vercel_ai_gateway",
21420
21420
  "max_input_tokens": 256000,
21421
21421
  "max_output_tokens": 8000,
21422
- "max_tokens": 256000,
21422
+ "max_tokens": 8000,
21423
21423
  "mode": "chat",
21424
21424
  "output_cost_per_token": 1e-05
21425
21425
  },
@@ -21437,7 +21437,7 @@
21437
21437
  "litellm_provider": "vercel_ai_gateway",
21438
21438
  "max_input_tokens": 128000,
21439
21439
  "max_output_tokens": 8192,
21440
- "max_tokens": 128000,
21440
+ "max_tokens": 8192,
21441
21441
  "mode": "chat",
21442
21442
  "output_cost_per_token": 1.5e-07
21443
21443
  },
@@ -21448,7 +21448,7 @@
21448
21448
  "litellm_provider": "vercel_ai_gateway",
21449
21449
  "max_input_tokens": 1047576,
21450
21450
  "max_output_tokens": 32768,
21451
- "max_tokens": 1047576,
21451
+ "max_tokens": 32768,
21452
21452
  "mode": "chat",
21453
21453
  "output_cost_per_token": 4e-07
21454
21454
  },
@@ -21459,7 +21459,7 @@
21459
21459
  "litellm_provider": "vercel_ai_gateway",
21460
21460
  "max_input_tokens": 200000,
21461
21461
  "max_output_tokens": 64000,
21462
- "max_tokens": 200000,
21462
+ "max_tokens": 64000,
21463
21463
  "mode": "chat",
21464
21464
  "output_cost_per_token": 1.5e-05
21465
21465
  },
@@ -21468,7 +21468,7 @@
21468
21468
  "litellm_provider": "vercel_ai_gateway",
21469
21469
  "max_input_tokens": 127000,
21470
21470
  "max_output_tokens": 8000,
21471
- "max_tokens": 127000,
21471
+ "max_tokens": 8000,
21472
21472
  "mode": "chat",
21473
21473
  "output_cost_per_token": 1e-06
21474
21474
  },
@@ -21477,7 +21477,7 @@
21477
21477
  "litellm_provider": "vercel_ai_gateway",
21478
21478
  "max_input_tokens": 131072,
21479
21479
  "max_output_tokens": 8192,
21480
- "max_tokens": 131072,
21480
+ "max_tokens": 8192,
21481
21481
  "mode": "chat",
21482
21482
  "output_cost_per_token": 6e-07
21483
21483
  },
@@ -21513,7 +21513,7 @@
21513
21513
  "litellm_provider": "vercel_ai_gateway",
21514
21514
  "max_input_tokens": 128000,
21515
21515
  "max_output_tokens": 8192,
21516
- "max_tokens": 128000,
21516
+ "max_tokens": 8192,
21517
21517
  "mode": "chat",
21518
21518
  "output_cost_per_token": 7.2e-07
21519
21519
  },
@@ -21522,7 +21522,7 @@
21522
21522
  "litellm_provider": "vercel_ai_gateway",
21523
21523
  "max_input_tokens": 128000,
21524
21524
  "max_output_tokens": 4096,
21525
- "max_tokens": 128000,
21525
+ "max_tokens": 4096,
21526
21526
  "mode": "chat",
21527
21527
  "output_cost_per_token": 1e-05
21528
21528
  },
@@ -21531,7 +21531,7 @@
21531
21531
  "litellm_provider": "vercel_ai_gateway",
21532
21532
  "max_input_tokens": 8192,
21533
21533
  "max_output_tokens": 4096,
21534
- "max_tokens": 8192,
21534
+ "max_tokens": 4096,
21535
21535
  "mode": "chat",
21536
21536
  "output_cost_per_token": 2e-06
21537
21537
  },
@@ -21551,7 +21551,7 @@
21551
21551
  "litellm_provider": "vercel_ai_gateway",
21552
21552
  "max_input_tokens": 200000,
21553
21553
  "max_output_tokens": 64000,
21554
- "max_tokens": 200000,
21554
+ "max_tokens": 64000,
21555
21555
  "mode": "chat",
21556
21556
  "output_cost_per_token": 1.5e-05
21557
21557
  },
@@ -21560,7 +21560,7 @@
21560
21560
  "litellm_provider": "vercel_ai_gateway",
21561
21561
  "max_input_tokens": 1048576,
21562
21562
  "max_output_tokens": 8192,
21563
- "max_tokens": 1048576,
21563
+ "max_tokens": 8192,
21564
21564
  "mode": "chat",
21565
21565
  "output_cost_per_token": 6e-07
21566
21566
  },
@@ -21569,7 +21569,7 @@
21569
21569
  "litellm_provider": "vercel_ai_gateway",
21570
21570
  "max_input_tokens": 128000,
21571
21571
  "max_output_tokens": 4000,
21572
- "max_tokens": 128000,
21572
+ "max_tokens": 4000,
21573
21573
  "mode": "chat",
21574
21574
  "output_cost_per_token": 1.5e-07
21575
21575
  },
@@ -21578,7 +21578,7 @@
21578
21578
  "litellm_provider": "vercel_ai_gateway",
21579
21579
  "max_input_tokens": 128000,
21580
21580
  "max_output_tokens": 64000,
21581
- "max_tokens": 128000,
21581
+ "max_tokens": 64000,
21582
21582
  "mode": "chat",
21583
21583
  "output_cost_per_token": 1.5e-06
21584
21584
  },
@@ -21587,7 +21587,7 @@
21587
21587
  "litellm_provider": "vercel_ai_gateway",
21588
21588
  "max_input_tokens": 131072,
21589
21589
  "max_output_tokens": 16384,
21590
- "max_tokens": 131072,
21590
+ "max_tokens": 16384,
21591
21591
  "mode": "chat",
21592
21592
  "output_cost_per_token": 2.2e-06
21593
21593
  },
@@ -21596,7 +21596,7 @@
21596
21596
  "litellm_provider": "vercel_ai_gateway",
21597
21597
  "max_input_tokens": 40960,
21598
21598
  "max_output_tokens": 16384,
21599
- "max_tokens": 40960,
21599
+ "max_tokens": 16384,
21600
21600
  "mode": "chat",
21601
21601
  "output_cost_per_token": 3e-07
21602
21602
  },
@@ -21607,7 +21607,7 @@
21607
21607
  "litellm_provider": "vercel_ai_gateway",
21608
21608
  "max_input_tokens": 1047576,
21609
21609
  "max_output_tokens": 32768,
21610
- "max_tokens": 1047576,
21610
+ "max_tokens": 32768,
21611
21611
  "mode": "chat",
21612
21612
  "output_cost_per_token": 8e-06
21613
21613
  },
@@ -21943,7 +21943,7 @@
21943
21943
  "litellm_provider": "dashscope",
21944
21944
  "max_input_tokens": 258048,
21945
21945
  "max_output_tokens": 65536,
21946
- "max_tokens": 262144,
21946
+ "max_tokens": 65536,
21947
21947
  "mode": "chat",
21948
21948
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
21949
21949
  "supports_function_calling": true,
@@ -21981,7 +21981,7 @@
21981
21981
  "litellm_provider": "dashscope",
21982
21982
  "max_input_tokens": 129024,
21983
21983
  "max_output_tokens": 16384,
21984
- "max_tokens": 131072,
21984
+ "max_tokens": 16384,
21985
21985
  "mode": "chat",
21986
21986
  "output_cost_per_token": 1.2e-06,
21987
21987
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
@@ -22023,7 +22023,7 @@
22023
22023
  "litellm_provider": "dashscope",
22024
22024
  "max_input_tokens": 1000000,
22025
22025
  "max_output_tokens": 16384,
22026
- "max_tokens": 1000000,
22026
+ "max_tokens": 16384,
22027
22027
  "mode": "chat",
22028
22028
  "output_cost_per_token": 1.5e-06,
22029
22029
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
@@ -22035,7 +22035,7 @@
22035
22035
  "litellm_provider": "dashscope",
22036
22036
  "max_input_tokens": 997952,
22037
22037
  "max_output_tokens": 65536,
22038
- "max_tokens": 1000000,
22038
+ "max_tokens": 65536,
22039
22039
  "mode": "chat",
22040
22040
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
22041
22041
  "supports_function_calling": true,
@@ -22084,7 +22084,7 @@
22084
22084
  "litellm_provider": "dashscope",
22085
22085
  "max_input_tokens": 997952,
22086
22086
  "max_output_tokens": 65536,
22087
- "max_tokens": 1000000,
22087
+ "max_tokens": 65536,
22088
22088
  "mode": "chat",
22089
22089
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
22090
22090
  "supports_function_calling": true,
@@ -22129,7 +22129,7 @@
22129
22129
  "litellm_provider": "dashscope",
22130
22130
  "max_input_tokens": 997952,
22131
22131
  "max_output_tokens": 65536,
22132
- "max_tokens": 1000000,
22132
+ "max_tokens": 65536,
22133
22133
  "mode": "chat",
22134
22134
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
22135
22135
  "supports_function_calling": true,
@@ -22178,7 +22178,7 @@
22178
22178
  "litellm_provider": "dashscope",
22179
22179
  "max_input_tokens": 997952,
22180
22180
  "max_output_tokens": 65536,
22181
- "max_tokens": 1000000,
22181
+ "max_tokens": 65536,
22182
22182
  "mode": "chat",
22183
22183
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
22184
22184
  "supports_function_calling": true,
@@ -22223,7 +22223,7 @@
22223
22223
  "litellm_provider": "dashscope",
22224
22224
  "max_input_tokens": 997952,
22225
22225
  "max_output_tokens": 32768,
22226
- "max_tokens": 1000000,
22226
+ "max_tokens": 32768,
22227
22227
  "mode": "chat",
22228
22228
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
22229
22229
  "supports_function_calling": true,
@@ -22254,7 +22254,7 @@
22254
22254
  "litellm_provider": "dashscope",
22255
22255
  "max_input_tokens": 997952,
22256
22256
  "max_output_tokens": 32768,
22257
- "max_tokens": 1000000,
22257
+ "max_tokens": 32768,
22258
22258
  "mode": "chat",
22259
22259
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
22260
22260
  "supports_function_calling": true,
@@ -22286,7 +22286,7 @@
22286
22286
  "litellm_provider": "dashscope",
22287
22287
  "max_input_tokens": 129024,
22288
22288
  "max_output_tokens": 16384,
22289
- "max_tokens": 131072,
22289
+ "max_tokens": 16384,
22290
22290
  "mode": "chat",
22291
22291
  "output_cost_per_reasoning_token": 4e-06,
22292
22292
  "output_cost_per_token": 1.2e-06,
@@ -22300,7 +22300,7 @@
22300
22300
  "litellm_provider": "dashscope",
22301
22301
  "max_input_tokens": 129024,
22302
22302
  "max_output_tokens": 16384,
22303
- "max_tokens": 131072,
22303
+ "max_tokens": 16384,
22304
22304
  "mode": "chat",
22305
22305
  "output_cost_per_reasoning_token": 4e-06,
22306
22306
  "output_cost_per_token": 1.2e-06,
@@ -22314,7 +22314,7 @@
22314
22314
  "litellm_provider": "dashscope",
22315
22315
  "max_input_tokens": 129024,
22316
22316
  "max_output_tokens": 8192,
22317
- "max_tokens": 131072,
22317
+ "max_tokens": 8192,
22318
22318
  "mode": "chat",
22319
22319
  "output_cost_per_token": 1.2e-06,
22320
22320
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
@@ -22326,7 +22326,7 @@
22326
22326
  "litellm_provider": "dashscope",
22327
22327
  "max_input_tokens": 997952,
22328
22328
  "max_output_tokens": 32768,
22329
- "max_tokens": 1000000,
22329
+ "max_tokens": 32768,
22330
22330
  "mode": "chat",
22331
22331
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
22332
22332
  "supports_function_calling": true,
@@ -22356,7 +22356,7 @@
22356
22356
  "litellm_provider": "dashscope",
22357
22357
  "max_input_tokens": 129024,
22358
22358
  "max_output_tokens": 16384,
22359
- "max_tokens": 131072,
22359
+ "max_tokens": 16384,
22360
22360
  "mode": "chat",
22361
22361
  "output_cost_per_reasoning_token": 5e-07,
22362
22362
  "output_cost_per_token": 2e-07,
@@ -22370,7 +22370,7 @@
22370
22370
  "litellm_provider": "dashscope",
22371
22371
  "max_input_tokens": 1000000,
22372
22372
  "max_output_tokens": 16384,
22373
- "max_tokens": 1000000,
22373
+ "max_tokens": 16384,
22374
22374
  "mode": "chat",
22375
22375
  "output_cost_per_reasoning_token": 5e-07,
22376
22376
  "output_cost_per_token": 2e-07,
@@ -22384,7 +22384,7 @@
22384
22384
  "litellm_provider": "dashscope",
22385
22385
  "max_input_tokens": 1000000,
22386
22386
  "max_output_tokens": 8192,
22387
- "max_tokens": 1000000,
22387
+ "max_tokens": 8192,
22388
22388
  "mode": "chat",
22389
22389
  "output_cost_per_token": 2e-07,
22390
22390
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
@@ -22397,7 +22397,7 @@
22397
22397
  "litellm_provider": "dashscope",
22398
22398
  "max_input_tokens": 98304,
22399
22399
  "max_output_tokens": 8192,
22400
- "max_tokens": 131072,
22400
+ "max_tokens": 8192,
22401
22401
  "mode": "chat",
22402
22402
  "output_cost_per_token": 2.4e-06,
22403
22403
  "source": "https://www.alibabacloud.com/help/en/model-studio/models",
@@ -22815,7 +22815,7 @@
22815
22815
  "litellm_provider": "openrouter",
22816
22816
  "max_input_tokens": 2000000,
22817
22817
  "max_output_tokens": 30000,
22818
- "max_tokens": 2000000,
22818
+ "max_tokens": 30000,
22819
22819
  "mode": "chat",
22820
22820
  "output_cost_per_token": 0,
22821
22821
  "source": "https://openrouter.ai/x-ai/grok-4-fast:free",
@@ -22916,7 +22916,7 @@
22916
22916
  "litellm_provider": "deepseek",
22917
22917
  "max_input_tokens": 131072,
22918
22918
  "max_output_tokens": 65536,
22919
- "max_tokens": 131072,
22919
+ "max_tokens": 65536,
22920
22920
  "mode": "chat",
22921
22921
  "output_cost_per_token": 1.7e-06,
22922
22922
  "source": "https://api-docs.deepseek.com/quick_start/pricing",
@@ -22937,7 +22937,7 @@
22937
22937
  "litellm_provider": "bedrock_converse",
22938
22938
  "max_input_tokens": 163840,
22939
22939
  "max_output_tokens": 81920,
22940
- "max_tokens": 163840,
22940
+ "max_tokens": 81920,
22941
22941
  "mode": "chat",
22942
22942
  "output_cost_per_token": 1.68e-06,
22943
22943
  "supports_function_calling": true,
@@ -22979,7 +22979,7 @@
22979
22979
  "litellm_provider": "bedrock_converse",
22980
22980
  "max_input_tokens": 262000,
22981
22981
  "max_output_tokens": 65536,
22982
- "max_tokens": 262144,
22982
+ "max_tokens": 65536,
22983
22983
  "mode": "chat",
22984
22984
  "output_cost_per_token": 1.8e-06,
22985
22985
  "supports_function_calling": true,
@@ -22991,7 +22991,7 @@
22991
22991
  "litellm_provider": "bedrock_converse",
22992
22992
  "max_input_tokens": 262144,
22993
22993
  "max_output_tokens": 131072,
22994
- "max_tokens": 262144,
22994
+ "max_tokens": 131072,
22995
22995
  "mode": "chat",
22996
22996
  "output_cost_per_token": 8.8e-07,
22997
22997
  "supports_function_calling": true,
@@ -23003,7 +23003,7 @@
23003
23003
  "litellm_provider": "bedrock_converse",
23004
23004
  "max_input_tokens": 262144,
23005
23005
  "max_output_tokens": 131072,
23006
- "max_tokens": 262144,
23006
+ "max_tokens": 131072,
23007
23007
  "mode": "chat",
23008
23008
  "output_cost_per_token": 6e-07,
23009
23009
  "supports_function_calling": true,
@@ -23015,7 +23015,7 @@
23015
23015
  "litellm_provider": "bedrock_converse",
23016
23016
  "max_input_tokens": 131072,
23017
23017
  "max_output_tokens": 16384,
23018
- "max_tokens": 131072,
23018
+ "max_tokens": 16384,
23019
23019
  "mode": "chat",
23020
23020
  "output_cost_per_token": 6e-07,
23021
23021
  "supports_function_calling": true,
@@ -23053,7 +23053,7 @@
23053
23053
  "litellm_provider": "vertex_ai-deepseek_models",
23054
23054
  "max_input_tokens": 163840,
23055
23055
  "max_output_tokens": 32768,
23056
- "max_tokens": 163840,
23056
+ "max_tokens": 32768,
23057
23057
  "mode": "chat",
23058
23058
  "output_cost_per_token": 5.4e-06,
23059
23059
  "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
@@ -23684,7 +23684,7 @@
23684
23684
  "lemonade/Qwen3-Coder-30B-A3B-Instruct-GGUF": {
23685
23685
  "input_cost_per_token": 0,
23686
23686
  "litellm_provider": "lemonade",
23687
- "max_tokens": 262144,
23687
+ "max_tokens": 32768,
23688
23688
  "max_input_tokens": 262144,
23689
23689
  "max_output_tokens": 32768,
23690
23690
  "mode": "chat",
@@ -23700,7 +23700,7 @@
23700
23700
  "litellm_provider": "groq",
23701
23701
  "max_input_tokens": 262144,
23702
23702
  "max_output_tokens": 16384,
23703
- "max_tokens": 278528,
23703
+ "max_tokens": 16384,
23704
23704
  "mode": "chat",
23705
23705
  "supports_function_calling": true,
23706
23706
  "supports_response_schema": true,
@@ -24496,7 +24496,7 @@
24496
24496
  "litellm_provider": "oci",
24497
24497
  "max_input_tokens": 128000,
24498
24498
  "max_output_tokens": 4000,
24499
- "max_tokens": 128000,
24499
+ "max_tokens": 4000,
24500
24500
  "mode": "chat",
24501
24501
  "output_cost_per_token": 1.56e-06,
24502
24502
  "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/",
@@ -24508,7 +24508,7 @@
24508
24508
  "litellm_provider": "oci",
24509
24509
  "max_input_tokens": 256000,
24510
24510
  "max_output_tokens": 4000,
24511
- "max_tokens": 256000,
24511
+ "max_tokens": 4000,
24512
24512
  "mode": "chat",
24513
24513
  "output_cost_per_token": 1.56e-06,
24514
24514
  "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/",
@@ -24520,7 +24520,7 @@
24520
24520
  "litellm_provider": "oci",
24521
24521
  "max_input_tokens": 128000,
24522
24522
  "max_output_tokens": 4000,
24523
- "max_tokens": 128000,
24523
+ "max_tokens": 4000,
24524
24524
  "mode": "chat",
24525
24525
  "output_cost_per_token": 1.56e-06,
24526
24526
  "source": "https://www.oracle.com/cloud/ai/generative-ai/pricing/",
@@ -24835,7 +24835,7 @@
24835
24835
  "lemonade/gpt-oss-20b-mxfp4-GGUF": {
24836
24836
  "input_cost_per_token": 0,
24837
24837
  "litellm_provider": "lemonade",
24838
- "max_tokens": 131072,
24838
+ "max_tokens": 32768,
24839
24839
  "max_input_tokens": 131072,
24840
24840
  "max_output_tokens": 32768,
24841
24841
  "mode": "chat",
@@ -24847,7 +24847,7 @@
24847
24847
  "lemonade/gpt-oss-120b-mxfp-GGUF": {
24848
24848
  "input_cost_per_token": 0,
24849
24849
  "litellm_provider": "lemonade",
24850
- "max_tokens": 131072,
24850
+ "max_tokens": 32768,
24851
24851
  "max_input_tokens": 131072,
24852
24852
  "max_output_tokens": 32768,
24853
24853
  "mode": "chat",
@@ -24859,7 +24859,7 @@
24859
24859
  "lemonade/Gemma-3-4b-it-GGUF": {
24860
24860
  "input_cost_per_token": 0,
24861
24861
  "litellm_provider": "lemonade",
24862
- "max_tokens": 128000,
24862
+ "max_tokens": 8192,
24863
24863
  "max_input_tokens": 128000,
24864
24864
  "max_output_tokens": 8192,
24865
24865
  "mode": "chat",
@@ -24871,7 +24871,7 @@
24871
24871
  "lemonade/Qwen3-4B-Instruct-2507-GGUF": {
24872
24872
  "input_cost_per_token": 0,
24873
24873
  "litellm_provider": "lemonade",
24874
- "max_tokens": 262144,
24874
+ "max_tokens": 32768,
24875
24875
  "max_input_tokens": 262144,
24876
24876
  "max_output_tokens": 32768,
24877
24877
  "mode": "chat",
@@ -25789,7 +25789,7 @@
25789
25789
  "litellm_provider": "azure",
25790
25790
  "max_input_tokens": 272000,
25791
25791
  "max_output_tokens": 128000,
25792
- "max_tokens": 400000,
25792
+ "max_tokens": 128000,
25793
25793
  "mode": "responses",
25794
25794
  "output_cost_per_token": 0.00012,
25795
25795
  "source": "https://learn.microsoft.com/en-us/azure/ai-foundry/foundry-models/concepts/models-sold-directly-by-azure?pivots=azure-openai&tabs=global-standard-aoai%2Cstandard-chat-completions%2Cglobal-standard#gpt-5",
@@ -25926,7 +25926,7 @@
25926
25926
  "litellm_provider": "openrouter",
25927
25927
  "max_input_tokens": 163840,
25928
25928
  "max_output_tokens": 163840,
25929
- "max_tokens": 8192,
25929
+ "max_tokens": 163840,
25930
25930
  "mode": "chat",
25931
25931
  "output_cost_per_token": 4e-07,
25932
25932
  "supports_assistant_prefill": true,
@@ -25940,7 +25940,7 @@
25940
25940
  "litellm_provider": "openrouter",
25941
25941
  "max_input_tokens": 204800,
25942
25942
  "max_output_tokens": 204800,
25943
- "max_tokens": 32768,
25943
+ "max_tokens": 204800,
25944
25944
  "mode": "chat",
25945
25945
  "output_cost_per_token": 1.02e-06,
25946
25946
  "supports_function_calling": true,
@@ -25953,7 +25953,7 @@
25953
25953
  "litellm_provider": "openrouter",
25954
25954
  "max_input_tokens": 202800,
25955
25955
  "max_output_tokens": 131000,
25956
- "max_tokens": 202800,
25956
+ "max_tokens": 131000,
25957
25957
  "mode": "chat",
25958
25958
  "output_cost_per_token": 1.75e-06,
25959
25959
  "source": "https://openrouter.ai/z-ai/glm-4.6",
@@ -25966,7 +25966,7 @@
25966
25966
  "litellm_provider": "openrouter",
25967
25967
  "max_input_tokens": 202800,
25968
25968
  "max_output_tokens": 131000,
25969
- "max_tokens": 202800,
25969
+ "max_tokens": 131000,
25970
25970
  "mode": "chat",
25971
25971
  "output_cost_per_token": 1.9e-06,
25972
25972
  "source": "https://openrouter.ai/z-ai/glm-4.6:exacto",
@@ -27739,7 +27739,7 @@
27739
27739
  "litellm_provider": "vertex_ai-language-models",
27740
27740
  "max_input_tokens": 65536,
27741
27741
  "max_output_tokens": 32768,
27742
- "max_tokens": 65536,
27742
+ "max_tokens": 32768,
27743
27743
  "mode": "image_generation",
27744
27744
  "output_cost_per_image": 0.134,
27745
27745
  "output_cost_per_image_token": 0.00012,
@@ -27773,7 +27773,7 @@
27773
27773
  "litellm_provider": "gemini",
27774
27774
  "max_input_tokens": 65536,
27775
27775
  "max_output_tokens": 32768,
27776
- "max_tokens": 65536,
27776
+ "max_tokens": 32768,
27777
27777
  "mode": "image_generation",
27778
27778
  "output_cost_per_image": 0.134,
27779
27779
  "output_cost_per_image_token": 0.00012,
@@ -27870,7 +27870,7 @@
27870
27870
  "litellm_provider": "vertex_ai-language-models",
27871
27871
  "max_input_tokens": 65536,
27872
27872
  "max_output_tokens": 32768,
27873
- "max_tokens": 65536,
27873
+ "max_tokens": 32768,
27874
27874
  "mode": "image_generation",
27875
27875
  "output_cost_per_image": 0.134,
27876
27876
  "output_cost_per_image_token": 0.00012,
@@ -28314,7 +28314,7 @@
28314
28314
  "litellm_provider": "publicai",
28315
28315
  "max_input_tokens": 8192,
28316
28316
  "max_output_tokens": 4096,
28317
- "max_tokens": 8192,
28317
+ "max_tokens": 4096,
28318
28318
  "mode": "chat",
28319
28319
  "output_cost_per_token": 0.0,
28320
28320
  "source": "https://platform.publicai.co/docs",
@@ -28326,7 +28326,7 @@
28326
28326
  "litellm_provider": "publicai",
28327
28327
  "max_input_tokens": 8192,
28328
28328
  "max_output_tokens": 4096,
28329
- "max_tokens": 8192,
28329
+ "max_tokens": 4096,
28330
28330
  "mode": "chat",
28331
28331
  "output_cost_per_token": 0.0,
28332
28332
  "source": "https://platform.publicai.co/docs",
@@ -28338,7 +28338,7 @@
28338
28338
  "litellm_provider": "publicai",
28339
28339
  "max_input_tokens": 8192,
28340
28340
  "max_output_tokens": 4096,
28341
- "max_tokens": 8192,
28341
+ "max_tokens": 4096,
28342
28342
  "mode": "chat",
28343
28343
  "output_cost_per_token": 0.0,
28344
28344
  "source": "https://platform.publicai.co/docs",
@@ -28350,7 +28350,7 @@
28350
28350
  "litellm_provider": "publicai",
28351
28351
  "max_input_tokens": 16384,
28352
28352
  "max_output_tokens": 4096,
28353
- "max_tokens": 16384,
28353
+ "max_tokens": 4096,
28354
28354
  "mode": "chat",
28355
28355
  "output_cost_per_token": 0.0,
28356
28356
  "source": "https://platform.publicai.co/docs",
@@ -28362,7 +28362,7 @@
28362
28362
  "litellm_provider": "publicai",
28363
28363
  "max_input_tokens": 8192,
28364
28364
  "max_output_tokens": 4096,
28365
- "max_tokens": 8192,
28365
+ "max_tokens": 4096,
28366
28366
  "mode": "chat",
28367
28367
  "output_cost_per_token": 0.0,
28368
28368
  "source": "https://platform.publicai.co/docs",
@@ -28374,7 +28374,7 @@
28374
28374
  "litellm_provider": "publicai",
28375
28375
  "max_input_tokens": 32768,
28376
28376
  "max_output_tokens": 4096,
28377
- "max_tokens": 32768,
28377
+ "max_tokens": 4096,
28378
28378
  "mode": "chat",
28379
28379
  "output_cost_per_token": 0.0,
28380
28380
  "source": "https://platform.publicai.co/docs",
@@ -28386,7 +28386,7 @@
28386
28386
  "litellm_provider": "publicai",
28387
28387
  "max_input_tokens": 32768,
28388
28388
  "max_output_tokens": 4096,
28389
- "max_tokens": 32768,
28389
+ "max_tokens": 4096,
28390
28390
  "mode": "chat",
28391
28391
  "output_cost_per_token": 0.0,
28392
28392
  "source": "https://platform.publicai.co/docs",
@@ -28398,7 +28398,7 @@
28398
28398
  "litellm_provider": "publicai",
28399
28399
  "max_input_tokens": 32768,
28400
28400
  "max_output_tokens": 4096,
28401
- "max_tokens": 32768,
28401
+ "max_tokens": 4096,
28402
28402
  "mode": "chat",
28403
28403
  "output_cost_per_token": 0.0,
28404
28404
  "source": "https://platform.publicai.co/docs",
@@ -28411,7 +28411,7 @@
28411
28411
  "litellm_provider": "publicai",
28412
28412
  "max_input_tokens": 32768,
28413
28413
  "max_output_tokens": 4096,
28414
- "max_tokens": 32768,
28414
+ "max_tokens": 4096,
28415
28415
  "mode": "chat",
28416
28416
  "output_cost_per_token": 0.0,
28417
28417
  "source": "https://platform.publicai.co/docs",
@@ -28424,7 +28424,7 @@
28424
28424
  "litellm_provider": "fireworks_ai",
28425
28425
  "max_input_tokens": 262144,
28426
28426
  "max_output_tokens": 32768,
28427
- "max_tokens": 262144,
28427
+ "max_tokens": 32768,
28428
28428
  "mode": "chat",
28429
28429
  "output_cost_per_token": 2.5e-06,
28430
28430
  "source": "https://app.fireworks.ai/models/fireworks/kimi-k2-instruct-0905",
@@ -28465,7 +28465,7 @@
28465
28465
  "litellm_provider": "databricks",
28466
28466
  "max_input_tokens": 200000,
28467
28467
  "max_output_tokens": 64000,
28468
- "max_tokens": 200000,
28468
+ "max_tokens": 64000,
28469
28469
  "metadata": {
28470
28470
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28471
28471
  },
@@ -28484,7 +28484,7 @@
28484
28484
  "litellm_provider": "databricks",
28485
28485
  "max_input_tokens": 200000,
28486
28486
  "max_output_tokens": 32000,
28487
- "max_tokens": 200000,
28487
+ "max_tokens": 32000,
28488
28488
  "metadata": {
28489
28489
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28490
28490
  },
@@ -28503,7 +28503,7 @@
28503
28503
  "litellm_provider": "databricks",
28504
28504
  "max_input_tokens": 200000,
28505
28505
  "max_output_tokens": 32000,
28506
- "max_tokens": 200000,
28506
+ "max_tokens": 32000,
28507
28507
  "metadata": {
28508
28508
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28509
28509
  },
@@ -28522,7 +28522,7 @@
28522
28522
  "litellm_provider": "databricks",
28523
28523
  "max_input_tokens": 200000,
28524
28524
  "max_output_tokens": 64000,
28525
- "max_tokens": 200000,
28525
+ "max_tokens": 64000,
28526
28526
  "metadata": {
28527
28527
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28528
28528
  },
@@ -28541,7 +28541,7 @@
28541
28541
  "litellm_provider": "databricks",
28542
28542
  "max_input_tokens": 200000,
28543
28543
  "max_output_tokens": 64000,
28544
- "max_tokens": 200000,
28544
+ "max_tokens": 64000,
28545
28545
  "metadata": {
28546
28546
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28547
28547
  },
@@ -28560,7 +28560,7 @@
28560
28560
  "litellm_provider": "databricks",
28561
28561
  "max_input_tokens": 200000,
28562
28562
  "max_output_tokens": 64000,
28563
- "max_tokens": 200000,
28563
+ "max_tokens": 64000,
28564
28564
  "metadata": {
28565
28565
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28566
28566
  },
@@ -28579,7 +28579,7 @@
28579
28579
  "litellm_provider": "databricks",
28580
28580
  "max_input_tokens": 200000,
28581
28581
  "max_output_tokens": 64000,
28582
- "max_tokens": 200000,
28582
+ "max_tokens": 64000,
28583
28583
  "metadata": {
28584
28584
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28585
28585
  },
@@ -28598,7 +28598,7 @@
28598
28598
  "litellm_provider": "databricks",
28599
28599
  "max_input_tokens": 1048576,
28600
28600
  "max_output_tokens": 65535,
28601
- "max_tokens": 1048576,
28601
+ "max_tokens": 65535,
28602
28602
  "metadata": {
28603
28603
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28604
28604
  },
@@ -28615,7 +28615,7 @@
28615
28615
  "litellm_provider": "databricks",
28616
28616
  "max_input_tokens": 1048576,
28617
28617
  "max_output_tokens": 65536,
28618
- "max_tokens": 1048576,
28618
+ "max_tokens": 65536,
28619
28619
  "metadata": {
28620
28620
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28621
28621
  },
@@ -28647,7 +28647,7 @@
28647
28647
  "litellm_provider": "databricks",
28648
28648
  "max_input_tokens": 400000,
28649
28649
  "max_output_tokens": 128000,
28650
- "max_tokens": 400000,
28650
+ "max_tokens": 128000,
28651
28651
  "metadata": {
28652
28652
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28653
28653
  },
@@ -28662,7 +28662,7 @@
28662
28662
  "litellm_provider": "databricks",
28663
28663
  "max_input_tokens": 400000,
28664
28664
  "max_output_tokens": 128000,
28665
- "max_tokens": 400000,
28665
+ "max_tokens": 128000,
28666
28666
  "metadata": {
28667
28667
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28668
28668
  },
@@ -28677,7 +28677,7 @@
28677
28677
  "litellm_provider": "databricks",
28678
28678
  "max_input_tokens": 400000,
28679
28679
  "max_output_tokens": 128000,
28680
- "max_tokens": 400000,
28680
+ "max_tokens": 128000,
28681
28681
  "metadata": {
28682
28682
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28683
28683
  },
@@ -28692,7 +28692,7 @@
28692
28692
  "litellm_provider": "databricks",
28693
28693
  "max_input_tokens": 400000,
28694
28694
  "max_output_tokens": 128000,
28695
- "max_tokens": 400000,
28695
+ "max_tokens": 128000,
28696
28696
  "metadata": {
28697
28697
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28698
28698
  },
@@ -28737,7 +28737,7 @@
28737
28737
  "litellm_provider": "databricks",
28738
28738
  "max_input_tokens": 200000,
28739
28739
  "max_output_tokens": 128000,
28740
- "max_tokens": 200000,
28740
+ "max_tokens": 128000,
28741
28741
  "metadata": {
28742
28742
  "notes": "Input/output cost per token is dbu cost * $0.070. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."
28743
28743
  },
@@ -28932,7 +28932,7 @@
28932
28932
  "litellm_provider": "deepseek",
28933
28933
  "max_input_tokens": 163840,
28934
28934
  "max_output_tokens": 163840,
28935
- "max_tokens": 8192,
28935
+ "max_tokens": 163840,
28936
28936
  "mode": "chat",
28937
28937
  "output_cost_per_token": 4e-07,
28938
28938
  "supports_assistant_prefill": true,
@@ -29019,7 +29019,7 @@
29019
29019
  "litellm_provider": "openrouter",
29020
29020
  "max_input_tokens": 163840,
29021
29021
  "max_output_tokens": 163840,
29022
- "max_tokens": 8192,
29022
+ "max_tokens": 163840,
29023
29023
  "mode": "chat",
29024
29024
  "output_cost_per_token": 4e-07,
29025
29025
  "supports_assistant_prefill": true,
@@ -31608,7 +31608,7 @@
31608
31608
  "litellm_provider": "vertex_ai-deepseek_models",
31609
31609
  "max_input_tokens": 163840,
31610
31610
  "max_output_tokens": 32768,
31611
- "max_tokens": 163840,
31611
+ "max_tokens": 32768,
31612
31612
  "mode": "chat",
31613
31613
  "output_cost_per_token": 1.68e-06,
31614
31614
  "output_cost_per_token_batches": 8.4e-07,
@@ -32214,7 +32214,7 @@
32214
32214
  "litellm_provider": "openrouter",
32215
32215
  "max_input_tokens": 262144,
32216
32216
  "max_output_tokens": 65536,
32217
- "max_tokens": 262144,
32217
+ "max_tokens": 65536,
32218
32218
  "mode": "chat",
32219
32219
  "output_cost_per_token": 6e-07,
32220
32220
  "supports_function_calling": true,
@@ -32285,7 +32285,7 @@
32285
32285
  "litellm_provider": "openrouter",
32286
32286
  "max_input_tokens": 400000,
32287
32287
  "max_output_tokens": 128000,
32288
- "max_tokens": 400000,
32288
+ "max_tokens": 128000,
32289
32289
  "mode": "chat",
32290
32290
  "output_cost_per_token": 1.4e-05,
32291
32291
  "supports_function_calling": true,
@@ -32301,7 +32301,7 @@
32301
32301
  "litellm_provider": "openrouter",
32302
32302
  "max_input_tokens": 128000,
32303
32303
  "max_output_tokens": 16384,
32304
- "max_tokens": 128000,
32304
+ "max_tokens": 16384,
32305
32305
  "mode": "chat",
32306
32306
  "output_cost_per_token": 1.4e-05,
32307
32307
  "supports_function_calling": true,
@@ -32315,7 +32315,7 @@
32315
32315
  "litellm_provider": "openrouter",
32316
32316
  "max_input_tokens": 400000,
32317
32317
  "max_output_tokens": 128000,
32318
- "max_tokens": 400000,
32318
+ "max_tokens": 128000,
32319
32319
  "mode": "chat",
32320
32320
  "output_cost_per_token": 0.000168,
32321
32321
  "supports_function_calling": true,
@@ -32329,7 +32329,7 @@
32329
32329
  "litellm_provider": "azure_ai",
32330
32330
  "max_input_tokens": 163840,
32331
32331
  "max_output_tokens": 163840,
32332
- "max_tokens": 8192,
32332
+ "max_tokens": 163840,
32333
32333
  "mode": "chat",
32334
32334
  "output_cost_per_token": 1.68e-06,
32335
32335
  "supports_assistant_prefill": true,
@@ -32343,7 +32343,7 @@
32343
32343
  "litellm_provider": "azure_ai",
32344
32344
  "max_input_tokens": 163840,
32345
32345
  "max_output_tokens": 163840,
32346
- "max_tokens": 8192,
32346
+ "max_tokens": 163840,
32347
32347
  "mode": "chat",
32348
32348
  "output_cost_per_token": 1.68e-06,
32349
32349
  "supports_assistant_prefill": true,
@@ -33980,5 +33980,1099 @@
33980
33980
  "supports_function_calling": true,
33981
33981
  "supports_reasoning": true,
33982
33982
  "supports_tool_choice": true
33983
+ },
33984
+ "novita/deepseek/deepseek-v3.2": {
33985
+ "litellm_provider": "novita",
33986
+ "mode": "chat",
33987
+ "input_cost_per_token": 0.00269,
33988
+ "output_cost_per_token": 0.004,
33989
+ "max_input_tokens": 163840,
33990
+ "max_output_tokens": 65536,
33991
+ "max_tokens": 65536,
33992
+ "supports_function_calling": true,
33993
+ "supports_parallel_function_calling": true,
33994
+ "supports_tool_choice": true,
33995
+ "supports_system_messages": true,
33996
+ "supports_response_schema": true,
33997
+ "cache_read_input_token_cost": 0.001345,
33998
+ "input_cost_per_token_cache_hit": 0.001345,
33999
+ "supports_reasoning": true
34000
+ },
34001
+ "novita/minimax/minimax-m2.1": {
34002
+ "litellm_provider": "novita",
34003
+ "mode": "chat",
34004
+ "input_cost_per_token": 0.003,
34005
+ "output_cost_per_token": 0.012,
34006
+ "max_input_tokens": 204800,
34007
+ "max_output_tokens": 131072,
34008
+ "max_tokens": 131072,
34009
+ "supports_function_calling": true,
34010
+ "supports_parallel_function_calling": true,
34011
+ "supports_tool_choice": true,
34012
+ "supports_system_messages": true,
34013
+ "supports_response_schema": true,
34014
+ "cache_read_input_token_cost": 0.0003,
34015
+ "input_cost_per_token_cache_hit": 0.0003,
34016
+ "supports_reasoning": true
34017
+ },
34018
+ "novita/zai-org/glm-4.7": {
34019
+ "litellm_provider": "novita",
34020
+ "mode": "chat",
34021
+ "input_cost_per_token": 0.006,
34022
+ "output_cost_per_token": 0.022,
34023
+ "max_input_tokens": 204800,
34024
+ "max_output_tokens": 131072,
34025
+ "max_tokens": 131072,
34026
+ "supports_function_calling": true,
34027
+ "supports_parallel_function_calling": true,
34028
+ "supports_tool_choice": true,
34029
+ "supports_system_messages": true,
34030
+ "supports_response_schema": true,
34031
+ "cache_read_input_token_cost": 0.0011,
34032
+ "input_cost_per_token_cache_hit": 0.0011,
34033
+ "supports_reasoning": true
34034
+ },
34035
+ "novita/xiaomimimo/mimo-v2-flash": {
34036
+ "litellm_provider": "novita",
34037
+ "mode": "chat",
34038
+ "input_cost_per_token": 0.001,
34039
+ "output_cost_per_token": 0.003,
34040
+ "max_input_tokens": 262144,
34041
+ "max_output_tokens": 32000,
34042
+ "max_tokens": 32000,
34043
+ "supports_function_calling": true,
34044
+ "supports_parallel_function_calling": true,
34045
+ "supports_tool_choice": true,
34046
+ "supports_system_messages": true,
34047
+ "supports_response_schema": true,
34048
+ "cache_read_input_token_cost": 0.0002,
34049
+ "input_cost_per_token_cache_hit": 0.0002,
34050
+ "supports_reasoning": true
34051
+ },
34052
+ "novita/zai-org/autoglm-phone-9b-multilingual": {
34053
+ "litellm_provider": "novita",
34054
+ "mode": "chat",
34055
+ "input_cost_per_token": 0.00035,
34056
+ "output_cost_per_token": 0.00138,
34057
+ "max_input_tokens": 65536,
34058
+ "max_output_tokens": 65536,
34059
+ "max_tokens": 65536,
34060
+ "supports_vision": true,
34061
+ "supports_system_messages": true
34062
+ },
34063
+ "novita/moonshotai/kimi-k2-thinking": {
34064
+ "litellm_provider": "novita",
34065
+ "mode": "chat",
34066
+ "input_cost_per_token": 0.0048,
34067
+ "output_cost_per_token": 0.02,
34068
+ "max_input_tokens": 262144,
34069
+ "max_output_tokens": 262144,
34070
+ "max_tokens": 262144,
34071
+ "supports_function_calling": true,
34072
+ "supports_parallel_function_calling": true,
34073
+ "supports_tool_choice": true,
34074
+ "supports_system_messages": true,
34075
+ "supports_response_schema": true,
34076
+ "supports_reasoning": true
34077
+ },
34078
+ "novita/minimax/minimax-m2": {
34079
+ "litellm_provider": "novita",
34080
+ "mode": "chat",
34081
+ "input_cost_per_token": 0.0024,
34082
+ "output_cost_per_token": 0.0096,
34083
+ "max_input_tokens": 204800,
34084
+ "max_output_tokens": 131072,
34085
+ "max_tokens": 131072,
34086
+ "supports_function_calling": true,
34087
+ "supports_parallel_function_calling": true,
34088
+ "supports_tool_choice": true,
34089
+ "supports_system_messages": true,
34090
+ "cache_read_input_token_cost": 0.00024,
34091
+ "input_cost_per_token_cache_hit": 0.00024,
34092
+ "supports_reasoning": true
34093
+ },
34094
+ "novita/paddlepaddle/paddleocr-vl": {
34095
+ "litellm_provider": "novita",
34096
+ "mode": "chat",
34097
+ "input_cost_per_token": 0.00016,
34098
+ "output_cost_per_token": 0.00016,
34099
+ "max_input_tokens": 16384,
34100
+ "max_output_tokens": 16384,
34101
+ "max_tokens": 16384,
34102
+ "supports_vision": true,
34103
+ "supports_system_messages": true
34104
+ },
34105
+ "novita/deepseek/deepseek-v3.2-exp": {
34106
+ "litellm_provider": "novita",
34107
+ "mode": "chat",
34108
+ "input_cost_per_token": 0.00216,
34109
+ "output_cost_per_token": 0.00328,
34110
+ "max_input_tokens": 163840,
34111
+ "max_output_tokens": 65536,
34112
+ "max_tokens": 65536,
34113
+ "supports_function_calling": true,
34114
+ "supports_parallel_function_calling": true,
34115
+ "supports_tool_choice": true,
34116
+ "supports_system_messages": true,
34117
+ "supports_response_schema": true,
34118
+ "supports_reasoning": true
34119
+ },
34120
+ "novita/qwen/qwen3-vl-235b-a22b-thinking": {
34121
+ "litellm_provider": "novita",
34122
+ "mode": "chat",
34123
+ "input_cost_per_token": 0.00784,
34124
+ "output_cost_per_token": 0.0316,
34125
+ "max_input_tokens": 131072,
34126
+ "max_output_tokens": 32768,
34127
+ "max_tokens": 32768,
34128
+ "supports_vision": true,
34129
+ "supports_system_messages": true,
34130
+ "supports_reasoning": true
34131
+ },
34132
+ "novita/zai-org/glm-4.6v": {
34133
+ "litellm_provider": "novita",
34134
+ "mode": "chat",
34135
+ "input_cost_per_token": 0.003,
34136
+ "output_cost_per_token": 0.009,
34137
+ "max_input_tokens": 131072,
34138
+ "max_output_tokens": 32768,
34139
+ "max_tokens": 32768,
34140
+ "supports_function_calling": true,
34141
+ "supports_parallel_function_calling": true,
34142
+ "supports_tool_choice": true,
34143
+ "supports_vision": true,
34144
+ "supports_system_messages": true,
34145
+ "supports_response_schema": true,
34146
+ "cache_read_input_token_cost": 0.00055,
34147
+ "input_cost_per_token_cache_hit": 0.00055,
34148
+ "supports_reasoning": true
34149
+ },
34150
+ "novita/zai-org/glm-4.6": {
34151
+ "litellm_provider": "novita",
34152
+ "mode": "chat",
34153
+ "input_cost_per_token": 0.0044,
34154
+ "output_cost_per_token": 0.0176,
34155
+ "max_input_tokens": 204800,
34156
+ "max_output_tokens": 131072,
34157
+ "max_tokens": 131072,
34158
+ "supports_function_calling": true,
34159
+ "supports_parallel_function_calling": true,
34160
+ "supports_tool_choice": true,
34161
+ "supports_system_messages": true,
34162
+ "supports_response_schema": true,
34163
+ "cache_read_input_token_cost": 0.00088,
34164
+ "input_cost_per_token_cache_hit": 0.00088,
34165
+ "supports_reasoning": true
34166
+ },
34167
+ "novita/qwen/qwen3-next-80b-a3b-instruct": {
34168
+ "litellm_provider": "novita",
34169
+ "mode": "chat",
34170
+ "input_cost_per_token": 0.0012,
34171
+ "output_cost_per_token": 0.012,
34172
+ "max_input_tokens": 131072,
34173
+ "max_output_tokens": 32768,
34174
+ "max_tokens": 32768,
34175
+ "supports_function_calling": true,
34176
+ "supports_parallel_function_calling": true,
34177
+ "supports_tool_choice": true,
34178
+ "supports_system_messages": true,
34179
+ "supports_response_schema": true
34180
+ },
34181
+ "novita/qwen/qwen3-next-80b-a3b-thinking": {
34182
+ "litellm_provider": "novita",
34183
+ "mode": "chat",
34184
+ "input_cost_per_token": 0.0012,
34185
+ "output_cost_per_token": 0.012,
34186
+ "max_input_tokens": 131072,
34187
+ "max_output_tokens": 32768,
34188
+ "max_tokens": 32768,
34189
+ "supports_function_calling": true,
34190
+ "supports_parallel_function_calling": true,
34191
+ "supports_tool_choice": true,
34192
+ "supports_system_messages": true,
34193
+ "supports_response_schema": true,
34194
+ "supports_reasoning": true
34195
+ },
34196
+ "novita/deepseek/deepseek-ocr": {
34197
+ "litellm_provider": "novita",
34198
+ "mode": "chat",
34199
+ "input_cost_per_token": 0.00024,
34200
+ "output_cost_per_token": 0.00024,
34201
+ "max_input_tokens": 8192,
34202
+ "max_output_tokens": 8192,
34203
+ "max_tokens": 8192,
34204
+ "supports_tool_choice": true,
34205
+ "supports_vision": true,
34206
+ "supports_system_messages": true,
34207
+ "supports_response_schema": true
34208
+ },
34209
+ "novita/deepseek/deepseek-v3.1-terminus": {
34210
+ "litellm_provider": "novita",
34211
+ "mode": "chat",
34212
+ "input_cost_per_token": 0.00216,
34213
+ "output_cost_per_token": 0.008,
34214
+ "max_input_tokens": 131072,
34215
+ "max_output_tokens": 32768,
34216
+ "max_tokens": 32768,
34217
+ "supports_function_calling": true,
34218
+ "supports_parallel_function_calling": true,
34219
+ "supports_tool_choice": true,
34220
+ "supports_system_messages": true,
34221
+ "supports_response_schema": true,
34222
+ "cache_read_input_token_cost": 0.00108,
34223
+ "input_cost_per_token_cache_hit": 0.00108,
34224
+ "supports_reasoning": true
34225
+ },
34226
+ "novita/qwen/qwen3-vl-235b-a22b-instruct": {
34227
+ "litellm_provider": "novita",
34228
+ "mode": "chat",
34229
+ "input_cost_per_token": 0.0024,
34230
+ "output_cost_per_token": 0.012,
34231
+ "max_input_tokens": 131072,
34232
+ "max_output_tokens": 32768,
34233
+ "max_tokens": 32768,
34234
+ "supports_function_calling": true,
34235
+ "supports_parallel_function_calling": true,
34236
+ "supports_tool_choice": true,
34237
+ "supports_vision": true,
34238
+ "supports_system_messages": true,
34239
+ "supports_response_schema": true
34240
+ },
34241
+ "novita/qwen/qwen3-max": {
34242
+ "litellm_provider": "novita",
34243
+ "mode": "chat",
34244
+ "input_cost_per_token": 0.01688,
34245
+ "output_cost_per_token": 0.0676,
34246
+ "max_input_tokens": 262144,
34247
+ "max_output_tokens": 65536,
34248
+ "max_tokens": 65536,
34249
+ "supports_function_calling": true,
34250
+ "supports_parallel_function_calling": true,
34251
+ "supports_tool_choice": true,
34252
+ "supports_system_messages": true,
34253
+ "supports_response_schema": true
34254
+ },
34255
+ "novita/skywork/r1v4-lite": {
34256
+ "litellm_provider": "novita",
34257
+ "mode": "chat",
34258
+ "input_cost_per_token": 0.002,
34259
+ "output_cost_per_token": 0.006,
34260
+ "max_input_tokens": 262144,
34261
+ "max_output_tokens": 65536,
34262
+ "max_tokens": 65536,
34263
+ "supports_tool_choice": true,
34264
+ "supports_vision": true,
34265
+ "supports_system_messages": true,
34266
+ "supports_response_schema": true
34267
+ },
34268
+ "novita/deepseek/deepseek-v3.1": {
34269
+ "litellm_provider": "novita",
34270
+ "mode": "chat",
34271
+ "input_cost_per_token": 0.00216,
34272
+ "output_cost_per_token": 0.008,
34273
+ "max_input_tokens": 131072,
34274
+ "max_output_tokens": 32768,
34275
+ "max_tokens": 32768,
34276
+ "supports_function_calling": true,
34277
+ "supports_parallel_function_calling": true,
34278
+ "supports_tool_choice": true,
34279
+ "supports_system_messages": true,
34280
+ "supports_response_schema": true,
34281
+ "cache_read_input_token_cost": 0.00108,
34282
+ "input_cost_per_token_cache_hit": 0.00108,
34283
+ "supports_reasoning": true
34284
+ },
34285
+ "novita/moonshotai/kimi-k2-0905": {
34286
+ "litellm_provider": "novita",
34287
+ "mode": "chat",
34288
+ "input_cost_per_token": 0.0048,
34289
+ "output_cost_per_token": 0.02,
34290
+ "max_input_tokens": 262144,
34291
+ "max_output_tokens": 262144,
34292
+ "max_tokens": 262144,
34293
+ "supports_function_calling": true,
34294
+ "supports_parallel_function_calling": true,
34295
+ "supports_tool_choice": true,
34296
+ "supports_system_messages": true,
34297
+ "supports_response_schema": true
34298
+ },
34299
+ "novita/qwen/qwen3-coder-480b-a35b-instruct": {
34300
+ "litellm_provider": "novita",
34301
+ "mode": "chat",
34302
+ "input_cost_per_token": 0.0024,
34303
+ "output_cost_per_token": 0.0104,
34304
+ "max_input_tokens": 262144,
34305
+ "max_output_tokens": 65536,
34306
+ "max_tokens": 65536,
34307
+ "supports_function_calling": true,
34308
+ "supports_parallel_function_calling": true,
34309
+ "supports_tool_choice": true,
34310
+ "supports_system_messages": true,
34311
+ "supports_response_schema": true
34312
+ },
34313
+ "novita/qwen/qwen3-coder-30b-a3b-instruct": {
34314
+ "litellm_provider": "novita",
34315
+ "mode": "chat",
34316
+ "input_cost_per_token": 0.0007,
34317
+ "output_cost_per_token": 0.0027,
34318
+ "max_input_tokens": 160000,
34319
+ "max_output_tokens": 32768,
34320
+ "max_tokens": 32768,
34321
+ "supports_function_calling": true,
34322
+ "supports_parallel_function_calling": true,
34323
+ "supports_tool_choice": true,
34324
+ "supports_system_messages": true,
34325
+ "supports_response_schema": true
34326
+ },
34327
+ "novita/openai/gpt-oss-120b": {
34328
+ "litellm_provider": "novita",
34329
+ "mode": "chat",
34330
+ "input_cost_per_token": 0.0004,
34331
+ "output_cost_per_token": 0.002,
34332
+ "max_input_tokens": 131072,
34333
+ "max_output_tokens": 32768,
34334
+ "max_tokens": 32768,
34335
+ "supports_function_calling": true,
34336
+ "supports_parallel_function_calling": true,
34337
+ "supports_tool_choice": true,
34338
+ "supports_vision": true,
34339
+ "supports_system_messages": true,
34340
+ "supports_response_schema": true,
34341
+ "supports_reasoning": true
34342
+ },
34343
+ "novita/moonshotai/kimi-k2-instruct": {
34344
+ "litellm_provider": "novita",
34345
+ "mode": "chat",
34346
+ "input_cost_per_token": 0.00456,
34347
+ "output_cost_per_token": 0.0184,
34348
+ "max_input_tokens": 131072,
34349
+ "max_output_tokens": 131072,
34350
+ "max_tokens": 131072,
34351
+ "supports_function_calling": true,
34352
+ "supports_parallel_function_calling": true,
34353
+ "supports_tool_choice": true,
34354
+ "supports_system_messages": true,
34355
+ "supports_response_schema": true
34356
+ },
34357
+ "novita/deepseek/deepseek-v3-0324": {
34358
+ "litellm_provider": "novita",
34359
+ "mode": "chat",
34360
+ "input_cost_per_token": 0.00216,
34361
+ "output_cost_per_token": 0.00896,
34362
+ "max_input_tokens": 163840,
34363
+ "max_output_tokens": 163840,
34364
+ "max_tokens": 163840,
34365
+ "supports_function_calling": true,
34366
+ "supports_parallel_function_calling": true,
34367
+ "supports_tool_choice": true,
34368
+ "supports_system_messages": true,
34369
+ "supports_response_schema": true,
34370
+ "cache_read_input_token_cost": 0.00108,
34371
+ "input_cost_per_token_cache_hit": 0.00108
34372
+ },
34373
+ "novita/zai-org/glm-4.5": {
34374
+ "litellm_provider": "novita",
34375
+ "mode": "chat",
34376
+ "input_cost_per_token": 0.0048,
34377
+ "output_cost_per_token": 0.0176,
34378
+ "max_input_tokens": 131072,
34379
+ "max_output_tokens": 98304,
34380
+ "max_tokens": 98304,
34381
+ "supports_function_calling": true,
34382
+ "supports_parallel_function_calling": true,
34383
+ "supports_tool_choice": true,
34384
+ "supports_system_messages": true,
34385
+ "cache_read_input_token_cost": 0.00088,
34386
+ "input_cost_per_token_cache_hit": 0.00088,
34387
+ "supports_reasoning": true
34388
+ },
34389
+ "novita/qwen/qwen3-235b-a22b-thinking-2507": {
34390
+ "litellm_provider": "novita",
34391
+ "mode": "chat",
34392
+ "input_cost_per_token": 0.0024,
34393
+ "output_cost_per_token": 0.024,
34394
+ "max_input_tokens": 131072,
34395
+ "max_output_tokens": 32768,
34396
+ "max_tokens": 32768,
34397
+ "supports_function_calling": true,
34398
+ "supports_parallel_function_calling": true,
34399
+ "supports_tool_choice": true,
34400
+ "supports_system_messages": true,
34401
+ "supports_reasoning": true
34402
+ },
34403
+ "novita/meta-llama/llama-3.1-8b-instruct": {
34404
+ "litellm_provider": "novita",
34405
+ "mode": "chat",
34406
+ "input_cost_per_token": 0.0002,
34407
+ "output_cost_per_token": 0.0005,
34408
+ "max_input_tokens": 16384,
34409
+ "max_output_tokens": 16384,
34410
+ "max_tokens": 16384,
34411
+ "supports_system_messages": true
34412
+ },
34413
+ "novita/google/gemma-3-12b-it": {
34414
+ "litellm_provider": "novita",
34415
+ "mode": "chat",
34416
+ "input_cost_per_token": 0.0004,
34417
+ "output_cost_per_token": 0.0008,
34418
+ "max_input_tokens": 131072,
34419
+ "max_output_tokens": 8192,
34420
+ "max_tokens": 8192,
34421
+ "supports_tool_choice": true,
34422
+ "supports_vision": true,
34423
+ "supports_system_messages": true,
34424
+ "supports_response_schema": true
34425
+ },
34426
+ "novita/zai-org/glm-4.5v": {
34427
+ "litellm_provider": "novita",
34428
+ "mode": "chat",
34429
+ "input_cost_per_token": 0.0048,
34430
+ "output_cost_per_token": 0.0144,
34431
+ "max_input_tokens": 65536,
34432
+ "max_output_tokens": 16384,
34433
+ "max_tokens": 16384,
34434
+ "supports_function_calling": true,
34435
+ "supports_parallel_function_calling": true,
34436
+ "supports_tool_choice": true,
34437
+ "supports_vision": true,
34438
+ "supports_system_messages": true,
34439
+ "supports_response_schema": true,
34440
+ "cache_read_input_token_cost": 0.00088,
34441
+ "input_cost_per_token_cache_hit": 0.00088,
34442
+ "supports_reasoning": true
34443
+ },
34444
+ "novita/openai/gpt-oss-20b": {
34445
+ "litellm_provider": "novita",
34446
+ "mode": "chat",
34447
+ "input_cost_per_token": 0.00032,
34448
+ "output_cost_per_token": 0.0012,
34449
+ "max_input_tokens": 131072,
34450
+ "max_output_tokens": 32768,
34451
+ "max_tokens": 32768,
34452
+ "supports_tool_choice": true,
34453
+ "supports_vision": true,
34454
+ "supports_system_messages": true,
34455
+ "supports_response_schema": true,
34456
+ "supports_reasoning": true
34457
+ },
34458
+ "novita/qwen/qwen3-235b-a22b-instruct-2507": {
34459
+ "litellm_provider": "novita",
34460
+ "mode": "chat",
34461
+ "input_cost_per_token": 0.00072,
34462
+ "output_cost_per_token": 0.00464,
34463
+ "max_input_tokens": 131072,
34464
+ "max_output_tokens": 16384,
34465
+ "max_tokens": 16384,
34466
+ "supports_function_calling": true,
34467
+ "supports_parallel_function_calling": true,
34468
+ "supports_tool_choice": true,
34469
+ "supports_system_messages": true,
34470
+ "supports_response_schema": true
34471
+ },
34472
+ "novita/deepseek/deepseek-r1-distill-qwen-14b": {
34473
+ "litellm_provider": "novita",
34474
+ "mode": "chat",
34475
+ "input_cost_per_token": 0.0012,
34476
+ "output_cost_per_token": 0.0012,
34477
+ "max_input_tokens": 32768,
34478
+ "max_output_tokens": 16384,
34479
+ "max_tokens": 16384,
34480
+ "supports_tool_choice": true,
34481
+ "supports_system_messages": true,
34482
+ "supports_response_schema": true,
34483
+ "supports_reasoning": true
34484
+ },
34485
+ "novita/meta-llama/llama-3.3-70b-instruct": {
34486
+ "litellm_provider": "novita",
34487
+ "mode": "chat",
34488
+ "input_cost_per_token": 0.00108,
34489
+ "output_cost_per_token": 0.0032,
34490
+ "max_input_tokens": 131072,
34491
+ "max_output_tokens": 120000,
34492
+ "max_tokens": 120000,
34493
+ "supports_function_calling": true,
34494
+ "supports_parallel_function_calling": true,
34495
+ "supports_tool_choice": true,
34496
+ "supports_system_messages": true
34497
+ },
34498
+ "novita/qwen/qwen-2.5-72b-instruct": {
34499
+ "litellm_provider": "novita",
34500
+ "mode": "chat",
34501
+ "input_cost_per_token": 0.00304,
34502
+ "output_cost_per_token": 0.0032,
34503
+ "max_input_tokens": 32000,
34504
+ "max_output_tokens": 8192,
34505
+ "max_tokens": 8192,
34506
+ "supports_function_calling": true,
34507
+ "supports_parallel_function_calling": true,
34508
+ "supports_tool_choice": true,
34509
+ "supports_system_messages": true,
34510
+ "supports_response_schema": true
34511
+ },
34512
+ "novita/mistralai/mistral-nemo": {
34513
+ "litellm_provider": "novita",
34514
+ "mode": "chat",
34515
+ "input_cost_per_token": 0.00032,
34516
+ "output_cost_per_token": 0.00136,
34517
+ "max_input_tokens": 60288,
34518
+ "max_output_tokens": 16000,
34519
+ "max_tokens": 16000,
34520
+ "supports_tool_choice": true,
34521
+ "supports_system_messages": true,
34522
+ "supports_response_schema": true
34523
+ },
34524
+ "novita/minimaxai/minimax-m1-80k": {
34525
+ "litellm_provider": "novita",
34526
+ "mode": "chat",
34527
+ "input_cost_per_token": 0.0044,
34528
+ "output_cost_per_token": 0.0176,
34529
+ "max_input_tokens": 1000000,
34530
+ "max_output_tokens": 40000,
34531
+ "max_tokens": 40000,
34532
+ "supports_function_calling": true,
34533
+ "supports_parallel_function_calling": true,
34534
+ "supports_tool_choice": true,
34535
+ "supports_system_messages": true,
34536
+ "supports_reasoning": true
34537
+ },
34538
+ "novita/deepseek/deepseek-r1-0528": {
34539
+ "litellm_provider": "novita",
34540
+ "mode": "chat",
34541
+ "input_cost_per_token": 0.0056,
34542
+ "output_cost_per_token": 0.02,
34543
+ "max_input_tokens": 163840,
34544
+ "max_output_tokens": 32768,
34545
+ "max_tokens": 32768,
34546
+ "supports_function_calling": true,
34547
+ "supports_parallel_function_calling": true,
34548
+ "supports_tool_choice": true,
34549
+ "supports_system_messages": true,
34550
+ "supports_response_schema": true,
34551
+ "cache_read_input_token_cost": 0.0028,
34552
+ "input_cost_per_token_cache_hit": 0.0028,
34553
+ "supports_reasoning": true
34554
+ },
34555
+ "novita/deepseek/deepseek-r1-distill-qwen-32b": {
34556
+ "litellm_provider": "novita",
34557
+ "mode": "chat",
34558
+ "input_cost_per_token": 0.0024,
34559
+ "output_cost_per_token": 0.0024,
34560
+ "max_input_tokens": 64000,
34561
+ "max_output_tokens": 32000,
34562
+ "max_tokens": 32000,
34563
+ "supports_tool_choice": true,
34564
+ "supports_system_messages": true,
34565
+ "supports_response_schema": true,
34566
+ "supports_reasoning": true
34567
+ },
34568
+ "novita/meta-llama/llama-3-8b-instruct": {
34569
+ "litellm_provider": "novita",
34570
+ "mode": "chat",
34571
+ "input_cost_per_token": 0.00032,
34572
+ "output_cost_per_token": 0.00032,
34573
+ "max_input_tokens": 8192,
34574
+ "max_output_tokens": 8192,
34575
+ "max_tokens": 8192,
34576
+ "supports_system_messages": true
34577
+ },
34578
+ "novita/microsoft/wizardlm-2-8x22b": {
34579
+ "litellm_provider": "novita",
34580
+ "mode": "chat",
34581
+ "input_cost_per_token": 0.00496,
34582
+ "output_cost_per_token": 0.00496,
34583
+ "max_input_tokens": 65535,
34584
+ "max_output_tokens": 8000,
34585
+ "max_tokens": 8000,
34586
+ "supports_system_messages": true
34587
+ },
34588
+ "novita/deepseek/deepseek-r1-0528-qwen3-8b": {
34589
+ "litellm_provider": "novita",
34590
+ "mode": "chat",
34591
+ "input_cost_per_token": 0.00048,
34592
+ "output_cost_per_token": 0.00072,
34593
+ "max_input_tokens": 128000,
34594
+ "max_output_tokens": 32000,
34595
+ "max_tokens": 32000,
34596
+ "supports_system_messages": true,
34597
+ "supports_reasoning": true
34598
+ },
34599
+ "novita/deepseek/deepseek-r1-distill-llama-70b": {
34600
+ "litellm_provider": "novita",
34601
+ "mode": "chat",
34602
+ "input_cost_per_token": 0.0064,
34603
+ "output_cost_per_token": 0.0064,
34604
+ "max_input_tokens": 8192,
34605
+ "max_output_tokens": 8192,
34606
+ "max_tokens": 8192,
34607
+ "supports_tool_choice": true,
34608
+ "supports_system_messages": true,
34609
+ "supports_response_schema": true,
34610
+ "supports_reasoning": true
34611
+ },
34612
+ "novita/meta-llama/llama-3-70b-instruct": {
34613
+ "litellm_provider": "novita",
34614
+ "mode": "chat",
34615
+ "input_cost_per_token": 0.0051,
34616
+ "output_cost_per_token": 0.0074,
34617
+ "max_input_tokens": 8192,
34618
+ "max_output_tokens": 8000,
34619
+ "max_tokens": 8000,
34620
+ "supports_tool_choice": true,
34621
+ "supports_system_messages": true,
34622
+ "supports_response_schema": true
34623
+ },
34624
+ "novita/qwen/qwen3-235b-a22b-fp8": {
34625
+ "litellm_provider": "novita",
34626
+ "mode": "chat",
34627
+ "input_cost_per_token": 0.0016,
34628
+ "output_cost_per_token": 0.0064,
34629
+ "max_input_tokens": 40960,
34630
+ "max_output_tokens": 20000,
34631
+ "max_tokens": 20000,
34632
+ "supports_system_messages": true,
34633
+ "supports_reasoning": true
34634
+ },
34635
+ "novita/meta-llama/llama-4-maverick-17b-128e-instruct-fp8": {
34636
+ "litellm_provider": "novita",
34637
+ "mode": "chat",
34638
+ "input_cost_per_token": 0.0016,
34639
+ "output_cost_per_token": 0.0072,
34640
+ "max_input_tokens": 1048576,
34641
+ "max_output_tokens": 8192,
34642
+ "max_tokens": 8192,
34643
+ "supports_vision": true,
34644
+ "supports_system_messages": true
34645
+ },
34646
+ "novita/meta-llama/llama-4-scout-17b-16e-instruct": {
34647
+ "litellm_provider": "novita",
34648
+ "mode": "chat",
34649
+ "input_cost_per_token": 0.0008,
34650
+ "output_cost_per_token": 0.004,
34651
+ "max_input_tokens": 131072,
34652
+ "max_output_tokens": 131072,
34653
+ "max_tokens": 131072,
34654
+ "supports_vision": true,
34655
+ "supports_system_messages": true
34656
+ },
34657
+ "novita/nousresearch/hermes-2-pro-llama-3-8b": {
34658
+ "litellm_provider": "novita",
34659
+ "mode": "chat",
34660
+ "input_cost_per_token": 0.0014,
34661
+ "output_cost_per_token": 0.0014,
34662
+ "max_input_tokens": 8192,
34663
+ "max_output_tokens": 8192,
34664
+ "max_tokens": 8192,
34665
+ "supports_tool_choice": true,
34666
+ "supports_system_messages": true,
34667
+ "supports_response_schema": true
34668
+ },
34669
+ "novita/qwen/qwen2.5-vl-72b-instruct": {
34670
+ "litellm_provider": "novita",
34671
+ "mode": "chat",
34672
+ "input_cost_per_token": 0.0064,
34673
+ "output_cost_per_token": 0.0064,
34674
+ "max_input_tokens": 32768,
34675
+ "max_output_tokens": 32768,
34676
+ "max_tokens": 32768,
34677
+ "supports_vision": true,
34678
+ "supports_system_messages": true
34679
+ },
34680
+ "novita/sao10k/l3-70b-euryale-v2.1": {
34681
+ "litellm_provider": "novita",
34682
+ "mode": "chat",
34683
+ "input_cost_per_token": 0.0148,
34684
+ "output_cost_per_token": 0.0148,
34685
+ "max_input_tokens": 8192,
34686
+ "max_output_tokens": 8192,
34687
+ "max_tokens": 8192,
34688
+ "supports_function_calling": true,
34689
+ "supports_parallel_function_calling": true,
34690
+ "supports_tool_choice": true,
34691
+ "supports_system_messages": true
34692
+ },
34693
+ "novita/baidu/ernie-4.5-21B-a3b-thinking": {
34694
+ "litellm_provider": "novita",
34695
+ "mode": "chat",
34696
+ "input_cost_per_token": 0.00056,
34697
+ "output_cost_per_token": 0.00224,
34698
+ "max_input_tokens": 131072,
34699
+ "max_output_tokens": 65536,
34700
+ "max_tokens": 65536,
34701
+ "supports_system_messages": true,
34702
+ "supports_reasoning": true
34703
+ },
34704
+ "novita/sao10k/l3-8b-lunaris": {
34705
+ "litellm_provider": "novita",
34706
+ "mode": "chat",
34707
+ "input_cost_per_token": 0.0005,
34708
+ "output_cost_per_token": 0.0005,
34709
+ "max_input_tokens": 8192,
34710
+ "max_output_tokens": 8192,
34711
+ "max_tokens": 8192,
34712
+ "supports_tool_choice": true,
34713
+ "supports_system_messages": true,
34714
+ "supports_response_schema": true
34715
+ },
34716
+ "novita/baichuan/baichuan-m2-32b": {
34717
+ "litellm_provider": "novita",
34718
+ "mode": "chat",
34719
+ "input_cost_per_token": 0.00056,
34720
+ "output_cost_per_token": 0.00056,
34721
+ "max_input_tokens": 131072,
34722
+ "max_output_tokens": 131072,
34723
+ "max_tokens": 131072,
34724
+ "supports_tool_choice": true,
34725
+ "supports_system_messages": true,
34726
+ "supports_response_schema": true,
34727
+ "supports_reasoning": true
34728
+ },
34729
+ "novita/thudm/glm-4.1v-9b-thinking": {
34730
+ "litellm_provider": "novita",
34731
+ "mode": "chat",
34732
+ "input_cost_per_token": 0.00028,
34733
+ "output_cost_per_token": 0.001104,
34734
+ "max_input_tokens": 65536,
34735
+ "max_output_tokens": 8000,
34736
+ "max_tokens": 8000,
34737
+ "supports_vision": true,
34738
+ "supports_system_messages": true,
34739
+ "supports_reasoning": true
34740
+ },
34741
+ "novita/baidu/ernie-4.5-vl-424b-a47b": {
34742
+ "litellm_provider": "novita",
34743
+ "mode": "chat",
34744
+ "input_cost_per_token": 0.00336,
34745
+ "output_cost_per_token": 0.01,
34746
+ "max_input_tokens": 123000,
34747
+ "max_output_tokens": 16000,
34748
+ "max_tokens": 16000,
34749
+ "supports_vision": true,
34750
+ "supports_system_messages": true,
34751
+ "supports_reasoning": true
34752
+ },
34753
+ "novita/baidu/ernie-4.5-300b-a47b-paddle": {
34754
+ "litellm_provider": "novita",
34755
+ "mode": "chat",
34756
+ "input_cost_per_token": 0.00224,
34757
+ "output_cost_per_token": 0.0088,
34758
+ "max_input_tokens": 123000,
34759
+ "max_output_tokens": 12000,
34760
+ "max_tokens": 12000,
34761
+ "supports_tool_choice": true,
34762
+ "supports_system_messages": true,
34763
+ "supports_response_schema": true
34764
+ },
34765
+ "novita/deepseek/deepseek-prover-v2-671b": {
34766
+ "litellm_provider": "novita",
34767
+ "mode": "chat",
34768
+ "input_cost_per_token": 0.0056,
34769
+ "output_cost_per_token": 0.02,
34770
+ "max_input_tokens": 160000,
34771
+ "max_output_tokens": 160000,
34772
+ "max_tokens": 160000,
34773
+ "supports_system_messages": true
34774
+ },
34775
+ "novita/qwen/qwen3-32b-fp8": {
34776
+ "litellm_provider": "novita",
34777
+ "mode": "chat",
34778
+ "input_cost_per_token": 0.0008,
34779
+ "output_cost_per_token": 0.0036,
34780
+ "max_input_tokens": 40960,
34781
+ "max_output_tokens": 20000,
34782
+ "max_tokens": 20000,
34783
+ "supports_system_messages": true,
34784
+ "supports_reasoning": true
34785
+ },
34786
+ "novita/qwen/qwen3-30b-a3b-fp8": {
34787
+ "litellm_provider": "novita",
34788
+ "mode": "chat",
34789
+ "input_cost_per_token": 0.00072,
34790
+ "output_cost_per_token": 0.0036,
34791
+ "max_input_tokens": 40960,
34792
+ "max_output_tokens": 20000,
34793
+ "max_tokens": 20000,
34794
+ "supports_system_messages": true,
34795
+ "supports_reasoning": true
34796
+ },
34797
+ "novita/google/gemma-3-27b-it": {
34798
+ "litellm_provider": "novita",
34799
+ "mode": "chat",
34800
+ "input_cost_per_token": 0.000952,
34801
+ "output_cost_per_token": 0.0016,
34802
+ "max_input_tokens": 98304,
34803
+ "max_output_tokens": 16384,
34804
+ "max_tokens": 16384,
34805
+ "supports_vision": true,
34806
+ "supports_system_messages": true
34807
+ },
34808
+ "novita/deepseek/deepseek-v3-turbo": {
34809
+ "litellm_provider": "novita",
34810
+ "mode": "chat",
34811
+ "input_cost_per_token": 0.0032,
34812
+ "output_cost_per_token": 0.0104,
34813
+ "max_input_tokens": 64000,
34814
+ "max_output_tokens": 16000,
34815
+ "max_tokens": 16000,
34816
+ "supports_function_calling": true,
34817
+ "supports_parallel_function_calling": true,
34818
+ "supports_tool_choice": true,
34819
+ "supports_system_messages": true
34820
+ },
34821
+ "novita/deepseek/deepseek-r1-turbo": {
34822
+ "litellm_provider": "novita",
34823
+ "mode": "chat",
34824
+ "input_cost_per_token": 0.0056,
34825
+ "output_cost_per_token": 0.02,
34826
+ "max_input_tokens": 64000,
34827
+ "max_output_tokens": 16000,
34828
+ "max_tokens": 16000,
34829
+ "supports_function_calling": true,
34830
+ "supports_parallel_function_calling": true,
34831
+ "supports_tool_choice": true,
34832
+ "supports_system_messages": true,
34833
+ "supports_reasoning": true
34834
+ },
34835
+ "novita/Sao10K/L3-8B-Stheno-v3.2": {
34836
+ "litellm_provider": "novita",
34837
+ "mode": "chat",
34838
+ "input_cost_per_token": 0.0005,
34839
+ "output_cost_per_token": 0.0005,
34840
+ "max_input_tokens": 8192,
34841
+ "max_output_tokens": 32000,
34842
+ "max_tokens": 32000,
34843
+ "supports_function_calling": true,
34844
+ "supports_parallel_function_calling": true,
34845
+ "supports_tool_choice": true,
34846
+ "supports_system_messages": true
34847
+ },
34848
+ "novita/gryphe/mythomax-l2-13b": {
34849
+ "litellm_provider": "novita",
34850
+ "mode": "chat",
34851
+ "input_cost_per_token": 0.00072,
34852
+ "output_cost_per_token": 0.00072,
34853
+ "max_input_tokens": 4096,
34854
+ "max_output_tokens": 3200,
34855
+ "max_tokens": 3200,
34856
+ "supports_system_messages": true
34857
+ },
34858
+ "novita/baidu/ernie-4.5-vl-28b-a3b-thinking": {
34859
+ "litellm_provider": "novita",
34860
+ "mode": "chat",
34861
+ "input_cost_per_token": 0.0039,
34862
+ "output_cost_per_token": 0.0039,
34863
+ "max_input_tokens": 131072,
34864
+ "max_output_tokens": 65536,
34865
+ "max_tokens": 65536,
34866
+ "supports_function_calling": true,
34867
+ "supports_parallel_function_calling": true,
34868
+ "supports_tool_choice": true,
34869
+ "supports_vision": true,
34870
+ "supports_system_messages": true,
34871
+ "supports_response_schema": true,
34872
+ "supports_reasoning": true
34873
+ },
34874
+ "novita/qwen/qwen3-vl-8b-instruct": {
34875
+ "litellm_provider": "novita",
34876
+ "mode": "chat",
34877
+ "input_cost_per_token": 0.00064,
34878
+ "output_cost_per_token": 0.004,
34879
+ "max_input_tokens": 131072,
34880
+ "max_output_tokens": 32768,
34881
+ "max_tokens": 32768,
34882
+ "supports_function_calling": true,
34883
+ "supports_parallel_function_calling": true,
34884
+ "supports_tool_choice": true,
34885
+ "supports_vision": true,
34886
+ "supports_system_messages": true,
34887
+ "supports_response_schema": true
34888
+ },
34889
+ "novita/zai-org/glm-4.5-air": {
34890
+ "litellm_provider": "novita",
34891
+ "mode": "chat",
34892
+ "input_cost_per_token": 0.00104,
34893
+ "output_cost_per_token": 0.0068,
34894
+ "max_input_tokens": 131072,
34895
+ "max_output_tokens": 98304,
34896
+ "max_tokens": 98304,
34897
+ "supports_function_calling": true,
34898
+ "supports_parallel_function_calling": true,
34899
+ "supports_tool_choice": true,
34900
+ "supports_system_messages": true,
34901
+ "supports_reasoning": true
34902
+ },
34903
+ "novita/qwen/qwen3-vl-30b-a3b-instruct": {
34904
+ "litellm_provider": "novita",
34905
+ "mode": "chat",
34906
+ "input_cost_per_token": 0.0016,
34907
+ "output_cost_per_token": 0.0056,
34908
+ "max_input_tokens": 131072,
34909
+ "max_output_tokens": 32768,
34910
+ "max_tokens": 32768,
34911
+ "supports_function_calling": true,
34912
+ "supports_parallel_function_calling": true,
34913
+ "supports_tool_choice": true,
34914
+ "supports_vision": true,
34915
+ "supports_system_messages": true,
34916
+ "supports_response_schema": true
34917
+ },
34918
+ "novita/qwen/qwen3-vl-30b-a3b-thinking": {
34919
+ "litellm_provider": "novita",
34920
+ "mode": "chat",
34921
+ "input_cost_per_token": 0.0016,
34922
+ "output_cost_per_token": 0.008,
34923
+ "max_input_tokens": 131072,
34924
+ "max_output_tokens": 32768,
34925
+ "max_tokens": 32768,
34926
+ "supports_function_calling": true,
34927
+ "supports_parallel_function_calling": true,
34928
+ "supports_tool_choice": true,
34929
+ "supports_vision": true,
34930
+ "supports_system_messages": true,
34931
+ "supports_response_schema": true
34932
+ },
34933
+ "novita/qwen/qwen-mt-plus": {
34934
+ "litellm_provider": "novita",
34935
+ "mode": "chat",
34936
+ "input_cost_per_token": 0.002,
34937
+ "output_cost_per_token": 0.006,
34938
+ "max_input_tokens": 16384,
34939
+ "max_output_tokens": 8192,
34940
+ "max_tokens": 8192,
34941
+ "supports_system_messages": true
34942
+ },
34943
+ "novita/baidu/ernie-4.5-vl-28b-a3b": {
34944
+ "litellm_provider": "novita",
34945
+ "mode": "chat",
34946
+ "input_cost_per_token": 0.00112,
34947
+ "output_cost_per_token": 0.00448,
34948
+ "max_input_tokens": 30000,
34949
+ "max_output_tokens": 8000,
34950
+ "max_tokens": 8000,
34951
+ "supports_function_calling": true,
34952
+ "supports_parallel_function_calling": true,
34953
+ "supports_tool_choice": true,
34954
+ "supports_vision": true,
34955
+ "supports_system_messages": true,
34956
+ "supports_reasoning": true
34957
+ },
34958
+ "novita/baidu/ernie-4.5-21B-a3b": {
34959
+ "litellm_provider": "novita",
34960
+ "mode": "chat",
34961
+ "input_cost_per_token": 0.00056,
34962
+ "output_cost_per_token": 0.00224,
34963
+ "max_input_tokens": 120000,
34964
+ "max_output_tokens": 8000,
34965
+ "max_tokens": 8000,
34966
+ "supports_function_calling": true,
34967
+ "supports_parallel_function_calling": true,
34968
+ "supports_tool_choice": true,
34969
+ "supports_system_messages": true
34970
+ },
34971
+ "novita/qwen/qwen3-8b-fp8": {
34972
+ "litellm_provider": "novita",
34973
+ "mode": "chat",
34974
+ "input_cost_per_token": 0.00028,
34975
+ "output_cost_per_token": 0.001104,
34976
+ "max_input_tokens": 128000,
34977
+ "max_output_tokens": 20000,
34978
+ "max_tokens": 20000,
34979
+ "supports_system_messages": true,
34980
+ "supports_reasoning": true
34981
+ },
34982
+ "novita/qwen/qwen3-4b-fp8": {
34983
+ "litellm_provider": "novita",
34984
+ "mode": "chat",
34985
+ "input_cost_per_token": 0.00024,
34986
+ "output_cost_per_token": 0.00024,
34987
+ "max_input_tokens": 128000,
34988
+ "max_output_tokens": 20000,
34989
+ "max_tokens": 20000,
34990
+ "supports_system_messages": true,
34991
+ "supports_reasoning": true
34992
+ },
34993
+ "novita/qwen/qwen2.5-7b-instruct": {
34994
+ "litellm_provider": "novita",
34995
+ "mode": "chat",
34996
+ "input_cost_per_token": 0.00056,
34997
+ "output_cost_per_token": 0.00056,
34998
+ "max_input_tokens": 32000,
34999
+ "max_output_tokens": 32000,
35000
+ "max_tokens": 32000,
35001
+ "supports_function_calling": true,
35002
+ "supports_parallel_function_calling": true,
35003
+ "supports_tool_choice": true,
35004
+ "supports_system_messages": true,
35005
+ "supports_response_schema": true
35006
+ },
35007
+ "novita/meta-llama/llama-3.2-3b-instruct": {
35008
+ "litellm_provider": "novita",
35009
+ "mode": "chat",
35010
+ "input_cost_per_token": 0.00024,
35011
+ "output_cost_per_token": 0.0004,
35012
+ "max_input_tokens": 32768,
35013
+ "max_output_tokens": 32000,
35014
+ "max_tokens": 32000,
35015
+ "supports_function_calling": true,
35016
+ "supports_parallel_function_calling": true,
35017
+ "supports_tool_choice": true,
35018
+ "supports_system_messages": true
35019
+ },
35020
+ "novita/sao10k/l31-70b-euryale-v2.2": {
35021
+ "litellm_provider": "novita",
35022
+ "mode": "chat",
35023
+ "input_cost_per_token": 0.0148,
35024
+ "output_cost_per_token": 0.0148,
35025
+ "max_input_tokens": 8192,
35026
+ "max_output_tokens": 8192,
35027
+ "max_tokens": 8192,
35028
+ "supports_function_calling": true,
35029
+ "supports_parallel_function_calling": true,
35030
+ "supports_tool_choice": true,
35031
+ "supports_system_messages": true
35032
+ },
35033
+ "novita/qwen/qwen3-embedding-0.6b": {
35034
+ "litellm_provider": "novita",
35035
+ "mode": "embedding",
35036
+ "input_cost_per_token": 0.00056,
35037
+ "output_cost_per_token": 0,
35038
+ "max_input_tokens": 32768,
35039
+ "max_output_tokens": 32768,
35040
+ "max_tokens": 32768
35041
+ },
35042
+ "novita/qwen/qwen3-embedding-8b": {
35043
+ "litellm_provider": "novita",
35044
+ "mode": "embedding",
35045
+ "input_cost_per_token": 0.00056,
35046
+ "output_cost_per_token": 0,
35047
+ "max_input_tokens": 32768,
35048
+ "max_output_tokens": 4096,
35049
+ "max_tokens": 4096
35050
+ },
35051
+ "novita/baai/bge-m3": {
35052
+ "litellm_provider": "novita",
35053
+ "mode": "embedding",
35054
+ "input_cost_per_token": 0.0001,
35055
+ "output_cost_per_token": 0.0001,
35056
+ "max_input_tokens": 8192,
35057
+ "max_output_tokens": 96000,
35058
+ "max_tokens": 96000
35059
+ },
35060
+ "novita/qwen/qwen3-reranker-8b": {
35061
+ "litellm_provider": "novita",
35062
+ "mode": "rerank",
35063
+ "input_cost_per_token": 0.0004,
35064
+ "output_cost_per_token": 0.0004,
35065
+ "max_input_tokens": 32768,
35066
+ "max_output_tokens": 4096,
35067
+ "max_tokens": 4096
35068
+ },
35069
+ "novita/baai/bge-reranker-v2-m3": {
35070
+ "litellm_provider": "novita",
35071
+ "mode": "rerank",
35072
+ "input_cost_per_token": 0.0001,
35073
+ "output_cost_per_token": 0.0001,
35074
+ "max_input_tokens": 8000,
35075
+ "max_output_tokens": 8000,
35076
+ "max_tokens": 8000
33983
35077
  }
33984
35078
  }