tokencostauto 0.1.449__py3-none-any.whl → 0.1.462__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -16617,14 +16617,14 @@
16617
16617
  "supports_tool_choice": true
16618
16618
  },
16619
16619
  "azure_ai/grok-3": {
16620
- "input_cost_per_token": 3.3e-06,
16620
+ "input_cost_per_token": 3e-06,
16621
16621
  "litellm_provider": "azure_ai",
16622
16622
  "max_input_tokens": 131072,
16623
16623
  "max_output_tokens": 131072,
16624
16624
  "max_tokens": 131072,
16625
16625
  "mode": "chat",
16626
- "output_cost_per_token": 1.65e-05,
16627
- "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/",
16626
+ "output_cost_per_token": 1.5e-05,
16627
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
16628
16628
  "supports_function_calling": true,
16629
16629
  "supports_response_schema": false,
16630
16630
  "supports_tool_choice": true,
@@ -16660,14 +16660,14 @@
16660
16660
  "supports_web_search": true
16661
16661
  },
16662
16662
  "azure_ai/grok-3-mini": {
16663
- "input_cost_per_token": 2.75e-07,
16663
+ "input_cost_per_token": 2.5e-07,
16664
16664
  "litellm_provider": "azure_ai",
16665
16665
  "max_input_tokens": 131072,
16666
16666
  "max_output_tokens": 131072,
16667
16667
  "max_tokens": 131072,
16668
16668
  "mode": "chat",
16669
- "output_cost_per_token": 1.38e-06,
16670
- "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/",
16669
+ "output_cost_per_token": 1.27e-06,
16670
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
16671
16671
  "supports_function_calling": true,
16672
16672
  "supports_reasoning": true,
16673
16673
  "supports_response_schema": false,
@@ -17766,8 +17766,8 @@
17766
17766
  "supports_tool_choice": true
17767
17767
  },
17768
17768
  "gemini/gemini-2.5-flash-lite": {
17769
- "cache_read_input_token_cost": 2.5e-08,
17770
- "input_cost_per_audio_token": 5e-07,
17769
+ "cache_read_input_token_cost": 1e-08,
17770
+ "input_cost_per_audio_token": 3e-07,
17771
17771
  "input_cost_per_token": 1e-07,
17772
17772
  "litellm_provider": "gemini",
17773
17773
  "max_audio_length_hours": 8.4,
@@ -17813,8 +17813,8 @@
17813
17813
  "tpm": 250000
17814
17814
  },
17815
17815
  "gemini-2.5-flash-lite": {
17816
- "cache_read_input_token_cost": 2.5e-08,
17817
- "input_cost_per_audio_token": 5e-07,
17816
+ "cache_read_input_token_cost": 1e-08,
17817
+ "input_cost_per_audio_token": 3e-07,
17818
17818
  "input_cost_per_token": 1e-07,
17819
17819
  "litellm_provider": "vertex_ai-language-models",
17820
17820
  "max_audio_length_hours": 8.4,
@@ -18179,7 +18179,7 @@
18179
18179
  "input_cost_per_token_flex": 6.25e-07,
18180
18180
  "input_cost_per_token_priority": 2.5e-06,
18181
18181
  "litellm_provider": "openai",
18182
- "max_input_tokens": 400000,
18182
+ "max_input_tokens": 272000,
18183
18183
  "max_output_tokens": 128000,
18184
18184
  "max_tokens": 128000,
18185
18185
  "mode": "chat",
@@ -18218,7 +18218,7 @@
18218
18218
  "input_cost_per_token_flex": 1.25e-07,
18219
18219
  "input_cost_per_token_priority": 4.5e-07,
18220
18220
  "litellm_provider": "openai",
18221
- "max_input_tokens": 400000,
18221
+ "max_input_tokens": 272000,
18222
18222
  "max_output_tokens": 128000,
18223
18223
  "max_tokens": 128000,
18224
18224
  "mode": "chat",
@@ -18256,7 +18256,7 @@
18256
18256
  "input_cost_per_token_flex": 2.5e-08,
18257
18257
  "input_cost_per_token_priority": 2.5e-06,
18258
18258
  "litellm_provider": "openai",
18259
- "max_input_tokens": 400000,
18259
+ "max_input_tokens": 272000,
18260
18260
  "max_output_tokens": 128000,
18261
18261
  "max_tokens": 128000,
18262
18262
  "mode": "chat",
@@ -18357,7 +18357,7 @@
18357
18357
  "input_cost_per_token_flex": 6.25e-07,
18358
18358
  "input_cost_per_token_priority": 2.5e-06,
18359
18359
  "litellm_provider": "openai",
18360
- "max_input_tokens": 400000,
18360
+ "max_input_tokens": 272000,
18361
18361
  "max_output_tokens": 128000,
18362
18362
  "max_tokens": 128000,
18363
18363
  "mode": "chat",
@@ -18396,7 +18396,7 @@
18396
18396
  "input_cost_per_token_flex": 1.25e-07,
18397
18397
  "input_cost_per_token_priority": 4.5e-07,
18398
18398
  "litellm_provider": "openai",
18399
- "max_input_tokens": 400000,
18399
+ "max_input_tokens": 272000,
18400
18400
  "max_output_tokens": 128000,
18401
18401
  "max_tokens": 128000,
18402
18402
  "mode": "chat",
@@ -18433,7 +18433,7 @@
18433
18433
  "input_cost_per_token": 5e-08,
18434
18434
  "input_cost_per_token_flex": 2.5e-08,
18435
18435
  "litellm_provider": "openai",
18436
- "max_input_tokens": 400000,
18436
+ "max_input_tokens": 272000,
18437
18437
  "max_output_tokens": 128000,
18438
18438
  "max_tokens": 128000,
18439
18439
  "mode": "chat",
@@ -18720,13 +18720,14 @@
18720
18720
  "supports_vision": true
18721
18721
  },
18722
18722
  "groq/openai/gpt-oss-20b": {
18723
- "input_cost_per_token": 1e-07,
18723
+ "cache_read_input_token_cost": 3.75e-08,
18724
+ "input_cost_per_token": 7.5e-08,
18724
18725
  "litellm_provider": "groq",
18725
18726
  "max_input_tokens": 131072,
18726
18727
  "max_output_tokens": 32768,
18727
18728
  "max_tokens": 32768,
18728
18729
  "mode": "chat",
18729
- "output_cost_per_token": 5e-07,
18730
+ "output_cost_per_token": 3e-07,
18730
18731
  "supports_function_calling": true,
18731
18732
  "supports_parallel_function_calling": true,
18732
18733
  "supports_reasoning": true,
@@ -18735,13 +18736,14 @@
18735
18736
  "supports_web_search": true
18736
18737
  },
18737
18738
  "groq/openai/gpt-oss-120b": {
18739
+ "cache_read_input_token_cost": 7.5e-08,
18738
18740
  "input_cost_per_token": 1.5e-07,
18739
18741
  "litellm_provider": "groq",
18740
18742
  "max_input_tokens": 131072,
18741
18743
  "max_output_tokens": 32766,
18742
18744
  "max_tokens": 32766,
18743
18745
  "mode": "chat",
18744
- "output_cost_per_token": 7.5e-07,
18746
+ "output_cost_per_token": 6e-07,
18745
18747
  "supports_function_calling": true,
18746
18748
  "supports_parallel_function_calling": true,
18747
18749
  "supports_reasoning": true,
@@ -20623,7 +20625,7 @@
20623
20625
  "cache_read_input_token_cost": 2.5e-08,
20624
20626
  "input_cost_per_token": 2.5e-07,
20625
20627
  "litellm_provider": "openrouter",
20626
- "max_input_tokens": 400000,
20628
+ "max_input_tokens": 272000,
20627
20629
  "max_output_tokens": 128000,
20628
20630
  "max_tokens": 128000,
20629
20631
  "mode": "chat",
@@ -20642,7 +20644,7 @@
20642
20644
  "cache_read_input_token_cost": 5e-09,
20643
20645
  "input_cost_per_token": 5e-08,
20644
20646
  "litellm_provider": "openrouter",
20645
- "max_input_tokens": 400000,
20647
+ "max_input_tokens": 272000,
20646
20648
  "max_output_tokens": 128000,
20647
20649
  "max_tokens": 128000,
20648
20650
  "mode": "chat",
@@ -22961,7 +22963,7 @@
22961
22963
  "cache_read_input_token_cost": 1.25e-07,
22962
22964
  "input_cost_per_token": 1.25e-06,
22963
22965
  "litellm_provider": "openai",
22964
- "max_input_tokens": 400000,
22966
+ "max_input_tokens": 272000,
22965
22967
  "max_output_tokens": 128000,
22966
22968
  "max_tokens": 128000,
22967
22969
  "mode": "responses",
@@ -23080,7 +23082,7 @@
23080
23082
  "supports_tool_choice": true
23081
23083
  },
23082
23084
  "gemini-2.5-flash-lite-preview-09-2025": {
23083
- "cache_read_input_token_cost": 2.5e-08,
23085
+ "cache_read_input_token_cost": 1e-08,
23084
23086
  "input_cost_per_audio_token": 3e-07,
23085
23087
  "input_cost_per_token": 1e-07,
23086
23088
  "litellm_provider": "vertex_ai-language-models",
@@ -23260,7 +23262,7 @@
23260
23262
  "supports_web_search": true
23261
23263
  },
23262
23264
  "gemini/gemini-2.5-flash-lite-preview-09-2025": {
23263
- "cache_read_input_token_cost": 2.5e-08,
23265
+ "cache_read_input_token_cost": 1e-08,
23264
23266
  "input_cost_per_audio_token": 3e-07,
23265
23267
  "input_cost_per_token": 1e-07,
23266
23268
  "litellm_provider": "gemini",
@@ -23451,7 +23453,7 @@
23451
23453
  "cache_read_input_token_cost": 1.25e-07,
23452
23454
  "input_cost_per_token": 1.25e-06,
23453
23455
  "litellm_provider": "openrouter",
23454
- "max_input_tokens": 400000,
23456
+ "max_input_tokens": 272000,
23455
23457
  "max_output_tokens": 128000,
23456
23458
  "max_tokens": 128000,
23457
23459
  "mode": "chat",
@@ -23470,7 +23472,7 @@
23470
23472
  "cache_read_input_token_cost": 1.25e-07,
23471
23473
  "input_cost_per_token": 1.25e-06,
23472
23474
  "litellm_provider": "openrouter",
23473
- "max_input_tokens": 400000,
23475
+ "max_input_tokens": 272000,
23474
23476
  "max_output_tokens": 128000,
23475
23477
  "max_tokens": 128000,
23476
23478
  "mode": "chat",
@@ -23720,22 +23722,22 @@
23720
23722
  "supports_tool_choice": true
23721
23723
  },
23722
23724
  "azure_ai/grok-4": {
23723
- "input_cost_per_token": 5.5e-06,
23725
+ "input_cost_per_token": 3e-06,
23724
23726
  "litellm_provider": "azure_ai",
23725
23727
  "max_input_tokens": 131072,
23726
23728
  "max_output_tokens": 131072,
23727
23729
  "max_tokens": 131072,
23728
23730
  "mode": "chat",
23729
- "output_cost_per_token": 2.75e-05,
23730
- "source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/",
23731
+ "output_cost_per_token": 1.5e-05,
23732
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
23731
23733
  "supports_function_calling": true,
23732
23734
  "supports_response_schema": true,
23733
23735
  "supports_tool_choice": true,
23734
23736
  "supports_web_search": true
23735
23737
  },
23736
23738
  "azure_ai/grok-4-fast-non-reasoning": {
23737
- "input_cost_per_token": 4.3e-07,
23738
- "output_cost_per_token": 1.73e-06,
23739
+ "input_cost_per_token": 2e-07,
23740
+ "output_cost_per_token": 5e-07,
23739
23741
  "litellm_provider": "azure_ai",
23740
23742
  "max_input_tokens": 131072,
23741
23743
  "max_output_tokens": 131072,
@@ -23747,28 +23749,28 @@
23747
23749
  "supports_web_search": true
23748
23750
  },
23749
23751
  "azure_ai/grok-4-fast-reasoning": {
23750
- "input_cost_per_token": 4.3e-07,
23751
- "output_cost_per_token": 1.73e-06,
23752
+ "input_cost_per_token": 2e-07,
23753
+ "output_cost_per_token": 5e-07,
23752
23754
  "litellm_provider": "azure_ai",
23753
23755
  "max_input_tokens": 131072,
23754
23756
  "max_output_tokens": 131072,
23755
23757
  "max_tokens": 131072,
23756
23758
  "mode": "chat",
23757
- "source": "https://techcommunity.microsoft.com/blog/azure-ai-foundry-blog/announcing-the-grok-4-fast-models-from-xai-now-available-in-azure-ai-foundry/4456701",
23759
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
23758
23760
  "supports_function_calling": true,
23759
23761
  "supports_response_schema": true,
23760
23762
  "supports_tool_choice": true,
23761
23763
  "supports_web_search": true
23762
23764
  },
23763
23765
  "azure_ai/grok-code-fast-1": {
23764
- "input_cost_per_token": 3.5e-06,
23766
+ "input_cost_per_token": 2e-07,
23765
23767
  "litellm_provider": "azure_ai",
23766
23768
  "max_input_tokens": 131072,
23767
23769
  "max_output_tokens": 131072,
23768
23770
  "max_tokens": 131072,
23769
23771
  "mode": "chat",
23770
- "output_cost_per_token": 1.75e-05,
23771
- "source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/",
23772
+ "output_cost_per_token": 1.5e-06,
23773
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
23772
23774
  "supports_function_calling": true,
23773
23775
  "supports_response_schema": true,
23774
23776
  "supports_tool_choice": true,
@@ -24274,7 +24276,7 @@
24274
24276
  "input_cost_per_token": 1.5e-05,
24275
24277
  "input_cost_per_token_batches": 7.5e-06,
24276
24278
  "litellm_provider": "openai",
24277
- "max_input_tokens": 400000,
24279
+ "max_input_tokens": 128000,
24278
24280
  "max_output_tokens": 272000,
24279
24281
  "max_tokens": 272000,
24280
24282
  "mode": "responses",
@@ -24409,7 +24411,7 @@
24409
24411
  "input_cost_per_token": 1.5e-05,
24410
24412
  "input_cost_per_token_batches": 7.5e-06,
24411
24413
  "litellm_provider": "openai",
24412
- "max_input_tokens": 400000,
24414
+ "max_input_tokens": 128000,
24413
24415
  "max_output_tokens": 272000,
24414
24416
  "max_tokens": 272000,
24415
24417
  "mode": "responses",
@@ -26095,7 +26097,7 @@
26095
26097
  "input_cost_per_token": 1.25e-06,
26096
26098
  "input_cost_per_token_priority": 2.5e-06,
26097
26099
  "litellm_provider": "openai",
26098
- "max_input_tokens": 400000,
26100
+ "max_input_tokens": 272000,
26099
26101
  "max_output_tokens": 128000,
26100
26102
  "max_tokens": 128000,
26101
26103
  "mode": "chat",
@@ -26131,7 +26133,7 @@
26131
26133
  "input_cost_per_token": 1.25e-06,
26132
26134
  "input_cost_per_token_priority": 2.5e-06,
26133
26135
  "litellm_provider": "openai",
26134
- "max_input_tokens": 400000,
26136
+ "max_input_tokens": 272000,
26135
26137
  "max_output_tokens": 128000,
26136
26138
  "max_tokens": 128000,
26137
26139
  "mode": "chat",
@@ -26202,7 +26204,7 @@
26202
26204
  "input_cost_per_token": 1.25e-06,
26203
26205
  "input_cost_per_token_priority": 2.5e-06,
26204
26206
  "litellm_provider": "openai",
26205
- "max_input_tokens": 400000,
26207
+ "max_input_tokens": 272000,
26206
26208
  "max_output_tokens": 128000,
26207
26209
  "max_tokens": 128000,
26208
26210
  "mode": "responses",
@@ -26235,7 +26237,7 @@
26235
26237
  "input_cost_per_token": 2.5e-07,
26236
26238
  "input_cost_per_token_priority": 4.5e-07,
26237
26239
  "litellm_provider": "openai",
26238
- "max_input_tokens": 400000,
26240
+ "max_input_tokens": 272000,
26239
26241
  "max_output_tokens": 128000,
26240
26242
  "max_tokens": 128000,
26241
26243
  "mode": "responses",
@@ -26992,7 +26994,7 @@
26992
26994
  "input_cost_per_token": 1.25e-06,
26993
26995
  "input_cost_per_token_priority": 2.5e-06,
26994
26996
  "litellm_provider": "azure",
26995
- "max_input_tokens": 400000,
26997
+ "max_input_tokens": 272000,
26996
26998
  "max_output_tokens": 128000,
26997
26999
  "max_tokens": 128000,
26998
27000
  "mode": "chat",
@@ -27063,7 +27065,7 @@
27063
27065
  "input_cost_per_token": 1.25e-06,
27064
27066
  "input_cost_per_token_priority": 2.5e-06,
27065
27067
  "litellm_provider": "azure",
27066
- "max_input_tokens": 400000,
27068
+ "max_input_tokens": 272000,
27067
27069
  "max_output_tokens": 128000,
27068
27070
  "max_tokens": 128000,
27069
27071
  "mode": "responses",
@@ -27096,7 +27098,7 @@
27096
27098
  "input_cost_per_token": 2.5e-07,
27097
27099
  "input_cost_per_token_priority": 4.5e-07,
27098
27100
  "litellm_provider": "azure",
27099
- "max_input_tokens": 400000,
27101
+ "max_input_tokens": 272000,
27100
27102
  "max_output_tokens": 128000,
27101
27103
  "max_tokens": 128000,
27102
27104
  "mode": "responses",
@@ -28134,15 +28136,15 @@
28134
28136
  "tool_use_system_prompt_tokens": 159
28135
28137
  },
28136
28138
  "us.anthropic.claude-opus-4-5-20251101-v1:0": {
28137
- "cache_creation_input_token_cost": 6.25e-06,
28138
- "cache_read_input_token_cost": 5e-07,
28139
- "input_cost_per_token": 5e-06,
28139
+ "cache_creation_input_token_cost": 6.875e-06,
28140
+ "cache_read_input_token_cost": 5.5e-07,
28141
+ "input_cost_per_token": 5.5e-06,
28140
28142
  "litellm_provider": "bedrock_converse",
28141
28143
  "max_input_tokens": 200000,
28142
28144
  "max_output_tokens": 64000,
28143
28145
  "max_tokens": 64000,
28144
28146
  "mode": "chat",
28145
- "output_cost_per_token": 2.5e-05,
28147
+ "output_cost_per_token": 2.75e-05,
28146
28148
  "search_context_cost_per_query": {
28147
28149
  "search_context_size_high": 0.01,
28148
28150
  "search_context_size_low": 0.01,
@@ -28658,7 +28660,7 @@
28658
28660
  "input_cost_per_token": 1.24999e-06,
28659
28661
  "input_dbu_cost_per_token": 1.7857e-05,
28660
28662
  "litellm_provider": "databricks",
28661
- "max_input_tokens": 400000,
28663
+ "max_input_tokens": 272000,
28662
28664
  "max_output_tokens": 128000,
28663
28665
  "max_tokens": 128000,
28664
28666
  "metadata": {
@@ -28673,7 +28675,7 @@
28673
28675
  "input_cost_per_token": 1.24999e-06,
28674
28676
  "input_dbu_cost_per_token": 1.7857e-05,
28675
28677
  "litellm_provider": "databricks",
28676
- "max_input_tokens": 400000,
28678
+ "max_input_tokens": 272000,
28677
28679
  "max_output_tokens": 128000,
28678
28680
  "max_tokens": 128000,
28679
28681
  "metadata": {
@@ -28688,7 +28690,7 @@
28688
28690
  "input_cost_per_token": 2.4997000000000006e-07,
28689
28691
  "input_dbu_cost_per_token": 3.571e-06,
28690
28692
  "litellm_provider": "databricks",
28691
- "max_input_tokens": 400000,
28693
+ "max_input_tokens": 272000,
28692
28694
  "max_output_tokens": 128000,
28693
28695
  "max_tokens": 128000,
28694
28696
  "metadata": {
@@ -28703,7 +28705,7 @@
28703
28705
  "input_cost_per_token": 4.998e-08,
28704
28706
  "input_dbu_cost_per_token": 7.14e-07,
28705
28707
  "litellm_provider": "databricks",
28706
- "max_input_tokens": 400000,
28708
+ "max_input_tokens": 272000,
28707
28709
  "max_output_tokens": 128000,
28708
28710
  "max_tokens": 128000,
28709
28711
  "metadata": {
@@ -29165,7 +29167,7 @@
29165
29167
  "cache_read_input_token_cost": 1.25e-07,
29166
29168
  "input_cost_per_token": 1.25e-06,
29167
29169
  "litellm_provider": "openai",
29168
- "max_input_tokens": 400000,
29170
+ "max_input_tokens": 272000,
29169
29171
  "max_output_tokens": 128000,
29170
29172
  "max_tokens": 128000,
29171
29173
  "mode": "responses",
@@ -29525,6 +29527,7 @@
29525
29527
  "supports_pdf_input": true
29526
29528
  },
29527
29529
  "cerebras/zai-glm-4.6": {
29530
+ "deprecation_date": "2026-01-20",
29528
29531
  "input_cost_per_token": 2.25e-06,
29529
29532
  "litellm_provider": "cerebras",
29530
29533
  "max_input_tokens": 128000,
@@ -31715,7 +31718,7 @@
31715
31718
  "input_cost_per_token": 1.75e-06,
31716
31719
  "input_cost_per_token_priority": 3.5e-06,
31717
31720
  "litellm_provider": "openai",
31718
- "max_input_tokens": 400000,
31721
+ "max_input_tokens": 272000,
31719
31722
  "max_output_tokens": 128000,
31720
31723
  "max_tokens": 128000,
31721
31724
  "mode": "chat",
@@ -31752,7 +31755,7 @@
31752
31755
  "input_cost_per_token": 1.75e-06,
31753
31756
  "input_cost_per_token_priority": 3.5e-06,
31754
31757
  "litellm_provider": "openai",
31755
- "max_input_tokens": 400000,
31758
+ "max_input_tokens": 272000,
31756
31759
  "max_output_tokens": 128000,
31757
31760
  "max_tokens": 128000,
31758
31761
  "mode": "chat",
@@ -31820,7 +31823,7 @@
31820
31823
  "gpt-5.2-pro": {
31821
31824
  "input_cost_per_token": 2.1e-05,
31822
31825
  "litellm_provider": "openai",
31823
- "max_input_tokens": 400000,
31826
+ "max_input_tokens": 272000,
31824
31827
  "max_output_tokens": 128000,
31825
31828
  "max_tokens": 128000,
31826
31829
  "mode": "responses",
@@ -31851,7 +31854,7 @@
31851
31854
  "gpt-5.2-pro-2025-12-11": {
31852
31855
  "input_cost_per_token": 2.1e-05,
31853
31856
  "litellm_provider": "openai",
31854
- "max_input_tokens": 400000,
31857
+ "max_input_tokens": 272000,
31855
31858
  "max_output_tokens": 128000,
31856
31859
  "max_tokens": 128000,
31857
31860
  "mode": "responses",
@@ -31960,7 +31963,7 @@
31960
31963
  "input_cost_per_token": 1.75e-06,
31961
31964
  "input_cost_per_token_priority": 3.5e-06,
31962
31965
  "litellm_provider": "azure",
31963
- "max_input_tokens": 400000,
31966
+ "max_input_tokens": 272000,
31964
31967
  "max_output_tokens": 128000,
31965
31968
  "max_tokens": 128000,
31966
31969
  "mode": "chat",
@@ -32296,7 +32299,7 @@
32296
32299
  "cache_read_input_token_cost": 1.75e-07,
32297
32300
  "input_cost_per_token": 1.75e-06,
32298
32301
  "litellm_provider": "openrouter",
32299
- "max_input_tokens": 400000,
32302
+ "max_input_tokens": 272000,
32300
32303
  "max_output_tokens": 128000,
32301
32304
  "max_tokens": 128000,
32302
32305
  "mode": "chat",
@@ -32326,7 +32329,7 @@
32326
32329
  "input_cost_per_image": 0,
32327
32330
  "input_cost_per_token": 2.1e-05,
32328
32331
  "litellm_provider": "openrouter",
32329
- "max_input_tokens": 400000,
32332
+ "max_input_tokens": 272000,
32330
32333
  "max_output_tokens": 128000,
32331
32334
  "max_tokens": 128000,
32332
32335
  "mode": "chat",
@@ -33982,13 +33985,13 @@
33982
33985
  ]
33983
33986
  },
33984
33987
  "vertex_ai/zai-org/glm-4.7-maas": {
33985
- "input_cost_per_token": 3e-07,
33988
+ "input_cost_per_token": 6e-07,
33986
33989
  "litellm_provider": "vertex_ai-zai_models",
33987
33990
  "max_input_tokens": 200000,
33988
33991
  "max_output_tokens": 128000,
33989
33992
  "max_tokens": 128000,
33990
33993
  "mode": "chat",
33991
- "output_cost_per_token": 1.2e-06,
33994
+ "output_cost_per_token": 2.2e-06,
33992
33995
  "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
33993
33996
  "supports_function_calling": true,
33994
33997
  "supports_reasoning": true,
@@ -35445,5 +35448,214 @@
35445
35448
  "supports_response_schema": true,
35446
35449
  "supports_audio_input": true,
35447
35450
  "supports_audio_output": true
35451
+ },
35452
+ "cerebras/zai-glm-4.7": {
35453
+ "input_cost_per_token": 2.25e-06,
35454
+ "litellm_provider": "cerebras",
35455
+ "max_input_tokens": 128000,
35456
+ "max_output_tokens": 128000,
35457
+ "max_tokens": 128000,
35458
+ "mode": "chat",
35459
+ "output_cost_per_token": 2.75e-06,
35460
+ "source": "https://www.cerebras.ai/pricing",
35461
+ "supports_function_calling": true,
35462
+ "supports_reasoning": true,
35463
+ "supports_tool_choice": true
35464
+ },
35465
+ "gpt-5.2-codex": {
35466
+ "cache_read_input_token_cost": 1.75e-07,
35467
+ "cache_read_input_token_cost_priority": 3.5e-07,
35468
+ "input_cost_per_token": 1.75e-06,
35469
+ "input_cost_per_token_priority": 3.5e-06,
35470
+ "litellm_provider": "openai",
35471
+ "max_input_tokens": 400000,
35472
+ "max_output_tokens": 128000,
35473
+ "max_tokens": 128000,
35474
+ "mode": "responses",
35475
+ "output_cost_per_token": 1.4e-05,
35476
+ "output_cost_per_token_priority": 2.8e-05,
35477
+ "supported_endpoints": [
35478
+ "/v1/responses"
35479
+ ],
35480
+ "supported_modalities": [
35481
+ "text",
35482
+ "image"
35483
+ ],
35484
+ "supported_output_modalities": [
35485
+ "text"
35486
+ ],
35487
+ "supports_function_calling": true,
35488
+ "supports_native_streaming": true,
35489
+ "supports_parallel_function_calling": true,
35490
+ "supports_pdf_input": true,
35491
+ "supports_prompt_caching": true,
35492
+ "supports_reasoning": true,
35493
+ "supports_response_schema": true,
35494
+ "supports_system_messages": false,
35495
+ "supports_tool_choice": true,
35496
+ "supports_vision": true
35497
+ },
35498
+ "openrouter/openai/gpt-5.2-codex": {
35499
+ "cache_read_input_token_cost": 1.75e-07,
35500
+ "input_cost_per_token": 1.75e-06,
35501
+ "litellm_provider": "openrouter",
35502
+ "max_input_tokens": 400000,
35503
+ "max_output_tokens": 128000,
35504
+ "max_tokens": 128000,
35505
+ "mode": "chat",
35506
+ "output_cost_per_token": 1.4e-05,
35507
+ "supported_modalities": [
35508
+ "text",
35509
+ "image"
35510
+ ],
35511
+ "supported_output_modalities": [
35512
+ "text"
35513
+ ],
35514
+ "supports_reasoning": true,
35515
+ "supports_tool_choice": true
35516
+ },
35517
+ "azure/gpt-5.2-codex": {
35518
+ "cache_read_input_token_cost": 1.75e-07,
35519
+ "input_cost_per_token": 1.75e-06,
35520
+ "litellm_provider": "azure",
35521
+ "max_input_tokens": 128000,
35522
+ "max_output_tokens": 16384,
35523
+ "max_tokens": 16384,
35524
+ "mode": "chat",
35525
+ "output_cost_per_token": 1.4e-05,
35526
+ "supported_endpoints": [
35527
+ "/v1/chat/completions",
35528
+ "/v1/responses"
35529
+ ],
35530
+ "supported_modalities": [
35531
+ "text",
35532
+ "image"
35533
+ ],
35534
+ "supported_output_modalities": [
35535
+ "text"
35536
+ ],
35537
+ "supports_function_calling": true,
35538
+ "supports_native_streaming": true,
35539
+ "supports_parallel_function_calling": true,
35540
+ "supports_pdf_input": true,
35541
+ "supports_prompt_caching": true,
35542
+ "supports_reasoning": true,
35543
+ "supports_response_schema": true,
35544
+ "supports_system_messages": true,
35545
+ "supports_tool_choice": true,
35546
+ "supports_vision": true
35547
+ },
35548
+ "dall-e-2": {
35549
+ "input_cost_per_image": 0.02,
35550
+ "litellm_provider": "openai",
35551
+ "mode": "image_generation",
35552
+ "supported_endpoints": [
35553
+ "/v1/images/generations",
35554
+ "/v1/images/edits",
35555
+ "/v1/images/variations"
35556
+ ]
35557
+ },
35558
+ "dall-e-3": {
35559
+ "input_cost_per_image": 0.04,
35560
+ "litellm_provider": "openai",
35561
+ "mode": "image_generation",
35562
+ "supported_endpoints": [
35563
+ "/v1/images/generations"
35564
+ ]
35565
+ },
35566
+ "gemini-2.5-computer-use-preview-10-2025": {
35567
+ "input_cost_per_token": 1.25e-06,
35568
+ "input_cost_per_token_above_200k_tokens": 2.5e-06,
35569
+ "litellm_provider": "vertex_ai-language-models",
35570
+ "max_images_per_prompt": 3000,
35571
+ "max_input_tokens": 128000,
35572
+ "max_output_tokens": 64000,
35573
+ "max_tokens": 64000,
35574
+ "mode": "chat",
35575
+ "output_cost_per_token": 1e-05,
35576
+ "output_cost_per_token_above_200k_tokens": 1.5e-05,
35577
+ "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/computer-use",
35578
+ "supported_modalities": [
35579
+ "text",
35580
+ "image"
35581
+ ],
35582
+ "supported_output_modalities": [
35583
+ "text"
35584
+ ],
35585
+ "supports_computer_use": true,
35586
+ "supports_function_calling": true,
35587
+ "supports_system_messages": true,
35588
+ "supports_tool_choice": true,
35589
+ "supports_vision": true
35590
+ },
35591
+ "chatgpt/gpt-5.2-codex": {
35592
+ "litellm_provider": "chatgpt",
35593
+ "max_input_tokens": 128000,
35594
+ "max_output_tokens": 128000,
35595
+ "max_tokens": 128000,
35596
+ "mode": "responses",
35597
+ "supported_endpoints": [
35598
+ "/v1/responses"
35599
+ ],
35600
+ "supports_function_calling": true,
35601
+ "supports_parallel_function_calling": true,
35602
+ "supports_response_schema": true,
35603
+ "supports_vision": true
35604
+ },
35605
+ "chatgpt/gpt-5.2": {
35606
+ "litellm_provider": "chatgpt",
35607
+ "max_input_tokens": 128000,
35608
+ "max_output_tokens": 64000,
35609
+ "max_tokens": 64000,
35610
+ "mode": "responses",
35611
+ "supported_endpoints": [
35612
+ "/v1/chat/completions",
35613
+ "/v1/responses"
35614
+ ],
35615
+ "supports_function_calling": true,
35616
+ "supports_parallel_function_calling": true,
35617
+ "supports_response_schema": true,
35618
+ "supports_vision": true
35619
+ },
35620
+ "chatgpt/gpt-5.1-codex-max": {
35621
+ "litellm_provider": "chatgpt",
35622
+ "max_input_tokens": 128000,
35623
+ "max_output_tokens": 128000,
35624
+ "max_tokens": 128000,
35625
+ "mode": "responses",
35626
+ "supported_endpoints": [
35627
+ "/v1/responses"
35628
+ ],
35629
+ "supports_function_calling": true,
35630
+ "supports_parallel_function_calling": true,
35631
+ "supports_response_schema": true,
35632
+ "supports_vision": true
35633
+ },
35634
+ "chatgpt/gpt-5.1-codex-mini": {
35635
+ "litellm_provider": "chatgpt",
35636
+ "max_input_tokens": 128000,
35637
+ "max_output_tokens": 64000,
35638
+ "max_tokens": 64000,
35639
+ "mode": "responses",
35640
+ "supported_endpoints": [
35641
+ "/v1/responses"
35642
+ ],
35643
+ "supports_function_calling": true,
35644
+ "supports_parallel_function_calling": true,
35645
+ "supports_response_schema": true,
35646
+ "supports_vision": true
35647
+ },
35648
+ "sarvam/sarvam-m": {
35649
+ "cache_creation_input_token_cost": 0,
35650
+ "cache_creation_input_token_cost_above_1hr": 0,
35651
+ "cache_read_input_token_cost": 0,
35652
+ "input_cost_per_token": 0,
35653
+ "litellm_provider": "sarvam",
35654
+ "max_input_tokens": 8192,
35655
+ "max_output_tokens": 32000,
35656
+ "max_tokens": 32000,
35657
+ "mode": "chat",
35658
+ "output_cost_per_token": 0,
35659
+ "supports_reasoning": true
35448
35660
  }
35449
35661
  }
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: tokencostauto
3
- Version: 0.1.449
3
+ Version: 0.1.462
4
4
  Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
5
5
  Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
6
6
  Project-URL: Homepage, https://github.com/madpin/tokencostaudo
@@ -0,0 +1,9 @@
1
+ tokencostauto/__init__.py,sha256=-4d_ryFH62SgNXPXA8vGPFZoAKtOBjnsg37EB_RkZG8,289
2
+ tokencostauto/constants.py,sha256=_82MlTkTrdrwzyRosQD7d3JdgNP9KAUM-cZo8DE00P0,3395
3
+ tokencostauto/costs.py,sha256=tXsgrTypq-dCHaHtoXcg2XepezWsAvZpl9gEsv_53iE,10679
4
+ tokencostauto/model_prices.json,sha256=aDFZZR6fnd6N4Fdc-zL3m6tsqmR8KiCelIqRuvx8oas,1265556
5
+ tokencostauto-0.1.462.dist-info/licenses/LICENSE,sha256=4PLv_CD6Ughnsvg_nM2XeTqGwVK6lQVR77kVWbPq-0U,1065
6
+ tokencostauto-0.1.462.dist-info/METADATA,sha256=bLqFKmPlxZSRF2VHZpr91xFCtaAPO1hdtx9yMt8TEbg,204076
7
+ tokencostauto-0.1.462.dist-info/WHEEL,sha256=qELbo2s1Yzl39ZmrAibXA2jjPLUYfnVhUNTlyF1rq0Y,92
8
+ tokencostauto-0.1.462.dist-info/top_level.txt,sha256=szZQTUJRotfIaeZCDsOgvofIkLt2ak88RP13oI51-TU,14
9
+ tokencostauto-0.1.462.dist-info/RECORD,,
@@ -1,5 +1,5 @@
1
1
  Wheel-Version: 1.0
2
- Generator: setuptools (80.9.0)
2
+ Generator: setuptools (80.10.1)
3
3
  Root-Is-Purelib: true
4
4
  Tag: py3-none-any
5
5
 
@@ -1,9 +0,0 @@
1
- tokencostauto/__init__.py,sha256=-4d_ryFH62SgNXPXA8vGPFZoAKtOBjnsg37EB_RkZG8,289
2
- tokencostauto/constants.py,sha256=_82MlTkTrdrwzyRosQD7d3JdgNP9KAUM-cZo8DE00P0,3395
3
- tokencostauto/costs.py,sha256=tXsgrTypq-dCHaHtoXcg2XepezWsAvZpl9gEsv_53iE,10679
4
- tokencostauto/model_prices.json,sha256=4MmQTNpNS3KNIbeLVss71S1D9YYkN2nLCmPIhBnMLl4,1258836
5
- tokencostauto-0.1.449.dist-info/licenses/LICENSE,sha256=4PLv_CD6Ughnsvg_nM2XeTqGwVK6lQVR77kVWbPq-0U,1065
6
- tokencostauto-0.1.449.dist-info/METADATA,sha256=Rw8U0zJEphYOT0W-Q0fuJbhuT5f91sQdzKXaXghGg-g,204076
7
- tokencostauto-0.1.449.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
8
- tokencostauto-0.1.449.dist-info/top_level.txt,sha256=szZQTUJRotfIaeZCDsOgvofIkLt2ak88RP13oI51-TU,14
9
- tokencostauto-0.1.449.dist-info/RECORD,,