tokencostauto 0.1.329__tar.gz → 0.1.333__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {tokencostauto-0.1.329/tokencostauto.egg-info → tokencostauto-0.1.333}/PKG-INFO +1 -1
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/pyproject.toml +1 -1
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/tokencostauto/model_prices.json +50 -11
- {tokencostauto-0.1.329 → tokencostauto-0.1.333/tokencostauto.egg-info}/PKG-INFO +1 -1
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/LICENSE +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/MANIFEST.in +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/README.md +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/setup.cfg +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/tests/test_costs.py +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/tokencostauto/__init__.py +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/tokencostauto/constants.py +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/tokencostauto/costs.py +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/tokencostauto.egg-info/SOURCES.txt +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/tokencostauto.egg-info/dependency_links.txt +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/tokencostauto.egg-info/requires.txt +0 -0
- {tokencostauto-0.1.329 → tokencostauto-0.1.333}/tokencostauto.egg-info/top_level.txt +0 -0
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: tokencostauto
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.333
|
|
4
4
|
Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
|
|
5
5
|
Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
|
|
6
6
|
Project-URL: Homepage, https://github.com/madpin/tokencostaudo
|
|
@@ -16267,7 +16267,6 @@
|
|
|
16267
16267
|
"output_cost_per_token": 1.5e-05,
|
|
16268
16268
|
"source": "https://docs.x.ai/docs/models",
|
|
16269
16269
|
"supports_function_calling": true,
|
|
16270
|
-
"supports_reasoning": true,
|
|
16271
16270
|
"supports_tool_choice": true,
|
|
16272
16271
|
"supports_web_search": true
|
|
16273
16272
|
},
|
|
@@ -16283,7 +16282,6 @@
|
|
|
16283
16282
|
"output_cost_per_token_above_128k_tokens": 3e-05,
|
|
16284
16283
|
"source": "https://docs.x.ai/docs/models",
|
|
16285
16284
|
"supports_function_calling": true,
|
|
16286
|
-
"supports_reasoning": true,
|
|
16287
16285
|
"supports_tool_choice": true,
|
|
16288
16286
|
"supports_web_search": true
|
|
16289
16287
|
},
|
|
@@ -16299,7 +16297,6 @@
|
|
|
16299
16297
|
"output_cost_per_token_above_128k_tokens": 3e-05,
|
|
16300
16298
|
"source": "https://docs.x.ai/docs/models",
|
|
16301
16299
|
"supports_function_calling": true,
|
|
16302
|
-
"supports_reasoning": true,
|
|
16303
16300
|
"supports_tool_choice": true,
|
|
16304
16301
|
"supports_web_search": true
|
|
16305
16302
|
},
|
|
@@ -22779,7 +22776,6 @@
|
|
|
22779
22776
|
"cache_read_input_token_cost": 5e-08,
|
|
22780
22777
|
"source": "https://docs.x.ai/docs/models",
|
|
22781
22778
|
"supports_function_calling": true,
|
|
22782
|
-
"supports_reasoning": true,
|
|
22783
22779
|
"supports_tool_choice": true,
|
|
22784
22780
|
"supports_web_search": true
|
|
22785
22781
|
},
|
|
@@ -23635,7 +23631,6 @@
|
|
|
23635
23631
|
"output_cost_per_token": 2.75e-05,
|
|
23636
23632
|
"source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/",
|
|
23637
23633
|
"supports_function_calling": true,
|
|
23638
|
-
"supports_reasoning": true,
|
|
23639
23634
|
"supports_response_schema": true,
|
|
23640
23635
|
"supports_tool_choice": true,
|
|
23641
23636
|
"supports_web_search": true
|
|
@@ -23663,7 +23658,6 @@
|
|
|
23663
23658
|
"mode": "chat",
|
|
23664
23659
|
"source": "https://techcommunity.microsoft.com/blog/azure-ai-foundry-blog/announcing-the-grok-4-fast-models-from-xai-now-available-in-azure-ai-foundry/4456701",
|
|
23665
23660
|
"supports_function_calling": true,
|
|
23666
|
-
"supports_reasoning": true,
|
|
23667
23661
|
"supports_response_schema": true,
|
|
23668
23662
|
"supports_tool_choice": true,
|
|
23669
23663
|
"supports_web_search": true
|
|
@@ -24883,19 +24877,21 @@
|
|
|
24883
24877
|
"supports_vision": false
|
|
24884
24878
|
},
|
|
24885
24879
|
"global.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
24886
|
-
"cache_creation_input_token_cost": 1.
|
|
24887
|
-
"cache_read_input_token_cost": 1e-07,
|
|
24888
|
-
"input_cost_per_token": 1e-06,
|
|
24880
|
+
"cache_creation_input_token_cost": 1.375e-06,
|
|
24881
|
+
"cache_read_input_token_cost": 1.1e-07,
|
|
24882
|
+
"input_cost_per_token": 1.1e-06,
|
|
24889
24883
|
"litellm_provider": "bedrock_converse",
|
|
24890
24884
|
"max_input_tokens": 200000,
|
|
24891
24885
|
"max_output_tokens": 8192,
|
|
24892
24886
|
"max_tokens": 8192,
|
|
24893
24887
|
"mode": "chat",
|
|
24894
|
-
"output_cost_per_token": 5e-06,
|
|
24888
|
+
"output_cost_per_token": 5.5e-06,
|
|
24889
|
+
"source": "https://aws.amazon.com/about-aws/whats-new/2025/10/claude-4-5-haiku-anthropic-amazon-bedrock",
|
|
24895
24890
|
"supports_assistant_prefill": true,
|
|
24896
24891
|
"supports_function_calling": true,
|
|
24897
24892
|
"supports_pdf_input": true,
|
|
24898
24893
|
"supports_prompt_caching": true,
|
|
24894
|
+
"supports_reasoning": true,
|
|
24899
24895
|
"supports_response_schema": true,
|
|
24900
24896
|
"supports_tool_choice": true,
|
|
24901
24897
|
"supports_vision": true,
|
|
@@ -25466,6 +25462,7 @@
|
|
|
25466
25462
|
"mode": "chat",
|
|
25467
25463
|
"output_cost_per_audio_token": 1.2e-05,
|
|
25468
25464
|
"output_cost_per_token": 2e-06,
|
|
25465
|
+
"rpm": 100000,
|
|
25469
25466
|
"source": "https://ai.google.dev/gemini-api/docs/pricing",
|
|
25470
25467
|
"supported_endpoints": [
|
|
25471
25468
|
"/v1/chat/completions",
|
|
@@ -25492,7 +25489,8 @@
|
|
|
25492
25489
|
"supports_tool_choice": true,
|
|
25493
25490
|
"supports_url_context": true,
|
|
25494
25491
|
"supports_vision": true,
|
|
25495
|
-
"supports_web_search": true
|
|
25492
|
+
"supports_web_search": true,
|
|
25493
|
+
"tpm": 8000000
|
|
25496
25494
|
},
|
|
25497
25495
|
"azure/gpt-image-1-mini": {
|
|
25498
25496
|
"input_cost_per_pixel": 8.0566406e-09,
|
|
@@ -25670,5 +25668,46 @@
|
|
|
25670
25668
|
"metadata": {
|
|
25671
25669
|
"notes": "SearXNG is an open-source metasearch engine. Free to use when self-hosted or using public instances."
|
|
25672
25670
|
}
|
|
25671
|
+
},
|
|
25672
|
+
"azure/gpt-5-pro": {
|
|
25673
|
+
"input_cost_per_token": 1.5e-05,
|
|
25674
|
+
"litellm_provider": "azure",
|
|
25675
|
+
"max_input_tokens": 272000,
|
|
25676
|
+
"max_output_tokens": 128000,
|
|
25677
|
+
"max_tokens": 400000,
|
|
25678
|
+
"mode": "responses",
|
|
25679
|
+
"output_cost_per_token": 0.00012,
|
|
25680
|
+
"source": "https://learn.microsoft.com/en-us/azure/ai-foundry/foundry-models/concepts/models-sold-directly-by-azure?pivots=azure-openai&tabs=global-standard-aoai%2Cstandard-chat-completions%2Cglobal-standard#gpt-5",
|
|
25681
|
+
"supported_endpoints": [
|
|
25682
|
+
"/v1/responses"
|
|
25683
|
+
],
|
|
25684
|
+
"supported_modalities": [
|
|
25685
|
+
"text",
|
|
25686
|
+
"image"
|
|
25687
|
+
],
|
|
25688
|
+
"supported_output_modalities": [
|
|
25689
|
+
"text"
|
|
25690
|
+
],
|
|
25691
|
+
"supports_function_calling": true,
|
|
25692
|
+
"supports_parallel_function_calling": true,
|
|
25693
|
+
"supports_pdf_input": true,
|
|
25694
|
+
"supports_prompt_caching": true,
|
|
25695
|
+
"supports_reasoning": true,
|
|
25696
|
+
"supports_response_schema": true,
|
|
25697
|
+
"supports_system_messages": true,
|
|
25698
|
+
"supports_tool_choice": true,
|
|
25699
|
+
"supports_vision": true
|
|
25700
|
+
},
|
|
25701
|
+
"vertex_ai/minimaxai/minimax-m2-maas": {
|
|
25702
|
+
"input_cost_per_token": 3e-07,
|
|
25703
|
+
"litellm_provider": "vertex_ai-minimax_models",
|
|
25704
|
+
"max_input_tokens": 196608,
|
|
25705
|
+
"max_output_tokens": 196608,
|
|
25706
|
+
"max_tokens": 196608,
|
|
25707
|
+
"mode": "chat",
|
|
25708
|
+
"output_cost_per_token": 1.2e-06,
|
|
25709
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
|
|
25710
|
+
"supports_function_calling": true,
|
|
25711
|
+
"supports_tool_choice": true
|
|
25673
25712
|
}
|
|
25674
25713
|
}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: tokencostauto
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.333
|
|
4
4
|
Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
|
|
5
5
|
Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
|
|
6
6
|
Project-URL: Homepage, https://github.com/madpin/tokencostaudo
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|