tokencostauto 0.1.455__tar.gz → 0.1.460__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {tokencostauto-0.1.455/tokencostauto.egg-info → tokencostauto-0.1.460}/PKG-INFO +1 -1
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/pyproject.toml +1 -1
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/tokencostauto/model_prices.json +109 -7
- {tokencostauto-0.1.455 → tokencostauto-0.1.460/tokencostauto.egg-info}/PKG-INFO +1 -1
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/LICENSE +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/MANIFEST.in +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/README.md +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/setup.cfg +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/tests/test_costs.py +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/tokencostauto/__init__.py +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/tokencostauto/constants.py +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/tokencostauto/costs.py +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/tokencostauto.egg-info/SOURCES.txt +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/tokencostauto.egg-info/dependency_links.txt +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/tokencostauto.egg-info/requires.txt +0 -0
- {tokencostauto-0.1.455 → tokencostauto-0.1.460}/tokencostauto.egg-info/top_level.txt +0 -0
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: tokencostauto
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.460
|
|
4
4
|
Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
|
|
5
5
|
Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
|
|
6
6
|
Project-URL: Homepage, https://github.com/madpin/tokencostaudo
|
|
@@ -18720,13 +18720,14 @@
|
|
|
18720
18720
|
"supports_vision": true
|
|
18721
18721
|
},
|
|
18722
18722
|
"groq/openai/gpt-oss-20b": {
|
|
18723
|
-
"
|
|
18723
|
+
"cache_read_input_token_cost": 3.75e-08,
|
|
18724
|
+
"input_cost_per_token": 7.5e-08,
|
|
18724
18725
|
"litellm_provider": "groq",
|
|
18725
18726
|
"max_input_tokens": 131072,
|
|
18726
18727
|
"max_output_tokens": 32768,
|
|
18727
18728
|
"max_tokens": 32768,
|
|
18728
18729
|
"mode": "chat",
|
|
18729
|
-
"output_cost_per_token":
|
|
18730
|
+
"output_cost_per_token": 3e-07,
|
|
18730
18731
|
"supports_function_calling": true,
|
|
18731
18732
|
"supports_parallel_function_calling": true,
|
|
18732
18733
|
"supports_reasoning": true,
|
|
@@ -18735,13 +18736,14 @@
|
|
|
18735
18736
|
"supports_web_search": true
|
|
18736
18737
|
},
|
|
18737
18738
|
"groq/openai/gpt-oss-120b": {
|
|
18739
|
+
"cache_read_input_token_cost": 7.5e-08,
|
|
18738
18740
|
"input_cost_per_token": 1.5e-07,
|
|
18739
18741
|
"litellm_provider": "groq",
|
|
18740
18742
|
"max_input_tokens": 131072,
|
|
18741
18743
|
"max_output_tokens": 32766,
|
|
18742
18744
|
"max_tokens": 32766,
|
|
18743
18745
|
"mode": "chat",
|
|
18744
|
-
"output_cost_per_token":
|
|
18746
|
+
"output_cost_per_token": 6e-07,
|
|
18745
18747
|
"supports_function_calling": true,
|
|
18746
18748
|
"supports_parallel_function_calling": true,
|
|
18747
18749
|
"supports_reasoning": true,
|
|
@@ -28134,15 +28136,15 @@
|
|
|
28134
28136
|
"tool_use_system_prompt_tokens": 159
|
|
28135
28137
|
},
|
|
28136
28138
|
"us.anthropic.claude-opus-4-5-20251101-v1:0": {
|
|
28137
|
-
"cache_creation_input_token_cost": 6.
|
|
28138
|
-
"cache_read_input_token_cost": 5e-07,
|
|
28139
|
-
"input_cost_per_token": 5e-06,
|
|
28139
|
+
"cache_creation_input_token_cost": 6.875e-06,
|
|
28140
|
+
"cache_read_input_token_cost": 5.5e-07,
|
|
28141
|
+
"input_cost_per_token": 5.5e-06,
|
|
28140
28142
|
"litellm_provider": "bedrock_converse",
|
|
28141
28143
|
"max_input_tokens": 200000,
|
|
28142
28144
|
"max_output_tokens": 64000,
|
|
28143
28145
|
"max_tokens": 64000,
|
|
28144
28146
|
"mode": "chat",
|
|
28145
|
-
"output_cost_per_token": 2.
|
|
28147
|
+
"output_cost_per_token": 2.75e-05,
|
|
28146
28148
|
"search_context_cost_per_query": {
|
|
28147
28149
|
"search_context_size_high": 0.01,
|
|
28148
28150
|
"search_context_size_low": 0.01,
|
|
@@ -35542,5 +35544,105 @@
|
|
|
35542
35544
|
"supports_system_messages": true,
|
|
35543
35545
|
"supports_tool_choice": true,
|
|
35544
35546
|
"supports_vision": true
|
|
35547
|
+
},
|
|
35548
|
+
"dall-e-2": {
|
|
35549
|
+
"input_cost_per_image": 0.02,
|
|
35550
|
+
"litellm_provider": "openai",
|
|
35551
|
+
"mode": "image_generation",
|
|
35552
|
+
"supported_endpoints": [
|
|
35553
|
+
"/v1/images/generations",
|
|
35554
|
+
"/v1/images/edits",
|
|
35555
|
+
"/v1/images/variations"
|
|
35556
|
+
]
|
|
35557
|
+
},
|
|
35558
|
+
"dall-e-3": {
|
|
35559
|
+
"input_cost_per_image": 0.04,
|
|
35560
|
+
"litellm_provider": "openai",
|
|
35561
|
+
"mode": "image_generation",
|
|
35562
|
+
"supported_endpoints": [
|
|
35563
|
+
"/v1/images/generations"
|
|
35564
|
+
]
|
|
35565
|
+
},
|
|
35566
|
+
"gemini-2.5-computer-use-preview-10-2025": {
|
|
35567
|
+
"input_cost_per_token": 1.25e-06,
|
|
35568
|
+
"input_cost_per_token_above_200k_tokens": 2.5e-06,
|
|
35569
|
+
"litellm_provider": "vertex_ai-language-models",
|
|
35570
|
+
"max_images_per_prompt": 3000,
|
|
35571
|
+
"max_input_tokens": 128000,
|
|
35572
|
+
"max_output_tokens": 64000,
|
|
35573
|
+
"max_tokens": 64000,
|
|
35574
|
+
"mode": "chat",
|
|
35575
|
+
"output_cost_per_token": 1e-05,
|
|
35576
|
+
"output_cost_per_token_above_200k_tokens": 1.5e-05,
|
|
35577
|
+
"source": "https://cloud.google.com/vertex-ai/generative-ai/docs/computer-use",
|
|
35578
|
+
"supported_modalities": [
|
|
35579
|
+
"text",
|
|
35580
|
+
"image"
|
|
35581
|
+
],
|
|
35582
|
+
"supported_output_modalities": [
|
|
35583
|
+
"text"
|
|
35584
|
+
],
|
|
35585
|
+
"supports_computer_use": true,
|
|
35586
|
+
"supports_function_calling": true,
|
|
35587
|
+
"supports_system_messages": true,
|
|
35588
|
+
"supports_tool_choice": true,
|
|
35589
|
+
"supports_vision": true
|
|
35590
|
+
},
|
|
35591
|
+
"chatgpt/gpt-5.2-codex": {
|
|
35592
|
+
"litellm_provider": "chatgpt",
|
|
35593
|
+
"max_input_tokens": 128000,
|
|
35594
|
+
"max_output_tokens": 128000,
|
|
35595
|
+
"max_tokens": 128000,
|
|
35596
|
+
"mode": "responses",
|
|
35597
|
+
"supported_endpoints": [
|
|
35598
|
+
"/v1/responses"
|
|
35599
|
+
],
|
|
35600
|
+
"supports_function_calling": true,
|
|
35601
|
+
"supports_parallel_function_calling": true,
|
|
35602
|
+
"supports_response_schema": true,
|
|
35603
|
+
"supports_vision": true
|
|
35604
|
+
},
|
|
35605
|
+
"chatgpt/gpt-5.2": {
|
|
35606
|
+
"litellm_provider": "chatgpt",
|
|
35607
|
+
"max_input_tokens": 128000,
|
|
35608
|
+
"max_output_tokens": 64000,
|
|
35609
|
+
"max_tokens": 64000,
|
|
35610
|
+
"mode": "responses",
|
|
35611
|
+
"supported_endpoints": [
|
|
35612
|
+
"/v1/chat/completions",
|
|
35613
|
+
"/v1/responses"
|
|
35614
|
+
],
|
|
35615
|
+
"supports_function_calling": true,
|
|
35616
|
+
"supports_parallel_function_calling": true,
|
|
35617
|
+
"supports_response_schema": true,
|
|
35618
|
+
"supports_vision": true
|
|
35619
|
+
},
|
|
35620
|
+
"chatgpt/gpt-5.1-codex-max": {
|
|
35621
|
+
"litellm_provider": "chatgpt",
|
|
35622
|
+
"max_input_tokens": 128000,
|
|
35623
|
+
"max_output_tokens": 128000,
|
|
35624
|
+
"max_tokens": 128000,
|
|
35625
|
+
"mode": "responses",
|
|
35626
|
+
"supported_endpoints": [
|
|
35627
|
+
"/v1/responses"
|
|
35628
|
+
],
|
|
35629
|
+
"supports_function_calling": true,
|
|
35630
|
+
"supports_parallel_function_calling": true,
|
|
35631
|
+
"supports_response_schema": true,
|
|
35632
|
+
"supports_vision": true
|
|
35633
|
+
},
|
|
35634
|
+
"chatgpt/gpt-5.1-codex-mini": {
|
|
35635
|
+
"litellm_provider": "chatgpt",
|
|
35636
|
+
"max_input_tokens": 128000,
|
|
35637
|
+
"max_output_tokens": 64000,
|
|
35638
|
+
"max_tokens": 64000,
|
|
35639
|
+
"mode": "responses",
|
|
35640
|
+
"supported_endpoints": [
|
|
35641
|
+
"/v1/responses"
|
|
35642
|
+
],
|
|
35643
|
+
"supports_function_calling": true,
|
|
35644
|
+
"supports_parallel_function_calling": true,
|
|
35645
|
+
"supports_response_schema": true,
|
|
35646
|
+
"supports_vision": true
|
|
35545
35647
|
}
|
|
35546
35648
|
}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: tokencostauto
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.460
|
|
4
4
|
Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
|
|
5
5
|
Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
|
|
6
6
|
Project-URL: Homepage, https://github.com/madpin/tokencostaudo
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|