tokencostauto 0.1.340__tar.gz → 0.1.344__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {tokencostauto-0.1.340/tokencostauto.egg-info → tokencostauto-0.1.344}/PKG-INFO +1 -1
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/pyproject.toml +1 -1
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/tokencostauto/model_prices.json +329 -6
- {tokencostauto-0.1.340 → tokencostauto-0.1.344/tokencostauto.egg-info}/PKG-INFO +1 -1
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/LICENSE +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/MANIFEST.in +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/README.md +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/setup.cfg +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/tests/test_costs.py +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/tokencostauto/__init__.py +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/tokencostauto/constants.py +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/tokencostauto/costs.py +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/tokencostauto.egg-info/SOURCES.txt +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/tokencostauto.egg-info/dependency_links.txt +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/tokencostauto.egg-info/requires.txt +0 -0
- {tokencostauto-0.1.340 → tokencostauto-0.1.344}/tokencostauto.egg-info/top_level.txt +0 -0
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: tokencostauto
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.344
|
|
4
4
|
Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
|
|
5
5
|
Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
|
|
6
6
|
Project-URL: Homepage, https://github.com/madpin/tokencostaudo
|
|
@@ -9266,6 +9266,7 @@
|
|
|
9266
9266
|
"supports_function_calling": false,
|
|
9267
9267
|
"supports_parallel_function_calling": true,
|
|
9268
9268
|
"supports_prompt_caching": true,
|
|
9269
|
+
"supports_reasoning": true,
|
|
9269
9270
|
"supports_response_schema": false,
|
|
9270
9271
|
"supports_system_messages": true,
|
|
9271
9272
|
"supports_tool_choice": true,
|
|
@@ -10442,6 +10443,7 @@
|
|
|
10442
10443
|
"supports_audio_output": true,
|
|
10443
10444
|
"supports_function_calling": true,
|
|
10444
10445
|
"supports_prompt_caching": true,
|
|
10446
|
+
"supports_reasoning": true,
|
|
10445
10447
|
"supports_response_schema": true,
|
|
10446
10448
|
"supports_system_messages": true,
|
|
10447
10449
|
"supports_tool_choice": true,
|
|
@@ -17812,15 +17814,16 @@
|
|
|
17812
17814
|
"supports_tool_choice": true
|
|
17813
17815
|
},
|
|
17814
17816
|
"openrouter/qwen/qwen3-coder": {
|
|
17815
|
-
"input_cost_per_token":
|
|
17817
|
+
"input_cost_per_token": 2.2e-07,
|
|
17816
17818
|
"litellm_provider": "openrouter",
|
|
17817
|
-
"max_input_tokens":
|
|
17818
|
-
"max_output_tokens":
|
|
17819
|
-
"max_tokens":
|
|
17819
|
+
"max_input_tokens": 262100,
|
|
17820
|
+
"max_output_tokens": 262100,
|
|
17821
|
+
"max_tokens": 262100,
|
|
17820
17822
|
"mode": "chat",
|
|
17821
|
-
"output_cost_per_token": 5e-
|
|
17823
|
+
"output_cost_per_token": 9.5e-07,
|
|
17822
17824
|
"source": "https://openrouter.ai/qwen/qwen3-coder",
|
|
17823
|
-
"supports_tool_choice": true
|
|
17825
|
+
"supports_tool_choice": true,
|
|
17826
|
+
"supports_function_calling": true
|
|
17824
17827
|
},
|
|
17825
17828
|
"gemini/imagen-4.0-generate-preview-06-06": {
|
|
17826
17829
|
"output_cost_per_image": 0.04,
|
|
@@ -24689,6 +24692,7 @@
|
|
|
24689
24692
|
"litellm_provider": "vertex_ai-language-models",
|
|
24690
24693
|
"max_audio_length_hours": 8.4,
|
|
24691
24694
|
"max_audio_per_prompt": 1,
|
|
24695
|
+
"supports_reasoning": false,
|
|
24692
24696
|
"max_images_per_prompt": 3000,
|
|
24693
24697
|
"max_input_tokens": 32768,
|
|
24694
24698
|
"max_output_tokens": 32768,
|
|
@@ -25804,5 +25808,324 @@
|
|
|
25804
25808
|
"supports_reasoning": true,
|
|
25805
25809
|
"supports_response_schema": true,
|
|
25806
25810
|
"supports_tool_choice": true
|
|
25811
|
+
},
|
|
25812
|
+
"openrouter/deepseek/deepseek-v3.2-exp": {
|
|
25813
|
+
"input_cost_per_token": 2e-07,
|
|
25814
|
+
"input_cost_per_token_cache_hit": 2e-08,
|
|
25815
|
+
"litellm_provider": "openrouter",
|
|
25816
|
+
"max_input_tokens": 163840,
|
|
25817
|
+
"max_output_tokens": 163840,
|
|
25818
|
+
"max_tokens": 8192,
|
|
25819
|
+
"mode": "chat",
|
|
25820
|
+
"output_cost_per_token": 4e-07,
|
|
25821
|
+
"supports_assistant_prefill": true,
|
|
25822
|
+
"supports_function_calling": true,
|
|
25823
|
+
"supports_prompt_caching": true,
|
|
25824
|
+
"supports_reasoning": false,
|
|
25825
|
+
"supports_tool_choice": true
|
|
25826
|
+
},
|
|
25827
|
+
"openrouter/minimax/minimax-m2": {
|
|
25828
|
+
"input_cost_per_token": 2.55e-07,
|
|
25829
|
+
"litellm_provider": "openrouter",
|
|
25830
|
+
"max_input_tokens": 204800,
|
|
25831
|
+
"max_output_tokens": 204800,
|
|
25832
|
+
"max_tokens": 32768,
|
|
25833
|
+
"mode": "chat",
|
|
25834
|
+
"output_cost_per_token": 1.02e-06,
|
|
25835
|
+
"supports_function_calling": true,
|
|
25836
|
+
"supports_prompt_caching": false,
|
|
25837
|
+
"supports_reasoning": true,
|
|
25838
|
+
"supports_tool_choice": true
|
|
25839
|
+
},
|
|
25840
|
+
"openrouter/z-ai/glm-4.6": {
|
|
25841
|
+
"input_cost_per_token": 4e-07,
|
|
25842
|
+
"litellm_provider": "openrouter",
|
|
25843
|
+
"max_input_tokens": 202800,
|
|
25844
|
+
"max_output_tokens": 131000,
|
|
25845
|
+
"max_tokens": 202800,
|
|
25846
|
+
"mode": "chat",
|
|
25847
|
+
"output_cost_per_token": 1.75e-06,
|
|
25848
|
+
"source": "https://openrouter.ai/z-ai/glm-4.6",
|
|
25849
|
+
"supports_function_calling": true,
|
|
25850
|
+
"supports_reasoning": true,
|
|
25851
|
+
"supports_tool_choice": true
|
|
25852
|
+
},
|
|
25853
|
+
"openrouter/z-ai/glm-4.6:exacto": {
|
|
25854
|
+
"input_cost_per_token": 4.5e-07,
|
|
25855
|
+
"litellm_provider": "openrouter",
|
|
25856
|
+
"max_input_tokens": 202800,
|
|
25857
|
+
"max_output_tokens": 131000,
|
|
25858
|
+
"max_tokens": 202800,
|
|
25859
|
+
"mode": "chat",
|
|
25860
|
+
"output_cost_per_token": 1.9e-06,
|
|
25861
|
+
"source": "https://openrouter.ai/z-ai/glm-4.6:exacto",
|
|
25862
|
+
"supports_function_calling": true,
|
|
25863
|
+
"supports_reasoning": true,
|
|
25864
|
+
"supports_tool_choice": true
|
|
25865
|
+
},
|
|
25866
|
+
"runwayml/gen4_turbo": {
|
|
25867
|
+
"litellm_provider": "runwayml",
|
|
25868
|
+
"mode": "video_generation",
|
|
25869
|
+
"output_cost_per_video_per_second": 0.05,
|
|
25870
|
+
"source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
25871
|
+
"supported_modalities": [
|
|
25872
|
+
"text",
|
|
25873
|
+
"image"
|
|
25874
|
+
],
|
|
25875
|
+
"supported_output_modalities": [
|
|
25876
|
+
"video"
|
|
25877
|
+
],
|
|
25878
|
+
"supported_resolutions": [
|
|
25879
|
+
"1280x720",
|
|
25880
|
+
"720x1280"
|
|
25881
|
+
],
|
|
25882
|
+
"comment": "5 credits per second @ $0.01 per credit = $0.05 per second"
|
|
25883
|
+
},
|
|
25884
|
+
"runwayml/gen4_aleph": {
|
|
25885
|
+
"litellm_provider": "runwayml",
|
|
25886
|
+
"mode": "video_generation",
|
|
25887
|
+
"output_cost_per_video_per_second": 0.15,
|
|
25888
|
+
"source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
25889
|
+
"supported_modalities": [
|
|
25890
|
+
"text",
|
|
25891
|
+
"image"
|
|
25892
|
+
],
|
|
25893
|
+
"supported_output_modalities": [
|
|
25894
|
+
"video"
|
|
25895
|
+
],
|
|
25896
|
+
"supported_resolutions": [
|
|
25897
|
+
"1280x720",
|
|
25898
|
+
"720x1280"
|
|
25899
|
+
],
|
|
25900
|
+
"comment": "15 credits per second @ $0.01 per credit = $0.15 per second"
|
|
25901
|
+
},
|
|
25902
|
+
"runwayml/gen3a_turbo": {
|
|
25903
|
+
"litellm_provider": "runwayml",
|
|
25904
|
+
"mode": "video_generation",
|
|
25905
|
+
"output_cost_per_video_per_second": 0.05,
|
|
25906
|
+
"source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
25907
|
+
"supported_modalities": [
|
|
25908
|
+
"text",
|
|
25909
|
+
"image"
|
|
25910
|
+
],
|
|
25911
|
+
"supported_output_modalities": [
|
|
25912
|
+
"video"
|
|
25913
|
+
],
|
|
25914
|
+
"supported_resolutions": [
|
|
25915
|
+
"1280x720",
|
|
25916
|
+
"720x1280"
|
|
25917
|
+
],
|
|
25918
|
+
"comment": "5 credits per second @ $0.01 per credit = $0.05 per second"
|
|
25919
|
+
},
|
|
25920
|
+
"runwayml/gen4_image": {
|
|
25921
|
+
"litellm_provider": "runwayml",
|
|
25922
|
+
"mode": "image_generation",
|
|
25923
|
+
"input_cost_per_image": 0.05,
|
|
25924
|
+
"output_cost_per_image": 0.05,
|
|
25925
|
+
"source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
25926
|
+
"supported_modalities": [
|
|
25927
|
+
"text",
|
|
25928
|
+
"image"
|
|
25929
|
+
],
|
|
25930
|
+
"supported_output_modalities": [
|
|
25931
|
+
"image"
|
|
25932
|
+
],
|
|
25933
|
+
"supported_resolutions": [
|
|
25934
|
+
"1280x720",
|
|
25935
|
+
"1920x1080"
|
|
25936
|
+
],
|
|
25937
|
+
"comment": "5 credits per 720p image or 8 credits per 1080p image @ $0.01 per credit. Using 5 credits ($0.05) as base cost"
|
|
25938
|
+
},
|
|
25939
|
+
"runwayml/gen4_image_turbo": {
|
|
25940
|
+
"litellm_provider": "runwayml",
|
|
25941
|
+
"mode": "image_generation",
|
|
25942
|
+
"input_cost_per_image": 0.02,
|
|
25943
|
+
"output_cost_per_image": 0.02,
|
|
25944
|
+
"source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
25945
|
+
"supported_modalities": [
|
|
25946
|
+
"text",
|
|
25947
|
+
"image"
|
|
25948
|
+
],
|
|
25949
|
+
"supported_output_modalities": [
|
|
25950
|
+
"image"
|
|
25951
|
+
],
|
|
25952
|
+
"supported_resolutions": [
|
|
25953
|
+
"1280x720",
|
|
25954
|
+
"1920x1080"
|
|
25955
|
+
],
|
|
25956
|
+
"comment": "2 credits per image (any resolution) @ $0.01 per credit = $0.02 per image"
|
|
25957
|
+
},
|
|
25958
|
+
"gpt-5.1": {
|
|
25959
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
25960
|
+
"cache_read_input_token_cost_priority": 2.5e-07,
|
|
25961
|
+
"input_cost_per_token": 1.25e-06,
|
|
25962
|
+
"input_cost_per_token_priority": 2.5e-06,
|
|
25963
|
+
"litellm_provider": "openai",
|
|
25964
|
+
"max_input_tokens": 272000,
|
|
25965
|
+
"max_output_tokens": 128000,
|
|
25966
|
+
"max_tokens": 128000,
|
|
25967
|
+
"mode": "chat",
|
|
25968
|
+
"output_cost_per_token": 1e-05,
|
|
25969
|
+
"output_cost_per_token_priority": 2e-05,
|
|
25970
|
+
"supported_endpoints": [
|
|
25971
|
+
"/v1/chat/completions",
|
|
25972
|
+
"/v1/responses"
|
|
25973
|
+
],
|
|
25974
|
+
"supported_modalities": [
|
|
25975
|
+
"text",
|
|
25976
|
+
"image"
|
|
25977
|
+
],
|
|
25978
|
+
"supported_output_modalities": [
|
|
25979
|
+
"text",
|
|
25980
|
+
"image"
|
|
25981
|
+
],
|
|
25982
|
+
"supports_function_calling": true,
|
|
25983
|
+
"supports_native_streaming": true,
|
|
25984
|
+
"supports_parallel_function_calling": true,
|
|
25985
|
+
"supports_pdf_input": true,
|
|
25986
|
+
"supports_prompt_caching": true,
|
|
25987
|
+
"supports_reasoning": true,
|
|
25988
|
+
"supports_response_schema": true,
|
|
25989
|
+
"supports_system_messages": true,
|
|
25990
|
+
"supports_tool_choice": true,
|
|
25991
|
+
"supports_service_tier": true,
|
|
25992
|
+
"supports_vision": true
|
|
25993
|
+
},
|
|
25994
|
+
"gpt-5.1-2025-11-13": {
|
|
25995
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
25996
|
+
"cache_read_input_token_cost_priority": 2.5e-07,
|
|
25997
|
+
"input_cost_per_token": 1.25e-06,
|
|
25998
|
+
"input_cost_per_token_priority": 2.5e-06,
|
|
25999
|
+
"litellm_provider": "openai",
|
|
26000
|
+
"max_input_tokens": 272000,
|
|
26001
|
+
"max_output_tokens": 128000,
|
|
26002
|
+
"max_tokens": 128000,
|
|
26003
|
+
"mode": "chat",
|
|
26004
|
+
"output_cost_per_token": 1e-05,
|
|
26005
|
+
"output_cost_per_token_priority": 2e-05,
|
|
26006
|
+
"supported_endpoints": [
|
|
26007
|
+
"/v1/chat/completions",
|
|
26008
|
+
"/v1/responses"
|
|
26009
|
+
],
|
|
26010
|
+
"supported_modalities": [
|
|
26011
|
+
"text",
|
|
26012
|
+
"image"
|
|
26013
|
+
],
|
|
26014
|
+
"supported_output_modalities": [
|
|
26015
|
+
"text",
|
|
26016
|
+
"image"
|
|
26017
|
+
],
|
|
26018
|
+
"supports_function_calling": true,
|
|
26019
|
+
"supports_native_streaming": true,
|
|
26020
|
+
"supports_parallel_function_calling": true,
|
|
26021
|
+
"supports_pdf_input": true,
|
|
26022
|
+
"supports_prompt_caching": true,
|
|
26023
|
+
"supports_reasoning": true,
|
|
26024
|
+
"supports_response_schema": true,
|
|
26025
|
+
"supports_system_messages": true,
|
|
26026
|
+
"supports_tool_choice": true,
|
|
26027
|
+
"supports_service_tier": true,
|
|
26028
|
+
"supports_vision": true
|
|
26029
|
+
},
|
|
26030
|
+
"gpt-5.1-chat-latest": {
|
|
26031
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
26032
|
+
"cache_read_input_token_cost_priority": 2.5e-07,
|
|
26033
|
+
"input_cost_per_token": 1.25e-06,
|
|
26034
|
+
"input_cost_per_token_priority": 2.5e-06,
|
|
26035
|
+
"litellm_provider": "openai",
|
|
26036
|
+
"max_input_tokens": 128000,
|
|
26037
|
+
"max_output_tokens": 16384,
|
|
26038
|
+
"max_tokens": 16384,
|
|
26039
|
+
"mode": "chat",
|
|
26040
|
+
"output_cost_per_token": 1e-05,
|
|
26041
|
+
"output_cost_per_token_priority": 2e-05,
|
|
26042
|
+
"supported_endpoints": [
|
|
26043
|
+
"/v1/chat/completions",
|
|
26044
|
+
"/v1/responses"
|
|
26045
|
+
],
|
|
26046
|
+
"supported_modalities": [
|
|
26047
|
+
"text",
|
|
26048
|
+
"image"
|
|
26049
|
+
],
|
|
26050
|
+
"supported_output_modalities": [
|
|
26051
|
+
"text",
|
|
26052
|
+
"image"
|
|
26053
|
+
],
|
|
26054
|
+
"supports_function_calling": false,
|
|
26055
|
+
"supports_native_streaming": true,
|
|
26056
|
+
"supports_parallel_function_calling": false,
|
|
26057
|
+
"supports_pdf_input": true,
|
|
26058
|
+
"supports_prompt_caching": true,
|
|
26059
|
+
"supports_reasoning": true,
|
|
26060
|
+
"supports_response_schema": true,
|
|
26061
|
+
"supports_system_messages": true,
|
|
26062
|
+
"supports_tool_choice": false,
|
|
26063
|
+
"supports_vision": true
|
|
26064
|
+
},
|
|
26065
|
+
"gpt-5.1-codex": {
|
|
26066
|
+
"cache_read_input_token_cost": 1.25e-07,
|
|
26067
|
+
"cache_read_input_token_cost_priority": 2.5e-07,
|
|
26068
|
+
"input_cost_per_token": 1.25e-06,
|
|
26069
|
+
"input_cost_per_token_priority": 2.5e-06,
|
|
26070
|
+
"litellm_provider": "openai",
|
|
26071
|
+
"max_input_tokens": 272000,
|
|
26072
|
+
"max_output_tokens": 128000,
|
|
26073
|
+
"max_tokens": 128000,
|
|
26074
|
+
"mode": "responses",
|
|
26075
|
+
"output_cost_per_token": 1e-05,
|
|
26076
|
+
"output_cost_per_token_priority": 2e-05,
|
|
26077
|
+
"supported_endpoints": [
|
|
26078
|
+
"/v1/responses"
|
|
26079
|
+
],
|
|
26080
|
+
"supported_modalities": [
|
|
26081
|
+
"text",
|
|
26082
|
+
"image"
|
|
26083
|
+
],
|
|
26084
|
+
"supported_output_modalities": [
|
|
26085
|
+
"text"
|
|
26086
|
+
],
|
|
26087
|
+
"supports_function_calling": true,
|
|
26088
|
+
"supports_native_streaming": true,
|
|
26089
|
+
"supports_parallel_function_calling": true,
|
|
26090
|
+
"supports_pdf_input": true,
|
|
26091
|
+
"supports_prompt_caching": true,
|
|
26092
|
+
"supports_reasoning": true,
|
|
26093
|
+
"supports_response_schema": true,
|
|
26094
|
+
"supports_system_messages": false,
|
|
26095
|
+
"supports_tool_choice": true,
|
|
26096
|
+
"supports_vision": true
|
|
26097
|
+
},
|
|
26098
|
+
"gpt-5.1-codex-mini": {
|
|
26099
|
+
"cache_read_input_token_cost": 2.5e-08,
|
|
26100
|
+
"cache_read_input_token_cost_priority": 4.5e-08,
|
|
26101
|
+
"input_cost_per_token": 2.5e-07,
|
|
26102
|
+
"input_cost_per_token_priority": 4.5e-07,
|
|
26103
|
+
"litellm_provider": "openai",
|
|
26104
|
+
"max_input_tokens": 272000,
|
|
26105
|
+
"max_output_tokens": 128000,
|
|
26106
|
+
"max_tokens": 128000,
|
|
26107
|
+
"mode": "responses",
|
|
26108
|
+
"output_cost_per_token": 2e-06,
|
|
26109
|
+
"output_cost_per_token_priority": 3.6e-06,
|
|
26110
|
+
"supported_endpoints": [
|
|
26111
|
+
"/v1/responses"
|
|
26112
|
+
],
|
|
26113
|
+
"supported_modalities": [
|
|
26114
|
+
"text",
|
|
26115
|
+
"image"
|
|
26116
|
+
],
|
|
26117
|
+
"supported_output_modalities": [
|
|
26118
|
+
"text"
|
|
26119
|
+
],
|
|
26120
|
+
"supports_function_calling": true,
|
|
26121
|
+
"supports_native_streaming": true,
|
|
26122
|
+
"supports_parallel_function_calling": true,
|
|
26123
|
+
"supports_pdf_input": true,
|
|
26124
|
+
"supports_prompt_caching": true,
|
|
26125
|
+
"supports_reasoning": true,
|
|
26126
|
+
"supports_response_schema": true,
|
|
26127
|
+
"supports_system_messages": false,
|
|
26128
|
+
"supports_tool_choice": true,
|
|
26129
|
+
"supports_vision": true
|
|
25807
26130
|
}
|
|
25808
26131
|
}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: tokencostauto
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.344
|
|
4
4
|
Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
|
|
5
5
|
Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
|
|
6
6
|
Project-URL: Homepage, https://github.com/madpin/tokencostaudo
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|