tokencostauto 0.1.340__tar.gz → 0.1.344__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: tokencostauto
3
- Version: 0.1.340
3
+ Version: 0.1.344
4
4
  Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
5
5
  Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
6
6
  Project-URL: Homepage, https://github.com/madpin/tokencostaudo
@@ -11,7 +11,7 @@ tokencostauto = ["model_prices.json"]
11
11
  [project]
12
12
 
13
13
  name = "tokencostauto"
14
- version = "0.1.340"
14
+ version = "0.1.344"
15
15
 
16
16
  authors = [
17
17
  { name = "Trisha Pan", email = "trishaepan@gmail.com" },
@@ -9266,6 +9266,7 @@
9266
9266
  "supports_function_calling": false,
9267
9267
  "supports_parallel_function_calling": true,
9268
9268
  "supports_prompt_caching": true,
9269
+ "supports_reasoning": true,
9269
9270
  "supports_response_schema": false,
9270
9271
  "supports_system_messages": true,
9271
9272
  "supports_tool_choice": true,
@@ -10442,6 +10443,7 @@
10442
10443
  "supports_audio_output": true,
10443
10444
  "supports_function_calling": true,
10444
10445
  "supports_prompt_caching": true,
10446
+ "supports_reasoning": true,
10445
10447
  "supports_response_schema": true,
10446
10448
  "supports_system_messages": true,
10447
10449
  "supports_tool_choice": true,
@@ -17812,15 +17814,16 @@
17812
17814
  "supports_tool_choice": true
17813
17815
  },
17814
17816
  "openrouter/qwen/qwen3-coder": {
17815
- "input_cost_per_token": 1e-06,
17817
+ "input_cost_per_token": 2.2e-07,
17816
17818
  "litellm_provider": "openrouter",
17817
- "max_input_tokens": 1000000,
17818
- "max_output_tokens": 1000000,
17819
- "max_tokens": 1000000,
17819
+ "max_input_tokens": 262100,
17820
+ "max_output_tokens": 262100,
17821
+ "max_tokens": 262100,
17820
17822
  "mode": "chat",
17821
- "output_cost_per_token": 5e-06,
17823
+ "output_cost_per_token": 9.5e-07,
17822
17824
  "source": "https://openrouter.ai/qwen/qwen3-coder",
17823
- "supports_tool_choice": true
17825
+ "supports_tool_choice": true,
17826
+ "supports_function_calling": true
17824
17827
  },
17825
17828
  "gemini/imagen-4.0-generate-preview-06-06": {
17826
17829
  "output_cost_per_image": 0.04,
@@ -24689,6 +24692,7 @@
24689
24692
  "litellm_provider": "vertex_ai-language-models",
24690
24693
  "max_audio_length_hours": 8.4,
24691
24694
  "max_audio_per_prompt": 1,
24695
+ "supports_reasoning": false,
24692
24696
  "max_images_per_prompt": 3000,
24693
24697
  "max_input_tokens": 32768,
24694
24698
  "max_output_tokens": 32768,
@@ -25804,5 +25808,324 @@
25804
25808
  "supports_reasoning": true,
25805
25809
  "supports_response_schema": true,
25806
25810
  "supports_tool_choice": true
25811
+ },
25812
+ "openrouter/deepseek/deepseek-v3.2-exp": {
25813
+ "input_cost_per_token": 2e-07,
25814
+ "input_cost_per_token_cache_hit": 2e-08,
25815
+ "litellm_provider": "openrouter",
25816
+ "max_input_tokens": 163840,
25817
+ "max_output_tokens": 163840,
25818
+ "max_tokens": 8192,
25819
+ "mode": "chat",
25820
+ "output_cost_per_token": 4e-07,
25821
+ "supports_assistant_prefill": true,
25822
+ "supports_function_calling": true,
25823
+ "supports_prompt_caching": true,
25824
+ "supports_reasoning": false,
25825
+ "supports_tool_choice": true
25826
+ },
25827
+ "openrouter/minimax/minimax-m2": {
25828
+ "input_cost_per_token": 2.55e-07,
25829
+ "litellm_provider": "openrouter",
25830
+ "max_input_tokens": 204800,
25831
+ "max_output_tokens": 204800,
25832
+ "max_tokens": 32768,
25833
+ "mode": "chat",
25834
+ "output_cost_per_token": 1.02e-06,
25835
+ "supports_function_calling": true,
25836
+ "supports_prompt_caching": false,
25837
+ "supports_reasoning": true,
25838
+ "supports_tool_choice": true
25839
+ },
25840
+ "openrouter/z-ai/glm-4.6": {
25841
+ "input_cost_per_token": 4e-07,
25842
+ "litellm_provider": "openrouter",
25843
+ "max_input_tokens": 202800,
25844
+ "max_output_tokens": 131000,
25845
+ "max_tokens": 202800,
25846
+ "mode": "chat",
25847
+ "output_cost_per_token": 1.75e-06,
25848
+ "source": "https://openrouter.ai/z-ai/glm-4.6",
25849
+ "supports_function_calling": true,
25850
+ "supports_reasoning": true,
25851
+ "supports_tool_choice": true
25852
+ },
25853
+ "openrouter/z-ai/glm-4.6:exacto": {
25854
+ "input_cost_per_token": 4.5e-07,
25855
+ "litellm_provider": "openrouter",
25856
+ "max_input_tokens": 202800,
25857
+ "max_output_tokens": 131000,
25858
+ "max_tokens": 202800,
25859
+ "mode": "chat",
25860
+ "output_cost_per_token": 1.9e-06,
25861
+ "source": "https://openrouter.ai/z-ai/glm-4.6:exacto",
25862
+ "supports_function_calling": true,
25863
+ "supports_reasoning": true,
25864
+ "supports_tool_choice": true
25865
+ },
25866
+ "runwayml/gen4_turbo": {
25867
+ "litellm_provider": "runwayml",
25868
+ "mode": "video_generation",
25869
+ "output_cost_per_video_per_second": 0.05,
25870
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
25871
+ "supported_modalities": [
25872
+ "text",
25873
+ "image"
25874
+ ],
25875
+ "supported_output_modalities": [
25876
+ "video"
25877
+ ],
25878
+ "supported_resolutions": [
25879
+ "1280x720",
25880
+ "720x1280"
25881
+ ],
25882
+ "comment": "5 credits per second @ $0.01 per credit = $0.05 per second"
25883
+ },
25884
+ "runwayml/gen4_aleph": {
25885
+ "litellm_provider": "runwayml",
25886
+ "mode": "video_generation",
25887
+ "output_cost_per_video_per_second": 0.15,
25888
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
25889
+ "supported_modalities": [
25890
+ "text",
25891
+ "image"
25892
+ ],
25893
+ "supported_output_modalities": [
25894
+ "video"
25895
+ ],
25896
+ "supported_resolutions": [
25897
+ "1280x720",
25898
+ "720x1280"
25899
+ ],
25900
+ "comment": "15 credits per second @ $0.01 per credit = $0.15 per second"
25901
+ },
25902
+ "runwayml/gen3a_turbo": {
25903
+ "litellm_provider": "runwayml",
25904
+ "mode": "video_generation",
25905
+ "output_cost_per_video_per_second": 0.05,
25906
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
25907
+ "supported_modalities": [
25908
+ "text",
25909
+ "image"
25910
+ ],
25911
+ "supported_output_modalities": [
25912
+ "video"
25913
+ ],
25914
+ "supported_resolutions": [
25915
+ "1280x720",
25916
+ "720x1280"
25917
+ ],
25918
+ "comment": "5 credits per second @ $0.01 per credit = $0.05 per second"
25919
+ },
25920
+ "runwayml/gen4_image": {
25921
+ "litellm_provider": "runwayml",
25922
+ "mode": "image_generation",
25923
+ "input_cost_per_image": 0.05,
25924
+ "output_cost_per_image": 0.05,
25925
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
25926
+ "supported_modalities": [
25927
+ "text",
25928
+ "image"
25929
+ ],
25930
+ "supported_output_modalities": [
25931
+ "image"
25932
+ ],
25933
+ "supported_resolutions": [
25934
+ "1280x720",
25935
+ "1920x1080"
25936
+ ],
25937
+ "comment": "5 credits per 720p image or 8 credits per 1080p image @ $0.01 per credit. Using 5 credits ($0.05) as base cost"
25938
+ },
25939
+ "runwayml/gen4_image_turbo": {
25940
+ "litellm_provider": "runwayml",
25941
+ "mode": "image_generation",
25942
+ "input_cost_per_image": 0.02,
25943
+ "output_cost_per_image": 0.02,
25944
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
25945
+ "supported_modalities": [
25946
+ "text",
25947
+ "image"
25948
+ ],
25949
+ "supported_output_modalities": [
25950
+ "image"
25951
+ ],
25952
+ "supported_resolutions": [
25953
+ "1280x720",
25954
+ "1920x1080"
25955
+ ],
25956
+ "comment": "2 credits per image (any resolution) @ $0.01 per credit = $0.02 per image"
25957
+ },
25958
+ "gpt-5.1": {
25959
+ "cache_read_input_token_cost": 1.25e-07,
25960
+ "cache_read_input_token_cost_priority": 2.5e-07,
25961
+ "input_cost_per_token": 1.25e-06,
25962
+ "input_cost_per_token_priority": 2.5e-06,
25963
+ "litellm_provider": "openai",
25964
+ "max_input_tokens": 272000,
25965
+ "max_output_tokens": 128000,
25966
+ "max_tokens": 128000,
25967
+ "mode": "chat",
25968
+ "output_cost_per_token": 1e-05,
25969
+ "output_cost_per_token_priority": 2e-05,
25970
+ "supported_endpoints": [
25971
+ "/v1/chat/completions",
25972
+ "/v1/responses"
25973
+ ],
25974
+ "supported_modalities": [
25975
+ "text",
25976
+ "image"
25977
+ ],
25978
+ "supported_output_modalities": [
25979
+ "text",
25980
+ "image"
25981
+ ],
25982
+ "supports_function_calling": true,
25983
+ "supports_native_streaming": true,
25984
+ "supports_parallel_function_calling": true,
25985
+ "supports_pdf_input": true,
25986
+ "supports_prompt_caching": true,
25987
+ "supports_reasoning": true,
25988
+ "supports_response_schema": true,
25989
+ "supports_system_messages": true,
25990
+ "supports_tool_choice": true,
25991
+ "supports_service_tier": true,
25992
+ "supports_vision": true
25993
+ },
25994
+ "gpt-5.1-2025-11-13": {
25995
+ "cache_read_input_token_cost": 1.25e-07,
25996
+ "cache_read_input_token_cost_priority": 2.5e-07,
25997
+ "input_cost_per_token": 1.25e-06,
25998
+ "input_cost_per_token_priority": 2.5e-06,
25999
+ "litellm_provider": "openai",
26000
+ "max_input_tokens": 272000,
26001
+ "max_output_tokens": 128000,
26002
+ "max_tokens": 128000,
26003
+ "mode": "chat",
26004
+ "output_cost_per_token": 1e-05,
26005
+ "output_cost_per_token_priority": 2e-05,
26006
+ "supported_endpoints": [
26007
+ "/v1/chat/completions",
26008
+ "/v1/responses"
26009
+ ],
26010
+ "supported_modalities": [
26011
+ "text",
26012
+ "image"
26013
+ ],
26014
+ "supported_output_modalities": [
26015
+ "text",
26016
+ "image"
26017
+ ],
26018
+ "supports_function_calling": true,
26019
+ "supports_native_streaming": true,
26020
+ "supports_parallel_function_calling": true,
26021
+ "supports_pdf_input": true,
26022
+ "supports_prompt_caching": true,
26023
+ "supports_reasoning": true,
26024
+ "supports_response_schema": true,
26025
+ "supports_system_messages": true,
26026
+ "supports_tool_choice": true,
26027
+ "supports_service_tier": true,
26028
+ "supports_vision": true
26029
+ },
26030
+ "gpt-5.1-chat-latest": {
26031
+ "cache_read_input_token_cost": 1.25e-07,
26032
+ "cache_read_input_token_cost_priority": 2.5e-07,
26033
+ "input_cost_per_token": 1.25e-06,
26034
+ "input_cost_per_token_priority": 2.5e-06,
26035
+ "litellm_provider": "openai",
26036
+ "max_input_tokens": 128000,
26037
+ "max_output_tokens": 16384,
26038
+ "max_tokens": 16384,
26039
+ "mode": "chat",
26040
+ "output_cost_per_token": 1e-05,
26041
+ "output_cost_per_token_priority": 2e-05,
26042
+ "supported_endpoints": [
26043
+ "/v1/chat/completions",
26044
+ "/v1/responses"
26045
+ ],
26046
+ "supported_modalities": [
26047
+ "text",
26048
+ "image"
26049
+ ],
26050
+ "supported_output_modalities": [
26051
+ "text",
26052
+ "image"
26053
+ ],
26054
+ "supports_function_calling": false,
26055
+ "supports_native_streaming": true,
26056
+ "supports_parallel_function_calling": false,
26057
+ "supports_pdf_input": true,
26058
+ "supports_prompt_caching": true,
26059
+ "supports_reasoning": true,
26060
+ "supports_response_schema": true,
26061
+ "supports_system_messages": true,
26062
+ "supports_tool_choice": false,
26063
+ "supports_vision": true
26064
+ },
26065
+ "gpt-5.1-codex": {
26066
+ "cache_read_input_token_cost": 1.25e-07,
26067
+ "cache_read_input_token_cost_priority": 2.5e-07,
26068
+ "input_cost_per_token": 1.25e-06,
26069
+ "input_cost_per_token_priority": 2.5e-06,
26070
+ "litellm_provider": "openai",
26071
+ "max_input_tokens": 272000,
26072
+ "max_output_tokens": 128000,
26073
+ "max_tokens": 128000,
26074
+ "mode": "responses",
26075
+ "output_cost_per_token": 1e-05,
26076
+ "output_cost_per_token_priority": 2e-05,
26077
+ "supported_endpoints": [
26078
+ "/v1/responses"
26079
+ ],
26080
+ "supported_modalities": [
26081
+ "text",
26082
+ "image"
26083
+ ],
26084
+ "supported_output_modalities": [
26085
+ "text"
26086
+ ],
26087
+ "supports_function_calling": true,
26088
+ "supports_native_streaming": true,
26089
+ "supports_parallel_function_calling": true,
26090
+ "supports_pdf_input": true,
26091
+ "supports_prompt_caching": true,
26092
+ "supports_reasoning": true,
26093
+ "supports_response_schema": true,
26094
+ "supports_system_messages": false,
26095
+ "supports_tool_choice": true,
26096
+ "supports_vision": true
26097
+ },
26098
+ "gpt-5.1-codex-mini": {
26099
+ "cache_read_input_token_cost": 2.5e-08,
26100
+ "cache_read_input_token_cost_priority": 4.5e-08,
26101
+ "input_cost_per_token": 2.5e-07,
26102
+ "input_cost_per_token_priority": 4.5e-07,
26103
+ "litellm_provider": "openai",
26104
+ "max_input_tokens": 272000,
26105
+ "max_output_tokens": 128000,
26106
+ "max_tokens": 128000,
26107
+ "mode": "responses",
26108
+ "output_cost_per_token": 2e-06,
26109
+ "output_cost_per_token_priority": 3.6e-06,
26110
+ "supported_endpoints": [
26111
+ "/v1/responses"
26112
+ ],
26113
+ "supported_modalities": [
26114
+ "text",
26115
+ "image"
26116
+ ],
26117
+ "supported_output_modalities": [
26118
+ "text"
26119
+ ],
26120
+ "supports_function_calling": true,
26121
+ "supports_native_streaming": true,
26122
+ "supports_parallel_function_calling": true,
26123
+ "supports_pdf_input": true,
26124
+ "supports_prompt_caching": true,
26125
+ "supports_reasoning": true,
26126
+ "supports_response_schema": true,
26127
+ "supports_system_messages": false,
26128
+ "supports_tool_choice": true,
26129
+ "supports_vision": true
25807
26130
  }
25808
26131
  }
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: tokencostauto
3
- Version: 0.1.340
3
+ Version: 0.1.344
4
4
  Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
5
5
  Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
6
6
  Project-URL: Homepage, https://github.com/madpin/tokencostaudo
File without changes