tokencostauto 0.1.452__py3-none-any.whl → 0.1.462__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -17766,8 +17766,8 @@
17766
17766
  "supports_tool_choice": true
17767
17767
  },
17768
17768
  "gemini/gemini-2.5-flash-lite": {
17769
- "cache_read_input_token_cost": 2.5e-08,
17770
- "input_cost_per_audio_token": 5e-07,
17769
+ "cache_read_input_token_cost": 1e-08,
17770
+ "input_cost_per_audio_token": 3e-07,
17771
17771
  "input_cost_per_token": 1e-07,
17772
17772
  "litellm_provider": "gemini",
17773
17773
  "max_audio_length_hours": 8.4,
@@ -17813,8 +17813,8 @@
17813
17813
  "tpm": 250000
17814
17814
  },
17815
17815
  "gemini-2.5-flash-lite": {
17816
- "cache_read_input_token_cost": 2.5e-08,
17817
- "input_cost_per_audio_token": 5e-07,
17816
+ "cache_read_input_token_cost": 1e-08,
17817
+ "input_cost_per_audio_token": 3e-07,
17818
17818
  "input_cost_per_token": 1e-07,
17819
17819
  "litellm_provider": "vertex_ai-language-models",
17820
17820
  "max_audio_length_hours": 8.4,
@@ -18720,13 +18720,14 @@
18720
18720
  "supports_vision": true
18721
18721
  },
18722
18722
  "groq/openai/gpt-oss-20b": {
18723
- "input_cost_per_token": 1e-07,
18723
+ "cache_read_input_token_cost": 3.75e-08,
18724
+ "input_cost_per_token": 7.5e-08,
18724
18725
  "litellm_provider": "groq",
18725
18726
  "max_input_tokens": 131072,
18726
18727
  "max_output_tokens": 32768,
18727
18728
  "max_tokens": 32768,
18728
18729
  "mode": "chat",
18729
- "output_cost_per_token": 5e-07,
18730
+ "output_cost_per_token": 3e-07,
18730
18731
  "supports_function_calling": true,
18731
18732
  "supports_parallel_function_calling": true,
18732
18733
  "supports_reasoning": true,
@@ -18735,13 +18736,14 @@
18735
18736
  "supports_web_search": true
18736
18737
  },
18737
18738
  "groq/openai/gpt-oss-120b": {
18739
+ "cache_read_input_token_cost": 7.5e-08,
18738
18740
  "input_cost_per_token": 1.5e-07,
18739
18741
  "litellm_provider": "groq",
18740
18742
  "max_input_tokens": 131072,
18741
18743
  "max_output_tokens": 32766,
18742
18744
  "max_tokens": 32766,
18743
18745
  "mode": "chat",
18744
- "output_cost_per_token": 7.5e-07,
18746
+ "output_cost_per_token": 6e-07,
18745
18747
  "supports_function_calling": true,
18746
18748
  "supports_parallel_function_calling": true,
18747
18749
  "supports_reasoning": true,
@@ -23080,7 +23082,7 @@
23080
23082
  "supports_tool_choice": true
23081
23083
  },
23082
23084
  "gemini-2.5-flash-lite-preview-09-2025": {
23083
- "cache_read_input_token_cost": 2.5e-08,
23085
+ "cache_read_input_token_cost": 1e-08,
23084
23086
  "input_cost_per_audio_token": 3e-07,
23085
23087
  "input_cost_per_token": 1e-07,
23086
23088
  "litellm_provider": "vertex_ai-language-models",
@@ -23260,7 +23262,7 @@
23260
23262
  "supports_web_search": true
23261
23263
  },
23262
23264
  "gemini/gemini-2.5-flash-lite-preview-09-2025": {
23263
- "cache_read_input_token_cost": 2.5e-08,
23265
+ "cache_read_input_token_cost": 1e-08,
23264
23266
  "input_cost_per_audio_token": 3e-07,
23265
23267
  "input_cost_per_token": 1e-07,
23266
23268
  "litellm_provider": "gemini",
@@ -28134,15 +28136,15 @@
28134
28136
  "tool_use_system_prompt_tokens": 159
28135
28137
  },
28136
28138
  "us.anthropic.claude-opus-4-5-20251101-v1:0": {
28137
- "cache_creation_input_token_cost": 6.25e-06,
28138
- "cache_read_input_token_cost": 5e-07,
28139
- "input_cost_per_token": 5e-06,
28139
+ "cache_creation_input_token_cost": 6.875e-06,
28140
+ "cache_read_input_token_cost": 5.5e-07,
28141
+ "input_cost_per_token": 5.5e-06,
28140
28142
  "litellm_provider": "bedrock_converse",
28141
28143
  "max_input_tokens": 200000,
28142
28144
  "max_output_tokens": 64000,
28143
28145
  "max_tokens": 64000,
28144
28146
  "mode": "chat",
28145
- "output_cost_per_token": 2.5e-05,
28147
+ "output_cost_per_token": 2.75e-05,
28146
28148
  "search_context_cost_per_query": {
28147
28149
  "search_context_size_high": 0.01,
28148
28150
  "search_context_size_low": 0.01,
@@ -33983,13 +33985,13 @@
33983
33985
  ]
33984
33986
  },
33985
33987
  "vertex_ai/zai-org/glm-4.7-maas": {
33986
- "input_cost_per_token": 3e-07,
33988
+ "input_cost_per_token": 6e-07,
33987
33989
  "litellm_provider": "vertex_ai-zai_models",
33988
33990
  "max_input_tokens": 200000,
33989
33991
  "max_output_tokens": 128000,
33990
33992
  "max_tokens": 128000,
33991
33993
  "mode": "chat",
33992
- "output_cost_per_token": 1.2e-06,
33994
+ "output_cost_per_token": 2.2e-06,
33993
33995
  "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
33994
33996
  "supports_function_calling": true,
33995
33997
  "supports_reasoning": true,
@@ -35511,5 +35513,149 @@
35511
35513
  ],
35512
35514
  "supports_reasoning": true,
35513
35515
  "supports_tool_choice": true
35516
+ },
35517
+ "azure/gpt-5.2-codex": {
35518
+ "cache_read_input_token_cost": 1.75e-07,
35519
+ "input_cost_per_token": 1.75e-06,
35520
+ "litellm_provider": "azure",
35521
+ "max_input_tokens": 128000,
35522
+ "max_output_tokens": 16384,
35523
+ "max_tokens": 16384,
35524
+ "mode": "chat",
35525
+ "output_cost_per_token": 1.4e-05,
35526
+ "supported_endpoints": [
35527
+ "/v1/chat/completions",
35528
+ "/v1/responses"
35529
+ ],
35530
+ "supported_modalities": [
35531
+ "text",
35532
+ "image"
35533
+ ],
35534
+ "supported_output_modalities": [
35535
+ "text"
35536
+ ],
35537
+ "supports_function_calling": true,
35538
+ "supports_native_streaming": true,
35539
+ "supports_parallel_function_calling": true,
35540
+ "supports_pdf_input": true,
35541
+ "supports_prompt_caching": true,
35542
+ "supports_reasoning": true,
35543
+ "supports_response_schema": true,
35544
+ "supports_system_messages": true,
35545
+ "supports_tool_choice": true,
35546
+ "supports_vision": true
35547
+ },
35548
+ "dall-e-2": {
35549
+ "input_cost_per_image": 0.02,
35550
+ "litellm_provider": "openai",
35551
+ "mode": "image_generation",
35552
+ "supported_endpoints": [
35553
+ "/v1/images/generations",
35554
+ "/v1/images/edits",
35555
+ "/v1/images/variations"
35556
+ ]
35557
+ },
35558
+ "dall-e-3": {
35559
+ "input_cost_per_image": 0.04,
35560
+ "litellm_provider": "openai",
35561
+ "mode": "image_generation",
35562
+ "supported_endpoints": [
35563
+ "/v1/images/generations"
35564
+ ]
35565
+ },
35566
+ "gemini-2.5-computer-use-preview-10-2025": {
35567
+ "input_cost_per_token": 1.25e-06,
35568
+ "input_cost_per_token_above_200k_tokens": 2.5e-06,
35569
+ "litellm_provider": "vertex_ai-language-models",
35570
+ "max_images_per_prompt": 3000,
35571
+ "max_input_tokens": 128000,
35572
+ "max_output_tokens": 64000,
35573
+ "max_tokens": 64000,
35574
+ "mode": "chat",
35575
+ "output_cost_per_token": 1e-05,
35576
+ "output_cost_per_token_above_200k_tokens": 1.5e-05,
35577
+ "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/computer-use",
35578
+ "supported_modalities": [
35579
+ "text",
35580
+ "image"
35581
+ ],
35582
+ "supported_output_modalities": [
35583
+ "text"
35584
+ ],
35585
+ "supports_computer_use": true,
35586
+ "supports_function_calling": true,
35587
+ "supports_system_messages": true,
35588
+ "supports_tool_choice": true,
35589
+ "supports_vision": true
35590
+ },
35591
+ "chatgpt/gpt-5.2-codex": {
35592
+ "litellm_provider": "chatgpt",
35593
+ "max_input_tokens": 128000,
35594
+ "max_output_tokens": 128000,
35595
+ "max_tokens": 128000,
35596
+ "mode": "responses",
35597
+ "supported_endpoints": [
35598
+ "/v1/responses"
35599
+ ],
35600
+ "supports_function_calling": true,
35601
+ "supports_parallel_function_calling": true,
35602
+ "supports_response_schema": true,
35603
+ "supports_vision": true
35604
+ },
35605
+ "chatgpt/gpt-5.2": {
35606
+ "litellm_provider": "chatgpt",
35607
+ "max_input_tokens": 128000,
35608
+ "max_output_tokens": 64000,
35609
+ "max_tokens": 64000,
35610
+ "mode": "responses",
35611
+ "supported_endpoints": [
35612
+ "/v1/chat/completions",
35613
+ "/v1/responses"
35614
+ ],
35615
+ "supports_function_calling": true,
35616
+ "supports_parallel_function_calling": true,
35617
+ "supports_response_schema": true,
35618
+ "supports_vision": true
35619
+ },
35620
+ "chatgpt/gpt-5.1-codex-max": {
35621
+ "litellm_provider": "chatgpt",
35622
+ "max_input_tokens": 128000,
35623
+ "max_output_tokens": 128000,
35624
+ "max_tokens": 128000,
35625
+ "mode": "responses",
35626
+ "supported_endpoints": [
35627
+ "/v1/responses"
35628
+ ],
35629
+ "supports_function_calling": true,
35630
+ "supports_parallel_function_calling": true,
35631
+ "supports_response_schema": true,
35632
+ "supports_vision": true
35633
+ },
35634
+ "chatgpt/gpt-5.1-codex-mini": {
35635
+ "litellm_provider": "chatgpt",
35636
+ "max_input_tokens": 128000,
35637
+ "max_output_tokens": 64000,
35638
+ "max_tokens": 64000,
35639
+ "mode": "responses",
35640
+ "supported_endpoints": [
35641
+ "/v1/responses"
35642
+ ],
35643
+ "supports_function_calling": true,
35644
+ "supports_parallel_function_calling": true,
35645
+ "supports_response_schema": true,
35646
+ "supports_vision": true
35647
+ },
35648
+ "sarvam/sarvam-m": {
35649
+ "cache_creation_input_token_cost": 0,
35650
+ "cache_creation_input_token_cost_above_1hr": 0,
35651
+ "cache_read_input_token_cost": 0,
35652
+ "input_cost_per_token": 0,
35653
+ "litellm_provider": "sarvam",
35654
+ "max_input_tokens": 8192,
35655
+ "max_output_tokens": 32000,
35656
+ "max_tokens": 32000,
35657
+ "mode": "chat",
35658
+ "output_cost_per_token": 0,
35659
+ "supports_reasoning": true
35514
35660
  }
35515
35661
  }
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: tokencostauto
3
- Version: 0.1.452
3
+ Version: 0.1.462
4
4
  Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
5
5
  Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
6
6
  Project-URL: Homepage, https://github.com/madpin/tokencostaudo
@@ -0,0 +1,9 @@
1
+ tokencostauto/__init__.py,sha256=-4d_ryFH62SgNXPXA8vGPFZoAKtOBjnsg37EB_RkZG8,289
2
+ tokencostauto/constants.py,sha256=_82MlTkTrdrwzyRosQD7d3JdgNP9KAUM-cZo8DE00P0,3395
3
+ tokencostauto/costs.py,sha256=tXsgrTypq-dCHaHtoXcg2XepezWsAvZpl9gEsv_53iE,10679
4
+ tokencostauto/model_prices.json,sha256=aDFZZR6fnd6N4Fdc-zL3m6tsqmR8KiCelIqRuvx8oas,1265556
5
+ tokencostauto-0.1.462.dist-info/licenses/LICENSE,sha256=4PLv_CD6Ughnsvg_nM2XeTqGwVK6lQVR77kVWbPq-0U,1065
6
+ tokencostauto-0.1.462.dist-info/METADATA,sha256=bLqFKmPlxZSRF2VHZpr91xFCtaAPO1hdtx9yMt8TEbg,204076
7
+ tokencostauto-0.1.462.dist-info/WHEEL,sha256=qELbo2s1Yzl39ZmrAibXA2jjPLUYfnVhUNTlyF1rq0Y,92
8
+ tokencostauto-0.1.462.dist-info/top_level.txt,sha256=szZQTUJRotfIaeZCDsOgvofIkLt2ak88RP13oI51-TU,14
9
+ tokencostauto-0.1.462.dist-info/RECORD,,
@@ -1,5 +1,5 @@
1
1
  Wheel-Version: 1.0
2
- Generator: setuptools (80.9.0)
2
+ Generator: setuptools (80.10.1)
3
3
  Root-Is-Purelib: true
4
4
  Tag: py3-none-any
5
5
 
@@ -1,9 +0,0 @@
1
- tokencostauto/__init__.py,sha256=-4d_ryFH62SgNXPXA8vGPFZoAKtOBjnsg37EB_RkZG8,289
2
- tokencostauto/constants.py,sha256=_82MlTkTrdrwzyRosQD7d3JdgNP9KAUM-cZo8DE00P0,3395
3
- tokencostauto/costs.py,sha256=tXsgrTypq-dCHaHtoXcg2XepezWsAvZpl9gEsv_53iE,10679
4
- tokencostauto/model_prices.json,sha256=7Bd0_RCuJMpHU_g_fgZC8Cjqk1SXSSghuKk-O66eq4E,1260764
5
- tokencostauto-0.1.452.dist-info/licenses/LICENSE,sha256=4PLv_CD6Ughnsvg_nM2XeTqGwVK6lQVR77kVWbPq-0U,1065
6
- tokencostauto-0.1.452.dist-info/METADATA,sha256=C5o1P9PWo59i_Q7pgkvqQ-zJgWrtbYapKEv9bIW8efw,204076
7
- tokencostauto-0.1.452.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
8
- tokencostauto-0.1.452.dist-info/top_level.txt,sha256=szZQTUJRotfIaeZCDsOgvofIkLt2ak88RP13oI51-TU,14
9
- tokencostauto-0.1.452.dist-info/RECORD,,