tokencostauto 0.1.449__tar.gz → 0.1.451__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: tokencostauto
3
- Version: 0.1.449
3
+ Version: 0.1.451
4
4
  Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
5
5
  Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
6
6
  Project-URL: Homepage, https://github.com/madpin/tokencostaudo
@@ -11,7 +11,7 @@ tokencostauto = ["model_prices.json"]
11
11
  [project]
12
12
 
13
13
  name = "tokencostauto"
14
- version = "0.1.449"
14
+ version = "0.1.451"
15
15
 
16
16
  authors = [
17
17
  { name = "Trisha Pan", email = "trishaepan@gmail.com" },
@@ -16617,14 +16617,14 @@
16617
16617
  "supports_tool_choice": true
16618
16618
  },
16619
16619
  "azure_ai/grok-3": {
16620
- "input_cost_per_token": 3.3e-06,
16620
+ "input_cost_per_token": 3e-06,
16621
16621
  "litellm_provider": "azure_ai",
16622
16622
  "max_input_tokens": 131072,
16623
16623
  "max_output_tokens": 131072,
16624
16624
  "max_tokens": 131072,
16625
16625
  "mode": "chat",
16626
- "output_cost_per_token": 1.65e-05,
16627
- "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/",
16626
+ "output_cost_per_token": 1.5e-05,
16627
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
16628
16628
  "supports_function_calling": true,
16629
16629
  "supports_response_schema": false,
16630
16630
  "supports_tool_choice": true,
@@ -16660,14 +16660,14 @@
16660
16660
  "supports_web_search": true
16661
16661
  },
16662
16662
  "azure_ai/grok-3-mini": {
16663
- "input_cost_per_token": 2.75e-07,
16663
+ "input_cost_per_token": 2.5e-07,
16664
16664
  "litellm_provider": "azure_ai",
16665
16665
  "max_input_tokens": 131072,
16666
16666
  "max_output_tokens": 131072,
16667
16667
  "max_tokens": 131072,
16668
16668
  "mode": "chat",
16669
- "output_cost_per_token": 1.38e-06,
16670
- "source": "https://devblogs.microsoft.com/foundry/announcing-grok-3-and-grok-3-mini-on-azure-ai-foundry/",
16669
+ "output_cost_per_token": 1.27e-06,
16670
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
16671
16671
  "supports_function_calling": true,
16672
16672
  "supports_reasoning": true,
16673
16673
  "supports_response_schema": false,
@@ -18179,7 +18179,7 @@
18179
18179
  "input_cost_per_token_flex": 6.25e-07,
18180
18180
  "input_cost_per_token_priority": 2.5e-06,
18181
18181
  "litellm_provider": "openai",
18182
- "max_input_tokens": 400000,
18182
+ "max_input_tokens": 272000,
18183
18183
  "max_output_tokens": 128000,
18184
18184
  "max_tokens": 128000,
18185
18185
  "mode": "chat",
@@ -18218,7 +18218,7 @@
18218
18218
  "input_cost_per_token_flex": 1.25e-07,
18219
18219
  "input_cost_per_token_priority": 4.5e-07,
18220
18220
  "litellm_provider": "openai",
18221
- "max_input_tokens": 400000,
18221
+ "max_input_tokens": 272000,
18222
18222
  "max_output_tokens": 128000,
18223
18223
  "max_tokens": 128000,
18224
18224
  "mode": "chat",
@@ -18256,7 +18256,7 @@
18256
18256
  "input_cost_per_token_flex": 2.5e-08,
18257
18257
  "input_cost_per_token_priority": 2.5e-06,
18258
18258
  "litellm_provider": "openai",
18259
- "max_input_tokens": 400000,
18259
+ "max_input_tokens": 272000,
18260
18260
  "max_output_tokens": 128000,
18261
18261
  "max_tokens": 128000,
18262
18262
  "mode": "chat",
@@ -18357,7 +18357,7 @@
18357
18357
  "input_cost_per_token_flex": 6.25e-07,
18358
18358
  "input_cost_per_token_priority": 2.5e-06,
18359
18359
  "litellm_provider": "openai",
18360
- "max_input_tokens": 400000,
18360
+ "max_input_tokens": 272000,
18361
18361
  "max_output_tokens": 128000,
18362
18362
  "max_tokens": 128000,
18363
18363
  "mode": "chat",
@@ -18396,7 +18396,7 @@
18396
18396
  "input_cost_per_token_flex": 1.25e-07,
18397
18397
  "input_cost_per_token_priority": 4.5e-07,
18398
18398
  "litellm_provider": "openai",
18399
- "max_input_tokens": 400000,
18399
+ "max_input_tokens": 272000,
18400
18400
  "max_output_tokens": 128000,
18401
18401
  "max_tokens": 128000,
18402
18402
  "mode": "chat",
@@ -18433,7 +18433,7 @@
18433
18433
  "input_cost_per_token": 5e-08,
18434
18434
  "input_cost_per_token_flex": 2.5e-08,
18435
18435
  "litellm_provider": "openai",
18436
- "max_input_tokens": 400000,
18436
+ "max_input_tokens": 272000,
18437
18437
  "max_output_tokens": 128000,
18438
18438
  "max_tokens": 128000,
18439
18439
  "mode": "chat",
@@ -20623,7 +20623,7 @@
20623
20623
  "cache_read_input_token_cost": 2.5e-08,
20624
20624
  "input_cost_per_token": 2.5e-07,
20625
20625
  "litellm_provider": "openrouter",
20626
- "max_input_tokens": 400000,
20626
+ "max_input_tokens": 272000,
20627
20627
  "max_output_tokens": 128000,
20628
20628
  "max_tokens": 128000,
20629
20629
  "mode": "chat",
@@ -20642,7 +20642,7 @@
20642
20642
  "cache_read_input_token_cost": 5e-09,
20643
20643
  "input_cost_per_token": 5e-08,
20644
20644
  "litellm_provider": "openrouter",
20645
- "max_input_tokens": 400000,
20645
+ "max_input_tokens": 272000,
20646
20646
  "max_output_tokens": 128000,
20647
20647
  "max_tokens": 128000,
20648
20648
  "mode": "chat",
@@ -22961,7 +22961,7 @@
22961
22961
  "cache_read_input_token_cost": 1.25e-07,
22962
22962
  "input_cost_per_token": 1.25e-06,
22963
22963
  "litellm_provider": "openai",
22964
- "max_input_tokens": 400000,
22964
+ "max_input_tokens": 272000,
22965
22965
  "max_output_tokens": 128000,
22966
22966
  "max_tokens": 128000,
22967
22967
  "mode": "responses",
@@ -23451,7 +23451,7 @@
23451
23451
  "cache_read_input_token_cost": 1.25e-07,
23452
23452
  "input_cost_per_token": 1.25e-06,
23453
23453
  "litellm_provider": "openrouter",
23454
- "max_input_tokens": 400000,
23454
+ "max_input_tokens": 272000,
23455
23455
  "max_output_tokens": 128000,
23456
23456
  "max_tokens": 128000,
23457
23457
  "mode": "chat",
@@ -23470,7 +23470,7 @@
23470
23470
  "cache_read_input_token_cost": 1.25e-07,
23471
23471
  "input_cost_per_token": 1.25e-06,
23472
23472
  "litellm_provider": "openrouter",
23473
- "max_input_tokens": 400000,
23473
+ "max_input_tokens": 272000,
23474
23474
  "max_output_tokens": 128000,
23475
23475
  "max_tokens": 128000,
23476
23476
  "mode": "chat",
@@ -23720,22 +23720,22 @@
23720
23720
  "supports_tool_choice": true
23721
23721
  },
23722
23722
  "azure_ai/grok-4": {
23723
- "input_cost_per_token": 5.5e-06,
23723
+ "input_cost_per_token": 3e-06,
23724
23724
  "litellm_provider": "azure_ai",
23725
23725
  "max_input_tokens": 131072,
23726
23726
  "max_output_tokens": 131072,
23727
23727
  "max_tokens": 131072,
23728
23728
  "mode": "chat",
23729
- "output_cost_per_token": 2.75e-05,
23730
- "source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/",
23729
+ "output_cost_per_token": 1.5e-05,
23730
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
23731
23731
  "supports_function_calling": true,
23732
23732
  "supports_response_schema": true,
23733
23733
  "supports_tool_choice": true,
23734
23734
  "supports_web_search": true
23735
23735
  },
23736
23736
  "azure_ai/grok-4-fast-non-reasoning": {
23737
- "input_cost_per_token": 4.3e-07,
23738
- "output_cost_per_token": 1.73e-06,
23737
+ "input_cost_per_token": 2e-07,
23738
+ "output_cost_per_token": 5e-07,
23739
23739
  "litellm_provider": "azure_ai",
23740
23740
  "max_input_tokens": 131072,
23741
23741
  "max_output_tokens": 131072,
@@ -23747,28 +23747,28 @@
23747
23747
  "supports_web_search": true
23748
23748
  },
23749
23749
  "azure_ai/grok-4-fast-reasoning": {
23750
- "input_cost_per_token": 4.3e-07,
23751
- "output_cost_per_token": 1.73e-06,
23750
+ "input_cost_per_token": 2e-07,
23751
+ "output_cost_per_token": 5e-07,
23752
23752
  "litellm_provider": "azure_ai",
23753
23753
  "max_input_tokens": 131072,
23754
23754
  "max_output_tokens": 131072,
23755
23755
  "max_tokens": 131072,
23756
23756
  "mode": "chat",
23757
- "source": "https://techcommunity.microsoft.com/blog/azure-ai-foundry-blog/announcing-the-grok-4-fast-models-from-xai-now-available-in-azure-ai-foundry/4456701",
23757
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
23758
23758
  "supports_function_calling": true,
23759
23759
  "supports_response_schema": true,
23760
23760
  "supports_tool_choice": true,
23761
23761
  "supports_web_search": true
23762
23762
  },
23763
23763
  "azure_ai/grok-code-fast-1": {
23764
- "input_cost_per_token": 3.5e-06,
23764
+ "input_cost_per_token": 2e-07,
23765
23765
  "litellm_provider": "azure_ai",
23766
23766
  "max_input_tokens": 131072,
23767
23767
  "max_output_tokens": 131072,
23768
23768
  "max_tokens": 131072,
23769
23769
  "mode": "chat",
23770
- "output_cost_per_token": 1.75e-05,
23771
- "source": "https://azure.microsoft.com/en-us/blog/grok-4-is-now-available-in-azure-ai-foundry-unlock-frontier-intelligence-and-business-ready-capabilities/",
23770
+ "output_cost_per_token": 1.5e-06,
23771
+ "source": "https://azure.microsoft.com/en-us/pricing/details/ai-foundry-models/grok/",
23772
23772
  "supports_function_calling": true,
23773
23773
  "supports_response_schema": true,
23774
23774
  "supports_tool_choice": true,
@@ -24274,7 +24274,7 @@
24274
24274
  "input_cost_per_token": 1.5e-05,
24275
24275
  "input_cost_per_token_batches": 7.5e-06,
24276
24276
  "litellm_provider": "openai",
24277
- "max_input_tokens": 400000,
24277
+ "max_input_tokens": 128000,
24278
24278
  "max_output_tokens": 272000,
24279
24279
  "max_tokens": 272000,
24280
24280
  "mode": "responses",
@@ -24409,7 +24409,7 @@
24409
24409
  "input_cost_per_token": 1.5e-05,
24410
24410
  "input_cost_per_token_batches": 7.5e-06,
24411
24411
  "litellm_provider": "openai",
24412
- "max_input_tokens": 400000,
24412
+ "max_input_tokens": 128000,
24413
24413
  "max_output_tokens": 272000,
24414
24414
  "max_tokens": 272000,
24415
24415
  "mode": "responses",
@@ -26095,7 +26095,7 @@
26095
26095
  "input_cost_per_token": 1.25e-06,
26096
26096
  "input_cost_per_token_priority": 2.5e-06,
26097
26097
  "litellm_provider": "openai",
26098
- "max_input_tokens": 400000,
26098
+ "max_input_tokens": 272000,
26099
26099
  "max_output_tokens": 128000,
26100
26100
  "max_tokens": 128000,
26101
26101
  "mode": "chat",
@@ -26131,7 +26131,7 @@
26131
26131
  "input_cost_per_token": 1.25e-06,
26132
26132
  "input_cost_per_token_priority": 2.5e-06,
26133
26133
  "litellm_provider": "openai",
26134
- "max_input_tokens": 400000,
26134
+ "max_input_tokens": 272000,
26135
26135
  "max_output_tokens": 128000,
26136
26136
  "max_tokens": 128000,
26137
26137
  "mode": "chat",
@@ -26202,7 +26202,7 @@
26202
26202
  "input_cost_per_token": 1.25e-06,
26203
26203
  "input_cost_per_token_priority": 2.5e-06,
26204
26204
  "litellm_provider": "openai",
26205
- "max_input_tokens": 400000,
26205
+ "max_input_tokens": 272000,
26206
26206
  "max_output_tokens": 128000,
26207
26207
  "max_tokens": 128000,
26208
26208
  "mode": "responses",
@@ -26235,7 +26235,7 @@
26235
26235
  "input_cost_per_token": 2.5e-07,
26236
26236
  "input_cost_per_token_priority": 4.5e-07,
26237
26237
  "litellm_provider": "openai",
26238
- "max_input_tokens": 400000,
26238
+ "max_input_tokens": 272000,
26239
26239
  "max_output_tokens": 128000,
26240
26240
  "max_tokens": 128000,
26241
26241
  "mode": "responses",
@@ -26992,7 +26992,7 @@
26992
26992
  "input_cost_per_token": 1.25e-06,
26993
26993
  "input_cost_per_token_priority": 2.5e-06,
26994
26994
  "litellm_provider": "azure",
26995
- "max_input_tokens": 400000,
26995
+ "max_input_tokens": 272000,
26996
26996
  "max_output_tokens": 128000,
26997
26997
  "max_tokens": 128000,
26998
26998
  "mode": "chat",
@@ -27063,7 +27063,7 @@
27063
27063
  "input_cost_per_token": 1.25e-06,
27064
27064
  "input_cost_per_token_priority": 2.5e-06,
27065
27065
  "litellm_provider": "azure",
27066
- "max_input_tokens": 400000,
27066
+ "max_input_tokens": 272000,
27067
27067
  "max_output_tokens": 128000,
27068
27068
  "max_tokens": 128000,
27069
27069
  "mode": "responses",
@@ -27096,7 +27096,7 @@
27096
27096
  "input_cost_per_token": 2.5e-07,
27097
27097
  "input_cost_per_token_priority": 4.5e-07,
27098
27098
  "litellm_provider": "azure",
27099
- "max_input_tokens": 400000,
27099
+ "max_input_tokens": 272000,
27100
27100
  "max_output_tokens": 128000,
27101
27101
  "max_tokens": 128000,
27102
27102
  "mode": "responses",
@@ -28658,7 +28658,7 @@
28658
28658
  "input_cost_per_token": 1.24999e-06,
28659
28659
  "input_dbu_cost_per_token": 1.7857e-05,
28660
28660
  "litellm_provider": "databricks",
28661
- "max_input_tokens": 400000,
28661
+ "max_input_tokens": 272000,
28662
28662
  "max_output_tokens": 128000,
28663
28663
  "max_tokens": 128000,
28664
28664
  "metadata": {
@@ -28673,7 +28673,7 @@
28673
28673
  "input_cost_per_token": 1.24999e-06,
28674
28674
  "input_dbu_cost_per_token": 1.7857e-05,
28675
28675
  "litellm_provider": "databricks",
28676
- "max_input_tokens": 400000,
28676
+ "max_input_tokens": 272000,
28677
28677
  "max_output_tokens": 128000,
28678
28678
  "max_tokens": 128000,
28679
28679
  "metadata": {
@@ -28688,7 +28688,7 @@
28688
28688
  "input_cost_per_token": 2.4997000000000006e-07,
28689
28689
  "input_dbu_cost_per_token": 3.571e-06,
28690
28690
  "litellm_provider": "databricks",
28691
- "max_input_tokens": 400000,
28691
+ "max_input_tokens": 272000,
28692
28692
  "max_output_tokens": 128000,
28693
28693
  "max_tokens": 128000,
28694
28694
  "metadata": {
@@ -28703,7 +28703,7 @@
28703
28703
  "input_cost_per_token": 4.998e-08,
28704
28704
  "input_dbu_cost_per_token": 7.14e-07,
28705
28705
  "litellm_provider": "databricks",
28706
- "max_input_tokens": 400000,
28706
+ "max_input_tokens": 272000,
28707
28707
  "max_output_tokens": 128000,
28708
28708
  "max_tokens": 128000,
28709
28709
  "metadata": {
@@ -29165,7 +29165,7 @@
29165
29165
  "cache_read_input_token_cost": 1.25e-07,
29166
29166
  "input_cost_per_token": 1.25e-06,
29167
29167
  "litellm_provider": "openai",
29168
- "max_input_tokens": 400000,
29168
+ "max_input_tokens": 272000,
29169
29169
  "max_output_tokens": 128000,
29170
29170
  "max_tokens": 128000,
29171
29171
  "mode": "responses",
@@ -29525,6 +29525,7 @@
29525
29525
  "supports_pdf_input": true
29526
29526
  },
29527
29527
  "cerebras/zai-glm-4.6": {
29528
+ "deprecation_date": "2026-01-20",
29528
29529
  "input_cost_per_token": 2.25e-06,
29529
29530
  "litellm_provider": "cerebras",
29530
29531
  "max_input_tokens": 128000,
@@ -31715,7 +31716,7 @@
31715
31716
  "input_cost_per_token": 1.75e-06,
31716
31717
  "input_cost_per_token_priority": 3.5e-06,
31717
31718
  "litellm_provider": "openai",
31718
- "max_input_tokens": 400000,
31719
+ "max_input_tokens": 272000,
31719
31720
  "max_output_tokens": 128000,
31720
31721
  "max_tokens": 128000,
31721
31722
  "mode": "chat",
@@ -31752,7 +31753,7 @@
31752
31753
  "input_cost_per_token": 1.75e-06,
31753
31754
  "input_cost_per_token_priority": 3.5e-06,
31754
31755
  "litellm_provider": "openai",
31755
- "max_input_tokens": 400000,
31756
+ "max_input_tokens": 272000,
31756
31757
  "max_output_tokens": 128000,
31757
31758
  "max_tokens": 128000,
31758
31759
  "mode": "chat",
@@ -31820,7 +31821,7 @@
31820
31821
  "gpt-5.2-pro": {
31821
31822
  "input_cost_per_token": 2.1e-05,
31822
31823
  "litellm_provider": "openai",
31823
- "max_input_tokens": 400000,
31824
+ "max_input_tokens": 272000,
31824
31825
  "max_output_tokens": 128000,
31825
31826
  "max_tokens": 128000,
31826
31827
  "mode": "responses",
@@ -31851,7 +31852,7 @@
31851
31852
  "gpt-5.2-pro-2025-12-11": {
31852
31853
  "input_cost_per_token": 2.1e-05,
31853
31854
  "litellm_provider": "openai",
31854
- "max_input_tokens": 400000,
31855
+ "max_input_tokens": 272000,
31855
31856
  "max_output_tokens": 128000,
31856
31857
  "max_tokens": 128000,
31857
31858
  "mode": "responses",
@@ -31960,7 +31961,7 @@
31960
31961
  "input_cost_per_token": 1.75e-06,
31961
31962
  "input_cost_per_token_priority": 3.5e-06,
31962
31963
  "litellm_provider": "azure",
31963
- "max_input_tokens": 400000,
31964
+ "max_input_tokens": 272000,
31964
31965
  "max_output_tokens": 128000,
31965
31966
  "max_tokens": 128000,
31966
31967
  "mode": "chat",
@@ -32296,7 +32297,7 @@
32296
32297
  "cache_read_input_token_cost": 1.75e-07,
32297
32298
  "input_cost_per_token": 1.75e-06,
32298
32299
  "litellm_provider": "openrouter",
32299
- "max_input_tokens": 400000,
32300
+ "max_input_tokens": 272000,
32300
32301
  "max_output_tokens": 128000,
32301
32302
  "max_tokens": 128000,
32302
32303
  "mode": "chat",
@@ -32326,7 +32327,7 @@
32326
32327
  "input_cost_per_image": 0,
32327
32328
  "input_cost_per_token": 2.1e-05,
32328
32329
  "litellm_provider": "openrouter",
32329
- "max_input_tokens": 400000,
32330
+ "max_input_tokens": 272000,
32330
32331
  "max_output_tokens": 128000,
32331
32332
  "max_tokens": 128000,
32332
32333
  "mode": "chat",
@@ -35445,5 +35446,70 @@
35445
35446
  "supports_response_schema": true,
35446
35447
  "supports_audio_input": true,
35447
35448
  "supports_audio_output": true
35449
+ },
35450
+ "cerebras/zai-glm-4.7": {
35451
+ "input_cost_per_token": 2.25e-06,
35452
+ "litellm_provider": "cerebras",
35453
+ "max_input_tokens": 128000,
35454
+ "max_output_tokens": 128000,
35455
+ "max_tokens": 128000,
35456
+ "mode": "chat",
35457
+ "output_cost_per_token": 2.75e-06,
35458
+ "source": "https://www.cerebras.ai/pricing",
35459
+ "supports_function_calling": true,
35460
+ "supports_reasoning": true,
35461
+ "supports_tool_choice": true
35462
+ },
35463
+ "gpt-5.2-codex": {
35464
+ "cache_read_input_token_cost": 1.75e-07,
35465
+ "cache_read_input_token_cost_priority": 3.5e-07,
35466
+ "input_cost_per_token": 1.75e-06,
35467
+ "input_cost_per_token_priority": 3.5e-06,
35468
+ "litellm_provider": "openai",
35469
+ "max_input_tokens": 400000,
35470
+ "max_output_tokens": 128000,
35471
+ "max_tokens": 128000,
35472
+ "mode": "responses",
35473
+ "output_cost_per_token": 1.4e-05,
35474
+ "output_cost_per_token_priority": 2.8e-05,
35475
+ "supported_endpoints": [
35476
+ "/v1/responses"
35477
+ ],
35478
+ "supported_modalities": [
35479
+ "text",
35480
+ "image"
35481
+ ],
35482
+ "supported_output_modalities": [
35483
+ "text"
35484
+ ],
35485
+ "supports_function_calling": true,
35486
+ "supports_native_streaming": true,
35487
+ "supports_parallel_function_calling": true,
35488
+ "supports_pdf_input": true,
35489
+ "supports_prompt_caching": true,
35490
+ "supports_reasoning": true,
35491
+ "supports_response_schema": true,
35492
+ "supports_system_messages": false,
35493
+ "supports_tool_choice": true,
35494
+ "supports_vision": true
35495
+ },
35496
+ "openrouter/openai/gpt-5.2-codex": {
35497
+ "cache_read_input_token_cost": 1.75e-07,
35498
+ "input_cost_per_token": 1.75e-06,
35499
+ "litellm_provider": "openrouter",
35500
+ "max_input_tokens": 400000,
35501
+ "max_output_tokens": 128000,
35502
+ "max_tokens": 128000,
35503
+ "mode": "chat",
35504
+ "output_cost_per_token": 1.4e-05,
35505
+ "supported_modalities": [
35506
+ "text",
35507
+ "image"
35508
+ ],
35509
+ "supported_output_modalities": [
35510
+ "text"
35511
+ ],
35512
+ "supports_reasoning": true,
35513
+ "supports_tool_choice": true
35448
35514
  }
35449
35515
  }
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: tokencostauto
3
- Version: 0.1.449
3
+ Version: 0.1.451
4
4
  Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
5
5
  Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
6
6
  Project-URL: Homepage, https://github.com/madpin/tokencostaudo
File without changes