tokencostauto 0.1.433__tar.gz → 0.1.437__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {tokencostauto-0.1.433/tokencostauto.egg-info → tokencostauto-0.1.437}/PKG-INFO +1 -1
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/pyproject.toml +1 -1
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/tokencostauto/model_prices.json +301 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437/tokencostauto.egg-info}/PKG-INFO +1 -1
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/LICENSE +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/MANIFEST.in +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/README.md +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/setup.cfg +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/tests/test_costs.py +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/tokencostauto/__init__.py +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/tokencostauto/constants.py +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/tokencostauto/costs.py +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/tokencostauto.egg-info/SOURCES.txt +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/tokencostauto.egg-info/dependency_links.txt +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/tokencostauto.egg-info/requires.txt +0 -0
- {tokencostauto-0.1.433 → tokencostauto-0.1.437}/tokencostauto.egg-info/top_level.txt +0 -0
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: tokencostauto
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.437
|
|
4
4
|
Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
|
|
5
5
|
Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
|
|
6
6
|
Project-URL: Homepage, https://github.com/madpin/tokencostaudo
|
|
@@ -33657,5 +33657,306 @@
|
|
|
33657
33657
|
"supports_reasoning": true,
|
|
33658
33658
|
"supports_tool_choice": true,
|
|
33659
33659
|
"source": "https://docs.z.ai/guides/overview/pricing"
|
|
33660
|
+
},
|
|
33661
|
+
"openrouter/google/gemini-3-flash-preview": {
|
|
33662
|
+
"cache_read_input_token_cost": 5e-08,
|
|
33663
|
+
"input_cost_per_audio_token": 1e-06,
|
|
33664
|
+
"input_cost_per_token": 5e-07,
|
|
33665
|
+
"litellm_provider": "openrouter",
|
|
33666
|
+
"max_audio_length_hours": 8.4,
|
|
33667
|
+
"max_audio_per_prompt": 1,
|
|
33668
|
+
"max_images_per_prompt": 3000,
|
|
33669
|
+
"max_input_tokens": 1048576,
|
|
33670
|
+
"max_output_tokens": 65535,
|
|
33671
|
+
"max_pdf_size_mb": 30,
|
|
33672
|
+
"max_tokens": 65535,
|
|
33673
|
+
"max_video_length": 1,
|
|
33674
|
+
"max_videos_per_prompt": 10,
|
|
33675
|
+
"mode": "chat",
|
|
33676
|
+
"output_cost_per_reasoning_token": 3e-06,
|
|
33677
|
+
"output_cost_per_token": 3e-06,
|
|
33678
|
+
"rpm": 2000,
|
|
33679
|
+
"source": "https://ai.google.dev/pricing/gemini-3",
|
|
33680
|
+
"supported_endpoints": [
|
|
33681
|
+
"/v1/chat/completions",
|
|
33682
|
+
"/v1/completions",
|
|
33683
|
+
"/v1/batch"
|
|
33684
|
+
],
|
|
33685
|
+
"supported_modalities": [
|
|
33686
|
+
"text",
|
|
33687
|
+
"image",
|
|
33688
|
+
"audio",
|
|
33689
|
+
"video"
|
|
33690
|
+
],
|
|
33691
|
+
"supported_output_modalities": [
|
|
33692
|
+
"text"
|
|
33693
|
+
],
|
|
33694
|
+
"supports_audio_output": false,
|
|
33695
|
+
"supports_function_calling": true,
|
|
33696
|
+
"supports_parallel_function_calling": true,
|
|
33697
|
+
"supports_pdf_input": true,
|
|
33698
|
+
"supports_prompt_caching": true,
|
|
33699
|
+
"supports_reasoning": true,
|
|
33700
|
+
"supports_response_schema": true,
|
|
33701
|
+
"supports_system_messages": true,
|
|
33702
|
+
"supports_tool_choice": true,
|
|
33703
|
+
"supports_url_context": true,
|
|
33704
|
+
"supports_vision": true,
|
|
33705
|
+
"supports_web_search": true,
|
|
33706
|
+
"tpm": 800000
|
|
33707
|
+
},
|
|
33708
|
+
"amazon.nova-2-multimodal-embeddings-v1:0": {
|
|
33709
|
+
"litellm_provider": "bedrock",
|
|
33710
|
+
"max_input_tokens": 8172,
|
|
33711
|
+
"max_tokens": 8172,
|
|
33712
|
+
"mode": "embedding",
|
|
33713
|
+
"input_cost_per_token": 1.35e-07,
|
|
33714
|
+
"input_cost_per_image": 6e-05,
|
|
33715
|
+
"input_cost_per_video_per_second": 0.0007,
|
|
33716
|
+
"input_cost_per_audio_per_second": 0.00014,
|
|
33717
|
+
"output_cost_per_token": 0.0,
|
|
33718
|
+
"output_vector_size": 3072,
|
|
33719
|
+
"source": "https://us-east-1.console.aws.amazon.com/bedrock/home?region=us-east-1#/model-catalog/serverless/amazon.nova-2-multimodal-embeddings-v1:0",
|
|
33720
|
+
"supports_embedding_image_input": true,
|
|
33721
|
+
"supports_image_input": true,
|
|
33722
|
+
"supports_video_input": true,
|
|
33723
|
+
"supports_audio_input": true
|
|
33724
|
+
},
|
|
33725
|
+
"gigachat/GigaChat-2-Lite": {
|
|
33726
|
+
"input_cost_per_token": 0.0,
|
|
33727
|
+
"litellm_provider": "gigachat",
|
|
33728
|
+
"max_input_tokens": 128000,
|
|
33729
|
+
"max_output_tokens": 8192,
|
|
33730
|
+
"max_tokens": 8192,
|
|
33731
|
+
"mode": "chat",
|
|
33732
|
+
"output_cost_per_token": 0.0,
|
|
33733
|
+
"supports_function_calling": true,
|
|
33734
|
+
"supports_system_messages": true
|
|
33735
|
+
},
|
|
33736
|
+
"gigachat/GigaChat-2-Max": {
|
|
33737
|
+
"input_cost_per_token": 0.0,
|
|
33738
|
+
"litellm_provider": "gigachat",
|
|
33739
|
+
"max_input_tokens": 128000,
|
|
33740
|
+
"max_output_tokens": 8192,
|
|
33741
|
+
"max_tokens": 8192,
|
|
33742
|
+
"mode": "chat",
|
|
33743
|
+
"output_cost_per_token": 0.0,
|
|
33744
|
+
"supports_function_calling": true,
|
|
33745
|
+
"supports_system_messages": true,
|
|
33746
|
+
"supports_vision": true
|
|
33747
|
+
},
|
|
33748
|
+
"gigachat/GigaChat-2-Pro": {
|
|
33749
|
+
"input_cost_per_token": 0.0,
|
|
33750
|
+
"litellm_provider": "gigachat",
|
|
33751
|
+
"max_input_tokens": 128000,
|
|
33752
|
+
"max_output_tokens": 8192,
|
|
33753
|
+
"max_tokens": 8192,
|
|
33754
|
+
"mode": "chat",
|
|
33755
|
+
"output_cost_per_token": 0.0,
|
|
33756
|
+
"supports_function_calling": true,
|
|
33757
|
+
"supports_system_messages": true,
|
|
33758
|
+
"supports_vision": true
|
|
33759
|
+
},
|
|
33760
|
+
"gigachat/Embeddings": {
|
|
33761
|
+
"input_cost_per_token": 0.0,
|
|
33762
|
+
"litellm_provider": "gigachat",
|
|
33763
|
+
"max_input_tokens": 512,
|
|
33764
|
+
"max_tokens": 512,
|
|
33765
|
+
"mode": "embedding",
|
|
33766
|
+
"output_cost_per_token": 0.0,
|
|
33767
|
+
"output_vector_size": 1024
|
|
33768
|
+
},
|
|
33769
|
+
"gigachat/Embeddings-2": {
|
|
33770
|
+
"input_cost_per_token": 0.0,
|
|
33771
|
+
"litellm_provider": "gigachat",
|
|
33772
|
+
"max_input_tokens": 512,
|
|
33773
|
+
"max_tokens": 512,
|
|
33774
|
+
"mode": "embedding",
|
|
33775
|
+
"output_cost_per_token": 0.0,
|
|
33776
|
+
"output_vector_size": 1024
|
|
33777
|
+
},
|
|
33778
|
+
"gigachat/EmbeddingsGigaR": {
|
|
33779
|
+
"input_cost_per_token": 0.0,
|
|
33780
|
+
"litellm_provider": "gigachat",
|
|
33781
|
+
"max_input_tokens": 4096,
|
|
33782
|
+
"max_tokens": 4096,
|
|
33783
|
+
"mode": "embedding",
|
|
33784
|
+
"output_cost_per_token": 0.0,
|
|
33785
|
+
"output_vector_size": 2560
|
|
33786
|
+
},
|
|
33787
|
+
"llamagate/llama-3.1-8b": {
|
|
33788
|
+
"max_tokens": 8192,
|
|
33789
|
+
"max_input_tokens": 131072,
|
|
33790
|
+
"max_output_tokens": 8192,
|
|
33791
|
+
"input_cost_per_token": 3e-08,
|
|
33792
|
+
"output_cost_per_token": 5e-08,
|
|
33793
|
+
"litellm_provider": "llamagate",
|
|
33794
|
+
"mode": "chat",
|
|
33795
|
+
"supports_function_calling": true,
|
|
33796
|
+
"supports_response_schema": true
|
|
33797
|
+
},
|
|
33798
|
+
"llamagate/llama-3.2-3b": {
|
|
33799
|
+
"max_tokens": 8192,
|
|
33800
|
+
"max_input_tokens": 131072,
|
|
33801
|
+
"max_output_tokens": 8192,
|
|
33802
|
+
"input_cost_per_token": 4e-08,
|
|
33803
|
+
"output_cost_per_token": 8e-08,
|
|
33804
|
+
"litellm_provider": "llamagate",
|
|
33805
|
+
"mode": "chat",
|
|
33806
|
+
"supports_function_calling": true,
|
|
33807
|
+
"supports_response_schema": true
|
|
33808
|
+
},
|
|
33809
|
+
"llamagate/mistral-7b-v0.3": {
|
|
33810
|
+
"max_tokens": 8192,
|
|
33811
|
+
"max_input_tokens": 32768,
|
|
33812
|
+
"max_output_tokens": 8192,
|
|
33813
|
+
"input_cost_per_token": 1e-07,
|
|
33814
|
+
"output_cost_per_token": 1.5e-07,
|
|
33815
|
+
"litellm_provider": "llamagate",
|
|
33816
|
+
"mode": "chat",
|
|
33817
|
+
"supports_function_calling": true,
|
|
33818
|
+
"supports_response_schema": true
|
|
33819
|
+
},
|
|
33820
|
+
"llamagate/qwen3-8b": {
|
|
33821
|
+
"max_tokens": 8192,
|
|
33822
|
+
"max_input_tokens": 32768,
|
|
33823
|
+
"max_output_tokens": 8192,
|
|
33824
|
+
"input_cost_per_token": 4e-08,
|
|
33825
|
+
"output_cost_per_token": 1.4e-07,
|
|
33826
|
+
"litellm_provider": "llamagate",
|
|
33827
|
+
"mode": "chat",
|
|
33828
|
+
"supports_function_calling": true,
|
|
33829
|
+
"supports_response_schema": true
|
|
33830
|
+
},
|
|
33831
|
+
"llamagate/dolphin3-8b": {
|
|
33832
|
+
"max_tokens": 8192,
|
|
33833
|
+
"max_input_tokens": 128000,
|
|
33834
|
+
"max_output_tokens": 8192,
|
|
33835
|
+
"input_cost_per_token": 8e-08,
|
|
33836
|
+
"output_cost_per_token": 1.5e-07,
|
|
33837
|
+
"litellm_provider": "llamagate",
|
|
33838
|
+
"mode": "chat",
|
|
33839
|
+
"supports_function_calling": true,
|
|
33840
|
+
"supports_response_schema": true
|
|
33841
|
+
},
|
|
33842
|
+
"llamagate/deepseek-r1-8b": {
|
|
33843
|
+
"max_tokens": 16384,
|
|
33844
|
+
"max_input_tokens": 65536,
|
|
33845
|
+
"max_output_tokens": 16384,
|
|
33846
|
+
"input_cost_per_token": 1e-07,
|
|
33847
|
+
"output_cost_per_token": 2e-07,
|
|
33848
|
+
"litellm_provider": "llamagate",
|
|
33849
|
+
"mode": "chat",
|
|
33850
|
+
"supports_function_calling": true,
|
|
33851
|
+
"supports_response_schema": true,
|
|
33852
|
+
"supports_reasoning": true
|
|
33853
|
+
},
|
|
33854
|
+
"llamagate/deepseek-r1-7b-qwen": {
|
|
33855
|
+
"max_tokens": 16384,
|
|
33856
|
+
"max_input_tokens": 131072,
|
|
33857
|
+
"max_output_tokens": 16384,
|
|
33858
|
+
"input_cost_per_token": 8e-08,
|
|
33859
|
+
"output_cost_per_token": 1.5e-07,
|
|
33860
|
+
"litellm_provider": "llamagate",
|
|
33861
|
+
"mode": "chat",
|
|
33862
|
+
"supports_function_calling": true,
|
|
33863
|
+
"supports_response_schema": true,
|
|
33864
|
+
"supports_reasoning": true
|
|
33865
|
+
},
|
|
33866
|
+
"llamagate/openthinker-7b": {
|
|
33867
|
+
"max_tokens": 8192,
|
|
33868
|
+
"max_input_tokens": 32768,
|
|
33869
|
+
"max_output_tokens": 8192,
|
|
33870
|
+
"input_cost_per_token": 8e-08,
|
|
33871
|
+
"output_cost_per_token": 1.5e-07,
|
|
33872
|
+
"litellm_provider": "llamagate",
|
|
33873
|
+
"mode": "chat",
|
|
33874
|
+
"supports_function_calling": true,
|
|
33875
|
+
"supports_response_schema": true,
|
|
33876
|
+
"supports_reasoning": true
|
|
33877
|
+
},
|
|
33878
|
+
"llamagate/qwen2.5-coder-7b": {
|
|
33879
|
+
"max_tokens": 8192,
|
|
33880
|
+
"max_input_tokens": 32768,
|
|
33881
|
+
"max_output_tokens": 8192,
|
|
33882
|
+
"input_cost_per_token": 6e-08,
|
|
33883
|
+
"output_cost_per_token": 1.2e-07,
|
|
33884
|
+
"litellm_provider": "llamagate",
|
|
33885
|
+
"mode": "chat",
|
|
33886
|
+
"supports_function_calling": true,
|
|
33887
|
+
"supports_response_schema": true
|
|
33888
|
+
},
|
|
33889
|
+
"llamagate/deepseek-coder-6.7b": {
|
|
33890
|
+
"max_tokens": 4096,
|
|
33891
|
+
"max_input_tokens": 16384,
|
|
33892
|
+
"max_output_tokens": 4096,
|
|
33893
|
+
"input_cost_per_token": 6e-08,
|
|
33894
|
+
"output_cost_per_token": 1.2e-07,
|
|
33895
|
+
"litellm_provider": "llamagate",
|
|
33896
|
+
"mode": "chat",
|
|
33897
|
+
"supports_function_calling": true,
|
|
33898
|
+
"supports_response_schema": true
|
|
33899
|
+
},
|
|
33900
|
+
"llamagate/codellama-7b": {
|
|
33901
|
+
"max_tokens": 4096,
|
|
33902
|
+
"max_input_tokens": 16384,
|
|
33903
|
+
"max_output_tokens": 4096,
|
|
33904
|
+
"input_cost_per_token": 6e-08,
|
|
33905
|
+
"output_cost_per_token": 1.2e-07,
|
|
33906
|
+
"litellm_provider": "llamagate",
|
|
33907
|
+
"mode": "chat",
|
|
33908
|
+
"supports_function_calling": true,
|
|
33909
|
+
"supports_response_schema": true
|
|
33910
|
+
},
|
|
33911
|
+
"llamagate/qwen3-vl-8b": {
|
|
33912
|
+
"max_tokens": 8192,
|
|
33913
|
+
"max_input_tokens": 32768,
|
|
33914
|
+
"max_output_tokens": 8192,
|
|
33915
|
+
"input_cost_per_token": 1.5e-07,
|
|
33916
|
+
"output_cost_per_token": 5.5e-07,
|
|
33917
|
+
"litellm_provider": "llamagate",
|
|
33918
|
+
"mode": "chat",
|
|
33919
|
+
"supports_function_calling": true,
|
|
33920
|
+
"supports_response_schema": true,
|
|
33921
|
+
"supports_vision": true
|
|
33922
|
+
},
|
|
33923
|
+
"llamagate/llava-7b": {
|
|
33924
|
+
"max_tokens": 2048,
|
|
33925
|
+
"max_input_tokens": 4096,
|
|
33926
|
+
"max_output_tokens": 2048,
|
|
33927
|
+
"input_cost_per_token": 1e-07,
|
|
33928
|
+
"output_cost_per_token": 2e-07,
|
|
33929
|
+
"litellm_provider": "llamagate",
|
|
33930
|
+
"mode": "chat",
|
|
33931
|
+
"supports_response_schema": true,
|
|
33932
|
+
"supports_vision": true
|
|
33933
|
+
},
|
|
33934
|
+
"llamagate/gemma3-4b": {
|
|
33935
|
+
"max_tokens": 8192,
|
|
33936
|
+
"max_input_tokens": 128000,
|
|
33937
|
+
"max_output_tokens": 8192,
|
|
33938
|
+
"input_cost_per_token": 3e-08,
|
|
33939
|
+
"output_cost_per_token": 8e-08,
|
|
33940
|
+
"litellm_provider": "llamagate",
|
|
33941
|
+
"mode": "chat",
|
|
33942
|
+
"supports_function_calling": true,
|
|
33943
|
+
"supports_response_schema": true,
|
|
33944
|
+
"supports_vision": true
|
|
33945
|
+
},
|
|
33946
|
+
"llamagate/nomic-embed-text": {
|
|
33947
|
+
"max_tokens": 8192,
|
|
33948
|
+
"max_input_tokens": 8192,
|
|
33949
|
+
"input_cost_per_token": 2e-08,
|
|
33950
|
+
"output_cost_per_token": 0,
|
|
33951
|
+
"litellm_provider": "llamagate",
|
|
33952
|
+
"mode": "embedding"
|
|
33953
|
+
},
|
|
33954
|
+
"llamagate/qwen3-embedding-8b": {
|
|
33955
|
+
"max_tokens": 40960,
|
|
33956
|
+
"max_input_tokens": 40960,
|
|
33957
|
+
"input_cost_per_token": 2e-08,
|
|
33958
|
+
"output_cost_per_token": 0,
|
|
33959
|
+
"litellm_provider": "llamagate",
|
|
33960
|
+
"mode": "embedding"
|
|
33660
33961
|
}
|
|
33661
33962
|
}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: tokencostauto
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.437
|
|
4
4
|
Summary: To calculate token and translated USD cost of string and message calls to OpenAI, for example when used by AI agents
|
|
5
5
|
Author-email: Trisha Pan <trishaepan@gmail.com>, Alex Reibman <areibman@gmail.com>, Pratyush Shukla <ps4534@nyu.edu>, Thiago MadPin <madpin@gmail.com>
|
|
6
6
|
Project-URL: Homepage, https://github.com/madpin/tokencostaudo
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|