openrouter-provider 1.0.0__py3-none-any.whl → 1.0.1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- llms.py +15 -3
- {openrouter_provider-1.0.0.dist-info → openrouter_provider-1.0.1.dist-info}/METADATA +1 -1
- {openrouter_provider-1.0.0.dist-info → openrouter_provider-1.0.1.dist-info}/RECORD +5 -5
- {openrouter_provider-1.0.0.dist-info → openrouter_provider-1.0.1.dist-info}/WHEEL +0 -0
- {openrouter_provider-1.0.0.dist-info → openrouter_provider-1.0.1.dist-info}/top_level.txt +0 -0
llms.py
CHANGED
|
@@ -15,16 +15,25 @@ gpt_4_1_mini = LLMModel(name='openai/gpt-4.1-mini', input_cost=0.4, output_cost=
|
|
|
15
15
|
gpt_4_1_nano = LLMModel(name='openai/gpt-4.1-nano', input_cost=0.1, output_cost=0.4)
|
|
16
16
|
o4_mini = LLMModel(name='openai/o4-mini', input_cost=1.1, output_cost=4.4)
|
|
17
17
|
o4_mini_high = LLMModel(name='openai/o4-mini-high', input_cost=1.1, output_cost=4.4)
|
|
18
|
-
o3 = LLMModel(name='openai/o3', input_cost=
|
|
18
|
+
o3 = LLMModel(name='openai/o3', input_cost=2, output_cost=8)
|
|
19
|
+
gpt_5_nano = LLMModel(name='openai/gpt-5-nano', input_cost=0.05, output_cost=0.4)
|
|
20
|
+
gpt_5_mini = LLMModel(name='openai/gpt-5-mini', input_cost=0.25, output_cost=2)
|
|
21
|
+
gpt_5 = LLMModel(name='openai/gpt-5', input_cost=1.25, output_cost=10)
|
|
22
|
+
gpt_oss_20B_free = LLMModel(name='openai/gpt-oss-20b:free', input_cost=0, output_cost=0)
|
|
23
|
+
gpt_oss_20B = LLMModel(name='openai/gpt-oss-20b', input_cost=0.06, output_cost=0.2)
|
|
24
|
+
gpt_oss_120B = LLMModel(name='openai/gpt-oss-120b', input_cost=0.25, output_cost=0.69)
|
|
19
25
|
|
|
20
26
|
# Anthropic
|
|
21
27
|
claude_3_7_sonnet = LLMModel(name='anthropic/claude-3.7-sonnet', input_cost=3.0, output_cost=15.0)
|
|
22
28
|
claude_3_7_sonnet_thinking = LLMModel(name='anthropic/claude-3.7-sonnet:thinking', input_cost=3.0, output_cost=15.0)
|
|
23
29
|
claude_3_5_haiku = LLMModel(name='anthropic/claude-3.5-haiku', input_cost=0.8, output_cost=4.0)
|
|
30
|
+
claude_4_sonnet = LLMModel(name='anthropic/claude-sonnet-4', input_cost=3.0, output_cost=15.0)
|
|
31
|
+
claude_4_opus = LLMModel(name='anthropic/claude-opus-4', input_cost=15, output_cost=75)
|
|
32
|
+
claude_4_1_opus = LLMModel(name='anthropic/claude-opus-4.1', input_cost=15, output_cost=75)
|
|
24
33
|
|
|
25
34
|
# Google
|
|
26
35
|
gemini_2_0_flash = LLMModel(name='google/gemini-2.0-flash-001', input_cost=0.1, output_cost=0.4)
|
|
27
|
-
|
|
36
|
+
gemini_2_5_flash_lite = LLMModel(name='google/gemini-2.5-flash-lite', input_cost=0.1, output_cost=0.4)
|
|
28
37
|
gemini_2_5_flash = LLMModel(name='google/gemini-2.5-flash-preview', input_cost=0.15, output_cost=0.60)
|
|
29
38
|
gemini_2_5_flash_thinking = LLMModel(name='google/gemini-2.5-flash-preview:thinking', input_cost=0.15, output_cost=3.5)
|
|
30
39
|
gemini_2_5_pro = LLMModel(name='google/gemini-2.5-pro-preview-03-25', input_cost=1.25, output_cost=10)
|
|
@@ -34,10 +43,12 @@ deepseek_v3_free = LLMModel(name='deepseek/deepseek-chat-v3-0324:free', input_co
|
|
|
34
43
|
deepseek_v3 = LLMModel(name='deepseek/deepseek-chat-v3-0324', input_cost=0.3, output_cost=1.2)
|
|
35
44
|
deepseek_r1_free = LLMModel(name='deepseek/deepseek-r1:free', input_cost=0, output_cost=0)
|
|
36
45
|
deepseek_r1 = LLMModel(name='deepseek/deepseek-r1', input_cost=0.5, output_cost=2.2)
|
|
46
|
+
deepseek_v3_1 = LLMModel(name='deepseek/deepseek-chat-v3.1', input_cost=0.55, output_cost=1.68)
|
|
37
47
|
|
|
38
48
|
# xAI
|
|
39
49
|
grok_3_mini = LLMModel(name='x-ai/grok-3-mini-beta', input_cost=0.3, output_cost=0.5)
|
|
40
50
|
grok_3 = LLMModel(name='x-ai/grok-3-beta', input_cost=3, output_cost=15)
|
|
51
|
+
grok_4 = LLMModel(name='x-ai/grok-4', input_cost=3, output_cost=15)
|
|
41
52
|
|
|
42
53
|
# Microsoft
|
|
43
54
|
mai_ds_r1_free = LLMModel(name="microsoft/mai-ds-r1:free", input_cost=0, output_cost=0)
|
|
@@ -45,4 +56,5 @@ mai_ds_r1_free = LLMModel(name="microsoft/mai-ds-r1:free", input_cost=0, output_
|
|
|
45
56
|
# Others
|
|
46
57
|
llama_4_maverick_free = LLMModel(name="meta-llama/llama-4-maverick:free", input_cost=0, output_cost=0)
|
|
47
58
|
llama_4_scout = LLMModel(name="meta-llama/llama-4-scout", input_cost=0.11, output_cost=0.34)
|
|
48
|
-
mistral_small_3_1_24B_free = LLMModel(name="mistralai/mistral-small-3.1-24b-instruct:free", input_cost=0, output_cost=0)
|
|
59
|
+
mistral_small_3_1_24B_free = LLMModel(name="mistralai/mistral-small-3.1-24b-instruct:free", input_cost=0, output_cost=0)
|
|
60
|
+
|
|
@@ -1,10 +1,10 @@
|
|
|
1
1
|
__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
2
|
-
llms.py,sha256
|
|
2
|
+
llms.py,sha256=IjG3WgmPDbNR8zvAOYDsZvSZNmDjYfPS4IpdtlI4OI8,3535
|
|
3
3
|
message.py,sha256=xTPzrgiVd651EeccUoxh5h1Xo4Xhu1VpAnb6NB-p9YY,2860
|
|
4
4
|
openrouter.py,sha256=zXa6huZ4QIzJpbk9GeX4fEKJsyYwZN09ZlmbkAef5ik,7857
|
|
5
5
|
openrouter_provider.py,sha256=dJe_kPENFfhzC9J7q4XyglpXYq-1KPFIwsgyIdHeml8,8021
|
|
6
6
|
tool.py,sha256=tUUNLosz1XhzPIwY1zHXWNM3ePs7hcVD1a_W5hWTCWk,1975
|
|
7
|
-
openrouter_provider-1.0.
|
|
8
|
-
openrouter_provider-1.0.
|
|
9
|
-
openrouter_provider-1.0.
|
|
10
|
-
openrouter_provider-1.0.
|
|
7
|
+
openrouter_provider-1.0.1.dist-info/METADATA,sha256=g4NGzvYX4bY8VU3c9gEAT8XWcfT0itTTY-HiOQVb3o4,11503
|
|
8
|
+
openrouter_provider-1.0.1.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
9
|
+
openrouter_provider-1.0.1.dist-info/top_level.txt,sha256=P1gRRxgxRCkgKpO2Eru8nR9svmkGy1YlKSVavfrNIFs,58
|
|
10
|
+
openrouter_provider-1.0.1.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|