@mariozechner/pi-ai 0.9.2 → 0.9.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/models.generated.js
CHANGED
|
@@ -104,6 +104,23 @@ export const MODELS = {
|
|
|
104
104
|
contextWindow: 200000,
|
|
105
105
|
maxTokens: 8192,
|
|
106
106
|
},
|
|
107
|
+
"claude-opus-4-5": {
|
|
108
|
+
id: "claude-opus-4-5",
|
|
109
|
+
name: "Claude Opus 4.5 (latest)",
|
|
110
|
+
api: "anthropic-messages",
|
|
111
|
+
provider: "anthropic",
|
|
112
|
+
baseUrl: "https://api.anthropic.com",
|
|
113
|
+
reasoning: true,
|
|
114
|
+
input: ["text", "image"],
|
|
115
|
+
cost: {
|
|
116
|
+
input: 5,
|
|
117
|
+
output: 25,
|
|
118
|
+
cacheRead: 1.5,
|
|
119
|
+
cacheWrite: 18.75,
|
|
120
|
+
},
|
|
121
|
+
contextWindow: 200000,
|
|
122
|
+
maxTokens: 64000,
|
|
123
|
+
},
|
|
107
124
|
"claude-3-opus-20240229": {
|
|
108
125
|
id: "claude-3-opus-20240229",
|
|
109
126
|
name: "Claude Opus 3",
|
|
@@ -1937,6 +1954,23 @@ export const MODELS = {
|
|
|
1937
1954
|
},
|
|
1938
1955
|
},
|
|
1939
1956
|
openrouter: {
|
|
1957
|
+
"anthropic/claude-opus-4.5": {
|
|
1958
|
+
id: "anthropic/claude-opus-4.5",
|
|
1959
|
+
name: "Anthropic: Claude Opus 4.5",
|
|
1960
|
+
api: "openai-completions",
|
|
1961
|
+
provider: "openrouter",
|
|
1962
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
1963
|
+
reasoning: true,
|
|
1964
|
+
input: ["text", "image"],
|
|
1965
|
+
cost: {
|
|
1966
|
+
input: 5,
|
|
1967
|
+
output: 25,
|
|
1968
|
+
cacheRead: 0.5,
|
|
1969
|
+
cacheWrite: 6.25,
|
|
1970
|
+
},
|
|
1971
|
+
contextWindow: 200000,
|
|
1972
|
+
maxTokens: 32000,
|
|
1973
|
+
},
|
|
1940
1974
|
"allenai/olmo-3-7b-instruct": {
|
|
1941
1975
|
id: "allenai/olmo-3-7b-instruct",
|
|
1942
1976
|
name: "AllenAI: Olmo 3 7B Instruct",
|
|
@@ -1954,23 +1988,6 @@ export const MODELS = {
|
|
|
1954
1988
|
contextWindow: 65536,
|
|
1955
1989
|
maxTokens: 65536,
|
|
1956
1990
|
},
|
|
1957
|
-
"x-ai/grok-4.1-fast": {
|
|
1958
|
-
id: "x-ai/grok-4.1-fast",
|
|
1959
|
-
name: "xAI: Grok 4.1 Fast",
|
|
1960
|
-
api: "openai-completions",
|
|
1961
|
-
provider: "openrouter",
|
|
1962
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
1963
|
-
reasoning: true,
|
|
1964
|
-
input: ["text", "image"],
|
|
1965
|
-
cost: {
|
|
1966
|
-
input: 0,
|
|
1967
|
-
output: 0,
|
|
1968
|
-
cacheRead: 0,
|
|
1969
|
-
cacheWrite: 0,
|
|
1970
|
-
},
|
|
1971
|
-
contextWindow: 2000000,
|
|
1972
|
-
maxTokens: 30000,
|
|
1973
|
-
},
|
|
1974
1991
|
"x-ai/grok-4.1-fast:free": {
|
|
1975
1992
|
id: "x-ai/grok-4.1-fast:free",
|
|
1976
1993
|
name: "xAI: Grok 4.1 Fast (free)",
|
|
@@ -2088,7 +2105,7 @@ export const MODELS = {
|
|
|
2088
2105
|
cacheWrite: 0,
|
|
2089
2106
|
},
|
|
2090
2107
|
contextWindow: 256000,
|
|
2091
|
-
maxTokens:
|
|
2108
|
+
maxTokens: 32768,
|
|
2092
2109
|
},
|
|
2093
2110
|
"moonshotai/kimi-k2-thinking": {
|
|
2094
2111
|
id: "moonshotai/kimi-k2-thinking",
|
|
@@ -3408,13 +3425,13 @@ export const MODELS = {
|
|
|
3408
3425
|
reasoning: true,
|
|
3409
3426
|
input: ["text"],
|
|
3410
3427
|
cost: {
|
|
3411
|
-
input:
|
|
3412
|
-
output:
|
|
3428
|
+
input: 0.38,
|
|
3429
|
+
output: 1.53,
|
|
3413
3430
|
cacheRead: 0,
|
|
3414
3431
|
cacheWrite: 0,
|
|
3415
3432
|
},
|
|
3416
3433
|
contextWindow: 262144,
|
|
3417
|
-
maxTokens:
|
|
3434
|
+
maxTokens: 262144,
|
|
3418
3435
|
},
|
|
3419
3436
|
"google/gemini-2.5-flash-lite": {
|
|
3420
3437
|
id: "google/gemini-2.5-flash-lite",
|
|
@@ -3850,8 +3867,8 @@ export const MODELS = {
|
|
|
3850
3867
|
reasoning: true,
|
|
3851
3868
|
input: ["text"],
|
|
3852
3869
|
cost: {
|
|
3853
|
-
input: 0.
|
|
3854
|
-
output: 0.
|
|
3870
|
+
input: 0.049999999999999996,
|
|
3871
|
+
output: 0.19999999999999998,
|
|
3855
3872
|
cacheRead: 0,
|
|
3856
3873
|
cacheWrite: 0,
|
|
3857
3874
|
},
|
|
@@ -4793,9 +4810,9 @@ export const MODELS = {
|
|
|
4793
4810
|
contextWindow: 32768,
|
|
4794
4811
|
maxTokens: 4096,
|
|
4795
4812
|
},
|
|
4796
|
-
"anthropic/claude-3.5-haiku
|
|
4797
|
-
id: "anthropic/claude-3.5-haiku
|
|
4798
|
-
name: "Anthropic: Claude 3.5 Haiku
|
|
4813
|
+
"anthropic/claude-3.5-haiku": {
|
|
4814
|
+
id: "anthropic/claude-3.5-haiku",
|
|
4815
|
+
name: "Anthropic: Claude 3.5 Haiku",
|
|
4799
4816
|
api: "openai-completions",
|
|
4800
4817
|
provider: "openrouter",
|
|
4801
4818
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -4810,9 +4827,9 @@ export const MODELS = {
|
|
|
4810
4827
|
contextWindow: 200000,
|
|
4811
4828
|
maxTokens: 8192,
|
|
4812
4829
|
},
|
|
4813
|
-
"anthropic/claude-3.5-haiku": {
|
|
4814
|
-
id: "anthropic/claude-3.5-haiku",
|
|
4815
|
-
name: "Anthropic: Claude 3.5 Haiku",
|
|
4830
|
+
"anthropic/claude-3.5-haiku-20241022": {
|
|
4831
|
+
id: "anthropic/claude-3.5-haiku-20241022",
|
|
4832
|
+
name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
|
|
4816
4833
|
api: "openai-completions",
|
|
4817
4834
|
provider: "openrouter",
|
|
4818
4835
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -4878,23 +4895,6 @@ export const MODELS = {
|
|
|
4878
4895
|
contextWindow: 131072,
|
|
4879
4896
|
maxTokens: 4096,
|
|
4880
4897
|
},
|
|
4881
|
-
"qwen/qwen-2.5-7b-instruct": {
|
|
4882
|
-
id: "qwen/qwen-2.5-7b-instruct",
|
|
4883
|
-
name: "Qwen: Qwen2.5 7B Instruct",
|
|
4884
|
-
api: "openai-completions",
|
|
4885
|
-
provider: "openrouter",
|
|
4886
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
4887
|
-
reasoning: false,
|
|
4888
|
-
input: ["text"],
|
|
4889
|
-
cost: {
|
|
4890
|
-
input: 0.04,
|
|
4891
|
-
output: 0.09999999999999999,
|
|
4892
|
-
cacheRead: 0,
|
|
4893
|
-
cacheWrite: 0,
|
|
4894
|
-
},
|
|
4895
|
-
contextWindow: 32768,
|
|
4896
|
-
maxTokens: 4096,
|
|
4897
|
-
},
|
|
4898
4898
|
"nvidia/llama-3.1-nemotron-70b-instruct": {
|
|
4899
4899
|
id: "nvidia/llama-3.1-nemotron-70b-instruct",
|
|
4900
4900
|
name: "NVIDIA: Llama 3.1 Nemotron 70B Instruct",
|
|
@@ -5082,23 +5082,6 @@ export const MODELS = {
|
|
|
5082
5082
|
contextWindow: 128000,
|
|
5083
5083
|
maxTokens: 16384,
|
|
5084
5084
|
},
|
|
5085
|
-
"meta-llama/llama-3.1-70b-instruct": {
|
|
5086
|
-
id: "meta-llama/llama-3.1-70b-instruct",
|
|
5087
|
-
name: "Meta: Llama 3.1 70B Instruct",
|
|
5088
|
-
api: "openai-completions",
|
|
5089
|
-
provider: "openrouter",
|
|
5090
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5091
|
-
reasoning: false,
|
|
5092
|
-
input: ["text"],
|
|
5093
|
-
cost: {
|
|
5094
|
-
input: 0.39999999999999997,
|
|
5095
|
-
output: 0.39999999999999997,
|
|
5096
|
-
cacheRead: 0,
|
|
5097
|
-
cacheWrite: 0,
|
|
5098
|
-
},
|
|
5099
|
-
contextWindow: 131072,
|
|
5100
|
-
maxTokens: 4096,
|
|
5101
|
-
},
|
|
5102
5085
|
"meta-llama/llama-3.1-8b-instruct": {
|
|
5103
5086
|
id: "meta-llama/llama-3.1-8b-instruct",
|
|
5104
5087
|
name: "Meta: Llama 3.1 8B Instruct",
|
|
@@ -5133,6 +5116,23 @@ export const MODELS = {
|
|
|
5133
5116
|
contextWindow: 130815,
|
|
5134
5117
|
maxTokens: 4096,
|
|
5135
5118
|
},
|
|
5119
|
+
"meta-llama/llama-3.1-70b-instruct": {
|
|
5120
|
+
id: "meta-llama/llama-3.1-70b-instruct",
|
|
5121
|
+
name: "Meta: Llama 3.1 70B Instruct",
|
|
5122
|
+
api: "openai-completions",
|
|
5123
|
+
provider: "openrouter",
|
|
5124
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
5125
|
+
reasoning: false,
|
|
5126
|
+
input: ["text"],
|
|
5127
|
+
cost: {
|
|
5128
|
+
input: 0.39999999999999997,
|
|
5129
|
+
output: 0.39999999999999997,
|
|
5130
|
+
cacheRead: 0,
|
|
5131
|
+
cacheWrite: 0,
|
|
5132
|
+
},
|
|
5133
|
+
contextWindow: 131072,
|
|
5134
|
+
maxTokens: 4096,
|
|
5135
|
+
},
|
|
5136
5136
|
"mistralai/mistral-nemo": {
|
|
5137
5137
|
id: "mistralai/mistral-nemo",
|
|
5138
5138
|
name: "Mistral: Mistral Nemo",
|
|
@@ -5150,9 +5150,9 @@ export const MODELS = {
|
|
|
5150
5150
|
contextWindow: 131072,
|
|
5151
5151
|
maxTokens: 16384,
|
|
5152
5152
|
},
|
|
5153
|
-
"openai/gpt-4o-mini
|
|
5154
|
-
id: "openai/gpt-4o-mini
|
|
5155
|
-
name: "OpenAI: GPT-4o-mini
|
|
5153
|
+
"openai/gpt-4o-mini": {
|
|
5154
|
+
id: "openai/gpt-4o-mini",
|
|
5155
|
+
name: "OpenAI: GPT-4o-mini",
|
|
5156
5156
|
api: "openai-completions",
|
|
5157
5157
|
provider: "openrouter",
|
|
5158
5158
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5167,9 +5167,9 @@ export const MODELS = {
|
|
|
5167
5167
|
contextWindow: 128000,
|
|
5168
5168
|
maxTokens: 16384,
|
|
5169
5169
|
},
|
|
5170
|
-
"openai/gpt-4o-mini": {
|
|
5171
|
-
id: "openai/gpt-4o-mini",
|
|
5172
|
-
name: "OpenAI: GPT-4o-mini",
|
|
5170
|
+
"openai/gpt-4o-mini-2024-07-18": {
|
|
5171
|
+
id: "openai/gpt-4o-mini-2024-07-18",
|
|
5172
|
+
name: "OpenAI: GPT-4o-mini (2024-07-18)",
|
|
5173
5173
|
api: "openai-completions",
|
|
5174
5174
|
provider: "openrouter",
|
|
5175
5175
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -5269,23 +5269,6 @@ export const MODELS = {
|
|
|
5269
5269
|
contextWindow: 128000,
|
|
5270
5270
|
maxTokens: 4096,
|
|
5271
5271
|
},
|
|
5272
|
-
"openai/gpt-4o-2024-05-13": {
|
|
5273
|
-
id: "openai/gpt-4o-2024-05-13",
|
|
5274
|
-
name: "OpenAI: GPT-4o (2024-05-13)",
|
|
5275
|
-
api: "openai-completions",
|
|
5276
|
-
provider: "openrouter",
|
|
5277
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
5278
|
-
reasoning: false,
|
|
5279
|
-
input: ["text", "image"],
|
|
5280
|
-
cost: {
|
|
5281
|
-
input: 5,
|
|
5282
|
-
output: 15,
|
|
5283
|
-
cacheRead: 0,
|
|
5284
|
-
cacheWrite: 0,
|
|
5285
|
-
},
|
|
5286
|
-
contextWindow: 128000,
|
|
5287
|
-
maxTokens: 4096,
|
|
5288
|
-
},
|
|
5289
5272
|
"openai/gpt-4o": {
|
|
5290
5273
|
id: "openai/gpt-4o",
|
|
5291
5274
|
name: "OpenAI: GPT-4o",
|
|
@@ -5320,22 +5303,22 @@ export const MODELS = {
|
|
|
5320
5303
|
contextWindow: 128000,
|
|
5321
5304
|
maxTokens: 64000,
|
|
5322
5305
|
},
|
|
5323
|
-
"
|
|
5324
|
-
id: "
|
|
5325
|
-
name: "
|
|
5306
|
+
"openai/gpt-4o-2024-05-13": {
|
|
5307
|
+
id: "openai/gpt-4o-2024-05-13",
|
|
5308
|
+
name: "OpenAI: GPT-4o (2024-05-13)",
|
|
5326
5309
|
api: "openai-completions",
|
|
5327
5310
|
provider: "openrouter",
|
|
5328
5311
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
5329
5312
|
reasoning: false,
|
|
5330
|
-
input: ["text"],
|
|
5313
|
+
input: ["text", "image"],
|
|
5331
5314
|
cost: {
|
|
5332
|
-
input:
|
|
5333
|
-
output:
|
|
5315
|
+
input: 5,
|
|
5316
|
+
output: 15,
|
|
5334
5317
|
cacheRead: 0,
|
|
5335
5318
|
cacheWrite: 0,
|
|
5336
5319
|
},
|
|
5337
|
-
contextWindow:
|
|
5338
|
-
maxTokens:
|
|
5320
|
+
contextWindow: 128000,
|
|
5321
|
+
maxTokens: 4096,
|
|
5339
5322
|
},
|
|
5340
5323
|
"meta-llama/llama-3-8b-instruct": {
|
|
5341
5324
|
id: "meta-llama/llama-3-8b-instruct",
|
|
@@ -5354,6 +5337,23 @@ export const MODELS = {
|
|
|
5354
5337
|
contextWindow: 8192,
|
|
5355
5338
|
maxTokens: 16384,
|
|
5356
5339
|
},
|
|
5340
|
+
"meta-llama/llama-3-70b-instruct": {
|
|
5341
|
+
id: "meta-llama/llama-3-70b-instruct",
|
|
5342
|
+
name: "Meta: Llama 3 70B Instruct",
|
|
5343
|
+
api: "openai-completions",
|
|
5344
|
+
provider: "openrouter",
|
|
5345
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
5346
|
+
reasoning: false,
|
|
5347
|
+
input: ["text"],
|
|
5348
|
+
cost: {
|
|
5349
|
+
input: 0.3,
|
|
5350
|
+
output: 0.39999999999999997,
|
|
5351
|
+
cacheRead: 0,
|
|
5352
|
+
cacheWrite: 0,
|
|
5353
|
+
},
|
|
5354
|
+
contextWindow: 8192,
|
|
5355
|
+
maxTokens: 16384,
|
|
5356
|
+
},
|
|
5357
5357
|
"mistralai/mixtral-8x22b-instruct": {
|
|
5358
5358
|
id: "mistralai/mixtral-8x22b-instruct",
|
|
5359
5359
|
name: "Mistral: Mixtral 8x22B Instruct",
|