@mariozechner/pi-ai 0.9.2 → 0.9.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -104,6 +104,23 @@ export const MODELS = {
104
104
  contextWindow: 200000,
105
105
  maxTokens: 8192,
106
106
  },
107
+ "claude-opus-4-5": {
108
+ id: "claude-opus-4-5",
109
+ name: "Claude Opus 4.5 (latest)",
110
+ api: "anthropic-messages",
111
+ provider: "anthropic",
112
+ baseUrl: "https://api.anthropic.com",
113
+ reasoning: true,
114
+ input: ["text", "image"],
115
+ cost: {
116
+ input: 5,
117
+ output: 25,
118
+ cacheRead: 1.5,
119
+ cacheWrite: 18.75,
120
+ },
121
+ contextWindow: 200000,
122
+ maxTokens: 64000,
123
+ },
107
124
  "claude-3-opus-20240229": {
108
125
  id: "claude-3-opus-20240229",
109
126
  name: "Claude Opus 3",
@@ -1937,6 +1954,23 @@ export const MODELS = {
1937
1954
  },
1938
1955
  },
1939
1956
  openrouter: {
1957
+ "anthropic/claude-opus-4.5": {
1958
+ id: "anthropic/claude-opus-4.5",
1959
+ name: "Anthropic: Claude Opus 4.5",
1960
+ api: "openai-completions",
1961
+ provider: "openrouter",
1962
+ baseUrl: "https://openrouter.ai/api/v1",
1963
+ reasoning: true,
1964
+ input: ["text", "image"],
1965
+ cost: {
1966
+ input: 5,
1967
+ output: 25,
1968
+ cacheRead: 0.5,
1969
+ cacheWrite: 6.25,
1970
+ },
1971
+ contextWindow: 200000,
1972
+ maxTokens: 32000,
1973
+ },
1940
1974
  "allenai/olmo-3-7b-instruct": {
1941
1975
  id: "allenai/olmo-3-7b-instruct",
1942
1976
  name: "AllenAI: Olmo 3 7B Instruct",
@@ -1954,23 +1988,6 @@ export const MODELS = {
1954
1988
  contextWindow: 65536,
1955
1989
  maxTokens: 65536,
1956
1990
  },
1957
- "x-ai/grok-4.1-fast": {
1958
- id: "x-ai/grok-4.1-fast",
1959
- name: "xAI: Grok 4.1 Fast",
1960
- api: "openai-completions",
1961
- provider: "openrouter",
1962
- baseUrl: "https://openrouter.ai/api/v1",
1963
- reasoning: true,
1964
- input: ["text", "image"],
1965
- cost: {
1966
- input: 0,
1967
- output: 0,
1968
- cacheRead: 0,
1969
- cacheWrite: 0,
1970
- },
1971
- contextWindow: 2000000,
1972
- maxTokens: 30000,
1973
- },
1974
1991
  "x-ai/grok-4.1-fast:free": {
1975
1992
  id: "x-ai/grok-4.1-fast:free",
1976
1993
  name: "xAI: Grok 4.1 Fast (free)",
@@ -2088,7 +2105,7 @@ export const MODELS = {
2088
2105
  cacheWrite: 0,
2089
2106
  },
2090
2107
  contextWindow: 256000,
2091
- maxTokens: 32000,
2108
+ maxTokens: 32768,
2092
2109
  },
2093
2110
  "moonshotai/kimi-k2-thinking": {
2094
2111
  id: "moonshotai/kimi-k2-thinking",
@@ -3408,13 +3425,13 @@ export const MODELS = {
3408
3425
  reasoning: true,
3409
3426
  input: ["text"],
3410
3427
  cost: {
3411
- input: 1,
3412
- output: 4,
3428
+ input: 0.38,
3429
+ output: 1.53,
3413
3430
  cacheRead: 0,
3414
3431
  cacheWrite: 0,
3415
3432
  },
3416
3433
  contextWindow: 262144,
3417
- maxTokens: 32768,
3434
+ maxTokens: 262144,
3418
3435
  },
3419
3436
  "google/gemini-2.5-flash-lite": {
3420
3437
  id: "google/gemini-2.5-flash-lite",
@@ -3850,8 +3867,8 @@ export const MODELS = {
3850
3867
  reasoning: true,
3851
3868
  input: ["text"],
3852
3869
  cost: {
3853
- input: 0.15,
3854
- output: 0.59,
3870
+ input: 0.049999999999999996,
3871
+ output: 0.19999999999999998,
3855
3872
  cacheRead: 0,
3856
3873
  cacheWrite: 0,
3857
3874
  },
@@ -4793,9 +4810,9 @@ export const MODELS = {
4793
4810
  contextWindow: 32768,
4794
4811
  maxTokens: 4096,
4795
4812
  },
4796
- "anthropic/claude-3.5-haiku-20241022": {
4797
- id: "anthropic/claude-3.5-haiku-20241022",
4798
- name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
4813
+ "anthropic/claude-3.5-haiku": {
4814
+ id: "anthropic/claude-3.5-haiku",
4815
+ name: "Anthropic: Claude 3.5 Haiku",
4799
4816
  api: "openai-completions",
4800
4817
  provider: "openrouter",
4801
4818
  baseUrl: "https://openrouter.ai/api/v1",
@@ -4810,9 +4827,9 @@ export const MODELS = {
4810
4827
  contextWindow: 200000,
4811
4828
  maxTokens: 8192,
4812
4829
  },
4813
- "anthropic/claude-3.5-haiku": {
4814
- id: "anthropic/claude-3.5-haiku",
4815
- name: "Anthropic: Claude 3.5 Haiku",
4830
+ "anthropic/claude-3.5-haiku-20241022": {
4831
+ id: "anthropic/claude-3.5-haiku-20241022",
4832
+ name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
4816
4833
  api: "openai-completions",
4817
4834
  provider: "openrouter",
4818
4835
  baseUrl: "https://openrouter.ai/api/v1",
@@ -4878,23 +4895,6 @@ export const MODELS = {
4878
4895
  contextWindow: 131072,
4879
4896
  maxTokens: 4096,
4880
4897
  },
4881
- "qwen/qwen-2.5-7b-instruct": {
4882
- id: "qwen/qwen-2.5-7b-instruct",
4883
- name: "Qwen: Qwen2.5 7B Instruct",
4884
- api: "openai-completions",
4885
- provider: "openrouter",
4886
- baseUrl: "https://openrouter.ai/api/v1",
4887
- reasoning: false,
4888
- input: ["text"],
4889
- cost: {
4890
- input: 0.04,
4891
- output: 0.09999999999999999,
4892
- cacheRead: 0,
4893
- cacheWrite: 0,
4894
- },
4895
- contextWindow: 32768,
4896
- maxTokens: 4096,
4897
- },
4898
4898
  "nvidia/llama-3.1-nemotron-70b-instruct": {
4899
4899
  id: "nvidia/llama-3.1-nemotron-70b-instruct",
4900
4900
  name: "NVIDIA: Llama 3.1 Nemotron 70B Instruct",
@@ -5082,23 +5082,6 @@ export const MODELS = {
5082
5082
  contextWindow: 128000,
5083
5083
  maxTokens: 16384,
5084
5084
  },
5085
- "meta-llama/llama-3.1-70b-instruct": {
5086
- id: "meta-llama/llama-3.1-70b-instruct",
5087
- name: "Meta: Llama 3.1 70B Instruct",
5088
- api: "openai-completions",
5089
- provider: "openrouter",
5090
- baseUrl: "https://openrouter.ai/api/v1",
5091
- reasoning: false,
5092
- input: ["text"],
5093
- cost: {
5094
- input: 0.39999999999999997,
5095
- output: 0.39999999999999997,
5096
- cacheRead: 0,
5097
- cacheWrite: 0,
5098
- },
5099
- contextWindow: 131072,
5100
- maxTokens: 4096,
5101
- },
5102
5085
  "meta-llama/llama-3.1-8b-instruct": {
5103
5086
  id: "meta-llama/llama-3.1-8b-instruct",
5104
5087
  name: "Meta: Llama 3.1 8B Instruct",
@@ -5133,6 +5116,23 @@ export const MODELS = {
5133
5116
  contextWindow: 130815,
5134
5117
  maxTokens: 4096,
5135
5118
  },
5119
+ "meta-llama/llama-3.1-70b-instruct": {
5120
+ id: "meta-llama/llama-3.1-70b-instruct",
5121
+ name: "Meta: Llama 3.1 70B Instruct",
5122
+ api: "openai-completions",
5123
+ provider: "openrouter",
5124
+ baseUrl: "https://openrouter.ai/api/v1",
5125
+ reasoning: false,
5126
+ input: ["text"],
5127
+ cost: {
5128
+ input: 0.39999999999999997,
5129
+ output: 0.39999999999999997,
5130
+ cacheRead: 0,
5131
+ cacheWrite: 0,
5132
+ },
5133
+ contextWindow: 131072,
5134
+ maxTokens: 4096,
5135
+ },
5136
5136
  "mistralai/mistral-nemo": {
5137
5137
  id: "mistralai/mistral-nemo",
5138
5138
  name: "Mistral: Mistral Nemo",
@@ -5150,9 +5150,9 @@ export const MODELS = {
5150
5150
  contextWindow: 131072,
5151
5151
  maxTokens: 16384,
5152
5152
  },
5153
- "openai/gpt-4o-mini-2024-07-18": {
5154
- id: "openai/gpt-4o-mini-2024-07-18",
5155
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
5153
+ "openai/gpt-4o-mini": {
5154
+ id: "openai/gpt-4o-mini",
5155
+ name: "OpenAI: GPT-4o-mini",
5156
5156
  api: "openai-completions",
5157
5157
  provider: "openrouter",
5158
5158
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5167,9 +5167,9 @@ export const MODELS = {
5167
5167
  contextWindow: 128000,
5168
5168
  maxTokens: 16384,
5169
5169
  },
5170
- "openai/gpt-4o-mini": {
5171
- id: "openai/gpt-4o-mini",
5172
- name: "OpenAI: GPT-4o-mini",
5170
+ "openai/gpt-4o-mini-2024-07-18": {
5171
+ id: "openai/gpt-4o-mini-2024-07-18",
5172
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
5173
5173
  api: "openai-completions",
5174
5174
  provider: "openrouter",
5175
5175
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5269,23 +5269,6 @@ export const MODELS = {
5269
5269
  contextWindow: 128000,
5270
5270
  maxTokens: 4096,
5271
5271
  },
5272
- "openai/gpt-4o-2024-05-13": {
5273
- id: "openai/gpt-4o-2024-05-13",
5274
- name: "OpenAI: GPT-4o (2024-05-13)",
5275
- api: "openai-completions",
5276
- provider: "openrouter",
5277
- baseUrl: "https://openrouter.ai/api/v1",
5278
- reasoning: false,
5279
- input: ["text", "image"],
5280
- cost: {
5281
- input: 5,
5282
- output: 15,
5283
- cacheRead: 0,
5284
- cacheWrite: 0,
5285
- },
5286
- contextWindow: 128000,
5287
- maxTokens: 4096,
5288
- },
5289
5272
  "openai/gpt-4o": {
5290
5273
  id: "openai/gpt-4o",
5291
5274
  name: "OpenAI: GPT-4o",
@@ -5320,22 +5303,22 @@ export const MODELS = {
5320
5303
  contextWindow: 128000,
5321
5304
  maxTokens: 64000,
5322
5305
  },
5323
- "meta-llama/llama-3-70b-instruct": {
5324
- id: "meta-llama/llama-3-70b-instruct",
5325
- name: "Meta: Llama 3 70B Instruct",
5306
+ "openai/gpt-4o-2024-05-13": {
5307
+ id: "openai/gpt-4o-2024-05-13",
5308
+ name: "OpenAI: GPT-4o (2024-05-13)",
5326
5309
  api: "openai-completions",
5327
5310
  provider: "openrouter",
5328
5311
  baseUrl: "https://openrouter.ai/api/v1",
5329
5312
  reasoning: false,
5330
- input: ["text"],
5313
+ input: ["text", "image"],
5331
5314
  cost: {
5332
- input: 0.3,
5333
- output: 0.39999999999999997,
5315
+ input: 5,
5316
+ output: 15,
5334
5317
  cacheRead: 0,
5335
5318
  cacheWrite: 0,
5336
5319
  },
5337
- contextWindow: 8192,
5338
- maxTokens: 16384,
5320
+ contextWindow: 128000,
5321
+ maxTokens: 4096,
5339
5322
  },
5340
5323
  "meta-llama/llama-3-8b-instruct": {
5341
5324
  id: "meta-llama/llama-3-8b-instruct",
@@ -5354,6 +5337,23 @@ export const MODELS = {
5354
5337
  contextWindow: 8192,
5355
5338
  maxTokens: 16384,
5356
5339
  },
5340
+ "meta-llama/llama-3-70b-instruct": {
5341
+ id: "meta-llama/llama-3-70b-instruct",
5342
+ name: "Meta: Llama 3 70B Instruct",
5343
+ api: "openai-completions",
5344
+ provider: "openrouter",
5345
+ baseUrl: "https://openrouter.ai/api/v1",
5346
+ reasoning: false,
5347
+ input: ["text"],
5348
+ cost: {
5349
+ input: 0.3,
5350
+ output: 0.39999999999999997,
5351
+ cacheRead: 0,
5352
+ cacheWrite: 0,
5353
+ },
5354
+ contextWindow: 8192,
5355
+ maxTokens: 16384,
5356
+ },
5357
5357
  "mistralai/mixtral-8x22b-instruct": {
5358
5358
  id: "mistralai/mixtral-8x22b-instruct",
5359
5359
  name: "Mistral: Mixtral 8x22B Instruct",