@mariozechner/pi-ai 0.18.8 → 0.19.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2770,7 +2770,7 @@ export const MODELS = {
2770
2770
  cacheWrite: 0,
2771
2771
  },
2772
2772
  contextWindow: 163840,
2773
- maxTokens: 65536,
2773
+ maxTokens: 163840,
2774
2774
  },
2775
2775
  "prime-intellect/intellect-3": {
2776
2776
  id: "prime-intellect/intellect-3",
@@ -3257,13 +3257,13 @@ export const MODELS = {
3257
3257
  reasoning: false,
3258
3258
  input: ["text", "image"],
3259
3259
  cost: {
3260
- input: 0.14,
3261
- output: 1,
3260
+ input: 0.15,
3261
+ output: 0.6,
3262
3262
  cacheRead: 0,
3263
3263
  cacheWrite: 0,
3264
3264
  },
3265
- contextWindow: 131072,
3266
- maxTokens: 131072,
3265
+ contextWindow: 262144,
3266
+ maxTokens: 4096,
3267
3267
  },
3268
3268
  "openai/gpt-5-pro": {
3269
3269
  id: "openai/gpt-5-pro",
@@ -5577,9 +5577,9 @@ export const MODELS = {
5577
5577
  contextWindow: 32768,
5578
5578
  maxTokens: 4096,
5579
5579
  },
5580
- "anthropic/claude-3.5-haiku": {
5581
- id: "anthropic/claude-3.5-haiku",
5582
- name: "Anthropic: Claude 3.5 Haiku",
5580
+ "anthropic/claude-3.5-haiku-20241022": {
5581
+ id: "anthropic/claude-3.5-haiku-20241022",
5582
+ name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
5583
5583
  api: "openai-completions",
5584
5584
  provider: "openrouter",
5585
5585
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5594,9 +5594,9 @@ export const MODELS = {
5594
5594
  contextWindow: 200000,
5595
5595
  maxTokens: 8192,
5596
5596
  },
5597
- "anthropic/claude-3.5-haiku-20241022": {
5598
- id: "anthropic/claude-3.5-haiku-20241022",
5599
- name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
5597
+ "anthropic/claude-3.5-haiku": {
5598
+ id: "anthropic/claude-3.5-haiku",
5599
+ name: "Anthropic: Claude 3.5 Haiku",
5600
5600
  api: "openai-completions",
5601
5601
  provider: "openrouter",
5602
5602
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5628,34 +5628,34 @@ export const MODELS = {
5628
5628
  contextWindow: 200000,
5629
5629
  maxTokens: 8192,
5630
5630
  },
5631
- "mistralai/ministral-3b": {
5632
- id: "mistralai/ministral-3b",
5633
- name: "Mistral: Ministral 3B",
5631
+ "mistralai/ministral-8b": {
5632
+ id: "mistralai/ministral-8b",
5633
+ name: "Mistral: Ministral 8B",
5634
5634
  api: "openai-completions",
5635
5635
  provider: "openrouter",
5636
5636
  baseUrl: "https://openrouter.ai/api/v1",
5637
5637
  reasoning: false,
5638
5638
  input: ["text"],
5639
5639
  cost: {
5640
- input: 0.04,
5641
- output: 0.04,
5640
+ input: 0.09999999999999999,
5641
+ output: 0.09999999999999999,
5642
5642
  cacheRead: 0,
5643
5643
  cacheWrite: 0,
5644
5644
  },
5645
5645
  contextWindow: 131072,
5646
5646
  maxTokens: 4096,
5647
5647
  },
5648
- "mistralai/ministral-8b": {
5649
- id: "mistralai/ministral-8b",
5650
- name: "Mistral: Ministral 8B",
5648
+ "mistralai/ministral-3b": {
5649
+ id: "mistralai/ministral-3b",
5650
+ name: "Mistral: Ministral 3B",
5651
5651
  api: "openai-completions",
5652
5652
  provider: "openrouter",
5653
5653
  baseUrl: "https://openrouter.ai/api/v1",
5654
5654
  reasoning: false,
5655
5655
  input: ["text"],
5656
5656
  cost: {
5657
- input: 0.09999999999999999,
5658
- output: 0.09999999999999999,
5657
+ input: 0.04,
5658
+ output: 0.04,
5659
5659
  cacheRead: 0,
5660
5660
  cacheWrite: 0,
5661
5661
  },
@@ -5832,39 +5832,39 @@ export const MODELS = {
5832
5832
  contextWindow: 128000,
5833
5833
  maxTokens: 16384,
5834
5834
  },
5835
- "meta-llama/llama-3.1-405b-instruct": {
5836
- id: "meta-llama/llama-3.1-405b-instruct",
5837
- name: "Meta: Llama 3.1 405B Instruct",
5835
+ "meta-llama/llama-3.1-8b-instruct": {
5836
+ id: "meta-llama/llama-3.1-8b-instruct",
5837
+ name: "Meta: Llama 3.1 8B Instruct",
5838
5838
  api: "openai-completions",
5839
5839
  provider: "openrouter",
5840
5840
  baseUrl: "https://openrouter.ai/api/v1",
5841
5841
  reasoning: false,
5842
5842
  input: ["text"],
5843
5843
  cost: {
5844
- input: 3.5,
5845
- output: 3.5,
5844
+ input: 0.02,
5845
+ output: 0.03,
5846
5846
  cacheRead: 0,
5847
5847
  cacheWrite: 0,
5848
5848
  },
5849
- contextWindow: 130815,
5850
- maxTokens: 4096,
5849
+ contextWindow: 131072,
5850
+ maxTokens: 16384,
5851
5851
  },
5852
- "meta-llama/llama-3.1-8b-instruct": {
5853
- id: "meta-llama/llama-3.1-8b-instruct",
5854
- name: "Meta: Llama 3.1 8B Instruct",
5852
+ "meta-llama/llama-3.1-405b-instruct": {
5853
+ id: "meta-llama/llama-3.1-405b-instruct",
5854
+ name: "Meta: Llama 3.1 405B Instruct",
5855
5855
  api: "openai-completions",
5856
5856
  provider: "openrouter",
5857
5857
  baseUrl: "https://openrouter.ai/api/v1",
5858
5858
  reasoning: false,
5859
5859
  input: ["text"],
5860
5860
  cost: {
5861
- input: 0.02,
5862
- output: 0.03,
5861
+ input: 3.5,
5862
+ output: 3.5,
5863
5863
  cacheRead: 0,
5864
5864
  cacheWrite: 0,
5865
5865
  },
5866
- contextWindow: 131072,
5867
- maxTokens: 16384,
5866
+ contextWindow: 130815,
5867
+ maxTokens: 4096,
5868
5868
  },
5869
5869
  "meta-llama/llama-3.1-70b-instruct": {
5870
5870
  id: "meta-llama/llama-3.1-70b-instruct",
@@ -5900,9 +5900,9 @@ export const MODELS = {
5900
5900
  contextWindow: 131072,
5901
5901
  maxTokens: 16384,
5902
5902
  },
5903
- "openai/gpt-4o-mini": {
5904
- id: "openai/gpt-4o-mini",
5905
- name: "OpenAI: GPT-4o-mini",
5903
+ "openai/gpt-4o-mini-2024-07-18": {
5904
+ id: "openai/gpt-4o-mini-2024-07-18",
5905
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
5906
5906
  api: "openai-completions",
5907
5907
  provider: "openrouter",
5908
5908
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5917,9 +5917,9 @@ export const MODELS = {
5917
5917
  contextWindow: 128000,
5918
5918
  maxTokens: 16384,
5919
5919
  },
5920
- "openai/gpt-4o-mini-2024-07-18": {
5921
- id: "openai/gpt-4o-mini-2024-07-18",
5922
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
5920
+ "openai/gpt-4o-mini": {
5921
+ id: "openai/gpt-4o-mini",
5922
+ name: "OpenAI: GPT-4o-mini",
5923
5923
  api: "openai-completions",
5924
5924
  provider: "openrouter",
5925
5925
  baseUrl: "https://openrouter.ai/api/v1",
@@ -6070,34 +6070,34 @@ export const MODELS = {
6070
6070
  contextWindow: 128000,
6071
6071
  maxTokens: 64000,
6072
6072
  },
6073
- "meta-llama/llama-3-8b-instruct": {
6074
- id: "meta-llama/llama-3-8b-instruct",
6075
- name: "Meta: Llama 3 8B Instruct",
6073
+ "meta-llama/llama-3-70b-instruct": {
6074
+ id: "meta-llama/llama-3-70b-instruct",
6075
+ name: "Meta: Llama 3 70B Instruct",
6076
6076
  api: "openai-completions",
6077
6077
  provider: "openrouter",
6078
6078
  baseUrl: "https://openrouter.ai/api/v1",
6079
6079
  reasoning: false,
6080
6080
  input: ["text"],
6081
6081
  cost: {
6082
- input: 0.03,
6083
- output: 0.06,
6082
+ input: 0.3,
6083
+ output: 0.39999999999999997,
6084
6084
  cacheRead: 0,
6085
6085
  cacheWrite: 0,
6086
6086
  },
6087
6087
  contextWindow: 8192,
6088
6088
  maxTokens: 16384,
6089
6089
  },
6090
- "meta-llama/llama-3-70b-instruct": {
6091
- id: "meta-llama/llama-3-70b-instruct",
6092
- name: "Meta: Llama 3 70B Instruct",
6090
+ "meta-llama/llama-3-8b-instruct": {
6091
+ id: "meta-llama/llama-3-8b-instruct",
6092
+ name: "Meta: Llama 3 8B Instruct",
6093
6093
  api: "openai-completions",
6094
6094
  provider: "openrouter",
6095
6095
  baseUrl: "https://openrouter.ai/api/v1",
6096
6096
  reasoning: false,
6097
6097
  input: ["text"],
6098
6098
  cost: {
6099
- input: 0.3,
6100
- output: 0.39999999999999997,
6099
+ input: 0.03,
6100
+ output: 0.06,
6101
6101
  cacheRead: 0,
6102
6102
  cacheWrite: 0,
6103
6103
  },
@@ -6291,26 +6291,26 @@ export const MODELS = {
6291
6291
  contextWindow: 16385,
6292
6292
  maxTokens: 4096,
6293
6293
  },
6294
- "openai/gpt-3.5-turbo": {
6295
- id: "openai/gpt-3.5-turbo",
6296
- name: "OpenAI: GPT-3.5 Turbo",
6294
+ "openai/gpt-4-0314": {
6295
+ id: "openai/gpt-4-0314",
6296
+ name: "OpenAI: GPT-4 (older v0314)",
6297
6297
  api: "openai-completions",
6298
6298
  provider: "openrouter",
6299
6299
  baseUrl: "https://openrouter.ai/api/v1",
6300
6300
  reasoning: false,
6301
6301
  input: ["text"],
6302
6302
  cost: {
6303
- input: 0.5,
6304
- output: 1.5,
6303
+ input: 30,
6304
+ output: 60,
6305
6305
  cacheRead: 0,
6306
6306
  cacheWrite: 0,
6307
6307
  },
6308
- contextWindow: 16385,
6308
+ contextWindow: 8191,
6309
6309
  maxTokens: 4096,
6310
6310
  },
6311
- "openai/gpt-4-0314": {
6312
- id: "openai/gpt-4-0314",
6313
- name: "OpenAI: GPT-4 (older v0314)",
6311
+ "openai/gpt-4": {
6312
+ id: "openai/gpt-4",
6313
+ name: "OpenAI: GPT-4",
6314
6314
  api: "openai-completions",
6315
6315
  provider: "openrouter",
6316
6316
  baseUrl: "https://openrouter.ai/api/v1",
@@ -6325,21 +6325,21 @@ export const MODELS = {
6325
6325
  contextWindow: 8191,
6326
6326
  maxTokens: 4096,
6327
6327
  },
6328
- "openai/gpt-4": {
6329
- id: "openai/gpt-4",
6330
- name: "OpenAI: GPT-4",
6328
+ "openai/gpt-3.5-turbo": {
6329
+ id: "openai/gpt-3.5-turbo",
6330
+ name: "OpenAI: GPT-3.5 Turbo",
6331
6331
  api: "openai-completions",
6332
6332
  provider: "openrouter",
6333
6333
  baseUrl: "https://openrouter.ai/api/v1",
6334
6334
  reasoning: false,
6335
6335
  input: ["text"],
6336
6336
  cost: {
6337
- input: 30,
6338
- output: 60,
6337
+ input: 0.5,
6338
+ output: 1.5,
6339
6339
  cacheRead: 0,
6340
6340
  cacheWrite: 0,
6341
6341
  },
6342
- contextWindow: 8191,
6342
+ contextWindow: 16385,
6343
6343
  maxTokens: 4096,
6344
6344
  },
6345
6345
  "openrouter/auto": {