@mariozechner/pi-ai 0.18.8 → 0.19.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2770,7 +2770,7 @@ export const MODELS = {
2770
2770
  cacheWrite: 0,
2771
2771
  },
2772
2772
  contextWindow: 163840,
2773
- maxTokens: 65536,
2773
+ maxTokens: 163840,
2774
2774
  },
2775
2775
  "prime-intellect/intellect-3": {
2776
2776
  id: "prime-intellect/intellect-3",
@@ -3257,13 +3257,13 @@ export const MODELS = {
3257
3257
  reasoning: false,
3258
3258
  input: ["text", "image"],
3259
3259
  cost: {
3260
- input: 0.14,
3261
- output: 1,
3260
+ input: 0.15,
3261
+ output: 0.6,
3262
3262
  cacheRead: 0,
3263
3263
  cacheWrite: 0,
3264
3264
  },
3265
- contextWindow: 131072,
3266
- maxTokens: 131072,
3265
+ contextWindow: 262144,
3266
+ maxTokens: 4096,
3267
3267
  },
3268
3268
  "openai/gpt-5-pro": {
3269
3269
  id: "openai/gpt-5-pro",
@@ -5577,9 +5577,9 @@ export const MODELS = {
5577
5577
  contextWindow: 32768,
5578
5578
  maxTokens: 4096,
5579
5579
  },
5580
- "anthropic/claude-3.5-haiku": {
5581
- id: "anthropic/claude-3.5-haiku",
5582
- name: "Anthropic: Claude 3.5 Haiku",
5580
+ "anthropic/claude-3.5-haiku-20241022": {
5581
+ id: "anthropic/claude-3.5-haiku-20241022",
5582
+ name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
5583
5583
  api: "openai-completions",
5584
5584
  provider: "openrouter",
5585
5585
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5594,9 +5594,9 @@ export const MODELS = {
5594
5594
  contextWindow: 200000,
5595
5595
  maxTokens: 8192,
5596
5596
  },
5597
- "anthropic/claude-3.5-haiku-20241022": {
5598
- id: "anthropic/claude-3.5-haiku-20241022",
5599
- name: "Anthropic: Claude 3.5 Haiku (2024-10-22)",
5597
+ "anthropic/claude-3.5-haiku": {
5598
+ id: "anthropic/claude-3.5-haiku",
5599
+ name: "Anthropic: Claude 3.5 Haiku",
5600
5600
  api: "openai-completions",
5601
5601
  provider: "openrouter",
5602
5602
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5628,34 +5628,34 @@ export const MODELS = {
5628
5628
  contextWindow: 200000,
5629
5629
  maxTokens: 8192,
5630
5630
  },
5631
- "mistralai/ministral-3b": {
5632
- id: "mistralai/ministral-3b",
5633
- name: "Mistral: Ministral 3B",
5631
+ "mistralai/ministral-8b": {
5632
+ id: "mistralai/ministral-8b",
5633
+ name: "Mistral: Ministral 8B",
5634
5634
  api: "openai-completions",
5635
5635
  provider: "openrouter",
5636
5636
  baseUrl: "https://openrouter.ai/api/v1",
5637
5637
  reasoning: false,
5638
5638
  input: ["text"],
5639
5639
  cost: {
5640
- input: 0.04,
5641
- output: 0.04,
5640
+ input: 0.09999999999999999,
5641
+ output: 0.09999999999999999,
5642
5642
  cacheRead: 0,
5643
5643
  cacheWrite: 0,
5644
5644
  },
5645
5645
  contextWindow: 131072,
5646
5646
  maxTokens: 4096,
5647
5647
  },
5648
- "mistralai/ministral-8b": {
5649
- id: "mistralai/ministral-8b",
5650
- name: "Mistral: Ministral 8B",
5648
+ "mistralai/ministral-3b": {
5649
+ id: "mistralai/ministral-3b",
5650
+ name: "Mistral: Ministral 3B",
5651
5651
  api: "openai-completions",
5652
5652
  provider: "openrouter",
5653
5653
  baseUrl: "https://openrouter.ai/api/v1",
5654
5654
  reasoning: false,
5655
5655
  input: ["text"],
5656
5656
  cost: {
5657
- input: 0.09999999999999999,
5658
- output: 0.09999999999999999,
5657
+ input: 0.04,
5658
+ output: 0.04,
5659
5659
  cacheRead: 0,
5660
5660
  cacheWrite: 0,
5661
5661
  },
@@ -5747,34 +5747,34 @@ export const MODELS = {
5747
5747
  contextWindow: 32768,
5748
5748
  maxTokens: 4096,
5749
5749
  },
5750
- "cohere/command-r-08-2024": {
5751
- id: "cohere/command-r-08-2024",
5752
- name: "Cohere: Command R (08-2024)",
5750
+ "cohere/command-r-plus-08-2024": {
5751
+ id: "cohere/command-r-plus-08-2024",
5752
+ name: "Cohere: Command R+ (08-2024)",
5753
5753
  api: "openai-completions",
5754
5754
  provider: "openrouter",
5755
5755
  baseUrl: "https://openrouter.ai/api/v1",
5756
5756
  reasoning: false,
5757
5757
  input: ["text"],
5758
5758
  cost: {
5759
- input: 0.15,
5760
- output: 0.6,
5759
+ input: 2.5,
5760
+ output: 10,
5761
5761
  cacheRead: 0,
5762
5762
  cacheWrite: 0,
5763
5763
  },
5764
5764
  contextWindow: 128000,
5765
5765
  maxTokens: 4000,
5766
5766
  },
5767
- "cohere/command-r-plus-08-2024": {
5768
- id: "cohere/command-r-plus-08-2024",
5769
- name: "Cohere: Command R+ (08-2024)",
5767
+ "cohere/command-r-08-2024": {
5768
+ id: "cohere/command-r-08-2024",
5769
+ name: "Cohere: Command R (08-2024)",
5770
5770
  api: "openai-completions",
5771
5771
  provider: "openrouter",
5772
5772
  baseUrl: "https://openrouter.ai/api/v1",
5773
5773
  reasoning: false,
5774
5774
  input: ["text"],
5775
5775
  cost: {
5776
- input: 2.5,
5777
- output: 10,
5776
+ input: 0.15,
5777
+ output: 0.6,
5778
5778
  cacheRead: 0,
5779
5779
  cacheWrite: 0,
5780
5780
  },
@@ -5832,23 +5832,6 @@ export const MODELS = {
5832
5832
  contextWindow: 128000,
5833
5833
  maxTokens: 16384,
5834
5834
  },
5835
- "meta-llama/llama-3.1-405b-instruct": {
5836
- id: "meta-llama/llama-3.1-405b-instruct",
5837
- name: "Meta: Llama 3.1 405B Instruct",
5838
- api: "openai-completions",
5839
- provider: "openrouter",
5840
- baseUrl: "https://openrouter.ai/api/v1",
5841
- reasoning: false,
5842
- input: ["text"],
5843
- cost: {
5844
- input: 3.5,
5845
- output: 3.5,
5846
- cacheRead: 0,
5847
- cacheWrite: 0,
5848
- },
5849
- contextWindow: 130815,
5850
- maxTokens: 4096,
5851
- },
5852
5835
  "meta-llama/llama-3.1-8b-instruct": {
5853
5836
  id: "meta-llama/llama-3.1-8b-instruct",
5854
5837
  name: "Meta: Llama 3.1 8B Instruct",
@@ -5883,6 +5866,23 @@ export const MODELS = {
5883
5866
  contextWindow: 131072,
5884
5867
  maxTokens: 4096,
5885
5868
  },
5869
+ "meta-llama/llama-3.1-405b-instruct": {
5870
+ id: "meta-llama/llama-3.1-405b-instruct",
5871
+ name: "Meta: Llama 3.1 405B Instruct",
5872
+ api: "openai-completions",
5873
+ provider: "openrouter",
5874
+ baseUrl: "https://openrouter.ai/api/v1",
5875
+ reasoning: false,
5876
+ input: ["text"],
5877
+ cost: {
5878
+ input: 3.5,
5879
+ output: 3.5,
5880
+ cacheRead: 0,
5881
+ cacheWrite: 0,
5882
+ },
5883
+ contextWindow: 130815,
5884
+ maxTokens: 4096,
5885
+ },
5886
5886
  "mistralai/mistral-nemo": {
5887
5887
  id: "mistralai/mistral-nemo",
5888
5888
  name: "Mistral: Mistral Nemo",
@@ -6070,34 +6070,34 @@ export const MODELS = {
6070
6070
  contextWindow: 128000,
6071
6071
  maxTokens: 64000,
6072
6072
  },
6073
- "meta-llama/llama-3-8b-instruct": {
6074
- id: "meta-llama/llama-3-8b-instruct",
6075
- name: "Meta: Llama 3 8B Instruct",
6073
+ "meta-llama/llama-3-70b-instruct": {
6074
+ id: "meta-llama/llama-3-70b-instruct",
6075
+ name: "Meta: Llama 3 70B Instruct",
6076
6076
  api: "openai-completions",
6077
6077
  provider: "openrouter",
6078
6078
  baseUrl: "https://openrouter.ai/api/v1",
6079
6079
  reasoning: false,
6080
6080
  input: ["text"],
6081
6081
  cost: {
6082
- input: 0.03,
6083
- output: 0.06,
6082
+ input: 0.3,
6083
+ output: 0.39999999999999997,
6084
6084
  cacheRead: 0,
6085
6085
  cacheWrite: 0,
6086
6086
  },
6087
6087
  contextWindow: 8192,
6088
6088
  maxTokens: 16384,
6089
6089
  },
6090
- "meta-llama/llama-3-70b-instruct": {
6091
- id: "meta-llama/llama-3-70b-instruct",
6092
- name: "Meta: Llama 3 70B Instruct",
6090
+ "meta-llama/llama-3-8b-instruct": {
6091
+ id: "meta-llama/llama-3-8b-instruct",
6092
+ name: "Meta: Llama 3 8B Instruct",
6093
6093
  api: "openai-completions",
6094
6094
  provider: "openrouter",
6095
6095
  baseUrl: "https://openrouter.ai/api/v1",
6096
6096
  reasoning: false,
6097
6097
  input: ["text"],
6098
6098
  cost: {
6099
- input: 0.3,
6100
- output: 0.39999999999999997,
6099
+ input: 0.03,
6100
+ output: 0.06,
6101
6101
  cacheRead: 0,
6102
6102
  cacheWrite: 0,
6103
6103
  },
@@ -6291,43 +6291,43 @@ export const MODELS = {
6291
6291
  contextWindow: 16385,
6292
6292
  maxTokens: 4096,
6293
6293
  },
6294
- "openai/gpt-3.5-turbo": {
6295
- id: "openai/gpt-3.5-turbo",
6296
- name: "OpenAI: GPT-3.5 Turbo",
6294
+ "openai/gpt-4": {
6295
+ id: "openai/gpt-4",
6296
+ name: "OpenAI: GPT-4",
6297
6297
  api: "openai-completions",
6298
6298
  provider: "openrouter",
6299
6299
  baseUrl: "https://openrouter.ai/api/v1",
6300
6300
  reasoning: false,
6301
6301
  input: ["text"],
6302
6302
  cost: {
6303
- input: 0.5,
6304
- output: 1.5,
6303
+ input: 30,
6304
+ output: 60,
6305
6305
  cacheRead: 0,
6306
6306
  cacheWrite: 0,
6307
6307
  },
6308
- contextWindow: 16385,
6308
+ contextWindow: 8191,
6309
6309
  maxTokens: 4096,
6310
6310
  },
6311
- "openai/gpt-4-0314": {
6312
- id: "openai/gpt-4-0314",
6313
- name: "OpenAI: GPT-4 (older v0314)",
6311
+ "openai/gpt-3.5-turbo": {
6312
+ id: "openai/gpt-3.5-turbo",
6313
+ name: "OpenAI: GPT-3.5 Turbo",
6314
6314
  api: "openai-completions",
6315
6315
  provider: "openrouter",
6316
6316
  baseUrl: "https://openrouter.ai/api/v1",
6317
6317
  reasoning: false,
6318
6318
  input: ["text"],
6319
6319
  cost: {
6320
- input: 30,
6321
- output: 60,
6320
+ input: 0.5,
6321
+ output: 1.5,
6322
6322
  cacheRead: 0,
6323
6323
  cacheWrite: 0,
6324
6324
  },
6325
- contextWindow: 8191,
6325
+ contextWindow: 16385,
6326
6326
  maxTokens: 4096,
6327
6327
  },
6328
- "openai/gpt-4": {
6329
- id: "openai/gpt-4",
6330
- name: "OpenAI: GPT-4",
6328
+ "openai/gpt-4-0314": {
6329
+ id: "openai/gpt-4-0314",
6330
+ name: "OpenAI: GPT-4 (older v0314)",
6331
6331
  api: "openai-completions",
6332
6332
  provider: "openrouter",
6333
6333
  baseUrl: "https://openrouter.ai/api/v1",