@mariozechner/pi-ai 0.7.1 → 0.7.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -4725,34 +4725,34 @@ export const MODELS = {
4725
4725
  contextWindow: 200000,
4726
4726
  maxTokens: 8192,
4727
4727
  },
4728
- "mistralai/ministral-8b": {
4729
- id: "mistralai/ministral-8b",
4730
- name: "Mistral: Ministral 8B",
4728
+ "mistralai/ministral-3b": {
4729
+ id: "mistralai/ministral-3b",
4730
+ name: "Mistral: Ministral 3B",
4731
4731
  api: "openai-completions",
4732
4732
  provider: "openrouter",
4733
4733
  baseUrl: "https://openrouter.ai/api/v1",
4734
4734
  reasoning: false,
4735
4735
  input: ["text"],
4736
4736
  cost: {
4737
- input: 0.09999999999999999,
4738
- output: 0.09999999999999999,
4737
+ input: 0.04,
4738
+ output: 0.04,
4739
4739
  cacheRead: 0,
4740
4740
  cacheWrite: 0,
4741
4741
  },
4742
4742
  contextWindow: 131072,
4743
4743
  maxTokens: 4096,
4744
4744
  },
4745
- "mistralai/ministral-3b": {
4746
- id: "mistralai/ministral-3b",
4747
- name: "Mistral: Ministral 3B",
4745
+ "mistralai/ministral-8b": {
4746
+ id: "mistralai/ministral-8b",
4747
+ name: "Mistral: Ministral 8B",
4748
4748
  api: "openai-completions",
4749
4749
  provider: "openrouter",
4750
4750
  baseUrl: "https://openrouter.ai/api/v1",
4751
4751
  reasoning: false,
4752
4752
  input: ["text"],
4753
4753
  cost: {
4754
- input: 0.04,
4755
- output: 0.04,
4754
+ input: 0.09999999999999999,
4755
+ output: 0.09999999999999999,
4756
4756
  cacheRead: 0,
4757
4757
  cacheWrite: 0,
4758
4758
  },
@@ -4963,23 +4963,6 @@ export const MODELS = {
4963
4963
  contextWindow: 128000,
4964
4964
  maxTokens: 16384,
4965
4965
  },
4966
- "meta-llama/llama-3.1-8b-instruct": {
4967
- id: "meta-llama/llama-3.1-8b-instruct",
4968
- name: "Meta: Llama 3.1 8B Instruct",
4969
- api: "openai-completions",
4970
- provider: "openrouter",
4971
- baseUrl: "https://openrouter.ai/api/v1",
4972
- reasoning: false,
4973
- input: ["text"],
4974
- cost: {
4975
- input: 0.02,
4976
- output: 0.03,
4977
- cacheRead: 0,
4978
- cacheWrite: 0,
4979
- },
4980
- contextWindow: 131072,
4981
- maxTokens: 16384,
4982
- },
4983
4966
  "meta-llama/llama-3.1-405b-instruct": {
4984
4967
  id: "meta-llama/llama-3.1-405b-instruct",
4985
4968
  name: "Meta: Llama 3.1 405B Instruct",
@@ -5014,6 +4997,23 @@ export const MODELS = {
5014
4997
  contextWindow: 131072,
5015
4998
  maxTokens: 4096,
5016
4999
  },
5000
+ "meta-llama/llama-3.1-8b-instruct": {
5001
+ id: "meta-llama/llama-3.1-8b-instruct",
5002
+ name: "Meta: Llama 3.1 8B Instruct",
5003
+ api: "openai-completions",
5004
+ provider: "openrouter",
5005
+ baseUrl: "https://openrouter.ai/api/v1",
5006
+ reasoning: false,
5007
+ input: ["text"],
5008
+ cost: {
5009
+ input: 0.02,
5010
+ output: 0.03,
5011
+ cacheRead: 0,
5012
+ cacheWrite: 0,
5013
+ },
5014
+ contextWindow: 131072,
5015
+ maxTokens: 16384,
5016
+ },
5017
5017
  "mistralai/mistral-nemo": {
5018
5018
  id: "mistralai/mistral-nemo",
5019
5019
  name: "Mistral: Mistral Nemo",
@@ -5031,9 +5031,9 @@ export const MODELS = {
5031
5031
  contextWindow: 131072,
5032
5032
  maxTokens: 16384,
5033
5033
  },
5034
- "openai/gpt-4o-mini-2024-07-18": {
5035
- id: "openai/gpt-4o-mini-2024-07-18",
5036
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
5034
+ "openai/gpt-4o-mini": {
5035
+ id: "openai/gpt-4o-mini",
5036
+ name: "OpenAI: GPT-4o-mini",
5037
5037
  api: "openai-completions",
5038
5038
  provider: "openrouter",
5039
5039
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5048,9 +5048,9 @@ export const MODELS = {
5048
5048
  contextWindow: 128000,
5049
5049
  maxTokens: 16384,
5050
5050
  },
5051
- "openai/gpt-4o-mini": {
5052
- id: "openai/gpt-4o-mini",
5053
- name: "OpenAI: GPT-4o-mini",
5051
+ "openai/gpt-4o-mini-2024-07-18": {
5052
+ id: "openai/gpt-4o-mini-2024-07-18",
5053
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
5054
5054
  api: "openai-completions",
5055
5055
  provider: "openrouter",
5056
5056
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5473,9 +5473,9 @@ export const MODELS = {
5473
5473
  contextWindow: 16385,
5474
5474
  maxTokens: 4096,
5475
5475
  },
5476
- "openai/gpt-4-0314": {
5477
- id: "openai/gpt-4-0314",
5478
- name: "OpenAI: GPT-4 (older v0314)",
5476
+ "openai/gpt-4": {
5477
+ id: "openai/gpt-4",
5478
+ name: "OpenAI: GPT-4",
5479
5479
  api: "openai-completions",
5480
5480
  provider: "openrouter",
5481
5481
  baseUrl: "https://openrouter.ai/api/v1",
@@ -5490,38 +5490,38 @@ export const MODELS = {
5490
5490
  contextWindow: 8191,
5491
5491
  maxTokens: 4096,
5492
5492
  },
5493
- "openai/gpt-4": {
5494
- id: "openai/gpt-4",
5495
- name: "OpenAI: GPT-4",
5493
+ "openai/gpt-3.5-turbo": {
5494
+ id: "openai/gpt-3.5-turbo",
5495
+ name: "OpenAI: GPT-3.5 Turbo",
5496
5496
  api: "openai-completions",
5497
5497
  provider: "openrouter",
5498
5498
  baseUrl: "https://openrouter.ai/api/v1",
5499
5499
  reasoning: false,
5500
5500
  input: ["text"],
5501
5501
  cost: {
5502
- input: 30,
5503
- output: 60,
5502
+ input: 0.5,
5503
+ output: 1.5,
5504
5504
  cacheRead: 0,
5505
5505
  cacheWrite: 0,
5506
5506
  },
5507
- contextWindow: 8191,
5507
+ contextWindow: 16385,
5508
5508
  maxTokens: 4096,
5509
5509
  },
5510
- "openai/gpt-3.5-turbo": {
5511
- id: "openai/gpt-3.5-turbo",
5512
- name: "OpenAI: GPT-3.5 Turbo",
5510
+ "openai/gpt-4-0314": {
5511
+ id: "openai/gpt-4-0314",
5512
+ name: "OpenAI: GPT-4 (older v0314)",
5513
5513
  api: "openai-completions",
5514
5514
  provider: "openrouter",
5515
5515
  baseUrl: "https://openrouter.ai/api/v1",
5516
5516
  reasoning: false,
5517
5517
  input: ["text"],
5518
5518
  cost: {
5519
- input: 0.5,
5520
- output: 1.5,
5519
+ input: 30,
5520
+ output: 60,
5521
5521
  cacheRead: 0,
5522
5522
  cacheWrite: 0,
5523
5523
  },
5524
- contextWindow: 16385,
5524
+ contextWindow: 8191,
5525
5525
  maxTokens: 4096,
5526
5526
  },
5527
5527
  },