@mariozechner/pi-ai 0.38.0 → 0.39.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2943,6 +2943,23 @@ export const MODELS = {
2943
2943
  contextWindow: 65536,
2944
2944
  maxTokens: 65536,
2945
2945
  },
2946
+ "allenai/olmo-3.1-32b-instruct": {
2947
+ id: "allenai/olmo-3.1-32b-instruct",
2948
+ name: "AllenAI: Olmo 3.1 32B Instruct",
2949
+ api: "openai-completions",
2950
+ provider: "openrouter",
2951
+ baseUrl: "https://openrouter.ai/api/v1",
2952
+ reasoning: false,
2953
+ input: ["text"],
2954
+ cost: {
2955
+ input: 0.19999999999999998,
2956
+ output: 0.6,
2957
+ cacheRead: 0,
2958
+ cacheWrite: 0,
2959
+ },
2960
+ contextWindow: 65536,
2961
+ maxTokens: 4096,
2962
+ },
2946
2963
  "amazon/nova-2-lite-v1": {
2947
2964
  id: "amazon/nova-2-lite-v1",
2948
2965
  name: "Amazon: Nova 2 Lite",
@@ -3819,13 +3836,13 @@ export const MODELS = {
3819
3836
  reasoning: false,
3820
3837
  input: ["text", "image"],
3821
3838
  cost: {
3822
- input: 0.036,
3823
- output: 0.064,
3839
+ input: 0.04,
3840
+ output: 0.15,
3824
3841
  cacheRead: 0,
3825
3842
  cacheWrite: 0,
3826
3843
  },
3827
- contextWindow: 131072,
3828
- maxTokens: 4096,
3844
+ contextWindow: 96000,
3845
+ maxTokens: 96000,
3829
3846
  },
3830
3847
  "google/gemma-3-27b-it:free": {
3831
3848
  id: "google/gemma-3-27b-it:free",
@@ -4125,9 +4142,9 @@ export const MODELS = {
4125
4142
  reasoning: true,
4126
4143
  input: ["text"],
4127
4144
  cost: {
4128
- input: 0.12,
4129
- output: 0.48,
4130
- cacheRead: 0,
4145
+ input: 0.28,
4146
+ output: 1.2,
4147
+ cacheRead: 0.14,
4131
4148
  cacheWrite: 0,
4132
4149
  },
4133
4150
  contextWindow: 196608,
@@ -4720,13 +4737,13 @@ export const MODELS = {
4720
4737
  reasoning: true,
4721
4738
  input: ["text"],
4722
4739
  cost: {
4723
- input: 0.32,
4724
- output: 0.48,
4740
+ input: 0.39999999999999997,
4741
+ output: 1.75,
4725
4742
  cacheRead: 0,
4726
4743
  cacheWrite: 0,
4727
4744
  },
4728
4745
  contextWindow: 262144,
4729
- maxTokens: 4096,
4746
+ maxTokens: 65535,
4730
4747
  },
4731
4748
  "nex-agi/deepseek-v3.1-nex-n1:free": {
4732
4749
  id: "nex-agi/deepseek-v3.1-nex-n1:free",
@@ -5502,8 +5519,8 @@ export const MODELS = {
5502
5519
  reasoning: true,
5503
5520
  input: ["text"],
5504
5521
  cost: {
5505
- input: 0.02,
5506
- output: 0.09999999999999999,
5522
+ input: 0.039,
5523
+ output: 0.19,
5507
5524
  cacheRead: 0,
5508
5525
  cacheWrite: 0,
5509
5526
  },
@@ -5553,13 +5570,13 @@ export const MODELS = {
5553
5570
  reasoning: true,
5554
5571
  input: ["text"],
5555
5572
  cost: {
5556
- input: 0.016,
5557
- output: 0.06,
5573
+ input: 0.02,
5574
+ output: 0.09999999999999999,
5558
5575
  cacheRead: 0,
5559
5576
  cacheWrite: 0,
5560
5577
  },
5561
5578
  contextWindow: 131072,
5562
- maxTokens: 4096,
5579
+ maxTokens: 131072,
5563
5580
  },
5564
5581
  "openai/gpt-oss-20b:free": {
5565
5582
  id: "openai/gpt-oss-20b:free",
@@ -5799,6 +5816,23 @@ export const MODELS = {
5799
5816
  contextWindow: 32768,
5800
5817
  maxTokens: 16384,
5801
5818
  },
5819
+ "qwen/qwen-2.5-7b-instruct": {
5820
+ id: "qwen/qwen-2.5-7b-instruct",
5821
+ name: "Qwen: Qwen2.5 7B Instruct",
5822
+ api: "openai-completions",
5823
+ provider: "openrouter",
5824
+ baseUrl: "https://openrouter.ai/api/v1",
5825
+ reasoning: false,
5826
+ input: ["text"],
5827
+ cost: {
5828
+ input: 0.04,
5829
+ output: 0.09999999999999999,
5830
+ cacheRead: 0,
5831
+ cacheWrite: 0,
5832
+ },
5833
+ contextWindow: 32768,
5834
+ maxTokens: 4096,
5835
+ },
5802
5836
  "qwen/qwen-max": {
5803
5837
  id: "qwen/qwen-max",
5804
5838
  name: "Qwen: Qwen-Max ",
@@ -6199,8 +6233,8 @@ export const MODELS = {
6199
6233
  reasoning: false,
6200
6234
  input: ["text"],
6201
6235
  cost: {
6202
- input: 0.06,
6203
- output: 0.6,
6236
+ input: 0.09,
6237
+ output: 1.1,
6204
6238
  cacheRead: 0,
6205
6239
  cacheWrite: 0,
6206
6240
  },
@@ -6233,8 +6267,8 @@ export const MODELS = {
6233
6267
  reasoning: false,
6234
6268
  input: ["text", "image"],
6235
6269
  cost: {
6236
- input: 0.12,
6237
- output: 0.56,
6270
+ input: 0.19999999999999998,
6271
+ output: 1.2,
6238
6272
  cacheRead: 0,
6239
6273
  cacheWrite: 0,
6240
6274
  },
@@ -6794,13 +6828,13 @@ export const MODELS = {
6794
6828
  reasoning: true,
6795
6829
  input: ["text"],
6796
6830
  cost: {
6797
- input: 0.16,
6798
- output: 0.7999999999999999,
6831
+ input: 0.39999999999999997,
6832
+ output: 1.5,
6799
6833
  cacheRead: 0,
6800
6834
  cacheWrite: 0,
6801
6835
  },
6802
6836
  contextWindow: 202752,
6803
- maxTokens: 4096,
6837
+ maxTokens: 65535,
6804
6838
  },
6805
6839
  },
6806
6840
  "xai": {