@mariozechner/pi-ai 0.27.1 → 0.27.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2871,6 +2871,23 @@ export const MODELS = {
2871
2871
  },
2872
2872
  },
2873
2873
  "openrouter": {
2874
+ "minimax/minimax-m2.1": {
2875
+ id: "minimax/minimax-m2.1",
2876
+ name: "MiniMax: MiniMax M2.1",
2877
+ api: "openai-completions",
2878
+ provider: "openrouter",
2879
+ baseUrl: "https://openrouter.ai/api/v1",
2880
+ reasoning: true,
2881
+ input: ["text"],
2882
+ cost: {
2883
+ input: 0.3,
2884
+ output: 1.2,
2885
+ cacheRead: 0,
2886
+ cacheWrite: 0,
2887
+ },
2888
+ contextWindow: 204800,
2889
+ maxTokens: 131072,
2890
+ },
2874
2891
  "z-ai/glm-4.7": {
2875
2892
  id: "z-ai/glm-4.7",
2876
2893
  name: "Z.AI: GLM 4.7",
@@ -2880,12 +2897,12 @@ export const MODELS = {
2880
2897
  reasoning: true,
2881
2898
  input: ["text"],
2882
2899
  cost: {
2883
- input: 0.11,
2884
- output: 2.2,
2885
- cacheRead: 0.11,
2900
+ input: 0.44,
2901
+ output: 1.74,
2902
+ cacheRead: 0,
2886
2903
  cacheWrite: 0,
2887
2904
  },
2888
- contextWindow: 200000,
2905
+ contextWindow: 202752,
2889
2906
  maxTokens: 131072,
2890
2907
  },
2891
2908
  "google/gemini-3-flash-preview": {
@@ -2971,7 +2988,7 @@ export const MODELS = {
2971
2988
  cacheWrite: 0,
2972
2989
  },
2973
2990
  contextWindow: 262144,
2974
- maxTokens: 4096,
2991
+ maxTokens: 262144,
2975
2992
  },
2976
2993
  "openai/gpt-5.2-chat": {
2977
2994
  id: "openai/gpt-5.2-chat",
@@ -3254,13 +3271,13 @@ export const MODELS = {
3254
3271
  reasoning: true,
3255
3272
  input: ["text"],
3256
3273
  cost: {
3257
- input: 0.23900000000000002,
3258
- output: 0.378,
3259
- cacheRead: 0.11,
3274
+ input: 0.26,
3275
+ output: 0.38,
3276
+ cacheRead: 0,
3260
3277
  cacheWrite: 0,
3261
3278
  },
3262
3279
  contextWindow: 163840,
3263
- maxTokens: 163840,
3280
+ maxTokens: 65536,
3264
3281
  },
3265
3282
  "prime-intellect/intellect-3": {
3266
3283
  id: "prime-intellect/intellect-3",
@@ -6101,34 +6118,34 @@ export const MODELS = {
6101
6118
  contextWindow: 200000,
6102
6119
  maxTokens: 8192,
6103
6120
  },
6104
- "mistralai/ministral-3b": {
6105
- id: "mistralai/ministral-3b",
6106
- name: "Mistral: Ministral 3B",
6121
+ "mistralai/ministral-8b": {
6122
+ id: "mistralai/ministral-8b",
6123
+ name: "Mistral: Ministral 8B",
6107
6124
  api: "openai-completions",
6108
6125
  provider: "openrouter",
6109
6126
  baseUrl: "https://openrouter.ai/api/v1",
6110
6127
  reasoning: false,
6111
6128
  input: ["text"],
6112
6129
  cost: {
6113
- input: 0.04,
6114
- output: 0.04,
6130
+ input: 0.09999999999999999,
6131
+ output: 0.09999999999999999,
6115
6132
  cacheRead: 0,
6116
6133
  cacheWrite: 0,
6117
6134
  },
6118
6135
  contextWindow: 131072,
6119
6136
  maxTokens: 4096,
6120
6137
  },
6121
- "mistralai/ministral-8b": {
6122
- id: "mistralai/ministral-8b",
6123
- name: "Mistral: Ministral 8B",
6138
+ "mistralai/ministral-3b": {
6139
+ id: "mistralai/ministral-3b",
6140
+ name: "Mistral: Ministral 3B",
6124
6141
  api: "openai-completions",
6125
6142
  provider: "openrouter",
6126
6143
  baseUrl: "https://openrouter.ai/api/v1",
6127
6144
  reasoning: false,
6128
6145
  input: ["text"],
6129
6146
  cost: {
6130
- input: 0.09999999999999999,
6131
- output: 0.09999999999999999,
6147
+ input: 0.04,
6148
+ output: 0.04,
6132
6149
  cacheRead: 0,
6133
6150
  cacheWrite: 0,
6134
6151
  },
@@ -6195,13 +6212,13 @@ export const MODELS = {
6195
6212
  reasoning: false,
6196
6213
  input: ["text"],
6197
6214
  cost: {
6198
- input: 0.07,
6199
- output: 0.26,
6215
+ input: 0.12,
6216
+ output: 0.39,
6200
6217
  cacheRead: 0,
6201
6218
  cacheWrite: 0,
6202
6219
  },
6203
6220
  contextWindow: 32768,
6204
- maxTokens: 32768,
6221
+ maxTokens: 16384,
6205
6222
  },
6206
6223
  "mistralai/pixtral-12b": {
6207
6224
  id: "mistralai/pixtral-12b",
@@ -6305,23 +6322,6 @@ export const MODELS = {
6305
6322
  contextWindow: 128000,
6306
6323
  maxTokens: 16384,
6307
6324
  },
6308
- "meta-llama/llama-3.1-70b-instruct": {
6309
- id: "meta-llama/llama-3.1-70b-instruct",
6310
- name: "Meta: Llama 3.1 70B Instruct",
6311
- api: "openai-completions",
6312
- provider: "openrouter",
6313
- baseUrl: "https://openrouter.ai/api/v1",
6314
- reasoning: false,
6315
- input: ["text"],
6316
- cost: {
6317
- input: 0.39999999999999997,
6318
- output: 0.39999999999999997,
6319
- cacheRead: 0,
6320
- cacheWrite: 0,
6321
- },
6322
- contextWindow: 131072,
6323
- maxTokens: 4096,
6324
- },
6325
6325
  "meta-llama/llama-3.1-405b-instruct": {
6326
6326
  id: "meta-llama/llama-3.1-405b-instruct",
6327
6327
  name: "Meta: Llama 3.1 405B Instruct",
@@ -6356,6 +6356,23 @@ export const MODELS = {
6356
6356
  contextWindow: 131072,
6357
6357
  maxTokens: 16384,
6358
6358
  },
6359
+ "meta-llama/llama-3.1-70b-instruct": {
6360
+ id: "meta-llama/llama-3.1-70b-instruct",
6361
+ name: "Meta: Llama 3.1 70B Instruct",
6362
+ api: "openai-completions",
6363
+ provider: "openrouter",
6364
+ baseUrl: "https://openrouter.ai/api/v1",
6365
+ reasoning: false,
6366
+ input: ["text"],
6367
+ cost: {
6368
+ input: 0.39999999999999997,
6369
+ output: 0.39999999999999997,
6370
+ cacheRead: 0,
6371
+ cacheWrite: 0,
6372
+ },
6373
+ contextWindow: 131072,
6374
+ maxTokens: 4096,
6375
+ },
6359
6376
  "mistralai/mistral-nemo": {
6360
6377
  id: "mistralai/mistral-nemo",
6361
6378
  name: "Mistral: Mistral Nemo",
@@ -6373,9 +6390,9 @@ export const MODELS = {
6373
6390
  contextWindow: 131072,
6374
6391
  maxTokens: 16384,
6375
6392
  },
6376
- "openai/gpt-4o-mini": {
6377
- id: "openai/gpt-4o-mini",
6378
- name: "OpenAI: GPT-4o-mini",
6393
+ "openai/gpt-4o-mini-2024-07-18": {
6394
+ id: "openai/gpt-4o-mini-2024-07-18",
6395
+ name: "OpenAI: GPT-4o-mini (2024-07-18)",
6379
6396
  api: "openai-completions",
6380
6397
  provider: "openrouter",
6381
6398
  baseUrl: "https://openrouter.ai/api/v1",
@@ -6390,9 +6407,9 @@ export const MODELS = {
6390
6407
  contextWindow: 128000,
6391
6408
  maxTokens: 16384,
6392
6409
  },
6393
- "openai/gpt-4o-mini-2024-07-18": {
6394
- id: "openai/gpt-4o-mini-2024-07-18",
6395
- name: "OpenAI: GPT-4o-mini (2024-07-18)",
6410
+ "openai/gpt-4o-mini": {
6411
+ id: "openai/gpt-4o-mini",
6412
+ name: "OpenAI: GPT-4o-mini",
6396
6413
  api: "openai-completions",
6397
6414
  provider: "openrouter",
6398
6415
  baseUrl: "https://openrouter.ai/api/v1",
@@ -6492,6 +6509,23 @@ export const MODELS = {
6492
6509
  contextWindow: 128000,
6493
6510
  maxTokens: 4096,
6494
6511
  },
6512
+ "openai/gpt-4o-2024-05-13": {
6513
+ id: "openai/gpt-4o-2024-05-13",
6514
+ name: "OpenAI: GPT-4o (2024-05-13)",
6515
+ api: "openai-completions",
6516
+ provider: "openrouter",
6517
+ baseUrl: "https://openrouter.ai/api/v1",
6518
+ reasoning: false,
6519
+ input: ["text", "image"],
6520
+ cost: {
6521
+ input: 5,
6522
+ output: 15,
6523
+ cacheRead: 0,
6524
+ cacheWrite: 0,
6525
+ },
6526
+ contextWindow: 128000,
6527
+ maxTokens: 4096,
6528
+ },
6495
6529
  "openai/gpt-4o": {
6496
6530
  id: "openai/gpt-4o",
6497
6531
  name: "OpenAI: GPT-4o",
@@ -6526,22 +6560,22 @@ export const MODELS = {
6526
6560
  contextWindow: 128000,
6527
6561
  maxTokens: 64000,
6528
6562
  },
6529
- "openai/gpt-4o-2024-05-13": {
6530
- id: "openai/gpt-4o-2024-05-13",
6531
- name: "OpenAI: GPT-4o (2024-05-13)",
6563
+ "meta-llama/llama-3-70b-instruct": {
6564
+ id: "meta-llama/llama-3-70b-instruct",
6565
+ name: "Meta: Llama 3 70B Instruct",
6532
6566
  api: "openai-completions",
6533
6567
  provider: "openrouter",
6534
6568
  baseUrl: "https://openrouter.ai/api/v1",
6535
6569
  reasoning: false,
6536
- input: ["text", "image"],
6570
+ input: ["text"],
6537
6571
  cost: {
6538
- input: 5,
6539
- output: 15,
6572
+ input: 0.3,
6573
+ output: 0.39999999999999997,
6540
6574
  cacheRead: 0,
6541
6575
  cacheWrite: 0,
6542
6576
  },
6543
- contextWindow: 128000,
6544
- maxTokens: 4096,
6577
+ contextWindow: 8192,
6578
+ maxTokens: 16384,
6545
6579
  },
6546
6580
  "meta-llama/llama-3-8b-instruct": {
6547
6581
  id: "meta-llama/llama-3-8b-instruct",
@@ -6560,23 +6594,6 @@ export const MODELS = {
6560
6594
  contextWindow: 8192,
6561
6595
  maxTokens: 16384,
6562
6596
  },
6563
- "meta-llama/llama-3-70b-instruct": {
6564
- id: "meta-llama/llama-3-70b-instruct",
6565
- name: "Meta: Llama 3 70B Instruct",
6566
- api: "openai-completions",
6567
- provider: "openrouter",
6568
- baseUrl: "https://openrouter.ai/api/v1",
6569
- reasoning: false,
6570
- input: ["text"],
6571
- cost: {
6572
- input: 0.3,
6573
- output: 0.39999999999999997,
6574
- cacheRead: 0,
6575
- cacheWrite: 0,
6576
- },
6577
- contextWindow: 8192,
6578
- maxTokens: 16384,
6579
- },
6580
6597
  "mistralai/mixtral-8x22b-instruct": {
6581
6598
  id: "mistralai/mixtral-8x22b-instruct",
6582
6599
  name: "Mistral: Mixtral 8x22B Instruct",
@@ -6662,38 +6679,38 @@ export const MODELS = {
6662
6679
  contextWindow: 128000,
6663
6680
  maxTokens: 4096,
6664
6681
  },
6665
- "openai/gpt-4-turbo-preview": {
6666
- id: "openai/gpt-4-turbo-preview",
6667
- name: "OpenAI: GPT-4 Turbo Preview",
6682
+ "openai/gpt-3.5-turbo-0613": {
6683
+ id: "openai/gpt-3.5-turbo-0613",
6684
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
6668
6685
  api: "openai-completions",
6669
6686
  provider: "openrouter",
6670
6687
  baseUrl: "https://openrouter.ai/api/v1",
6671
6688
  reasoning: false,
6672
6689
  input: ["text"],
6673
6690
  cost: {
6674
- input: 10,
6675
- output: 30,
6691
+ input: 1,
6692
+ output: 2,
6676
6693
  cacheRead: 0,
6677
6694
  cacheWrite: 0,
6678
6695
  },
6679
- contextWindow: 128000,
6696
+ contextWindow: 4095,
6680
6697
  maxTokens: 4096,
6681
6698
  },
6682
- "openai/gpt-3.5-turbo-0613": {
6683
- id: "openai/gpt-3.5-turbo-0613",
6684
- name: "OpenAI: GPT-3.5 Turbo (older v0613)",
6699
+ "openai/gpt-4-turbo-preview": {
6700
+ id: "openai/gpt-4-turbo-preview",
6701
+ name: "OpenAI: GPT-4 Turbo Preview",
6685
6702
  api: "openai-completions",
6686
6703
  provider: "openrouter",
6687
6704
  baseUrl: "https://openrouter.ai/api/v1",
6688
6705
  reasoning: false,
6689
6706
  input: ["text"],
6690
6707
  cost: {
6691
- input: 1,
6692
- output: 2,
6708
+ input: 10,
6709
+ output: 30,
6693
6710
  cacheRead: 0,
6694
6711
  cacheWrite: 0,
6695
6712
  },
6696
- contextWindow: 4095,
6713
+ contextWindow: 128000,
6697
6714
  maxTokens: 4096,
6698
6715
  },
6699
6716
  "mistralai/mistral-tiny": {
@@ -6764,9 +6781,9 @@ export const MODELS = {
6764
6781
  contextWindow: 16385,
6765
6782
  maxTokens: 4096,
6766
6783
  },
6767
- "openai/gpt-4-0314": {
6768
- id: "openai/gpt-4-0314",
6769
- name: "OpenAI: GPT-4 (older v0314)",
6784
+ "openai/gpt-4": {
6785
+ id: "openai/gpt-4",
6786
+ name: "OpenAI: GPT-4",
6770
6787
  api: "openai-completions",
6771
6788
  provider: "openrouter",
6772
6789
  baseUrl: "https://openrouter.ai/api/v1",
@@ -6781,38 +6798,38 @@ export const MODELS = {
6781
6798
  contextWindow: 8191,
6782
6799
  maxTokens: 4096,
6783
6800
  },
6784
- "openai/gpt-3.5-turbo": {
6785
- id: "openai/gpt-3.5-turbo",
6786
- name: "OpenAI: GPT-3.5 Turbo",
6801
+ "openai/gpt-4-0314": {
6802
+ id: "openai/gpt-4-0314",
6803
+ name: "OpenAI: GPT-4 (older v0314)",
6787
6804
  api: "openai-completions",
6788
6805
  provider: "openrouter",
6789
6806
  baseUrl: "https://openrouter.ai/api/v1",
6790
6807
  reasoning: false,
6791
6808
  input: ["text"],
6792
6809
  cost: {
6793
- input: 0.5,
6794
- output: 1.5,
6810
+ input: 30,
6811
+ output: 60,
6795
6812
  cacheRead: 0,
6796
6813
  cacheWrite: 0,
6797
6814
  },
6798
- contextWindow: 16385,
6815
+ contextWindow: 8191,
6799
6816
  maxTokens: 4096,
6800
6817
  },
6801
- "openai/gpt-4": {
6802
- id: "openai/gpt-4",
6803
- name: "OpenAI: GPT-4",
6818
+ "openai/gpt-3.5-turbo": {
6819
+ id: "openai/gpt-3.5-turbo",
6820
+ name: "OpenAI: GPT-3.5 Turbo",
6804
6821
  api: "openai-completions",
6805
6822
  provider: "openrouter",
6806
6823
  baseUrl: "https://openrouter.ai/api/v1",
6807
6824
  reasoning: false,
6808
6825
  input: ["text"],
6809
6826
  cost: {
6810
- input: 30,
6811
- output: 60,
6827
+ input: 0.5,
6828
+ output: 1.5,
6812
6829
  cacheRead: 0,
6813
6830
  cacheWrite: 0,
6814
6831
  },
6815
- contextWindow: 8191,
6832
+ contextWindow: 16385,
6816
6833
  maxTokens: 4096,
6817
6834
  },
6818
6835
  "openrouter/auto": {