@mariozechner/pi-ai 0.18.1 → 0.18.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -839,6 +839,23 @@ export const MODELS = {
839
839
  contextWindow: 200000,
840
840
  maxTokens: 100000,
841
841
  },
842
+ "gpt-5.2-pro": {
843
+ id: "gpt-5.2-pro",
844
+ name: "GPT-5.2 Pro",
845
+ api: "openai-responses",
846
+ provider: "openai",
847
+ baseUrl: "https://api.openai.com/v1",
848
+ reasoning: true,
849
+ input: ["text", "image"],
850
+ cost: {
851
+ input: 21,
852
+ output: 168,
853
+ cacheRead: 0,
854
+ cacheWrite: 0,
855
+ },
856
+ contextWindow: 400000,
857
+ maxTokens: 128000,
858
+ },
842
859
  "gpt-4-turbo": {
843
860
  id: "gpt-4-turbo",
844
861
  name: "GPT-4 Turbo",
@@ -890,6 +907,23 @@ export const MODELS = {
890
907
  contextWindow: 200000,
891
908
  maxTokens: 100000,
892
909
  },
910
+ "gpt-5.2-chat-latest": {
911
+ id: "gpt-5.2-chat-latest",
912
+ name: "GPT-5.2 Chat",
913
+ api: "openai-responses",
914
+ provider: "openai",
915
+ baseUrl: "https://api.openai.com/v1",
916
+ reasoning: true,
917
+ input: ["text", "image"],
918
+ cost: {
919
+ input: 1.75,
920
+ output: 14,
921
+ cacheRead: 0.175,
922
+ cacheWrite: 0,
923
+ },
924
+ contextWindow: 128000,
925
+ maxTokens: 16384,
926
+ },
893
927
  "gpt-5.1": {
894
928
  id: "gpt-5.1",
895
929
  name: "GPT-5.1",
@@ -1179,6 +1213,23 @@ export const MODELS = {
1179
1213
  contextWindow: 400000,
1180
1214
  maxTokens: 272000,
1181
1215
  },
1216
+ "gpt-5.2": {
1217
+ id: "gpt-5.2",
1218
+ name: "GPT-5.2",
1219
+ api: "openai-responses",
1220
+ provider: "openai",
1221
+ baseUrl: "https://api.openai.com/v1",
1222
+ reasoning: true,
1223
+ input: ["text", "image"],
1224
+ cost: {
1225
+ input: 1.75,
1226
+ output: 14,
1227
+ cacheRead: 0.175,
1228
+ cacheWrite: 0,
1229
+ },
1230
+ contextWindow: 400000,
1231
+ maxTokens: 128000,
1232
+ },
1182
1233
  "gpt-5.1-chat-latest": {
1183
1234
  id: "gpt-5.1-chat-latest",
1184
1235
  name: "GPT-5.1 Chat",
@@ -2090,6 +2141,23 @@ export const MODELS = {
2090
2141
  contextWindow: 128000,
2091
2142
  maxTokens: 128000,
2092
2143
  },
2144
+ "mistral-small-2506": {
2145
+ id: "mistral-small-2506",
2146
+ name: "Mistral Small 3.2",
2147
+ api: "openai-completions",
2148
+ provider: "mistral",
2149
+ baseUrl: "https://api.mistral.ai/v1",
2150
+ reasoning: false,
2151
+ input: ["text", "image"],
2152
+ cost: {
2153
+ input: 0.1,
2154
+ output: 0.3,
2155
+ cacheRead: 0,
2156
+ cacheWrite: 0,
2157
+ },
2158
+ contextWindow: 128000,
2159
+ maxTokens: 16384,
2160
+ },
2093
2161
  "ministral-3b-latest": {
2094
2162
  id: "ministral-3b-latest",
2095
2163
  name: "Ministral 3B",
@@ -2398,6 +2466,57 @@ export const MODELS = {
2398
2466
  },
2399
2467
  },
2400
2468
  openrouter: {
2469
+ "openai/gpt-5.2-chat": {
2470
+ id: "openai/gpt-5.2-chat",
2471
+ name: "OpenAI: GPT-5.2 Chat",
2472
+ api: "openai-completions",
2473
+ provider: "openrouter",
2474
+ baseUrl: "https://openrouter.ai/api/v1",
2475
+ reasoning: false,
2476
+ input: ["text", "image"],
2477
+ cost: {
2478
+ input: 1.75,
2479
+ output: 14,
2480
+ cacheRead: 0.175,
2481
+ cacheWrite: 0,
2482
+ },
2483
+ contextWindow: 128000,
2484
+ maxTokens: 16384,
2485
+ },
2486
+ "openai/gpt-5.2-pro": {
2487
+ id: "openai/gpt-5.2-pro",
2488
+ name: "OpenAI: GPT-5.2 Pro",
2489
+ api: "openai-completions",
2490
+ provider: "openrouter",
2491
+ baseUrl: "https://openrouter.ai/api/v1",
2492
+ reasoning: true,
2493
+ input: ["text", "image"],
2494
+ cost: {
2495
+ input: 21,
2496
+ output: 168,
2497
+ cacheRead: 0,
2498
+ cacheWrite: 0,
2499
+ },
2500
+ contextWindow: 400000,
2501
+ maxTokens: 128000,
2502
+ },
2503
+ "openai/gpt-5.2": {
2504
+ id: "openai/gpt-5.2",
2505
+ name: "OpenAI: GPT-5.2",
2506
+ api: "openai-completions",
2507
+ provider: "openrouter",
2508
+ baseUrl: "https://openrouter.ai/api/v1",
2509
+ reasoning: true,
2510
+ input: ["text", "image"],
2511
+ cost: {
2512
+ input: 1.75,
2513
+ output: 14,
2514
+ cacheRead: 0.175,
2515
+ cacheWrite: 0,
2516
+ },
2517
+ contextWindow: 400000,
2518
+ maxTokens: 128000,
2519
+ },
2401
2520
  "mistralai/devstral-2512:free": {
2402
2521
  id: "mistralai/devstral-2512:free",
2403
2522
  name: "Mistral: Devstral 2 2512 (free)",
@@ -2415,6 +2534,23 @@ export const MODELS = {
2415
2534
  contextWindow: 262144,
2416
2535
  maxTokens: 4096,
2417
2536
  },
2537
+ "mistralai/devstral-2512": {
2538
+ id: "mistralai/devstral-2512",
2539
+ name: "Mistral: Devstral 2 2512",
2540
+ api: "openai-completions",
2541
+ provider: "openrouter",
2542
+ baseUrl: "https://openrouter.ai/api/v1",
2543
+ reasoning: false,
2544
+ input: ["text"],
2545
+ cost: {
2546
+ input: 0.15,
2547
+ output: 0.6,
2548
+ cacheRead: 0,
2549
+ cacheWrite: 0,
2550
+ },
2551
+ contextWindow: 262144,
2552
+ maxTokens: 65536,
2553
+ },
2418
2554
  "relace/relace-search": {
2419
2555
  id: "relace/relace-search",
2420
2556
  name: "Relace: Relace Search",
@@ -2449,6 +2585,23 @@ export const MODELS = {
2449
2585
  contextWindow: 131072,
2450
2586
  maxTokens: 24000,
2451
2587
  },
2588
+ "nex-agi/deepseek-v3.1-nex-n1:free": {
2589
+ id: "nex-agi/deepseek-v3.1-nex-n1:free",
2590
+ name: "Nex AGI: DeepSeek V3.1 Nex N1 (free)",
2591
+ api: "openai-completions",
2592
+ provider: "openrouter",
2593
+ baseUrl: "https://openrouter.ai/api/v1",
2594
+ reasoning: false,
2595
+ input: ["text"],
2596
+ cost: {
2597
+ input: 0,
2598
+ output: 0,
2599
+ cacheRead: 0,
2600
+ cacheWrite: 0,
2601
+ },
2602
+ contextWindow: 131072,
2603
+ maxTokens: 163840,
2604
+ },
2452
2605
  "openai/gpt-5.1-codex-max": {
2453
2606
  id: "openai/gpt-5.1-codex-max",
2454
2607
  name: "OpenAI: GPT-5.1-Codex-Max",
@@ -2611,9 +2764,9 @@ export const MODELS = {
2611
2764
  reasoning: true,
2612
2765
  input: ["text"],
2613
2766
  cost: {
2614
- input: 0.26,
2615
- output: 0.39,
2616
- cacheRead: 0.19999999999999998,
2767
+ input: 0.25,
2768
+ output: 0.38,
2769
+ cacheRead: 0.19,
2617
2770
  cacheWrite: 0,
2618
2771
  },
2619
2772
  contextWindow: 163840,
@@ -3155,13 +3308,13 @@ export const MODELS = {
3155
3308
  reasoning: true,
3156
3309
  input: ["text"],
3157
3310
  cost: {
3158
- input: 0.43,
3159
- output: 1.75,
3160
- cacheRead: 0.0799999993,
3311
+ input: 0.44,
3312
+ output: 1.76,
3313
+ cacheRead: 0,
3161
3314
  cacheWrite: 0,
3162
3315
  },
3163
- contextWindow: 202752,
3164
- maxTokens: 4096,
3316
+ contextWindow: 204800,
3317
+ maxTokens: 131072,
3165
3318
  },
3166
3319
  "anthropic/claude-sonnet-4.5": {
3167
3320
  id: "anthropic/claude-sonnet-4.5",
@@ -3444,13 +3597,13 @@ export const MODELS = {
3444
3597
  reasoning: false,
3445
3598
  input: ["text"],
3446
3599
  cost: {
3447
- input: 0.09999999999999999,
3448
- output: 0.7999999999999999,
3600
+ input: 0.09,
3601
+ output: 1.1,
3449
3602
  cacheRead: 0,
3450
3603
  cacheWrite: 0,
3451
3604
  },
3452
3605
  contextWindow: 262144,
3453
- maxTokens: 262144,
3606
+ maxTokens: 4096,
3454
3607
  },
3455
3608
  "meituan/longcat-flash-chat:free": {
3456
3609
  id: "meituan/longcat-flash-chat:free",
@@ -4189,7 +4342,7 @@ export const MODELS = {
4189
4342
  api: "openai-completions",
4190
4343
  provider: "openrouter",
4191
4344
  baseUrl: "https://openrouter.ai/api/v1",
4192
- reasoning: true,
4345
+ reasoning: false,
4193
4346
  input: ["text"],
4194
4347
  cost: {
4195
4348
  input: 0.071,
@@ -5280,13 +5433,13 @@ export const MODELS = {
5280
5433
  reasoning: false,
5281
5434
  input: ["text"],
5282
5435
  cost: {
5283
- input: 0.108,
5436
+ input: 0.09999999999999999,
5284
5437
  output: 0.32,
5285
5438
  cacheRead: 0,
5286
5439
  cacheWrite: 0,
5287
5440
  },
5288
5441
  contextWindow: 131072,
5289
- maxTokens: 120000,
5442
+ maxTokens: 16384,
5290
5443
  },
5291
5444
  "amazon/nova-lite-v1": {
5292
5445
  id: "amazon/nova-lite-v1",
@@ -5475,34 +5628,34 @@ export const MODELS = {
5475
5628
  contextWindow: 200000,
5476
5629
  maxTokens: 8192,
5477
5630
  },
5478
- "mistralai/ministral-3b": {
5479
- id: "mistralai/ministral-3b",
5480
- name: "Mistral: Ministral 3B",
5631
+ "mistralai/ministral-8b": {
5632
+ id: "mistralai/ministral-8b",
5633
+ name: "Mistral: Ministral 8B",
5481
5634
  api: "openai-completions",
5482
5635
  provider: "openrouter",
5483
5636
  baseUrl: "https://openrouter.ai/api/v1",
5484
5637
  reasoning: false,
5485
5638
  input: ["text"],
5486
5639
  cost: {
5487
- input: 0.04,
5488
- output: 0.04,
5640
+ input: 0.09999999999999999,
5641
+ output: 0.09999999999999999,
5489
5642
  cacheRead: 0,
5490
5643
  cacheWrite: 0,
5491
5644
  },
5492
5645
  contextWindow: 131072,
5493
5646
  maxTokens: 4096,
5494
5647
  },
5495
- "mistralai/ministral-8b": {
5496
- id: "mistralai/ministral-8b",
5497
- name: "Mistral: Ministral 8B",
5648
+ "mistralai/ministral-3b": {
5649
+ id: "mistralai/ministral-3b",
5650
+ name: "Mistral: Ministral 3B",
5498
5651
  api: "openai-completions",
5499
5652
  provider: "openrouter",
5500
5653
  baseUrl: "https://openrouter.ai/api/v1",
5501
5654
  reasoning: false,
5502
5655
  input: ["text"],
5503
5656
  cost: {
5504
- input: 0.09999999999999999,
5505
- output: 0.09999999999999999,
5657
+ input: 0.04,
5658
+ output: 0.04,
5506
5659
  cacheRead: 0,
5507
5660
  cacheWrite: 0,
5508
5661
  },
@@ -5679,6 +5832,23 @@ export const MODELS = {
5679
5832
  contextWindow: 128000,
5680
5833
  maxTokens: 16384,
5681
5834
  },
5835
+ "meta-llama/llama-3.1-8b-instruct": {
5836
+ id: "meta-llama/llama-3.1-8b-instruct",
5837
+ name: "Meta: Llama 3.1 8B Instruct",
5838
+ api: "openai-completions",
5839
+ provider: "openrouter",
5840
+ baseUrl: "https://openrouter.ai/api/v1",
5841
+ reasoning: false,
5842
+ input: ["text"],
5843
+ cost: {
5844
+ input: 0.02,
5845
+ output: 0.03,
5846
+ cacheRead: 0,
5847
+ cacheWrite: 0,
5848
+ },
5849
+ contextWindow: 131072,
5850
+ maxTokens: 16384,
5851
+ },
5682
5852
  "meta-llama/llama-3.1-405b-instruct": {
5683
5853
  id: "meta-llama/llama-3.1-405b-instruct",
5684
5854
  name: "Meta: Llama 3.1 405B Instruct",
@@ -5713,23 +5883,6 @@ export const MODELS = {
5713
5883
  contextWindow: 131072,
5714
5884
  maxTokens: 4096,
5715
5885
  },
5716
- "meta-llama/llama-3.1-8b-instruct": {
5717
- id: "meta-llama/llama-3.1-8b-instruct",
5718
- name: "Meta: Llama 3.1 8B Instruct",
5719
- api: "openai-completions",
5720
- provider: "openrouter",
5721
- baseUrl: "https://openrouter.ai/api/v1",
5722
- reasoning: false,
5723
- input: ["text"],
5724
- cost: {
5725
- input: 0.02,
5726
- output: 0.03,
5727
- cacheRead: 0,
5728
- cacheWrite: 0,
5729
- },
5730
- contextWindow: 131072,
5731
- maxTokens: 16384,
5732
- },
5733
5886
  "mistralai/mistral-nemo": {
5734
5887
  id: "mistralai/mistral-nemo",
5735
5888
  name: "Mistral: Mistral Nemo",
@@ -5866,6 +6019,23 @@ export const MODELS = {
5866
6019
  contextWindow: 128000,
5867
6020
  maxTokens: 4096,
5868
6021
  },
6022
+ "openai/gpt-4o-2024-05-13": {
6023
+ id: "openai/gpt-4o-2024-05-13",
6024
+ name: "OpenAI: GPT-4o (2024-05-13)",
6025
+ api: "openai-completions",
6026
+ provider: "openrouter",
6027
+ baseUrl: "https://openrouter.ai/api/v1",
6028
+ reasoning: false,
6029
+ input: ["text", "image"],
6030
+ cost: {
6031
+ input: 5,
6032
+ output: 15,
6033
+ cacheRead: 0,
6034
+ cacheWrite: 0,
6035
+ },
6036
+ contextWindow: 128000,
6037
+ maxTokens: 4096,
6038
+ },
5869
6039
  "openai/gpt-4o": {
5870
6040
  id: "openai/gpt-4o",
5871
6041
  name: "OpenAI: GPT-4o",
@@ -5900,22 +6070,22 @@ export const MODELS = {
5900
6070
  contextWindow: 128000,
5901
6071
  maxTokens: 64000,
5902
6072
  },
5903
- "openai/gpt-4o-2024-05-13": {
5904
- id: "openai/gpt-4o-2024-05-13",
5905
- name: "OpenAI: GPT-4o (2024-05-13)",
6073
+ "meta-llama/llama-3-70b-instruct": {
6074
+ id: "meta-llama/llama-3-70b-instruct",
6075
+ name: "Meta: Llama 3 70B Instruct",
5906
6076
  api: "openai-completions",
5907
6077
  provider: "openrouter",
5908
6078
  baseUrl: "https://openrouter.ai/api/v1",
5909
6079
  reasoning: false,
5910
- input: ["text", "image"],
6080
+ input: ["text"],
5911
6081
  cost: {
5912
- input: 5,
5913
- output: 15,
6082
+ input: 0.3,
6083
+ output: 0.39999999999999997,
5914
6084
  cacheRead: 0,
5915
6085
  cacheWrite: 0,
5916
6086
  },
5917
- contextWindow: 128000,
5918
- maxTokens: 4096,
6087
+ contextWindow: 8192,
6088
+ maxTokens: 16384,
5919
6089
  },
5920
6090
  "meta-llama/llama-3-8b-instruct": {
5921
6091
  id: "meta-llama/llama-3-8b-instruct",
@@ -5934,23 +6104,6 @@ export const MODELS = {
5934
6104
  contextWindow: 8192,
5935
6105
  maxTokens: 16384,
5936
6106
  },
5937
- "meta-llama/llama-3-70b-instruct": {
5938
- id: "meta-llama/llama-3-70b-instruct",
5939
- name: "Meta: Llama 3 70B Instruct",
5940
- api: "openai-completions",
5941
- provider: "openrouter",
5942
- baseUrl: "https://openrouter.ai/api/v1",
5943
- reasoning: false,
5944
- input: ["text"],
5945
- cost: {
5946
- input: 0.3,
5947
- output: 0.39999999999999997,
5948
- cacheRead: 0,
5949
- cacheWrite: 0,
5950
- },
5951
- contextWindow: 8192,
5952
- maxTokens: 16384,
5953
- },
5954
6107
  "mistralai/mixtral-8x22b-instruct": {
5955
6108
  id: "mistralai/mixtral-8x22b-instruct",
5956
6109
  name: "Mistral: Mixtral 8x22B Instruct",
@@ -6036,38 +6189,38 @@ export const MODELS = {
6036
6189
  contextWindow: 128000,
6037
6190
  maxTokens: 4096,
6038
6191
  },
6039
- "openai/gpt-4-turbo-preview": {
6040
- id: "openai/gpt-4-turbo-preview",
6041
- name: "OpenAI: GPT-4 Turbo Preview",
6192
+ "openai/gpt-3.5-turbo-0613": {
6193
+ id: "openai/gpt-3.5-turbo-0613",
6194
+ name: "OpenAI: GPT-3.5 Turbo (older v0613)",
6042
6195
  api: "openai-completions",
6043
6196
  provider: "openrouter",
6044
6197
  baseUrl: "https://openrouter.ai/api/v1",
6045
6198
  reasoning: false,
6046
6199
  input: ["text"],
6047
6200
  cost: {
6048
- input: 10,
6049
- output: 30,
6201
+ input: 1,
6202
+ output: 2,
6050
6203
  cacheRead: 0,
6051
6204
  cacheWrite: 0,
6052
6205
  },
6053
- contextWindow: 128000,
6206
+ contextWindow: 4095,
6054
6207
  maxTokens: 4096,
6055
6208
  },
6056
- "openai/gpt-3.5-turbo-0613": {
6057
- id: "openai/gpt-3.5-turbo-0613",
6058
- name: "OpenAI: GPT-3.5 Turbo (older v0613)",
6209
+ "openai/gpt-4-turbo-preview": {
6210
+ id: "openai/gpt-4-turbo-preview",
6211
+ name: "OpenAI: GPT-4 Turbo Preview",
6059
6212
  api: "openai-completions",
6060
6213
  provider: "openrouter",
6061
6214
  baseUrl: "https://openrouter.ai/api/v1",
6062
6215
  reasoning: false,
6063
6216
  input: ["text"],
6064
6217
  cost: {
6065
- input: 1,
6066
- output: 2,
6218
+ input: 10,
6219
+ output: 30,
6067
6220
  cacheRead: 0,
6068
6221
  cacheWrite: 0,
6069
6222
  },
6070
- contextWindow: 4095,
6223
+ contextWindow: 128000,
6071
6224
  maxTokens: 4096,
6072
6225
  },
6073
6226
  "mistralai/mistral-tiny": {
@@ -6138,9 +6291,9 @@ export const MODELS = {
6138
6291
  contextWindow: 16385,
6139
6292
  maxTokens: 4096,
6140
6293
  },
6141
- "openai/gpt-4": {
6142
- id: "openai/gpt-4",
6143
- name: "OpenAI: GPT-4",
6294
+ "openai/gpt-4-0314": {
6295
+ id: "openai/gpt-4-0314",
6296
+ name: "OpenAI: GPT-4 (older v0314)",
6144
6297
  api: "openai-completions",
6145
6298
  provider: "openrouter",
6146
6299
  baseUrl: "https://openrouter.ai/api/v1",
@@ -6155,38 +6308,38 @@ export const MODELS = {
6155
6308
  contextWindow: 8191,
6156
6309
  maxTokens: 4096,
6157
6310
  },
6158
- "openai/gpt-3.5-turbo": {
6159
- id: "openai/gpt-3.5-turbo",
6160
- name: "OpenAI: GPT-3.5 Turbo",
6311
+ "openai/gpt-4": {
6312
+ id: "openai/gpt-4",
6313
+ name: "OpenAI: GPT-4",
6161
6314
  api: "openai-completions",
6162
6315
  provider: "openrouter",
6163
6316
  baseUrl: "https://openrouter.ai/api/v1",
6164
6317
  reasoning: false,
6165
6318
  input: ["text"],
6166
6319
  cost: {
6167
- input: 0.5,
6168
- output: 1.5,
6320
+ input: 30,
6321
+ output: 60,
6169
6322
  cacheRead: 0,
6170
6323
  cacheWrite: 0,
6171
6324
  },
6172
- contextWindow: 16385,
6325
+ contextWindow: 8191,
6173
6326
  maxTokens: 4096,
6174
6327
  },
6175
- "openai/gpt-4-0314": {
6176
- id: "openai/gpt-4-0314",
6177
- name: "OpenAI: GPT-4 (older v0314)",
6328
+ "openai/gpt-3.5-turbo": {
6329
+ id: "openai/gpt-3.5-turbo",
6330
+ name: "OpenAI: GPT-3.5 Turbo",
6178
6331
  api: "openai-completions",
6179
6332
  provider: "openrouter",
6180
6333
  baseUrl: "https://openrouter.ai/api/v1",
6181
6334
  reasoning: false,
6182
6335
  input: ["text"],
6183
6336
  cost: {
6184
- input: 30,
6185
- output: 60,
6337
+ input: 0.5,
6338
+ output: 1.5,
6186
6339
  cacheRead: 0,
6187
6340
  cacheWrite: 0,
6188
6341
  },
6189
- contextWindow: 8191,
6342
+ contextWindow: 16385,
6190
6343
  maxTokens: 4096,
6191
6344
  },
6192
6345
  "openrouter/auto": {