@mariozechner/pi-ai 0.70.2 → 0.70.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3702,6 +3702,29 @@ export declare const MODELS: {
3702
3702
  contextWindow: number;
3703
3703
  maxTokens: number;
3704
3704
  };
3705
+ readonly "gpt-5.5": {
3706
+ id: string;
3707
+ name: string;
3708
+ api: "openai-responses";
3709
+ provider: string;
3710
+ baseUrl: string;
3711
+ headers: {
3712
+ "User-Agent": string;
3713
+ "Editor-Version": string;
3714
+ "Editor-Plugin-Version": string;
3715
+ "Copilot-Integration-Id": string;
3716
+ };
3717
+ reasoning: true;
3718
+ input: ("image" | "text")[];
3719
+ cost: {
3720
+ input: number;
3721
+ output: number;
3722
+ cacheRead: number;
3723
+ cacheWrite: number;
3724
+ };
3725
+ contextWindow: number;
3726
+ maxTokens: number;
3727
+ };
3705
3728
  readonly "grok-code-fast-1": {
3706
3729
  id: string;
3707
3730
  name: string;
@@ -5240,6 +5263,26 @@ export declare const MODELS: {
5240
5263
  contextWindow: number;
5241
5264
  maxTokens: number;
5242
5265
  };
5266
+ readonly "deepseek-ai/DeepSeek-V4-Pro": {
5267
+ id: string;
5268
+ name: string;
5269
+ api: "openai-completions";
5270
+ provider: string;
5271
+ baseUrl: string;
5272
+ compat: {
5273
+ supportsDeveloperRole: false;
5274
+ };
5275
+ reasoning: true;
5276
+ input: "text"[];
5277
+ cost: {
5278
+ input: number;
5279
+ output: number;
5280
+ cacheRead: number;
5281
+ cacheWrite: number;
5282
+ };
5283
+ contextWindow: number;
5284
+ maxTokens: number;
5285
+ };
5243
5286
  readonly "moonshotai/Kimi-K2-Instruct": {
5244
5287
  id: string;
5245
5288
  name: string;
@@ -6888,23 +6931,6 @@ export declare const MODELS: {
6888
6931
  contextWindow: number;
6889
6932
  maxTokens: number;
6890
6933
  };
6891
- readonly "claude-3-5-haiku": {
6892
- id: string;
6893
- name: string;
6894
- api: "anthropic-messages";
6895
- provider: string;
6896
- baseUrl: string;
6897
- reasoning: false;
6898
- input: ("image" | "text")[];
6899
- cost: {
6900
- input: number;
6901
- output: number;
6902
- cacheRead: number;
6903
- cacheWrite: number;
6904
- };
6905
- contextWindow: number;
6906
- maxTokens: number;
6907
- };
6908
6934
  readonly "claude-haiku-4-5": {
6909
6935
  id: string;
6910
6936
  name: string;
@@ -7347,6 +7373,40 @@ export declare const MODELS: {
7347
7373
  contextWindow: number;
7348
7374
  maxTokens: number;
7349
7375
  };
7376
+ readonly "gpt-5.5": {
7377
+ id: string;
7378
+ name: string;
7379
+ api: "openai-responses";
7380
+ provider: string;
7381
+ baseUrl: string;
7382
+ reasoning: true;
7383
+ input: ("image" | "text")[];
7384
+ cost: {
7385
+ input: number;
7386
+ output: number;
7387
+ cacheRead: number;
7388
+ cacheWrite: number;
7389
+ };
7390
+ contextWindow: number;
7391
+ maxTokens: number;
7392
+ };
7393
+ readonly "gpt-5.5-pro": {
7394
+ id: string;
7395
+ name: string;
7396
+ api: "openai-responses";
7397
+ provider: string;
7398
+ baseUrl: string;
7399
+ reasoning: true;
7400
+ input: ("image" | "text")[];
7401
+ cost: {
7402
+ input: number;
7403
+ output: number;
7404
+ cacheRead: number;
7405
+ cacheWrite: number;
7406
+ };
7407
+ contextWindow: number;
7408
+ maxTokens: number;
7409
+ };
7350
7410
  readonly "hy3-preview-free": {
7351
7411
  id: string;
7352
7412
  name: string;
@@ -7519,6 +7579,46 @@ export declare const MODELS: {
7519
7579
  };
7520
7580
  };
7521
7581
  readonly "opencode-go": {
7582
+ readonly "deepseek-v4-flash": {
7583
+ id: string;
7584
+ name: string;
7585
+ api: "openai-completions";
7586
+ provider: string;
7587
+ baseUrl: string;
7588
+ compat: {
7589
+ requiresReasoningContentOnAssistantMessages: true;
7590
+ };
7591
+ reasoning: true;
7592
+ input: "text"[];
7593
+ cost: {
7594
+ input: number;
7595
+ output: number;
7596
+ cacheRead: number;
7597
+ cacheWrite: number;
7598
+ };
7599
+ contextWindow: number;
7600
+ maxTokens: number;
7601
+ };
7602
+ readonly "deepseek-v4-pro": {
7603
+ id: string;
7604
+ name: string;
7605
+ api: "openai-completions";
7606
+ provider: string;
7607
+ baseUrl: string;
7608
+ compat: {
7609
+ requiresReasoningContentOnAssistantMessages: true;
7610
+ };
7611
+ reasoning: true;
7612
+ input: "text"[];
7613
+ cost: {
7614
+ input: number;
7615
+ output: number;
7616
+ cacheRead: number;
7617
+ cacheWrite: number;
7618
+ };
7619
+ contextWindow: number;
7620
+ maxTokens: number;
7621
+ };
7522
7622
  readonly "glm-5": {
7523
7623
  id: string;
7524
7624
  name: string;
@@ -8462,6 +8562,9 @@ export declare const MODELS: {
8462
8562
  api: "openai-completions";
8463
8563
  provider: string;
8464
8564
  baseUrl: string;
8565
+ compat: {
8566
+ requiresReasoningContentOnAssistantMessages: true;
8567
+ };
8465
8568
  reasoning: true;
8466
8569
  input: "text"[];
8467
8570
  cost: {
@@ -8479,6 +8582,9 @@ export declare const MODELS: {
8479
8582
  api: "openai-completions";
8480
8583
  provider: string;
8481
8584
  baseUrl: string;
8585
+ compat: {
8586
+ requiresReasoningContentOnAssistantMessages: true;
8587
+ };
8482
8588
  reasoning: true;
8483
8589
  input: "text"[];
8484
8590
  cost: {
@@ -8711,6 +8817,40 @@ export declare const MODELS: {
8711
8817
  contextWindow: number;
8712
8818
  maxTokens: number;
8713
8819
  };
8820
+ readonly "google/gemma-3-12b-it": {
8821
+ id: string;
8822
+ name: string;
8823
+ api: "openai-completions";
8824
+ provider: string;
8825
+ baseUrl: string;
8826
+ reasoning: false;
8827
+ input: ("image" | "text")[];
8828
+ cost: {
8829
+ input: number;
8830
+ output: number;
8831
+ cacheRead: number;
8832
+ cacheWrite: number;
8833
+ };
8834
+ contextWindow: number;
8835
+ maxTokens: number;
8836
+ };
8837
+ readonly "google/gemma-3-27b-it": {
8838
+ id: string;
8839
+ name: string;
8840
+ api: "openai-completions";
8841
+ provider: string;
8842
+ baseUrl: string;
8843
+ reasoning: false;
8844
+ input: ("image" | "text")[];
8845
+ cost: {
8846
+ input: number;
8847
+ output: number;
8848
+ cacheRead: number;
8849
+ cacheWrite: number;
8850
+ };
8851
+ contextWindow: number;
8852
+ maxTokens: number;
8853
+ };
8714
8854
  readonly "google/gemma-4-26b-a4b-it": {
8715
8855
  id: string;
8716
8856
  name: string;
@@ -10326,6 +10466,40 @@ export declare const MODELS: {
10326
10466
  contextWindow: number;
10327
10467
  maxTokens: number;
10328
10468
  };
10469
+ readonly "openai/gpt-5.5": {
10470
+ id: string;
10471
+ name: string;
10472
+ api: "openai-completions";
10473
+ provider: string;
10474
+ baseUrl: string;
10475
+ reasoning: true;
10476
+ input: ("image" | "text")[];
10477
+ cost: {
10478
+ input: number;
10479
+ output: number;
10480
+ cacheRead: number;
10481
+ cacheWrite: number;
10482
+ };
10483
+ contextWindow: number;
10484
+ maxTokens: number;
10485
+ };
10486
+ readonly "openai/gpt-5.5-pro": {
10487
+ id: string;
10488
+ name: string;
10489
+ api: "openai-completions";
10490
+ provider: string;
10491
+ baseUrl: string;
10492
+ reasoning: true;
10493
+ input: ("image" | "text")[];
10494
+ cost: {
10495
+ input: number;
10496
+ output: number;
10497
+ cacheRead: number;
10498
+ cacheWrite: number;
10499
+ };
10500
+ contextWindow: number;
10501
+ maxTokens: number;
10502
+ };
10329
10503
  readonly "openai/gpt-audio": {
10330
10504
  id: string;
10331
10505
  name: string;
@@ -11363,6 +11537,74 @@ export declare const MODELS: {
11363
11537
  contextWindow: number;
11364
11538
  maxTokens: number;
11365
11539
  };
11540
+ readonly "qwen/qwen3.5-plus-20260420": {
11541
+ id: string;
11542
+ name: string;
11543
+ api: "openai-completions";
11544
+ provider: string;
11545
+ baseUrl: string;
11546
+ reasoning: true;
11547
+ input: ("image" | "text")[];
11548
+ cost: {
11549
+ input: number;
11550
+ output: number;
11551
+ cacheRead: number;
11552
+ cacheWrite: number;
11553
+ };
11554
+ contextWindow: number;
11555
+ maxTokens: number;
11556
+ };
11557
+ readonly "qwen/qwen3.6-27b": {
11558
+ id: string;
11559
+ name: string;
11560
+ api: "openai-completions";
11561
+ provider: string;
11562
+ baseUrl: string;
11563
+ reasoning: true;
11564
+ input: ("image" | "text")[];
11565
+ cost: {
11566
+ input: number;
11567
+ output: number;
11568
+ cacheRead: number;
11569
+ cacheWrite: number;
11570
+ };
11571
+ contextWindow: number;
11572
+ maxTokens: number;
11573
+ };
11574
+ readonly "qwen/qwen3.6-flash": {
11575
+ id: string;
11576
+ name: string;
11577
+ api: "openai-completions";
11578
+ provider: string;
11579
+ baseUrl: string;
11580
+ reasoning: true;
11581
+ input: ("image" | "text")[];
11582
+ cost: {
11583
+ input: number;
11584
+ output: number;
11585
+ cacheRead: number;
11586
+ cacheWrite: number;
11587
+ };
11588
+ contextWindow: number;
11589
+ maxTokens: number;
11590
+ };
11591
+ readonly "qwen/qwen3.6-max-preview": {
11592
+ id: string;
11593
+ name: string;
11594
+ api: "openai-completions";
11595
+ provider: string;
11596
+ baseUrl: string;
11597
+ reasoning: true;
11598
+ input: "text"[];
11599
+ cost: {
11600
+ input: number;
11601
+ output: number;
11602
+ cacheRead: number;
11603
+ cacheWrite: number;
11604
+ };
11605
+ contextWindow: number;
11606
+ maxTokens: number;
11607
+ };
11366
11608
  readonly "qwen/qwen3.6-plus": {
11367
11609
  id: string;
11368
11610
  name: string;
@@ -14017,6 +14259,40 @@ export declare const MODELS: {
14017
14259
  contextWindow: number;
14018
14260
  maxTokens: number;
14019
14261
  };
14262
+ readonly "openai/gpt-5.5": {
14263
+ id: string;
14264
+ name: string;
14265
+ api: "anthropic-messages";
14266
+ provider: string;
14267
+ baseUrl: string;
14268
+ reasoning: true;
14269
+ input: ("image" | "text")[];
14270
+ cost: {
14271
+ input: number;
14272
+ output: number;
14273
+ cacheRead: number;
14274
+ cacheWrite: number;
14275
+ };
14276
+ contextWindow: number;
14277
+ maxTokens: number;
14278
+ };
14279
+ readonly "openai/gpt-5.5-pro": {
14280
+ id: string;
14281
+ name: string;
14282
+ api: "anthropic-messages";
14283
+ provider: string;
14284
+ baseUrl: string;
14285
+ reasoning: true;
14286
+ input: ("image" | "text")[];
14287
+ cost: {
14288
+ input: number;
14289
+ output: number;
14290
+ cacheRead: number;
14291
+ cacheWrite: number;
14292
+ };
14293
+ contextWindow: number;
14294
+ maxTokens: number;
14295
+ };
14020
14296
  readonly "openai/gpt-oss-20b": {
14021
14297
  id: string;
14022
14298
  name: string;
@@ -15143,27 +15419,6 @@ export declare const MODELS: {
15143
15419
  };
15144
15420
  };
15145
15421
  readonly zai: {
15146
- readonly "glm-4.5": {
15147
- id: string;
15148
- name: string;
15149
- api: "openai-completions";
15150
- provider: string;
15151
- baseUrl: string;
15152
- compat: {
15153
- supportsDeveloperRole: false;
15154
- thinkingFormat: "zai";
15155
- };
15156
- reasoning: true;
15157
- input: "text"[];
15158
- cost: {
15159
- input: number;
15160
- output: number;
15161
- cacheRead: number;
15162
- cacheWrite: number;
15163
- };
15164
- contextWindow: number;
15165
- maxTokens: number;
15166
- };
15167
15422
  readonly "glm-4.5-air": {
15168
15423
  id: string;
15169
15424
  name: string;
@@ -15185,92 +15440,6 @@ export declare const MODELS: {
15185
15440
  contextWindow: number;
15186
15441
  maxTokens: number;
15187
15442
  };
15188
- readonly "glm-4.5-flash": {
15189
- id: string;
15190
- name: string;
15191
- api: "openai-completions";
15192
- provider: string;
15193
- baseUrl: string;
15194
- compat: {
15195
- supportsDeveloperRole: false;
15196
- thinkingFormat: "zai";
15197
- };
15198
- reasoning: true;
15199
- input: "text"[];
15200
- cost: {
15201
- input: number;
15202
- output: number;
15203
- cacheRead: number;
15204
- cacheWrite: number;
15205
- };
15206
- contextWindow: number;
15207
- maxTokens: number;
15208
- };
15209
- readonly "glm-4.5v": {
15210
- id: string;
15211
- name: string;
15212
- api: "openai-completions";
15213
- provider: string;
15214
- baseUrl: string;
15215
- compat: {
15216
- supportsDeveloperRole: false;
15217
- thinkingFormat: "zai";
15218
- };
15219
- reasoning: true;
15220
- input: ("image" | "text")[];
15221
- cost: {
15222
- input: number;
15223
- output: number;
15224
- cacheRead: number;
15225
- cacheWrite: number;
15226
- };
15227
- contextWindow: number;
15228
- maxTokens: number;
15229
- };
15230
- readonly "glm-4.6": {
15231
- id: string;
15232
- name: string;
15233
- api: "openai-completions";
15234
- provider: string;
15235
- baseUrl: string;
15236
- compat: {
15237
- supportsDeveloperRole: false;
15238
- thinkingFormat: "zai";
15239
- zaiToolStream: true;
15240
- };
15241
- reasoning: true;
15242
- input: "text"[];
15243
- cost: {
15244
- input: number;
15245
- output: number;
15246
- cacheRead: number;
15247
- cacheWrite: number;
15248
- };
15249
- contextWindow: number;
15250
- maxTokens: number;
15251
- };
15252
- readonly "glm-4.6v": {
15253
- id: string;
15254
- name: string;
15255
- api: "openai-completions";
15256
- provider: string;
15257
- baseUrl: string;
15258
- compat: {
15259
- supportsDeveloperRole: false;
15260
- thinkingFormat: "zai";
15261
- zaiToolStream: true;
15262
- };
15263
- reasoning: true;
15264
- input: ("image" | "text")[];
15265
- cost: {
15266
- input: number;
15267
- output: number;
15268
- cacheRead: number;
15269
- cacheWrite: number;
15270
- };
15271
- contextWindow: number;
15272
- maxTokens: number;
15273
- };
15274
15443
  readonly "glm-4.7": {
15275
15444
  id: string;
15276
15445
  name: string;
@@ -15293,72 +15462,6 @@ export declare const MODELS: {
15293
15462
  contextWindow: number;
15294
15463
  maxTokens: number;
15295
15464
  };
15296
- readonly "glm-4.7-flash": {
15297
- id: string;
15298
- name: string;
15299
- api: "openai-completions";
15300
- provider: string;
15301
- baseUrl: string;
15302
- compat: {
15303
- supportsDeveloperRole: false;
15304
- thinkingFormat: "zai";
15305
- zaiToolStream: true;
15306
- };
15307
- reasoning: true;
15308
- input: "text"[];
15309
- cost: {
15310
- input: number;
15311
- output: number;
15312
- cacheRead: number;
15313
- cacheWrite: number;
15314
- };
15315
- contextWindow: number;
15316
- maxTokens: number;
15317
- };
15318
- readonly "glm-4.7-flashx": {
15319
- id: string;
15320
- name: string;
15321
- api: "openai-completions";
15322
- provider: string;
15323
- baseUrl: string;
15324
- compat: {
15325
- supportsDeveloperRole: false;
15326
- thinkingFormat: "zai";
15327
- zaiToolStream: true;
15328
- };
15329
- reasoning: true;
15330
- input: "text"[];
15331
- cost: {
15332
- input: number;
15333
- output: number;
15334
- cacheRead: number;
15335
- cacheWrite: number;
15336
- };
15337
- contextWindow: number;
15338
- maxTokens: number;
15339
- };
15340
- readonly "glm-5": {
15341
- id: string;
15342
- name: string;
15343
- api: "openai-completions";
15344
- provider: string;
15345
- baseUrl: string;
15346
- compat: {
15347
- supportsDeveloperRole: false;
15348
- thinkingFormat: "zai";
15349
- zaiToolStream: true;
15350
- };
15351
- reasoning: true;
15352
- input: "text"[];
15353
- cost: {
15354
- input: number;
15355
- output: number;
15356
- cacheRead: number;
15357
- cacheWrite: number;
15358
- };
15359
- contextWindow: number;
15360
- maxTokens: number;
15361
- };
15362
15465
  readonly "glm-5-turbo": {
15363
15466
  id: string;
15364
15467
  name: string;
@@ -15403,28 +15506,6 @@ export declare const MODELS: {
15403
15506
  contextWindow: number;
15404
15507
  maxTokens: number;
15405
15508
  };
15406
- readonly "glm-5v-turbo": {
15407
- id: string;
15408
- name: string;
15409
- api: "openai-completions";
15410
- provider: string;
15411
- baseUrl: string;
15412
- compat: {
15413
- supportsDeveloperRole: false;
15414
- thinkingFormat: "zai";
15415
- zaiToolStream: true;
15416
- };
15417
- reasoning: true;
15418
- input: ("image" | "text")[];
15419
- cost: {
15420
- input: number;
15421
- output: number;
15422
- cacheRead: number;
15423
- cacheWrite: number;
15424
- };
15425
- contextWindow: number;
15426
- maxTokens: number;
15427
- };
15428
15509
  };
15429
15510
  };
15430
15511
  //# sourceMappingURL=models.generated.d.ts.map