@draht/ai 2026.4.23 → 2026.4.25-1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (49) hide show
  1. package/README.md +4 -4
  2. package/dist/models.d.ts +7 -1
  3. package/dist/models.d.ts.map +1 -1
  4. package/dist/models.generated.d.ts +360 -209
  5. package/dist/models.generated.d.ts.map +1 -1
  6. package/dist/models.generated.js +374 -195
  7. package/dist/models.generated.js.map +1 -1
  8. package/dist/models.js +20 -2
  9. package/dist/models.js.map +1 -1
  10. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  11. package/dist/providers/amazon-bedrock.js +9 -5
  12. package/dist/providers/amazon-bedrock.js.map +1 -1
  13. package/dist/providers/anthropic.d.ts.map +1 -1
  14. package/dist/providers/anthropic.js +6 -3
  15. package/dist/providers/anthropic.js.map +1 -1
  16. package/dist/providers/azure-openai-responses.d.ts.map +1 -1
  17. package/dist/providers/azure-openai-responses.js +2 -2
  18. package/dist/providers/azure-openai-responses.js.map +1 -1
  19. package/dist/providers/google-gemini-cli.d.ts.map +1 -1
  20. package/dist/providers/google-gemini-cli.js.map +1 -1
  21. package/dist/providers/google-vertex.d.ts.map +1 -1
  22. package/dist/providers/google-vertex.js.map +1 -1
  23. package/dist/providers/google.d.ts.map +1 -1
  24. package/dist/providers/google.js.map +1 -1
  25. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  26. package/dist/providers/openai-codex-responses.js +2 -2
  27. package/dist/providers/openai-codex-responses.js.map +1 -1
  28. package/dist/providers/openai-completions.d.ts.map +1 -1
  29. package/dist/providers/openai-completions.js +16 -10
  30. package/dist/providers/openai-completions.js.map +1 -1
  31. package/dist/providers/openai-responses.d.ts.map +1 -1
  32. package/dist/providers/openai-responses.js +2 -2
  33. package/dist/providers/openai-responses.js.map +1 -1
  34. package/dist/providers/simple-options.d.ts +6 -1
  35. package/dist/providers/simple-options.d.ts.map +1 -1
  36. package/dist/providers/simple-options.js +10 -1
  37. package/dist/providers/simple-options.js.map +1 -1
  38. package/dist/types.d.ts +1 -1
  39. package/dist/types.d.ts.map +1 -1
  40. package/dist/types.js.map +1 -1
  41. package/dist/utils/oauth/index.d.ts +1 -0
  42. package/dist/utils/oauth/index.d.ts.map +1 -1
  43. package/dist/utils/oauth/index.js +4 -0
  44. package/dist/utils/oauth/index.js.map +1 -1
  45. package/dist/utils/oauth/opencode-go.d.ts +23 -0
  46. package/dist/utils/oauth/opencode-go.d.ts.map +1 -0
  47. package/dist/utils/oauth/opencode-go.js +46 -0
  48. package/dist/utils/oauth/opencode-go.js.map +1 -0
  49. package/package.json +1 -1
@@ -323,6 +323,40 @@ export declare const MODELS: {
323
323
  contextWindow: number;
324
324
  maxTokens: number;
325
325
  };
326
+ readonly "au.anthropic.claude-opus-4-6-v1": {
327
+ id: string;
328
+ name: string;
329
+ api: "bedrock-converse-stream";
330
+ provider: string;
331
+ baseUrl: string;
332
+ reasoning: true;
333
+ input: ("image" | "text")[];
334
+ cost: {
335
+ input: number;
336
+ output: number;
337
+ cacheRead: number;
338
+ cacheWrite: number;
339
+ };
340
+ contextWindow: number;
341
+ maxTokens: number;
342
+ };
343
+ readonly "au.anthropic.claude-sonnet-4-6": {
344
+ id: string;
345
+ name: string;
346
+ api: "bedrock-converse-stream";
347
+ provider: string;
348
+ baseUrl: string;
349
+ reasoning: true;
350
+ input: ("image" | "text")[];
351
+ cost: {
352
+ input: number;
353
+ output: number;
354
+ cacheRead: number;
355
+ cacheWrite: number;
356
+ };
357
+ contextWindow: number;
358
+ maxTokens: number;
359
+ };
326
360
  readonly "deepseek.r1-v1:0": {
327
361
  id: string;
328
362
  name: string;
@@ -2486,6 +2520,23 @@ export declare const MODELS: {
2486
2520
  contextWindow: number;
2487
2521
  maxTokens: number;
2488
2522
  };
2523
+ readonly "gpt-5.5": {
2524
+ id: string;
2525
+ name: string;
2526
+ api: "azure-openai-responses";
2527
+ provider: string;
2528
+ baseUrl: string;
2529
+ reasoning: true;
2530
+ input: ("image" | "text")[];
2531
+ cost: {
2532
+ input: number;
2533
+ output: number;
2534
+ cacheRead: number;
2535
+ cacheWrite: number;
2536
+ };
2537
+ contextWindow: number;
2538
+ maxTokens: number;
2539
+ };
2489
2540
  readonly o1: {
2490
2541
  id: string;
2491
2542
  name: string;
@@ -3276,6 +3327,29 @@ export declare const MODELS: {
3276
3327
  contextWindow: number;
3277
3328
  maxTokens: number;
3278
3329
  };
3330
+ readonly "gpt-5.5": {
3331
+ id: string;
3332
+ name: string;
3333
+ api: "openai-responses";
3334
+ provider: string;
3335
+ baseUrl: string;
3336
+ headers: {
3337
+ "User-Agent": string;
3338
+ "Editor-Version": string;
3339
+ "Editor-Plugin-Version": string;
3340
+ "Copilot-Integration-Id": string;
3341
+ };
3342
+ reasoning: true;
3343
+ input: ("image" | "text")[];
3344
+ cost: {
3345
+ input: number;
3346
+ output: number;
3347
+ cacheRead: number;
3348
+ cacheWrite: number;
3349
+ };
3350
+ contextWindow: number;
3351
+ maxTokens: number;
3352
+ };
3279
3353
  readonly "grok-code-fast-1": {
3280
3354
  id: string;
3281
3355
  name: string;
@@ -6092,6 +6166,23 @@ export declare const MODELS: {
6092
6166
  contextWindow: number;
6093
6167
  maxTokens: number;
6094
6168
  };
6169
+ readonly "gpt-5.5": {
6170
+ id: string;
6171
+ name: string;
6172
+ api: "openai-responses";
6173
+ provider: string;
6174
+ baseUrl: string;
6175
+ reasoning: true;
6176
+ input: ("image" | "text")[];
6177
+ cost: {
6178
+ input: number;
6179
+ output: number;
6180
+ cacheRead: number;
6181
+ cacheWrite: number;
6182
+ };
6183
+ contextWindow: number;
6184
+ maxTokens: number;
6185
+ };
6095
6186
  readonly o1: {
6096
6187
  id: string;
6097
6188
  name: string;
@@ -6861,6 +6952,40 @@ export declare const MODELS: {
6861
6952
  contextWindow: number;
6862
6953
  maxTokens: number;
6863
6954
  };
6955
+ readonly "gpt-5.5": {
6956
+ id: string;
6957
+ name: string;
6958
+ api: "openai-responses";
6959
+ provider: string;
6960
+ baseUrl: string;
6961
+ reasoning: true;
6962
+ input: ("image" | "text")[];
6963
+ cost: {
6964
+ input: number;
6965
+ output: number;
6966
+ cacheRead: number;
6967
+ cacheWrite: number;
6968
+ };
6969
+ contextWindow: number;
6970
+ maxTokens: number;
6971
+ };
6972
+ readonly "gpt-5.5-pro": {
6973
+ id: string;
6974
+ name: string;
6975
+ api: "openai-responses";
6976
+ provider: string;
6977
+ baseUrl: string;
6978
+ reasoning: true;
6979
+ input: ("image" | "text")[];
6980
+ cost: {
6981
+ input: number;
6982
+ output: number;
6983
+ cacheRead: number;
6984
+ cacheWrite: number;
6985
+ };
6986
+ contextWindow: number;
6987
+ maxTokens: number;
6988
+ };
6864
6989
  readonly "hy3-preview-free": {
6865
6990
  id: string;
6866
6991
  name: string;
@@ -7033,6 +7158,40 @@ export declare const MODELS: {
7033
7158
  };
7034
7159
  };
7035
7160
  readonly "opencode-go": {
7161
+ readonly "deepseek-v4-flash": {
7162
+ id: string;
7163
+ name: string;
7164
+ api: "openai-completions";
7165
+ provider: string;
7166
+ baseUrl: string;
7167
+ reasoning: true;
7168
+ input: "text"[];
7169
+ cost: {
7170
+ input: number;
7171
+ output: number;
7172
+ cacheRead: number;
7173
+ cacheWrite: number;
7174
+ };
7175
+ contextWindow: number;
7176
+ maxTokens: number;
7177
+ };
7178
+ readonly "deepseek-v4-pro": {
7179
+ id: string;
7180
+ name: string;
7181
+ api: "openai-completions";
7182
+ provider: string;
7183
+ baseUrl: string;
7184
+ reasoning: true;
7185
+ input: "text"[];
7186
+ cost: {
7187
+ input: number;
7188
+ output: number;
7189
+ cacheRead: number;
7190
+ cacheWrite: number;
7191
+ };
7192
+ contextWindow: number;
7193
+ maxTokens: number;
7194
+ };
7036
7195
  readonly "glm-5": {
7037
7196
  id: string;
7038
7197
  name: string;
@@ -7970,6 +8129,40 @@ export declare const MODELS: {
7970
8129
  contextWindow: number;
7971
8130
  maxTokens: number;
7972
8131
  };
8132
+ readonly "deepseek/deepseek-v4-flash": {
8133
+ id: string;
8134
+ name: string;
8135
+ api: "openai-completions";
8136
+ provider: string;
8137
+ baseUrl: string;
8138
+ reasoning: true;
8139
+ input: "text"[];
8140
+ cost: {
8141
+ input: number;
8142
+ output: number;
8143
+ cacheRead: number;
8144
+ cacheWrite: number;
8145
+ };
8146
+ contextWindow: number;
8147
+ maxTokens: number;
8148
+ };
8149
+ readonly "deepseek/deepseek-v4-pro": {
8150
+ id: string;
8151
+ name: string;
8152
+ api: "openai-completions";
8153
+ provider: string;
8154
+ baseUrl: string;
8155
+ reasoning: true;
8156
+ input: "text"[];
8157
+ cost: {
8158
+ input: number;
8159
+ output: number;
8160
+ cacheRead: number;
8161
+ cacheWrite: number;
8162
+ };
8163
+ contextWindow: number;
8164
+ maxTokens: number;
8165
+ };
7973
8166
  readonly "essentialai/rnj-1-instruct": {
7974
8167
  id: string;
7975
8168
  name: string;
@@ -8191,6 +8384,40 @@ export declare const MODELS: {
8191
8384
  contextWindow: number;
8192
8385
  maxTokens: number;
8193
8386
  };
8387
+ readonly "google/gemma-3-12b-it": {
8388
+ id: string;
8389
+ name: string;
8390
+ api: "openai-completions";
8391
+ provider: string;
8392
+ baseUrl: string;
8393
+ reasoning: false;
8394
+ input: ("image" | "text")[];
8395
+ cost: {
8396
+ input: number;
8397
+ output: number;
8398
+ cacheRead: number;
8399
+ cacheWrite: number;
8400
+ };
8401
+ contextWindow: number;
8402
+ maxTokens: number;
8403
+ };
8404
+ readonly "google/gemma-3-27b-it": {
8405
+ id: string;
8406
+ name: string;
8407
+ api: "openai-completions";
8408
+ provider: string;
8409
+ baseUrl: string;
8410
+ reasoning: false;
8411
+ input: ("image" | "text")[];
8412
+ cost: {
8413
+ input: number;
8414
+ output: number;
8415
+ cacheRead: number;
8416
+ cacheWrite: number;
8417
+ };
8418
+ contextWindow: number;
8419
+ maxTokens: number;
8420
+ };
8194
8421
  readonly "google/gemma-4-26b-a4b-it": {
8195
8422
  id: string;
8196
8423
  name: string;
@@ -8276,6 +8503,23 @@ export declare const MODELS: {
8276
8503
  contextWindow: number;
8277
8504
  maxTokens: number;
8278
8505
  };
8506
+ readonly "inclusionai/ling-2.6-1t:free": {
8507
+ id: string;
8508
+ name: string;
8509
+ api: "openai-completions";
8510
+ provider: string;
8511
+ baseUrl: string;
8512
+ reasoning: false;
8513
+ input: "text"[];
8514
+ cost: {
8515
+ input: number;
8516
+ output: number;
8517
+ cacheRead: number;
8518
+ cacheWrite: number;
8519
+ };
8520
+ contextWindow: number;
8521
+ maxTokens: number;
8522
+ };
8279
8523
  readonly "inclusionai/ling-2.6-flash:free": {
8280
8524
  id: string;
8281
8525
  name: string;
@@ -9789,14 +10033,14 @@ export declare const MODELS: {
9789
10033
  contextWindow: number;
9790
10034
  maxTokens: number;
9791
10035
  };
9792
- readonly "openai/gpt-audio": {
10036
+ readonly "openai/gpt-5.5": {
9793
10037
  id: string;
9794
10038
  name: string;
9795
10039
  api: "openai-completions";
9796
10040
  provider: string;
9797
10041
  baseUrl: string;
9798
- reasoning: false;
9799
- input: "text"[];
10042
+ reasoning: true;
10043
+ input: ("image" | "text")[];
9800
10044
  cost: {
9801
10045
  input: number;
9802
10046
  output: number;
@@ -9806,14 +10050,14 @@ export declare const MODELS: {
9806
10050
  contextWindow: number;
9807
10051
  maxTokens: number;
9808
10052
  };
9809
- readonly "openai/gpt-audio-mini": {
10053
+ readonly "openai/gpt-5.5-pro": {
9810
10054
  id: string;
9811
10055
  name: string;
9812
10056
  api: "openai-completions";
9813
10057
  provider: string;
9814
10058
  baseUrl: string;
9815
- reasoning: false;
9816
- input: "text"[];
10059
+ reasoning: true;
10060
+ input: ("image" | "text")[];
9817
10061
  cost: {
9818
10062
  input: number;
9819
10063
  output: number;
@@ -9823,13 +10067,13 @@ export declare const MODELS: {
9823
10067
  contextWindow: number;
9824
10068
  maxTokens: number;
9825
10069
  };
9826
- readonly "openai/gpt-oss-120b": {
10070
+ readonly "openai/gpt-audio": {
9827
10071
  id: string;
9828
10072
  name: string;
9829
10073
  api: "openai-completions";
9830
10074
  provider: string;
9831
10075
  baseUrl: string;
9832
- reasoning: true;
10076
+ reasoning: false;
9833
10077
  input: "text"[];
9834
10078
  cost: {
9835
10079
  input: number;
@@ -9840,13 +10084,13 @@ export declare const MODELS: {
9840
10084
  contextWindow: number;
9841
10085
  maxTokens: number;
9842
10086
  };
9843
- readonly "openai/gpt-oss-120b:free": {
10087
+ readonly "openai/gpt-audio-mini": {
9844
10088
  id: string;
9845
10089
  name: string;
9846
10090
  api: "openai-completions";
9847
10091
  provider: string;
9848
10092
  baseUrl: string;
9849
- reasoning: true;
10093
+ reasoning: false;
9850
10094
  input: "text"[];
9851
10095
  cost: {
9852
10096
  input: number;
@@ -9857,7 +10101,7 @@ export declare const MODELS: {
9857
10101
  contextWindow: number;
9858
10102
  maxTokens: number;
9859
10103
  };
9860
- readonly "openai/gpt-oss-20b": {
10104
+ readonly "openai/gpt-oss-120b": {
9861
10105
  id: string;
9862
10106
  name: string;
9863
10107
  api: "openai-completions";
@@ -9874,7 +10118,7 @@ export declare const MODELS: {
9874
10118
  contextWindow: number;
9875
10119
  maxTokens: number;
9876
10120
  };
9877
- readonly "openai/gpt-oss-20b:free": {
10121
+ readonly "openai/gpt-oss-120b:free": {
9878
10122
  id: string;
9879
10123
  name: string;
9880
10124
  api: "openai-completions";
@@ -9891,7 +10135,7 @@ export declare const MODELS: {
9891
10135
  contextWindow: number;
9892
10136
  maxTokens: number;
9893
10137
  };
9894
- readonly "openai/gpt-oss-safeguard-20b": {
10138
+ readonly "openai/gpt-oss-20b": {
9895
10139
  id: string;
9896
10140
  name: string;
9897
10141
  api: "openai-completions";
@@ -9908,7 +10152,41 @@ export declare const MODELS: {
9908
10152
  contextWindow: number;
9909
10153
  maxTokens: number;
9910
10154
  };
9911
- readonly "openai/o1": {
10155
+ readonly "openai/gpt-oss-20b:free": {
10156
+ id: string;
10157
+ name: string;
10158
+ api: "openai-completions";
10159
+ provider: string;
10160
+ baseUrl: string;
10161
+ reasoning: true;
10162
+ input: "text"[];
10163
+ cost: {
10164
+ input: number;
10165
+ output: number;
10166
+ cacheRead: number;
10167
+ cacheWrite: number;
10168
+ };
10169
+ contextWindow: number;
10170
+ maxTokens: number;
10171
+ };
10172
+ readonly "openai/gpt-oss-safeguard-20b": {
10173
+ id: string;
10174
+ name: string;
10175
+ api: "openai-completions";
10176
+ provider: string;
10177
+ baseUrl: string;
10178
+ reasoning: true;
10179
+ input: "text"[];
10180
+ cost: {
10181
+ input: number;
10182
+ output: number;
10183
+ cacheRead: number;
10184
+ cacheWrite: number;
10185
+ };
10186
+ contextWindow: number;
10187
+ maxTokens: number;
10188
+ };
10189
+ readonly "openai/o1": {
9912
10190
  id: string;
9913
10191
  name: string;
9914
10192
  api: "openai-completions";
@@ -12171,6 +12449,40 @@ export declare const MODELS: {
12171
12449
  contextWindow: number;
12172
12450
  maxTokens: number;
12173
12451
  };
12452
+ readonly "deepseek/deepseek-v4-flash": {
12453
+ id: string;
12454
+ name: string;
12455
+ api: "anthropic-messages";
12456
+ provider: string;
12457
+ baseUrl: string;
12458
+ reasoning: true;
12459
+ input: "text"[];
12460
+ cost: {
12461
+ input: number;
12462
+ output: number;
12463
+ cacheRead: number;
12464
+ cacheWrite: number;
12465
+ };
12466
+ contextWindow: number;
12467
+ maxTokens: number;
12468
+ };
12469
+ readonly "deepseek/deepseek-v4-pro": {
12470
+ id: string;
12471
+ name: string;
12472
+ api: "anthropic-messages";
12473
+ provider: string;
12474
+ baseUrl: string;
12475
+ reasoning: true;
12476
+ input: "text"[];
12477
+ cost: {
12478
+ input: number;
12479
+ output: number;
12480
+ cacheRead: number;
12481
+ cacheWrite: number;
12482
+ };
12483
+ contextWindow: number;
12484
+ maxTokens: number;
12485
+ };
12174
12486
  readonly "google/gemini-2.0-flash": {
12175
12487
  id: string;
12176
12488
  name: string;
@@ -13446,6 +13758,40 @@ export declare const MODELS: {
13446
13758
  contextWindow: number;
13447
13759
  maxTokens: number;
13448
13760
  };
13761
+ readonly "openai/gpt-5.5": {
13762
+ id: string;
13763
+ name: string;
13764
+ api: "anthropic-messages";
13765
+ provider: string;
13766
+ baseUrl: string;
13767
+ reasoning: true;
13768
+ input: ("image" | "text")[];
13769
+ cost: {
13770
+ input: number;
13771
+ output: number;
13772
+ cacheRead: number;
13773
+ cacheWrite: number;
13774
+ };
13775
+ contextWindow: number;
13776
+ maxTokens: number;
13777
+ };
13778
+ readonly "openai/gpt-5.5-pro": {
13779
+ id: string;
13780
+ name: string;
13781
+ api: "anthropic-messages";
13782
+ provider: string;
13783
+ baseUrl: string;
13784
+ reasoning: true;
13785
+ input: ("image" | "text")[];
13786
+ cost: {
13787
+ input: number;
13788
+ output: number;
13789
+ cacheRead: number;
13790
+ cacheWrite: number;
13791
+ };
13792
+ contextWindow: number;
13793
+ maxTokens: number;
13794
+ };
13449
13795
  readonly "openai/gpt-oss-20b": {
13450
13796
  id: string;
13451
13797
  name: string;
@@ -14572,27 +14918,6 @@ export declare const MODELS: {
14572
14918
  };
14573
14919
  };
14574
14920
  readonly zai: {
14575
- readonly "glm-4.5": {
14576
- id: string;
14577
- name: string;
14578
- api: "openai-completions";
14579
- provider: string;
14580
- baseUrl: string;
14581
- compat: {
14582
- supportsDeveloperRole: false;
14583
- thinkingFormat: "zai";
14584
- };
14585
- reasoning: true;
14586
- input: "text"[];
14587
- cost: {
14588
- input: number;
14589
- output: number;
14590
- cacheRead: number;
14591
- cacheWrite: number;
14592
- };
14593
- contextWindow: number;
14594
- maxTokens: number;
14595
- };
14596
14921
  readonly "glm-4.5-air": {
14597
14922
  id: string;
14598
14923
  name: string;
@@ -14614,92 +14939,6 @@ export declare const MODELS: {
14614
14939
  contextWindow: number;
14615
14940
  maxTokens: number;
14616
14941
  };
14617
- readonly "glm-4.5-flash": {
14618
- id: string;
14619
- name: string;
14620
- api: "openai-completions";
14621
- provider: string;
14622
- baseUrl: string;
14623
- compat: {
14624
- supportsDeveloperRole: false;
14625
- thinkingFormat: "zai";
14626
- };
14627
- reasoning: true;
14628
- input: "text"[];
14629
- cost: {
14630
- input: number;
14631
- output: number;
14632
- cacheRead: number;
14633
- cacheWrite: number;
14634
- };
14635
- contextWindow: number;
14636
- maxTokens: number;
14637
- };
14638
- readonly "glm-4.5v": {
14639
- id: string;
14640
- name: string;
14641
- api: "openai-completions";
14642
- provider: string;
14643
- baseUrl: string;
14644
- compat: {
14645
- supportsDeveloperRole: false;
14646
- thinkingFormat: "zai";
14647
- };
14648
- reasoning: true;
14649
- input: ("image" | "text")[];
14650
- cost: {
14651
- input: number;
14652
- output: number;
14653
- cacheRead: number;
14654
- cacheWrite: number;
14655
- };
14656
- contextWindow: number;
14657
- maxTokens: number;
14658
- };
14659
- readonly "glm-4.6": {
14660
- id: string;
14661
- name: string;
14662
- api: "openai-completions";
14663
- provider: string;
14664
- baseUrl: string;
14665
- compat: {
14666
- supportsDeveloperRole: false;
14667
- thinkingFormat: "zai";
14668
- zaiToolStream: true;
14669
- };
14670
- reasoning: true;
14671
- input: "text"[];
14672
- cost: {
14673
- input: number;
14674
- output: number;
14675
- cacheRead: number;
14676
- cacheWrite: number;
14677
- };
14678
- contextWindow: number;
14679
- maxTokens: number;
14680
- };
14681
- readonly "glm-4.6v": {
14682
- id: string;
14683
- name: string;
14684
- api: "openai-completions";
14685
- provider: string;
14686
- baseUrl: string;
14687
- compat: {
14688
- supportsDeveloperRole: false;
14689
- thinkingFormat: "zai";
14690
- zaiToolStream: true;
14691
- };
14692
- reasoning: true;
14693
- input: ("image" | "text")[];
14694
- cost: {
14695
- input: number;
14696
- output: number;
14697
- cacheRead: number;
14698
- cacheWrite: number;
14699
- };
14700
- contextWindow: number;
14701
- maxTokens: number;
14702
- };
14703
14942
  readonly "glm-4.7": {
14704
14943
  id: string;
14705
14944
  name: string;
@@ -14722,72 +14961,6 @@ export declare const MODELS: {
14722
14961
  contextWindow: number;
14723
14962
  maxTokens: number;
14724
14963
  };
14725
- readonly "glm-4.7-flash": {
14726
- id: string;
14727
- name: string;
14728
- api: "openai-completions";
14729
- provider: string;
14730
- baseUrl: string;
14731
- compat: {
14732
- supportsDeveloperRole: false;
14733
- thinkingFormat: "zai";
14734
- zaiToolStream: true;
14735
- };
14736
- reasoning: true;
14737
- input: "text"[];
14738
- cost: {
14739
- input: number;
14740
- output: number;
14741
- cacheRead: number;
14742
- cacheWrite: number;
14743
- };
14744
- contextWindow: number;
14745
- maxTokens: number;
14746
- };
14747
- readonly "glm-4.7-flashx": {
14748
- id: string;
14749
- name: string;
14750
- api: "openai-completions";
14751
- provider: string;
14752
- baseUrl: string;
14753
- compat: {
14754
- supportsDeveloperRole: false;
14755
- thinkingFormat: "zai";
14756
- zaiToolStream: true;
14757
- };
14758
- reasoning: true;
14759
- input: "text"[];
14760
- cost: {
14761
- input: number;
14762
- output: number;
14763
- cacheRead: number;
14764
- cacheWrite: number;
14765
- };
14766
- contextWindow: number;
14767
- maxTokens: number;
14768
- };
14769
- readonly "glm-5": {
14770
- id: string;
14771
- name: string;
14772
- api: "openai-completions";
14773
- provider: string;
14774
- baseUrl: string;
14775
- compat: {
14776
- supportsDeveloperRole: false;
14777
- thinkingFormat: "zai";
14778
- zaiToolStream: true;
14779
- };
14780
- reasoning: true;
14781
- input: "text"[];
14782
- cost: {
14783
- input: number;
14784
- output: number;
14785
- cacheRead: number;
14786
- cacheWrite: number;
14787
- };
14788
- contextWindow: number;
14789
- maxTokens: number;
14790
- };
14791
14964
  readonly "glm-5-turbo": {
14792
14965
  id: string;
14793
14966
  name: string;
@@ -14832,28 +15005,6 @@ export declare const MODELS: {
14832
15005
  contextWindow: number;
14833
15006
  maxTokens: number;
14834
15007
  };
14835
- readonly "glm-5v-turbo": {
14836
- id: string;
14837
- name: string;
14838
- api: "openai-completions";
14839
- provider: string;
14840
- baseUrl: string;
14841
- compat: {
14842
- supportsDeveloperRole: false;
14843
- thinkingFormat: "zai";
14844
- zaiToolStream: true;
14845
- };
14846
- reasoning: true;
14847
- input: ("image" | "text")[];
14848
- cost: {
14849
- input: number;
14850
- output: number;
14851
- cacheRead: number;
14852
- cacheWrite: number;
14853
- };
14854
- contextWindow: number;
14855
- maxTokens: number;
14856
- };
14857
15008
  };
14858
15009
  };
14859
15010
  //# sourceMappingURL=models.generated.d.ts.map