@draht/ai 2026.4.23 → 2026.4.25
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +4 -4
- package/dist/cli.js +0 -0
- package/dist/models.d.ts +7 -1
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +360 -209
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +374 -195
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +20 -2
- package/dist/models.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +9 -5
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +6 -3
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +2 -2
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/google-gemini-cli.d.ts.map +1 -1
- package/dist/providers/google-gemini-cli.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/google.d.ts.map +1 -1
- package/dist/providers/google.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +2 -2
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +16 -10
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +2 -2
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/simple-options.d.ts +6 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +10 -1
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/types.d.ts +1 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/oauth/index.d.ts +1 -0
- package/dist/utils/oauth/index.d.ts.map +1 -1
- package/dist/utils/oauth/index.js +4 -0
- package/dist/utils/oauth/index.js.map +1 -1
- package/dist/utils/oauth/opencode-go.d.ts +23 -0
- package/dist/utils/oauth/opencode-go.d.ts.map +1 -0
- package/dist/utils/oauth/opencode-go.js +46 -0
- package/dist/utils/oauth/opencode-go.js.map +1 -0
- package/package.json +1 -1
|
@@ -323,6 +323,40 @@ export declare const MODELS: {
|
|
|
323
323
|
contextWindow: number;
|
|
324
324
|
maxTokens: number;
|
|
325
325
|
};
|
|
326
|
+
readonly "au.anthropic.claude-opus-4-6-v1": {
|
|
327
|
+
id: string;
|
|
328
|
+
name: string;
|
|
329
|
+
api: "bedrock-converse-stream";
|
|
330
|
+
provider: string;
|
|
331
|
+
baseUrl: string;
|
|
332
|
+
reasoning: true;
|
|
333
|
+
input: ("image" | "text")[];
|
|
334
|
+
cost: {
|
|
335
|
+
input: number;
|
|
336
|
+
output: number;
|
|
337
|
+
cacheRead: number;
|
|
338
|
+
cacheWrite: number;
|
|
339
|
+
};
|
|
340
|
+
contextWindow: number;
|
|
341
|
+
maxTokens: number;
|
|
342
|
+
};
|
|
343
|
+
readonly "au.anthropic.claude-sonnet-4-6": {
|
|
344
|
+
id: string;
|
|
345
|
+
name: string;
|
|
346
|
+
api: "bedrock-converse-stream";
|
|
347
|
+
provider: string;
|
|
348
|
+
baseUrl: string;
|
|
349
|
+
reasoning: true;
|
|
350
|
+
input: ("image" | "text")[];
|
|
351
|
+
cost: {
|
|
352
|
+
input: number;
|
|
353
|
+
output: number;
|
|
354
|
+
cacheRead: number;
|
|
355
|
+
cacheWrite: number;
|
|
356
|
+
};
|
|
357
|
+
contextWindow: number;
|
|
358
|
+
maxTokens: number;
|
|
359
|
+
};
|
|
326
360
|
readonly "deepseek.r1-v1:0": {
|
|
327
361
|
id: string;
|
|
328
362
|
name: string;
|
|
@@ -2486,6 +2520,23 @@ export declare const MODELS: {
|
|
|
2486
2520
|
contextWindow: number;
|
|
2487
2521
|
maxTokens: number;
|
|
2488
2522
|
};
|
|
2523
|
+
readonly "gpt-5.5": {
|
|
2524
|
+
id: string;
|
|
2525
|
+
name: string;
|
|
2526
|
+
api: "azure-openai-responses";
|
|
2527
|
+
provider: string;
|
|
2528
|
+
baseUrl: string;
|
|
2529
|
+
reasoning: true;
|
|
2530
|
+
input: ("image" | "text")[];
|
|
2531
|
+
cost: {
|
|
2532
|
+
input: number;
|
|
2533
|
+
output: number;
|
|
2534
|
+
cacheRead: number;
|
|
2535
|
+
cacheWrite: number;
|
|
2536
|
+
};
|
|
2537
|
+
contextWindow: number;
|
|
2538
|
+
maxTokens: number;
|
|
2539
|
+
};
|
|
2489
2540
|
readonly o1: {
|
|
2490
2541
|
id: string;
|
|
2491
2542
|
name: string;
|
|
@@ -3276,6 +3327,29 @@ export declare const MODELS: {
|
|
|
3276
3327
|
contextWindow: number;
|
|
3277
3328
|
maxTokens: number;
|
|
3278
3329
|
};
|
|
3330
|
+
readonly "gpt-5.5": {
|
|
3331
|
+
id: string;
|
|
3332
|
+
name: string;
|
|
3333
|
+
api: "openai-responses";
|
|
3334
|
+
provider: string;
|
|
3335
|
+
baseUrl: string;
|
|
3336
|
+
headers: {
|
|
3337
|
+
"User-Agent": string;
|
|
3338
|
+
"Editor-Version": string;
|
|
3339
|
+
"Editor-Plugin-Version": string;
|
|
3340
|
+
"Copilot-Integration-Id": string;
|
|
3341
|
+
};
|
|
3342
|
+
reasoning: true;
|
|
3343
|
+
input: ("image" | "text")[];
|
|
3344
|
+
cost: {
|
|
3345
|
+
input: number;
|
|
3346
|
+
output: number;
|
|
3347
|
+
cacheRead: number;
|
|
3348
|
+
cacheWrite: number;
|
|
3349
|
+
};
|
|
3350
|
+
contextWindow: number;
|
|
3351
|
+
maxTokens: number;
|
|
3352
|
+
};
|
|
3279
3353
|
readonly "grok-code-fast-1": {
|
|
3280
3354
|
id: string;
|
|
3281
3355
|
name: string;
|
|
@@ -6092,6 +6166,23 @@ export declare const MODELS: {
|
|
|
6092
6166
|
contextWindow: number;
|
|
6093
6167
|
maxTokens: number;
|
|
6094
6168
|
};
|
|
6169
|
+
readonly "gpt-5.5": {
|
|
6170
|
+
id: string;
|
|
6171
|
+
name: string;
|
|
6172
|
+
api: "openai-responses";
|
|
6173
|
+
provider: string;
|
|
6174
|
+
baseUrl: string;
|
|
6175
|
+
reasoning: true;
|
|
6176
|
+
input: ("image" | "text")[];
|
|
6177
|
+
cost: {
|
|
6178
|
+
input: number;
|
|
6179
|
+
output: number;
|
|
6180
|
+
cacheRead: number;
|
|
6181
|
+
cacheWrite: number;
|
|
6182
|
+
};
|
|
6183
|
+
contextWindow: number;
|
|
6184
|
+
maxTokens: number;
|
|
6185
|
+
};
|
|
6095
6186
|
readonly o1: {
|
|
6096
6187
|
id: string;
|
|
6097
6188
|
name: string;
|
|
@@ -6861,6 +6952,40 @@ export declare const MODELS: {
|
|
|
6861
6952
|
contextWindow: number;
|
|
6862
6953
|
maxTokens: number;
|
|
6863
6954
|
};
|
|
6955
|
+
readonly "gpt-5.5": {
|
|
6956
|
+
id: string;
|
|
6957
|
+
name: string;
|
|
6958
|
+
api: "openai-responses";
|
|
6959
|
+
provider: string;
|
|
6960
|
+
baseUrl: string;
|
|
6961
|
+
reasoning: true;
|
|
6962
|
+
input: ("image" | "text")[];
|
|
6963
|
+
cost: {
|
|
6964
|
+
input: number;
|
|
6965
|
+
output: number;
|
|
6966
|
+
cacheRead: number;
|
|
6967
|
+
cacheWrite: number;
|
|
6968
|
+
};
|
|
6969
|
+
contextWindow: number;
|
|
6970
|
+
maxTokens: number;
|
|
6971
|
+
};
|
|
6972
|
+
readonly "gpt-5.5-pro": {
|
|
6973
|
+
id: string;
|
|
6974
|
+
name: string;
|
|
6975
|
+
api: "openai-responses";
|
|
6976
|
+
provider: string;
|
|
6977
|
+
baseUrl: string;
|
|
6978
|
+
reasoning: true;
|
|
6979
|
+
input: ("image" | "text")[];
|
|
6980
|
+
cost: {
|
|
6981
|
+
input: number;
|
|
6982
|
+
output: number;
|
|
6983
|
+
cacheRead: number;
|
|
6984
|
+
cacheWrite: number;
|
|
6985
|
+
};
|
|
6986
|
+
contextWindow: number;
|
|
6987
|
+
maxTokens: number;
|
|
6988
|
+
};
|
|
6864
6989
|
readonly "hy3-preview-free": {
|
|
6865
6990
|
id: string;
|
|
6866
6991
|
name: string;
|
|
@@ -7033,6 +7158,40 @@ export declare const MODELS: {
|
|
|
7033
7158
|
};
|
|
7034
7159
|
};
|
|
7035
7160
|
readonly "opencode-go": {
|
|
7161
|
+
readonly "deepseek-v4-flash": {
|
|
7162
|
+
id: string;
|
|
7163
|
+
name: string;
|
|
7164
|
+
api: "openai-completions";
|
|
7165
|
+
provider: string;
|
|
7166
|
+
baseUrl: string;
|
|
7167
|
+
reasoning: true;
|
|
7168
|
+
input: "text"[];
|
|
7169
|
+
cost: {
|
|
7170
|
+
input: number;
|
|
7171
|
+
output: number;
|
|
7172
|
+
cacheRead: number;
|
|
7173
|
+
cacheWrite: number;
|
|
7174
|
+
};
|
|
7175
|
+
contextWindow: number;
|
|
7176
|
+
maxTokens: number;
|
|
7177
|
+
};
|
|
7178
|
+
readonly "deepseek-v4-pro": {
|
|
7179
|
+
id: string;
|
|
7180
|
+
name: string;
|
|
7181
|
+
api: "openai-completions";
|
|
7182
|
+
provider: string;
|
|
7183
|
+
baseUrl: string;
|
|
7184
|
+
reasoning: true;
|
|
7185
|
+
input: "text"[];
|
|
7186
|
+
cost: {
|
|
7187
|
+
input: number;
|
|
7188
|
+
output: number;
|
|
7189
|
+
cacheRead: number;
|
|
7190
|
+
cacheWrite: number;
|
|
7191
|
+
};
|
|
7192
|
+
contextWindow: number;
|
|
7193
|
+
maxTokens: number;
|
|
7194
|
+
};
|
|
7036
7195
|
readonly "glm-5": {
|
|
7037
7196
|
id: string;
|
|
7038
7197
|
name: string;
|
|
@@ -7970,6 +8129,40 @@ export declare const MODELS: {
|
|
|
7970
8129
|
contextWindow: number;
|
|
7971
8130
|
maxTokens: number;
|
|
7972
8131
|
};
|
|
8132
|
+
readonly "deepseek/deepseek-v4-flash": {
|
|
8133
|
+
id: string;
|
|
8134
|
+
name: string;
|
|
8135
|
+
api: "openai-completions";
|
|
8136
|
+
provider: string;
|
|
8137
|
+
baseUrl: string;
|
|
8138
|
+
reasoning: true;
|
|
8139
|
+
input: "text"[];
|
|
8140
|
+
cost: {
|
|
8141
|
+
input: number;
|
|
8142
|
+
output: number;
|
|
8143
|
+
cacheRead: number;
|
|
8144
|
+
cacheWrite: number;
|
|
8145
|
+
};
|
|
8146
|
+
contextWindow: number;
|
|
8147
|
+
maxTokens: number;
|
|
8148
|
+
};
|
|
8149
|
+
readonly "deepseek/deepseek-v4-pro": {
|
|
8150
|
+
id: string;
|
|
8151
|
+
name: string;
|
|
8152
|
+
api: "openai-completions";
|
|
8153
|
+
provider: string;
|
|
8154
|
+
baseUrl: string;
|
|
8155
|
+
reasoning: true;
|
|
8156
|
+
input: "text"[];
|
|
8157
|
+
cost: {
|
|
8158
|
+
input: number;
|
|
8159
|
+
output: number;
|
|
8160
|
+
cacheRead: number;
|
|
8161
|
+
cacheWrite: number;
|
|
8162
|
+
};
|
|
8163
|
+
contextWindow: number;
|
|
8164
|
+
maxTokens: number;
|
|
8165
|
+
};
|
|
7973
8166
|
readonly "essentialai/rnj-1-instruct": {
|
|
7974
8167
|
id: string;
|
|
7975
8168
|
name: string;
|
|
@@ -8191,6 +8384,40 @@ export declare const MODELS: {
|
|
|
8191
8384
|
contextWindow: number;
|
|
8192
8385
|
maxTokens: number;
|
|
8193
8386
|
};
|
|
8387
|
+
readonly "google/gemma-3-12b-it": {
|
|
8388
|
+
id: string;
|
|
8389
|
+
name: string;
|
|
8390
|
+
api: "openai-completions";
|
|
8391
|
+
provider: string;
|
|
8392
|
+
baseUrl: string;
|
|
8393
|
+
reasoning: false;
|
|
8394
|
+
input: ("image" | "text")[];
|
|
8395
|
+
cost: {
|
|
8396
|
+
input: number;
|
|
8397
|
+
output: number;
|
|
8398
|
+
cacheRead: number;
|
|
8399
|
+
cacheWrite: number;
|
|
8400
|
+
};
|
|
8401
|
+
contextWindow: number;
|
|
8402
|
+
maxTokens: number;
|
|
8403
|
+
};
|
|
8404
|
+
readonly "google/gemma-3-27b-it": {
|
|
8405
|
+
id: string;
|
|
8406
|
+
name: string;
|
|
8407
|
+
api: "openai-completions";
|
|
8408
|
+
provider: string;
|
|
8409
|
+
baseUrl: string;
|
|
8410
|
+
reasoning: false;
|
|
8411
|
+
input: ("image" | "text")[];
|
|
8412
|
+
cost: {
|
|
8413
|
+
input: number;
|
|
8414
|
+
output: number;
|
|
8415
|
+
cacheRead: number;
|
|
8416
|
+
cacheWrite: number;
|
|
8417
|
+
};
|
|
8418
|
+
contextWindow: number;
|
|
8419
|
+
maxTokens: number;
|
|
8420
|
+
};
|
|
8194
8421
|
readonly "google/gemma-4-26b-a4b-it": {
|
|
8195
8422
|
id: string;
|
|
8196
8423
|
name: string;
|
|
@@ -8276,6 +8503,23 @@ export declare const MODELS: {
|
|
|
8276
8503
|
contextWindow: number;
|
|
8277
8504
|
maxTokens: number;
|
|
8278
8505
|
};
|
|
8506
|
+
readonly "inclusionai/ling-2.6-1t:free": {
|
|
8507
|
+
id: string;
|
|
8508
|
+
name: string;
|
|
8509
|
+
api: "openai-completions";
|
|
8510
|
+
provider: string;
|
|
8511
|
+
baseUrl: string;
|
|
8512
|
+
reasoning: false;
|
|
8513
|
+
input: "text"[];
|
|
8514
|
+
cost: {
|
|
8515
|
+
input: number;
|
|
8516
|
+
output: number;
|
|
8517
|
+
cacheRead: number;
|
|
8518
|
+
cacheWrite: number;
|
|
8519
|
+
};
|
|
8520
|
+
contextWindow: number;
|
|
8521
|
+
maxTokens: number;
|
|
8522
|
+
};
|
|
8279
8523
|
readonly "inclusionai/ling-2.6-flash:free": {
|
|
8280
8524
|
id: string;
|
|
8281
8525
|
name: string;
|
|
@@ -9789,14 +10033,14 @@ export declare const MODELS: {
|
|
|
9789
10033
|
contextWindow: number;
|
|
9790
10034
|
maxTokens: number;
|
|
9791
10035
|
};
|
|
9792
|
-
readonly "openai/gpt-
|
|
10036
|
+
readonly "openai/gpt-5.5": {
|
|
9793
10037
|
id: string;
|
|
9794
10038
|
name: string;
|
|
9795
10039
|
api: "openai-completions";
|
|
9796
10040
|
provider: string;
|
|
9797
10041
|
baseUrl: string;
|
|
9798
|
-
reasoning:
|
|
9799
|
-
input: "text"[];
|
|
10042
|
+
reasoning: true;
|
|
10043
|
+
input: ("image" | "text")[];
|
|
9800
10044
|
cost: {
|
|
9801
10045
|
input: number;
|
|
9802
10046
|
output: number;
|
|
@@ -9806,14 +10050,14 @@ export declare const MODELS: {
|
|
|
9806
10050
|
contextWindow: number;
|
|
9807
10051
|
maxTokens: number;
|
|
9808
10052
|
};
|
|
9809
|
-
readonly "openai/gpt-
|
|
10053
|
+
readonly "openai/gpt-5.5-pro": {
|
|
9810
10054
|
id: string;
|
|
9811
10055
|
name: string;
|
|
9812
10056
|
api: "openai-completions";
|
|
9813
10057
|
provider: string;
|
|
9814
10058
|
baseUrl: string;
|
|
9815
|
-
reasoning:
|
|
9816
|
-
input: "text"[];
|
|
10059
|
+
reasoning: true;
|
|
10060
|
+
input: ("image" | "text")[];
|
|
9817
10061
|
cost: {
|
|
9818
10062
|
input: number;
|
|
9819
10063
|
output: number;
|
|
@@ -9823,13 +10067,13 @@ export declare const MODELS: {
|
|
|
9823
10067
|
contextWindow: number;
|
|
9824
10068
|
maxTokens: number;
|
|
9825
10069
|
};
|
|
9826
|
-
readonly "openai/gpt-
|
|
10070
|
+
readonly "openai/gpt-audio": {
|
|
9827
10071
|
id: string;
|
|
9828
10072
|
name: string;
|
|
9829
10073
|
api: "openai-completions";
|
|
9830
10074
|
provider: string;
|
|
9831
10075
|
baseUrl: string;
|
|
9832
|
-
reasoning:
|
|
10076
|
+
reasoning: false;
|
|
9833
10077
|
input: "text"[];
|
|
9834
10078
|
cost: {
|
|
9835
10079
|
input: number;
|
|
@@ -9840,13 +10084,13 @@ export declare const MODELS: {
|
|
|
9840
10084
|
contextWindow: number;
|
|
9841
10085
|
maxTokens: number;
|
|
9842
10086
|
};
|
|
9843
|
-
readonly "openai/gpt-
|
|
10087
|
+
readonly "openai/gpt-audio-mini": {
|
|
9844
10088
|
id: string;
|
|
9845
10089
|
name: string;
|
|
9846
10090
|
api: "openai-completions";
|
|
9847
10091
|
provider: string;
|
|
9848
10092
|
baseUrl: string;
|
|
9849
|
-
reasoning:
|
|
10093
|
+
reasoning: false;
|
|
9850
10094
|
input: "text"[];
|
|
9851
10095
|
cost: {
|
|
9852
10096
|
input: number;
|
|
@@ -9857,7 +10101,7 @@ export declare const MODELS: {
|
|
|
9857
10101
|
contextWindow: number;
|
|
9858
10102
|
maxTokens: number;
|
|
9859
10103
|
};
|
|
9860
|
-
readonly "openai/gpt-oss-
|
|
10104
|
+
readonly "openai/gpt-oss-120b": {
|
|
9861
10105
|
id: string;
|
|
9862
10106
|
name: string;
|
|
9863
10107
|
api: "openai-completions";
|
|
@@ -9874,7 +10118,7 @@ export declare const MODELS: {
|
|
|
9874
10118
|
contextWindow: number;
|
|
9875
10119
|
maxTokens: number;
|
|
9876
10120
|
};
|
|
9877
|
-
readonly "openai/gpt-oss-
|
|
10121
|
+
readonly "openai/gpt-oss-120b:free": {
|
|
9878
10122
|
id: string;
|
|
9879
10123
|
name: string;
|
|
9880
10124
|
api: "openai-completions";
|
|
@@ -9891,7 +10135,7 @@ export declare const MODELS: {
|
|
|
9891
10135
|
contextWindow: number;
|
|
9892
10136
|
maxTokens: number;
|
|
9893
10137
|
};
|
|
9894
|
-
readonly "openai/gpt-oss-
|
|
10138
|
+
readonly "openai/gpt-oss-20b": {
|
|
9895
10139
|
id: string;
|
|
9896
10140
|
name: string;
|
|
9897
10141
|
api: "openai-completions";
|
|
@@ -9908,7 +10152,41 @@ export declare const MODELS: {
|
|
|
9908
10152
|
contextWindow: number;
|
|
9909
10153
|
maxTokens: number;
|
|
9910
10154
|
};
|
|
9911
|
-
readonly "openai/
|
|
10155
|
+
readonly "openai/gpt-oss-20b:free": {
|
|
10156
|
+
id: string;
|
|
10157
|
+
name: string;
|
|
10158
|
+
api: "openai-completions";
|
|
10159
|
+
provider: string;
|
|
10160
|
+
baseUrl: string;
|
|
10161
|
+
reasoning: true;
|
|
10162
|
+
input: "text"[];
|
|
10163
|
+
cost: {
|
|
10164
|
+
input: number;
|
|
10165
|
+
output: number;
|
|
10166
|
+
cacheRead: number;
|
|
10167
|
+
cacheWrite: number;
|
|
10168
|
+
};
|
|
10169
|
+
contextWindow: number;
|
|
10170
|
+
maxTokens: number;
|
|
10171
|
+
};
|
|
10172
|
+
readonly "openai/gpt-oss-safeguard-20b": {
|
|
10173
|
+
id: string;
|
|
10174
|
+
name: string;
|
|
10175
|
+
api: "openai-completions";
|
|
10176
|
+
provider: string;
|
|
10177
|
+
baseUrl: string;
|
|
10178
|
+
reasoning: true;
|
|
10179
|
+
input: "text"[];
|
|
10180
|
+
cost: {
|
|
10181
|
+
input: number;
|
|
10182
|
+
output: number;
|
|
10183
|
+
cacheRead: number;
|
|
10184
|
+
cacheWrite: number;
|
|
10185
|
+
};
|
|
10186
|
+
contextWindow: number;
|
|
10187
|
+
maxTokens: number;
|
|
10188
|
+
};
|
|
10189
|
+
readonly "openai/o1": {
|
|
9912
10190
|
id: string;
|
|
9913
10191
|
name: string;
|
|
9914
10192
|
api: "openai-completions";
|
|
@@ -12171,6 +12449,40 @@ export declare const MODELS: {
|
|
|
12171
12449
|
contextWindow: number;
|
|
12172
12450
|
maxTokens: number;
|
|
12173
12451
|
};
|
|
12452
|
+
readonly "deepseek/deepseek-v4-flash": {
|
|
12453
|
+
id: string;
|
|
12454
|
+
name: string;
|
|
12455
|
+
api: "anthropic-messages";
|
|
12456
|
+
provider: string;
|
|
12457
|
+
baseUrl: string;
|
|
12458
|
+
reasoning: true;
|
|
12459
|
+
input: "text"[];
|
|
12460
|
+
cost: {
|
|
12461
|
+
input: number;
|
|
12462
|
+
output: number;
|
|
12463
|
+
cacheRead: number;
|
|
12464
|
+
cacheWrite: number;
|
|
12465
|
+
};
|
|
12466
|
+
contextWindow: number;
|
|
12467
|
+
maxTokens: number;
|
|
12468
|
+
};
|
|
12469
|
+
readonly "deepseek/deepseek-v4-pro": {
|
|
12470
|
+
id: string;
|
|
12471
|
+
name: string;
|
|
12472
|
+
api: "anthropic-messages";
|
|
12473
|
+
provider: string;
|
|
12474
|
+
baseUrl: string;
|
|
12475
|
+
reasoning: true;
|
|
12476
|
+
input: "text"[];
|
|
12477
|
+
cost: {
|
|
12478
|
+
input: number;
|
|
12479
|
+
output: number;
|
|
12480
|
+
cacheRead: number;
|
|
12481
|
+
cacheWrite: number;
|
|
12482
|
+
};
|
|
12483
|
+
contextWindow: number;
|
|
12484
|
+
maxTokens: number;
|
|
12485
|
+
};
|
|
12174
12486
|
readonly "google/gemini-2.0-flash": {
|
|
12175
12487
|
id: string;
|
|
12176
12488
|
name: string;
|
|
@@ -13446,6 +13758,40 @@ export declare const MODELS: {
|
|
|
13446
13758
|
contextWindow: number;
|
|
13447
13759
|
maxTokens: number;
|
|
13448
13760
|
};
|
|
13761
|
+
readonly "openai/gpt-5.5": {
|
|
13762
|
+
id: string;
|
|
13763
|
+
name: string;
|
|
13764
|
+
api: "anthropic-messages";
|
|
13765
|
+
provider: string;
|
|
13766
|
+
baseUrl: string;
|
|
13767
|
+
reasoning: true;
|
|
13768
|
+
input: ("image" | "text")[];
|
|
13769
|
+
cost: {
|
|
13770
|
+
input: number;
|
|
13771
|
+
output: number;
|
|
13772
|
+
cacheRead: number;
|
|
13773
|
+
cacheWrite: number;
|
|
13774
|
+
};
|
|
13775
|
+
contextWindow: number;
|
|
13776
|
+
maxTokens: number;
|
|
13777
|
+
};
|
|
13778
|
+
readonly "openai/gpt-5.5-pro": {
|
|
13779
|
+
id: string;
|
|
13780
|
+
name: string;
|
|
13781
|
+
api: "anthropic-messages";
|
|
13782
|
+
provider: string;
|
|
13783
|
+
baseUrl: string;
|
|
13784
|
+
reasoning: true;
|
|
13785
|
+
input: ("image" | "text")[];
|
|
13786
|
+
cost: {
|
|
13787
|
+
input: number;
|
|
13788
|
+
output: number;
|
|
13789
|
+
cacheRead: number;
|
|
13790
|
+
cacheWrite: number;
|
|
13791
|
+
};
|
|
13792
|
+
contextWindow: number;
|
|
13793
|
+
maxTokens: number;
|
|
13794
|
+
};
|
|
13449
13795
|
readonly "openai/gpt-oss-20b": {
|
|
13450
13796
|
id: string;
|
|
13451
13797
|
name: string;
|
|
@@ -14572,27 +14918,6 @@ export declare const MODELS: {
|
|
|
14572
14918
|
};
|
|
14573
14919
|
};
|
|
14574
14920
|
readonly zai: {
|
|
14575
|
-
readonly "glm-4.5": {
|
|
14576
|
-
id: string;
|
|
14577
|
-
name: string;
|
|
14578
|
-
api: "openai-completions";
|
|
14579
|
-
provider: string;
|
|
14580
|
-
baseUrl: string;
|
|
14581
|
-
compat: {
|
|
14582
|
-
supportsDeveloperRole: false;
|
|
14583
|
-
thinkingFormat: "zai";
|
|
14584
|
-
};
|
|
14585
|
-
reasoning: true;
|
|
14586
|
-
input: "text"[];
|
|
14587
|
-
cost: {
|
|
14588
|
-
input: number;
|
|
14589
|
-
output: number;
|
|
14590
|
-
cacheRead: number;
|
|
14591
|
-
cacheWrite: number;
|
|
14592
|
-
};
|
|
14593
|
-
contextWindow: number;
|
|
14594
|
-
maxTokens: number;
|
|
14595
|
-
};
|
|
14596
14921
|
readonly "glm-4.5-air": {
|
|
14597
14922
|
id: string;
|
|
14598
14923
|
name: string;
|
|
@@ -14614,92 +14939,6 @@ export declare const MODELS: {
|
|
|
14614
14939
|
contextWindow: number;
|
|
14615
14940
|
maxTokens: number;
|
|
14616
14941
|
};
|
|
14617
|
-
readonly "glm-4.5-flash": {
|
|
14618
|
-
id: string;
|
|
14619
|
-
name: string;
|
|
14620
|
-
api: "openai-completions";
|
|
14621
|
-
provider: string;
|
|
14622
|
-
baseUrl: string;
|
|
14623
|
-
compat: {
|
|
14624
|
-
supportsDeveloperRole: false;
|
|
14625
|
-
thinkingFormat: "zai";
|
|
14626
|
-
};
|
|
14627
|
-
reasoning: true;
|
|
14628
|
-
input: "text"[];
|
|
14629
|
-
cost: {
|
|
14630
|
-
input: number;
|
|
14631
|
-
output: number;
|
|
14632
|
-
cacheRead: number;
|
|
14633
|
-
cacheWrite: number;
|
|
14634
|
-
};
|
|
14635
|
-
contextWindow: number;
|
|
14636
|
-
maxTokens: number;
|
|
14637
|
-
};
|
|
14638
|
-
readonly "glm-4.5v": {
|
|
14639
|
-
id: string;
|
|
14640
|
-
name: string;
|
|
14641
|
-
api: "openai-completions";
|
|
14642
|
-
provider: string;
|
|
14643
|
-
baseUrl: string;
|
|
14644
|
-
compat: {
|
|
14645
|
-
supportsDeveloperRole: false;
|
|
14646
|
-
thinkingFormat: "zai";
|
|
14647
|
-
};
|
|
14648
|
-
reasoning: true;
|
|
14649
|
-
input: ("image" | "text")[];
|
|
14650
|
-
cost: {
|
|
14651
|
-
input: number;
|
|
14652
|
-
output: number;
|
|
14653
|
-
cacheRead: number;
|
|
14654
|
-
cacheWrite: number;
|
|
14655
|
-
};
|
|
14656
|
-
contextWindow: number;
|
|
14657
|
-
maxTokens: number;
|
|
14658
|
-
};
|
|
14659
|
-
readonly "glm-4.6": {
|
|
14660
|
-
id: string;
|
|
14661
|
-
name: string;
|
|
14662
|
-
api: "openai-completions";
|
|
14663
|
-
provider: string;
|
|
14664
|
-
baseUrl: string;
|
|
14665
|
-
compat: {
|
|
14666
|
-
supportsDeveloperRole: false;
|
|
14667
|
-
thinkingFormat: "zai";
|
|
14668
|
-
zaiToolStream: true;
|
|
14669
|
-
};
|
|
14670
|
-
reasoning: true;
|
|
14671
|
-
input: "text"[];
|
|
14672
|
-
cost: {
|
|
14673
|
-
input: number;
|
|
14674
|
-
output: number;
|
|
14675
|
-
cacheRead: number;
|
|
14676
|
-
cacheWrite: number;
|
|
14677
|
-
};
|
|
14678
|
-
contextWindow: number;
|
|
14679
|
-
maxTokens: number;
|
|
14680
|
-
};
|
|
14681
|
-
readonly "glm-4.6v": {
|
|
14682
|
-
id: string;
|
|
14683
|
-
name: string;
|
|
14684
|
-
api: "openai-completions";
|
|
14685
|
-
provider: string;
|
|
14686
|
-
baseUrl: string;
|
|
14687
|
-
compat: {
|
|
14688
|
-
supportsDeveloperRole: false;
|
|
14689
|
-
thinkingFormat: "zai";
|
|
14690
|
-
zaiToolStream: true;
|
|
14691
|
-
};
|
|
14692
|
-
reasoning: true;
|
|
14693
|
-
input: ("image" | "text")[];
|
|
14694
|
-
cost: {
|
|
14695
|
-
input: number;
|
|
14696
|
-
output: number;
|
|
14697
|
-
cacheRead: number;
|
|
14698
|
-
cacheWrite: number;
|
|
14699
|
-
};
|
|
14700
|
-
contextWindow: number;
|
|
14701
|
-
maxTokens: number;
|
|
14702
|
-
};
|
|
14703
14942
|
readonly "glm-4.7": {
|
|
14704
14943
|
id: string;
|
|
14705
14944
|
name: string;
|
|
@@ -14722,72 +14961,6 @@ export declare const MODELS: {
|
|
|
14722
14961
|
contextWindow: number;
|
|
14723
14962
|
maxTokens: number;
|
|
14724
14963
|
};
|
|
14725
|
-
readonly "glm-4.7-flash": {
|
|
14726
|
-
id: string;
|
|
14727
|
-
name: string;
|
|
14728
|
-
api: "openai-completions";
|
|
14729
|
-
provider: string;
|
|
14730
|
-
baseUrl: string;
|
|
14731
|
-
compat: {
|
|
14732
|
-
supportsDeveloperRole: false;
|
|
14733
|
-
thinkingFormat: "zai";
|
|
14734
|
-
zaiToolStream: true;
|
|
14735
|
-
};
|
|
14736
|
-
reasoning: true;
|
|
14737
|
-
input: "text"[];
|
|
14738
|
-
cost: {
|
|
14739
|
-
input: number;
|
|
14740
|
-
output: number;
|
|
14741
|
-
cacheRead: number;
|
|
14742
|
-
cacheWrite: number;
|
|
14743
|
-
};
|
|
14744
|
-
contextWindow: number;
|
|
14745
|
-
maxTokens: number;
|
|
14746
|
-
};
|
|
14747
|
-
readonly "glm-4.7-flashx": {
|
|
14748
|
-
id: string;
|
|
14749
|
-
name: string;
|
|
14750
|
-
api: "openai-completions";
|
|
14751
|
-
provider: string;
|
|
14752
|
-
baseUrl: string;
|
|
14753
|
-
compat: {
|
|
14754
|
-
supportsDeveloperRole: false;
|
|
14755
|
-
thinkingFormat: "zai";
|
|
14756
|
-
zaiToolStream: true;
|
|
14757
|
-
};
|
|
14758
|
-
reasoning: true;
|
|
14759
|
-
input: "text"[];
|
|
14760
|
-
cost: {
|
|
14761
|
-
input: number;
|
|
14762
|
-
output: number;
|
|
14763
|
-
cacheRead: number;
|
|
14764
|
-
cacheWrite: number;
|
|
14765
|
-
};
|
|
14766
|
-
contextWindow: number;
|
|
14767
|
-
maxTokens: number;
|
|
14768
|
-
};
|
|
14769
|
-
readonly "glm-5": {
|
|
14770
|
-
id: string;
|
|
14771
|
-
name: string;
|
|
14772
|
-
api: "openai-completions";
|
|
14773
|
-
provider: string;
|
|
14774
|
-
baseUrl: string;
|
|
14775
|
-
compat: {
|
|
14776
|
-
supportsDeveloperRole: false;
|
|
14777
|
-
thinkingFormat: "zai";
|
|
14778
|
-
zaiToolStream: true;
|
|
14779
|
-
};
|
|
14780
|
-
reasoning: true;
|
|
14781
|
-
input: "text"[];
|
|
14782
|
-
cost: {
|
|
14783
|
-
input: number;
|
|
14784
|
-
output: number;
|
|
14785
|
-
cacheRead: number;
|
|
14786
|
-
cacheWrite: number;
|
|
14787
|
-
};
|
|
14788
|
-
contextWindow: number;
|
|
14789
|
-
maxTokens: number;
|
|
14790
|
-
};
|
|
14791
14964
|
readonly "glm-5-turbo": {
|
|
14792
14965
|
id: string;
|
|
14793
14966
|
name: string;
|
|
@@ -14832,28 +15005,6 @@ export declare const MODELS: {
|
|
|
14832
15005
|
contextWindow: number;
|
|
14833
15006
|
maxTokens: number;
|
|
14834
15007
|
};
|
|
14835
|
-
readonly "glm-5v-turbo": {
|
|
14836
|
-
id: string;
|
|
14837
|
-
name: string;
|
|
14838
|
-
api: "openai-completions";
|
|
14839
|
-
provider: string;
|
|
14840
|
-
baseUrl: string;
|
|
14841
|
-
compat: {
|
|
14842
|
-
supportsDeveloperRole: false;
|
|
14843
|
-
thinkingFormat: "zai";
|
|
14844
|
-
zaiToolStream: true;
|
|
14845
|
-
};
|
|
14846
|
-
reasoning: true;
|
|
14847
|
-
input: ("image" | "text")[];
|
|
14848
|
-
cost: {
|
|
14849
|
-
input: number;
|
|
14850
|
-
output: number;
|
|
14851
|
-
cacheRead: number;
|
|
14852
|
-
cacheWrite: number;
|
|
14853
|
-
};
|
|
14854
|
-
contextWindow: number;
|
|
14855
|
-
maxTokens: number;
|
|
14856
|
-
};
|
|
14857
15008
|
};
|
|
14858
15009
|
};
|
|
14859
15010
|
//# sourceMappingURL=models.generated.d.ts.map
|