@avadisabelle/ava-pi-ai 0.64.1 → 0.64.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +379 -97
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +498 -223
- package/dist/models.generated.js.map +1 -1
- package/package.json +1 -1
|
@@ -255,6 +255,23 @@ export declare const MODELS: {
|
|
|
255
255
|
contextWindow: number;
|
|
256
256
|
maxTokens: number;
|
|
257
257
|
};
|
|
258
|
+
readonly "anthropic.claude-opus-4-7": {
|
|
259
|
+
id: string;
|
|
260
|
+
name: string;
|
|
261
|
+
api: "bedrock-converse-stream";
|
|
262
|
+
provider: string;
|
|
263
|
+
baseUrl: string;
|
|
264
|
+
reasoning: true;
|
|
265
|
+
input: ("image" | "text")[];
|
|
266
|
+
cost: {
|
|
267
|
+
input: number;
|
|
268
|
+
output: number;
|
|
269
|
+
cacheRead: number;
|
|
270
|
+
cacheWrite: number;
|
|
271
|
+
};
|
|
272
|
+
contextWindow: number;
|
|
273
|
+
maxTokens: number;
|
|
274
|
+
};
|
|
258
275
|
readonly "anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
259
276
|
id: string;
|
|
260
277
|
name: string;
|
|
@@ -408,6 +425,23 @@ export declare const MODELS: {
|
|
|
408
425
|
contextWindow: number;
|
|
409
426
|
maxTokens: number;
|
|
410
427
|
};
|
|
428
|
+
readonly "eu.anthropic.claude-opus-4-7": {
|
|
429
|
+
id: string;
|
|
430
|
+
name: string;
|
|
431
|
+
api: "bedrock-converse-stream";
|
|
432
|
+
provider: string;
|
|
433
|
+
baseUrl: string;
|
|
434
|
+
reasoning: true;
|
|
435
|
+
input: ("image" | "text")[];
|
|
436
|
+
cost: {
|
|
437
|
+
input: number;
|
|
438
|
+
output: number;
|
|
439
|
+
cacheRead: number;
|
|
440
|
+
cacheWrite: number;
|
|
441
|
+
};
|
|
442
|
+
contextWindow: number;
|
|
443
|
+
maxTokens: number;
|
|
444
|
+
};
|
|
411
445
|
readonly "eu.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
412
446
|
id: string;
|
|
413
447
|
name: string;
|
|
@@ -510,6 +544,23 @@ export declare const MODELS: {
|
|
|
510
544
|
contextWindow: number;
|
|
511
545
|
maxTokens: number;
|
|
512
546
|
};
|
|
547
|
+
readonly "global.anthropic.claude-opus-4-7": {
|
|
548
|
+
id: string;
|
|
549
|
+
name: string;
|
|
550
|
+
api: "bedrock-converse-stream";
|
|
551
|
+
provider: string;
|
|
552
|
+
baseUrl: string;
|
|
553
|
+
reasoning: true;
|
|
554
|
+
input: ("image" | "text")[];
|
|
555
|
+
cost: {
|
|
556
|
+
input: number;
|
|
557
|
+
output: number;
|
|
558
|
+
cacheRead: number;
|
|
559
|
+
cacheWrite: number;
|
|
560
|
+
};
|
|
561
|
+
contextWindow: number;
|
|
562
|
+
maxTokens: number;
|
|
563
|
+
};
|
|
513
564
|
readonly "global.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
514
565
|
id: string;
|
|
515
566
|
name: string;
|
|
@@ -1207,6 +1258,23 @@ export declare const MODELS: {
|
|
|
1207
1258
|
contextWindow: number;
|
|
1208
1259
|
maxTokens: number;
|
|
1209
1260
|
};
|
|
1261
|
+
readonly "qwen.qwen3-coder-next": {
|
|
1262
|
+
id: string;
|
|
1263
|
+
name: string;
|
|
1264
|
+
api: "bedrock-converse-stream";
|
|
1265
|
+
provider: string;
|
|
1266
|
+
baseUrl: string;
|
|
1267
|
+
reasoning: true;
|
|
1268
|
+
input: "text"[];
|
|
1269
|
+
cost: {
|
|
1270
|
+
input: number;
|
|
1271
|
+
output: number;
|
|
1272
|
+
cacheRead: number;
|
|
1273
|
+
cacheWrite: number;
|
|
1274
|
+
};
|
|
1275
|
+
contextWindow: number;
|
|
1276
|
+
maxTokens: number;
|
|
1277
|
+
};
|
|
1210
1278
|
readonly "qwen.qwen3-next-80b-a3b": {
|
|
1211
1279
|
id: string;
|
|
1212
1280
|
name: string;
|
|
@@ -1326,6 +1394,23 @@ export declare const MODELS: {
|
|
|
1326
1394
|
contextWindow: number;
|
|
1327
1395
|
maxTokens: number;
|
|
1328
1396
|
};
|
|
1397
|
+
readonly "us.anthropic.claude-opus-4-7": {
|
|
1398
|
+
id: string;
|
|
1399
|
+
name: string;
|
|
1400
|
+
api: "bedrock-converse-stream";
|
|
1401
|
+
provider: string;
|
|
1402
|
+
baseUrl: string;
|
|
1403
|
+
reasoning: true;
|
|
1404
|
+
input: ("image" | "text")[];
|
|
1405
|
+
cost: {
|
|
1406
|
+
input: number;
|
|
1407
|
+
output: number;
|
|
1408
|
+
cacheRead: number;
|
|
1409
|
+
cacheWrite: number;
|
|
1410
|
+
};
|
|
1411
|
+
contextWindow: number;
|
|
1412
|
+
maxTokens: number;
|
|
1413
|
+
};
|
|
1329
1414
|
readonly "us.anthropic.claude-sonnet-4-20250514-v1:0": {
|
|
1330
1415
|
id: string;
|
|
1331
1416
|
name: string;
|
|
@@ -1753,6 +1838,23 @@ export declare const MODELS: {
|
|
|
1753
1838
|
contextWindow: number;
|
|
1754
1839
|
maxTokens: number;
|
|
1755
1840
|
};
|
|
1841
|
+
readonly "claude-opus-4-7": {
|
|
1842
|
+
id: string;
|
|
1843
|
+
name: string;
|
|
1844
|
+
api: "anthropic-messages";
|
|
1845
|
+
provider: string;
|
|
1846
|
+
baseUrl: string;
|
|
1847
|
+
reasoning: true;
|
|
1848
|
+
input: ("image" | "text")[];
|
|
1849
|
+
cost: {
|
|
1850
|
+
input: number;
|
|
1851
|
+
output: number;
|
|
1852
|
+
cacheRead: number;
|
|
1853
|
+
cacheWrite: number;
|
|
1854
|
+
};
|
|
1855
|
+
contextWindow: number;
|
|
1856
|
+
maxTokens: number;
|
|
1857
|
+
};
|
|
1756
1858
|
readonly "claude-sonnet-4-0": {
|
|
1757
1859
|
id: string;
|
|
1758
1860
|
name: string;
|
|
@@ -3623,7 +3725,7 @@ export declare const MODELS: {
|
|
|
3623
3725
|
contextWindow: number;
|
|
3624
3726
|
maxTokens: number;
|
|
3625
3727
|
};
|
|
3626
|
-
readonly "gemma-4-26b": {
|
|
3728
|
+
readonly "gemma-4-26b-it": {
|
|
3627
3729
|
id: string;
|
|
3628
3730
|
name: string;
|
|
3629
3731
|
api: "google-generative-ai";
|
|
@@ -3640,7 +3742,7 @@ export declare const MODELS: {
|
|
|
3640
3742
|
contextWindow: number;
|
|
3641
3743
|
maxTokens: number;
|
|
3642
3744
|
};
|
|
3643
|
-
readonly "gemma-4-31b": {
|
|
3745
|
+
readonly "gemma-4-31b-it": {
|
|
3644
3746
|
id: string;
|
|
3645
3747
|
name: string;
|
|
3646
3748
|
api: "google-generative-ai";
|
|
@@ -4472,6 +4574,26 @@ export declare const MODELS: {
|
|
|
4472
4574
|
contextWindow: number;
|
|
4473
4575
|
maxTokens: number;
|
|
4474
4576
|
};
|
|
4577
|
+
readonly "MiniMaxAI/MiniMax-M2.7": {
|
|
4578
|
+
id: string;
|
|
4579
|
+
name: string;
|
|
4580
|
+
api: "openai-completions";
|
|
4581
|
+
provider: string;
|
|
4582
|
+
baseUrl: string;
|
|
4583
|
+
compat: {
|
|
4584
|
+
supportsDeveloperRole: false;
|
|
4585
|
+
};
|
|
4586
|
+
reasoning: true;
|
|
4587
|
+
input: "text"[];
|
|
4588
|
+
cost: {
|
|
4589
|
+
input: number;
|
|
4590
|
+
output: number;
|
|
4591
|
+
cacheRead: number;
|
|
4592
|
+
cacheWrite: number;
|
|
4593
|
+
};
|
|
4594
|
+
contextWindow: number;
|
|
4595
|
+
maxTokens: number;
|
|
4596
|
+
};
|
|
4475
4597
|
readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
|
|
4476
4598
|
id: string;
|
|
4477
4599
|
name: string;
|
|
@@ -4792,6 +4914,26 @@ export declare const MODELS: {
|
|
|
4792
4914
|
contextWindow: number;
|
|
4793
4915
|
maxTokens: number;
|
|
4794
4916
|
};
|
|
4917
|
+
readonly "zai-org/GLM-5.1": {
|
|
4918
|
+
id: string;
|
|
4919
|
+
name: string;
|
|
4920
|
+
api: "openai-completions";
|
|
4921
|
+
provider: string;
|
|
4922
|
+
baseUrl: string;
|
|
4923
|
+
compat: {
|
|
4924
|
+
supportsDeveloperRole: false;
|
|
4925
|
+
};
|
|
4926
|
+
reasoning: true;
|
|
4927
|
+
input: "text"[];
|
|
4928
|
+
cost: {
|
|
4929
|
+
input: number;
|
|
4930
|
+
output: number;
|
|
4931
|
+
cacheRead: number;
|
|
4932
|
+
cacheWrite: number;
|
|
4933
|
+
};
|
|
4934
|
+
contextWindow: number;
|
|
4935
|
+
maxTokens: number;
|
|
4936
|
+
};
|
|
4795
4937
|
};
|
|
4796
4938
|
readonly "kimi-coding": {
|
|
4797
4939
|
readonly k2p5: {
|
|
@@ -6472,6 +6614,23 @@ export declare const MODELS: {
|
|
|
6472
6614
|
contextWindow: number;
|
|
6473
6615
|
maxTokens: number;
|
|
6474
6616
|
};
|
|
6617
|
+
readonly "claude-opus-4-7": {
|
|
6618
|
+
id: string;
|
|
6619
|
+
name: string;
|
|
6620
|
+
api: "anthropic-messages";
|
|
6621
|
+
provider: string;
|
|
6622
|
+
baseUrl: string;
|
|
6623
|
+
reasoning: true;
|
|
6624
|
+
input: ("image" | "text")[];
|
|
6625
|
+
cost: {
|
|
6626
|
+
input: number;
|
|
6627
|
+
output: number;
|
|
6628
|
+
cacheRead: number;
|
|
6629
|
+
cacheWrite: number;
|
|
6630
|
+
};
|
|
6631
|
+
contextWindow: number;
|
|
6632
|
+
maxTokens: number;
|
|
6633
|
+
};
|
|
6475
6634
|
readonly "claude-sonnet-4": {
|
|
6476
6635
|
id: string;
|
|
6477
6636
|
name: string;
|
|
@@ -6574,6 +6733,23 @@ export declare const MODELS: {
|
|
|
6574
6733
|
contextWindow: number;
|
|
6575
6734
|
maxTokens: number;
|
|
6576
6735
|
};
|
|
6736
|
+
readonly "glm-5.1": {
|
|
6737
|
+
id: string;
|
|
6738
|
+
name: string;
|
|
6739
|
+
api: "openai-completions";
|
|
6740
|
+
provider: string;
|
|
6741
|
+
baseUrl: string;
|
|
6742
|
+
reasoning: true;
|
|
6743
|
+
input: "text"[];
|
|
6744
|
+
cost: {
|
|
6745
|
+
input: number;
|
|
6746
|
+
output: number;
|
|
6747
|
+
cacheRead: number;
|
|
6748
|
+
cacheWrite: number;
|
|
6749
|
+
};
|
|
6750
|
+
contextWindow: number;
|
|
6751
|
+
maxTokens: number;
|
|
6752
|
+
};
|
|
6577
6753
|
readonly "gpt-5": {
|
|
6578
6754
|
id: string;
|
|
6579
6755
|
name: string;
|
|
@@ -6880,14 +7056,31 @@ export declare const MODELS: {
|
|
|
6880
7056
|
contextWindow: number;
|
|
6881
7057
|
maxTokens: number;
|
|
6882
7058
|
};
|
|
6883
|
-
readonly "qwen3.
|
|
7059
|
+
readonly "qwen3.5-plus": {
|
|
6884
7060
|
id: string;
|
|
6885
7061
|
name: string;
|
|
6886
7062
|
api: "openai-completions";
|
|
6887
7063
|
provider: string;
|
|
6888
7064
|
baseUrl: string;
|
|
6889
7065
|
reasoning: true;
|
|
6890
|
-
input: "text"[];
|
|
7066
|
+
input: ("image" | "text")[];
|
|
7067
|
+
cost: {
|
|
7068
|
+
input: number;
|
|
7069
|
+
output: number;
|
|
7070
|
+
cacheRead: number;
|
|
7071
|
+
cacheWrite: number;
|
|
7072
|
+
};
|
|
7073
|
+
contextWindow: number;
|
|
7074
|
+
maxTokens: number;
|
|
7075
|
+
};
|
|
7076
|
+
readonly "qwen3.6-plus": {
|
|
7077
|
+
id: string;
|
|
7078
|
+
name: string;
|
|
7079
|
+
api: "openai-completions";
|
|
7080
|
+
provider: string;
|
|
7081
|
+
baseUrl: string;
|
|
7082
|
+
reasoning: true;
|
|
7083
|
+
input: ("image" | "text")[];
|
|
6891
7084
|
cost: {
|
|
6892
7085
|
input: number;
|
|
6893
7086
|
output: number;
|
|
@@ -6916,6 +7109,23 @@ export declare const MODELS: {
|
|
|
6916
7109
|
contextWindow: number;
|
|
6917
7110
|
maxTokens: number;
|
|
6918
7111
|
};
|
|
7112
|
+
readonly "glm-5.1": {
|
|
7113
|
+
id: string;
|
|
7114
|
+
name: string;
|
|
7115
|
+
api: "openai-completions";
|
|
7116
|
+
provider: string;
|
|
7117
|
+
baseUrl: string;
|
|
7118
|
+
reasoning: true;
|
|
7119
|
+
input: "text"[];
|
|
7120
|
+
cost: {
|
|
7121
|
+
input: number;
|
|
7122
|
+
output: number;
|
|
7123
|
+
cacheRead: number;
|
|
7124
|
+
cacheWrite: number;
|
|
7125
|
+
};
|
|
7126
|
+
contextWindow: number;
|
|
7127
|
+
maxTokens: number;
|
|
7128
|
+
};
|
|
6919
7129
|
readonly "kimi-k2.5": {
|
|
6920
7130
|
id: string;
|
|
6921
7131
|
name: string;
|
|
@@ -6970,7 +7180,7 @@ export declare const MODELS: {
|
|
|
6970
7180
|
readonly "minimax-m2.5": {
|
|
6971
7181
|
id: string;
|
|
6972
7182
|
name: string;
|
|
6973
|
-
api: "
|
|
7183
|
+
api: "anthropic-messages";
|
|
6974
7184
|
provider: string;
|
|
6975
7185
|
baseUrl: string;
|
|
6976
7186
|
reasoning: true;
|
|
@@ -7001,6 +7211,40 @@ export declare const MODELS: {
|
|
|
7001
7211
|
contextWindow: number;
|
|
7002
7212
|
maxTokens: number;
|
|
7003
7213
|
};
|
|
7214
|
+
readonly "qwen3.5-plus": {
|
|
7215
|
+
id: string;
|
|
7216
|
+
name: string;
|
|
7217
|
+
api: "openai-completions";
|
|
7218
|
+
provider: string;
|
|
7219
|
+
baseUrl: string;
|
|
7220
|
+
reasoning: true;
|
|
7221
|
+
input: ("image" | "text")[];
|
|
7222
|
+
cost: {
|
|
7223
|
+
input: number;
|
|
7224
|
+
output: number;
|
|
7225
|
+
cacheRead: number;
|
|
7226
|
+
cacheWrite: number;
|
|
7227
|
+
};
|
|
7228
|
+
contextWindow: number;
|
|
7229
|
+
maxTokens: number;
|
|
7230
|
+
};
|
|
7231
|
+
readonly "qwen3.6-plus": {
|
|
7232
|
+
id: string;
|
|
7233
|
+
name: string;
|
|
7234
|
+
api: "openai-completions";
|
|
7235
|
+
provider: string;
|
|
7236
|
+
baseUrl: string;
|
|
7237
|
+
reasoning: true;
|
|
7238
|
+
input: ("image" | "text")[];
|
|
7239
|
+
cost: {
|
|
7240
|
+
input: number;
|
|
7241
|
+
output: number;
|
|
7242
|
+
cacheRead: number;
|
|
7243
|
+
cacheWrite: number;
|
|
7244
|
+
};
|
|
7245
|
+
contextWindow: number;
|
|
7246
|
+
maxTokens: number;
|
|
7247
|
+
};
|
|
7004
7248
|
};
|
|
7005
7249
|
readonly openrouter: {
|
|
7006
7250
|
readonly "ai21/jamba-large-1.7": {
|
|
@@ -7292,7 +7536,7 @@ export declare const MODELS: {
|
|
|
7292
7536
|
contextWindow: number;
|
|
7293
7537
|
maxTokens: number;
|
|
7294
7538
|
};
|
|
7295
|
-
readonly "anthropic/claude-
|
|
7539
|
+
readonly "anthropic/claude-opus-4.6-fast": {
|
|
7296
7540
|
id: string;
|
|
7297
7541
|
name: string;
|
|
7298
7542
|
api: "openai-completions";
|
|
@@ -7309,7 +7553,7 @@ export declare const MODELS: {
|
|
|
7309
7553
|
contextWindow: number;
|
|
7310
7554
|
maxTokens: number;
|
|
7311
7555
|
};
|
|
7312
|
-
readonly "anthropic/claude-
|
|
7556
|
+
readonly "anthropic/claude-opus-4.7": {
|
|
7313
7557
|
id: string;
|
|
7314
7558
|
name: string;
|
|
7315
7559
|
api: "openai-completions";
|
|
@@ -7326,7 +7570,7 @@ export declare const MODELS: {
|
|
|
7326
7570
|
contextWindow: number;
|
|
7327
7571
|
maxTokens: number;
|
|
7328
7572
|
};
|
|
7329
|
-
readonly "anthropic/claude-sonnet-4
|
|
7573
|
+
readonly "anthropic/claude-sonnet-4": {
|
|
7330
7574
|
id: string;
|
|
7331
7575
|
name: string;
|
|
7332
7576
|
api: "openai-completions";
|
|
@@ -7343,14 +7587,14 @@ export declare const MODELS: {
|
|
|
7343
7587
|
contextWindow: number;
|
|
7344
7588
|
maxTokens: number;
|
|
7345
7589
|
};
|
|
7346
|
-
readonly "
|
|
7590
|
+
readonly "anthropic/claude-sonnet-4.5": {
|
|
7347
7591
|
id: string;
|
|
7348
7592
|
name: string;
|
|
7349
7593
|
api: "openai-completions";
|
|
7350
7594
|
provider: string;
|
|
7351
7595
|
baseUrl: string;
|
|
7352
|
-
reasoning:
|
|
7353
|
-
input: "text"[];
|
|
7596
|
+
reasoning: true;
|
|
7597
|
+
input: ("image" | "text")[];
|
|
7354
7598
|
cost: {
|
|
7355
7599
|
input: number;
|
|
7356
7600
|
output: number;
|
|
@@ -7360,13 +7604,30 @@ export declare const MODELS: {
|
|
|
7360
7604
|
contextWindow: number;
|
|
7361
7605
|
maxTokens: number;
|
|
7362
7606
|
};
|
|
7363
|
-
readonly "
|
|
7607
|
+
readonly "anthropic/claude-sonnet-4.6": {
|
|
7364
7608
|
id: string;
|
|
7365
7609
|
name: string;
|
|
7366
7610
|
api: "openai-completions";
|
|
7367
7611
|
provider: string;
|
|
7368
7612
|
baseUrl: string;
|
|
7369
7613
|
reasoning: true;
|
|
7614
|
+
input: ("image" | "text")[];
|
|
7615
|
+
cost: {
|
|
7616
|
+
input: number;
|
|
7617
|
+
output: number;
|
|
7618
|
+
cacheRead: number;
|
|
7619
|
+
cacheWrite: number;
|
|
7620
|
+
};
|
|
7621
|
+
contextWindow: number;
|
|
7622
|
+
maxTokens: number;
|
|
7623
|
+
};
|
|
7624
|
+
readonly "arcee-ai/trinity-large-preview:free": {
|
|
7625
|
+
id: string;
|
|
7626
|
+
name: string;
|
|
7627
|
+
api: "openai-completions";
|
|
7628
|
+
provider: string;
|
|
7629
|
+
baseUrl: string;
|
|
7630
|
+
reasoning: false;
|
|
7370
7631
|
input: "text"[];
|
|
7371
7632
|
cost: {
|
|
7372
7633
|
input: number;
|
|
@@ -7377,7 +7638,7 @@ export declare const MODELS: {
|
|
|
7377
7638
|
contextWindow: number;
|
|
7378
7639
|
maxTokens: number;
|
|
7379
7640
|
};
|
|
7380
|
-
readonly "arcee-ai/trinity-
|
|
7641
|
+
readonly "arcee-ai/trinity-large-thinking": {
|
|
7381
7642
|
id: string;
|
|
7382
7643
|
name: string;
|
|
7383
7644
|
api: "openai-completions";
|
|
@@ -7394,7 +7655,7 @@ export declare const MODELS: {
|
|
|
7394
7655
|
contextWindow: number;
|
|
7395
7656
|
maxTokens: number;
|
|
7396
7657
|
};
|
|
7397
|
-
readonly "arcee-ai/trinity-mini
|
|
7658
|
+
readonly "arcee-ai/trinity-mini": {
|
|
7398
7659
|
id: string;
|
|
7399
7660
|
name: string;
|
|
7400
7661
|
api: "openai-completions";
|
|
@@ -7955,7 +8216,7 @@ export declare const MODELS: {
|
|
|
7955
8216
|
contextWindow: number;
|
|
7956
8217
|
maxTokens: number;
|
|
7957
8218
|
};
|
|
7958
|
-
readonly "google/gemma-4-
|
|
8219
|
+
readonly "google/gemma-4-26b-a4b-it:free": {
|
|
7959
8220
|
id: string;
|
|
7960
8221
|
name: string;
|
|
7961
8222
|
api: "openai-completions";
|
|
@@ -7972,14 +8233,14 @@ export declare const MODELS: {
|
|
|
7972
8233
|
contextWindow: number;
|
|
7973
8234
|
maxTokens: number;
|
|
7974
8235
|
};
|
|
7975
|
-
readonly "
|
|
8236
|
+
readonly "google/gemma-4-31b-it": {
|
|
7976
8237
|
id: string;
|
|
7977
8238
|
name: string;
|
|
7978
8239
|
api: "openai-completions";
|
|
7979
8240
|
provider: string;
|
|
7980
8241
|
baseUrl: string;
|
|
7981
|
-
reasoning:
|
|
7982
|
-
input: "text"[];
|
|
8242
|
+
reasoning: true;
|
|
8243
|
+
input: ("image" | "text")[];
|
|
7983
8244
|
cost: {
|
|
7984
8245
|
input: number;
|
|
7985
8246
|
output: number;
|
|
@@ -7989,14 +8250,14 @@ export declare const MODELS: {
|
|
|
7989
8250
|
contextWindow: number;
|
|
7990
8251
|
maxTokens: number;
|
|
7991
8252
|
};
|
|
7992
|
-
readonly "
|
|
8253
|
+
readonly "google/gemma-4-31b-it:free": {
|
|
7993
8254
|
id: string;
|
|
7994
8255
|
name: string;
|
|
7995
8256
|
api: "openai-completions";
|
|
7996
8257
|
provider: string;
|
|
7997
8258
|
baseUrl: string;
|
|
7998
8259
|
reasoning: true;
|
|
7999
|
-
input: "text"[];
|
|
8260
|
+
input: ("image" | "text")[];
|
|
8000
8261
|
cost: {
|
|
8001
8262
|
input: number;
|
|
8002
8263
|
output: number;
|
|
@@ -8006,13 +8267,13 @@ export declare const MODELS: {
|
|
|
8006
8267
|
contextWindow: number;
|
|
8007
8268
|
maxTokens: number;
|
|
8008
8269
|
};
|
|
8009
|
-
readonly "inception/mercury-
|
|
8270
|
+
readonly "inception/mercury-2": {
|
|
8010
8271
|
id: string;
|
|
8011
8272
|
name: string;
|
|
8012
8273
|
api: "openai-completions";
|
|
8013
8274
|
provider: string;
|
|
8014
8275
|
baseUrl: string;
|
|
8015
|
-
reasoning:
|
|
8276
|
+
reasoning: true;
|
|
8016
8277
|
input: "text"[];
|
|
8017
8278
|
cost: {
|
|
8018
8279
|
input: number;
|
|
@@ -8040,23 +8301,6 @@ export declare const MODELS: {
|
|
|
8040
8301
|
contextWindow: number;
|
|
8041
8302
|
maxTokens: number;
|
|
8042
8303
|
};
|
|
8043
|
-
readonly "meituan/longcat-flash-chat": {
|
|
8044
|
-
id: string;
|
|
8045
|
-
name: string;
|
|
8046
|
-
api: "openai-completions";
|
|
8047
|
-
provider: string;
|
|
8048
|
-
baseUrl: string;
|
|
8049
|
-
reasoning: false;
|
|
8050
|
-
input: "text"[];
|
|
8051
|
-
cost: {
|
|
8052
|
-
input: number;
|
|
8053
|
-
output: number;
|
|
8054
|
-
cacheRead: number;
|
|
8055
|
-
cacheWrite: number;
|
|
8056
|
-
};
|
|
8057
|
-
contextWindow: number;
|
|
8058
|
-
maxTokens: number;
|
|
8059
|
-
};
|
|
8060
8304
|
readonly "meta-llama/llama-3-8b-instruct": {
|
|
8061
8305
|
id: string;
|
|
8062
8306
|
name: string;
|
|
@@ -9876,6 +10120,23 @@ export declare const MODELS: {
|
|
|
9876
10120
|
contextWindow: number;
|
|
9877
10121
|
maxTokens: number;
|
|
9878
10122
|
};
|
|
10123
|
+
readonly "openrouter/elephant-alpha": {
|
|
10124
|
+
id: string;
|
|
10125
|
+
name: string;
|
|
10126
|
+
api: "openai-completions";
|
|
10127
|
+
provider: string;
|
|
10128
|
+
baseUrl: string;
|
|
10129
|
+
reasoning: false;
|
|
10130
|
+
input: "text"[];
|
|
10131
|
+
cost: {
|
|
10132
|
+
input: number;
|
|
10133
|
+
output: number;
|
|
10134
|
+
cacheRead: number;
|
|
10135
|
+
cacheWrite: number;
|
|
10136
|
+
};
|
|
10137
|
+
contextWindow: number;
|
|
10138
|
+
maxTokens: number;
|
|
10139
|
+
};
|
|
9879
10140
|
readonly "openrouter/free": {
|
|
9880
10141
|
id: string;
|
|
9881
10142
|
name: string;
|
|
@@ -10624,7 +10885,7 @@ export declare const MODELS: {
|
|
|
10624
10885
|
contextWindow: number;
|
|
10625
10886
|
maxTokens: number;
|
|
10626
10887
|
};
|
|
10627
|
-
readonly "qwen/qwen3.6-plus
|
|
10888
|
+
readonly "qwen/qwen3.6-plus": {
|
|
10628
10889
|
id: string;
|
|
10629
10890
|
name: string;
|
|
10630
10891
|
api: "openai-completions";
|
|
@@ -10743,23 +11004,6 @@ export declare const MODELS: {
|
|
|
10743
11004
|
contextWindow: number;
|
|
10744
11005
|
maxTokens: number;
|
|
10745
11006
|
};
|
|
10746
|
-
readonly "stepfun/step-3.5-flash:free": {
|
|
10747
|
-
id: string;
|
|
10748
|
-
name: string;
|
|
10749
|
-
api: "openai-completions";
|
|
10750
|
-
provider: string;
|
|
10751
|
-
baseUrl: string;
|
|
10752
|
-
reasoning: true;
|
|
10753
|
-
input: "text"[];
|
|
10754
|
-
cost: {
|
|
10755
|
-
input: number;
|
|
10756
|
-
output: number;
|
|
10757
|
-
cacheRead: number;
|
|
10758
|
-
cacheWrite: number;
|
|
10759
|
-
};
|
|
10760
|
-
contextWindow: number;
|
|
10761
|
-
maxTokens: number;
|
|
10762
|
-
};
|
|
10763
11007
|
readonly "thedrummer/rocinante-12b": {
|
|
10764
11008
|
id: string;
|
|
10765
11009
|
name: string;
|
|
@@ -11219,6 +11463,23 @@ export declare const MODELS: {
|
|
|
11219
11463
|
contextWindow: number;
|
|
11220
11464
|
maxTokens: number;
|
|
11221
11465
|
};
|
|
11466
|
+
readonly "z-ai/glm-5.1": {
|
|
11467
|
+
id: string;
|
|
11468
|
+
name: string;
|
|
11469
|
+
api: "openai-completions";
|
|
11470
|
+
provider: string;
|
|
11471
|
+
baseUrl: string;
|
|
11472
|
+
reasoning: true;
|
|
11473
|
+
input: "text"[];
|
|
11474
|
+
cost: {
|
|
11475
|
+
input: number;
|
|
11476
|
+
output: number;
|
|
11477
|
+
cacheRead: number;
|
|
11478
|
+
cacheWrite: number;
|
|
11479
|
+
};
|
|
11480
|
+
contextWindow: number;
|
|
11481
|
+
maxTokens: number;
|
|
11482
|
+
};
|
|
11222
11483
|
readonly "z-ai/glm-5v-turbo": {
|
|
11223
11484
|
id: string;
|
|
11224
11485
|
name: string;
|
|
@@ -11646,6 +11907,23 @@ export declare const MODELS: {
|
|
|
11646
11907
|
contextWindow: number;
|
|
11647
11908
|
maxTokens: number;
|
|
11648
11909
|
};
|
|
11910
|
+
readonly "anthropic/claude-opus-4.7": {
|
|
11911
|
+
id: string;
|
|
11912
|
+
name: string;
|
|
11913
|
+
api: "anthropic-messages";
|
|
11914
|
+
provider: string;
|
|
11915
|
+
baseUrl: string;
|
|
11916
|
+
reasoning: true;
|
|
11917
|
+
input: ("image" | "text")[];
|
|
11918
|
+
cost: {
|
|
11919
|
+
input: number;
|
|
11920
|
+
output: number;
|
|
11921
|
+
cacheRead: number;
|
|
11922
|
+
cacheWrite: number;
|
|
11923
|
+
};
|
|
11924
|
+
contextWindow: number;
|
|
11925
|
+
maxTokens: number;
|
|
11926
|
+
};
|
|
11649
11927
|
readonly "anthropic/claude-sonnet-4": {
|
|
11650
11928
|
id: string;
|
|
11651
11929
|
name: string;
|
|
@@ -13125,23 +13403,6 @@ export declare const MODELS: {
|
|
|
13125
13403
|
contextWindow: number;
|
|
13126
13404
|
maxTokens: number;
|
|
13127
13405
|
};
|
|
13128
|
-
readonly "openai/gpt-oss-120b": {
|
|
13129
|
-
id: string;
|
|
13130
|
-
name: string;
|
|
13131
|
-
api: "anthropic-messages";
|
|
13132
|
-
provider: string;
|
|
13133
|
-
baseUrl: string;
|
|
13134
|
-
reasoning: true;
|
|
13135
|
-
input: "text"[];
|
|
13136
|
-
cost: {
|
|
13137
|
-
input: number;
|
|
13138
|
-
output: number;
|
|
13139
|
-
cacheRead: number;
|
|
13140
|
-
cacheWrite: number;
|
|
13141
|
-
};
|
|
13142
|
-
contextWindow: number;
|
|
13143
|
-
maxTokens: number;
|
|
13144
|
-
};
|
|
13145
13406
|
readonly "openai/gpt-oss-20b": {
|
|
13146
13407
|
id: string;
|
|
13147
13408
|
name: string;
|
|
@@ -13329,23 +13590,6 @@ export declare const MODELS: {
|
|
|
13329
13590
|
contextWindow: number;
|
|
13330
13591
|
maxTokens: number;
|
|
13331
13592
|
};
|
|
13332
|
-
readonly "xai/grok-2-vision": {
|
|
13333
|
-
id: string;
|
|
13334
|
-
name: string;
|
|
13335
|
-
api: "anthropic-messages";
|
|
13336
|
-
provider: string;
|
|
13337
|
-
baseUrl: string;
|
|
13338
|
-
reasoning: false;
|
|
13339
|
-
input: ("image" | "text")[];
|
|
13340
|
-
cost: {
|
|
13341
|
-
input: number;
|
|
13342
|
-
output: number;
|
|
13343
|
-
cacheRead: number;
|
|
13344
|
-
cacheWrite: number;
|
|
13345
|
-
};
|
|
13346
|
-
contextWindow: number;
|
|
13347
|
-
maxTokens: number;
|
|
13348
|
-
};
|
|
13349
13593
|
readonly "xai/grok-3": {
|
|
13350
13594
|
id: string;
|
|
13351
13595
|
name: string;
|
|
@@ -13438,7 +13682,7 @@ export declare const MODELS: {
|
|
|
13438
13682
|
provider: string;
|
|
13439
13683
|
baseUrl: string;
|
|
13440
13684
|
reasoning: false;
|
|
13441
|
-
input: "text"[];
|
|
13685
|
+
input: ("image" | "text")[];
|
|
13442
13686
|
cost: {
|
|
13443
13687
|
input: number;
|
|
13444
13688
|
output: number;
|
|
@@ -13455,7 +13699,7 @@ export declare const MODELS: {
|
|
|
13455
13699
|
provider: string;
|
|
13456
13700
|
baseUrl: string;
|
|
13457
13701
|
reasoning: true;
|
|
13458
|
-
input: "text"[];
|
|
13702
|
+
input: ("image" | "text")[];
|
|
13459
13703
|
cost: {
|
|
13460
13704
|
input: number;
|
|
13461
13705
|
output: number;
|
|
@@ -13472,7 +13716,7 @@ export declare const MODELS: {
|
|
|
13472
13716
|
provider: string;
|
|
13473
13717
|
baseUrl: string;
|
|
13474
13718
|
reasoning: false;
|
|
13475
|
-
input: "text"[];
|
|
13719
|
+
input: ("image" | "text")[];
|
|
13476
13720
|
cost: {
|
|
13477
13721
|
input: number;
|
|
13478
13722
|
output: number;
|
|
@@ -13489,7 +13733,7 @@ export declare const MODELS: {
|
|
|
13489
13733
|
provider: string;
|
|
13490
13734
|
baseUrl: string;
|
|
13491
13735
|
reasoning: true;
|
|
13492
|
-
input: "text"[];
|
|
13736
|
+
input: ("image" | "text")[];
|
|
13493
13737
|
cost: {
|
|
13494
13738
|
input: number;
|
|
13495
13739
|
output: number;
|
|
@@ -13506,7 +13750,7 @@ export declare const MODELS: {
|
|
|
13506
13750
|
provider: string;
|
|
13507
13751
|
baseUrl: string;
|
|
13508
13752
|
reasoning: true;
|
|
13509
|
-
input: "text"[];
|
|
13753
|
+
input: ("image" | "text")[];
|
|
13510
13754
|
cost: {
|
|
13511
13755
|
input: number;
|
|
13512
13756
|
output: number;
|
|
@@ -13523,7 +13767,7 @@ export declare const MODELS: {
|
|
|
13523
13767
|
provider: string;
|
|
13524
13768
|
baseUrl: string;
|
|
13525
13769
|
reasoning: true;
|
|
13526
|
-
input: "text"[];
|
|
13770
|
+
input: ("image" | "text")[];
|
|
13527
13771
|
cost: {
|
|
13528
13772
|
input: number;
|
|
13529
13773
|
output: number;
|
|
@@ -13839,6 +14083,23 @@ export declare const MODELS: {
|
|
|
13839
14083
|
contextWindow: number;
|
|
13840
14084
|
maxTokens: number;
|
|
13841
14085
|
};
|
|
14086
|
+
readonly "zai/glm-5.1": {
|
|
14087
|
+
id: string;
|
|
14088
|
+
name: string;
|
|
14089
|
+
api: "anthropic-messages";
|
|
14090
|
+
provider: string;
|
|
14091
|
+
baseUrl: string;
|
|
14092
|
+
reasoning: true;
|
|
14093
|
+
input: ("image" | "text")[];
|
|
14094
|
+
cost: {
|
|
14095
|
+
input: number;
|
|
14096
|
+
output: number;
|
|
14097
|
+
cacheRead: number;
|
|
14098
|
+
cacheWrite: number;
|
|
14099
|
+
};
|
|
14100
|
+
contextWindow: number;
|
|
14101
|
+
maxTokens: number;
|
|
14102
|
+
};
|
|
13842
14103
|
readonly "zai/glm-5v-turbo": {
|
|
13843
14104
|
id: string;
|
|
13844
14105
|
name: string;
|
|
@@ -14499,6 +14760,27 @@ export declare const MODELS: {
|
|
|
14499
14760
|
contextWindow: number;
|
|
14500
14761
|
maxTokens: number;
|
|
14501
14762
|
};
|
|
14763
|
+
readonly "glm-5.1": {
|
|
14764
|
+
id: string;
|
|
14765
|
+
name: string;
|
|
14766
|
+
api: "openai-completions";
|
|
14767
|
+
provider: string;
|
|
14768
|
+
baseUrl: string;
|
|
14769
|
+
compat: {
|
|
14770
|
+
supportsDeveloperRole: false;
|
|
14771
|
+
thinkingFormat: "zai";
|
|
14772
|
+
};
|
|
14773
|
+
reasoning: true;
|
|
14774
|
+
input: "text"[];
|
|
14775
|
+
cost: {
|
|
14776
|
+
input: number;
|
|
14777
|
+
output: number;
|
|
14778
|
+
cacheRead: number;
|
|
14779
|
+
cacheWrite: number;
|
|
14780
|
+
};
|
|
14781
|
+
contextWindow: number;
|
|
14782
|
+
maxTokens: number;
|
|
14783
|
+
};
|
|
14502
14784
|
readonly "glm-5v-turbo": {
|
|
14503
14785
|
id: string;
|
|
14504
14786
|
name: string;
|