@mariozechner/pi-ai 0.63.2 → 0.65.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +86 -0
- package/dist/index.d.ts +1 -0
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +1 -0
- package/dist/index.js.map +1 -1
- package/dist/models.generated.d.ts +342 -58
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +420 -147
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +35 -7
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/faux.d.ts +56 -0
- package/dist/providers/faux.d.ts.map +1 -0
- package/dist/providers/faux.js +367 -0
- package/dist/providers/faux.js.map +1 -0
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +5 -0
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +12 -0
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/types.d.ts +2 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/overflow.d.ts +1 -1
- package/dist/utils/overflow.d.ts.map +1 -1
- package/dist/utils/overflow.js +23 -10
- package/dist/utils/overflow.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -1551,23 +1551,6 @@ export const MODELS = {
|
|
|
1551
1551
|
contextWindow: 200000,
|
|
1552
1552
|
maxTokens: 64000,
|
|
1553
1553
|
},
|
|
1554
|
-
"claude-3-7-sonnet-latest": {
|
|
1555
|
-
id: "claude-3-7-sonnet-latest",
|
|
1556
|
-
name: "Claude Sonnet 3.7 (latest)",
|
|
1557
|
-
api: "anthropic-messages",
|
|
1558
|
-
provider: "anthropic",
|
|
1559
|
-
baseUrl: "https://api.anthropic.com",
|
|
1560
|
-
reasoning: true,
|
|
1561
|
-
input: ["text", "image"],
|
|
1562
|
-
cost: {
|
|
1563
|
-
input: 3,
|
|
1564
|
-
output: 15,
|
|
1565
|
-
cacheRead: 0.3,
|
|
1566
|
-
cacheWrite: 3.75,
|
|
1567
|
-
},
|
|
1568
|
-
contextWindow: 200000,
|
|
1569
|
-
maxTokens: 64000,
|
|
1570
|
-
},
|
|
1571
1554
|
"claude-3-haiku-20240307": {
|
|
1572
1555
|
id: "claude-3-haiku-20240307",
|
|
1573
1556
|
name: "Claude Haiku 3",
|
|
@@ -2301,6 +2284,23 @@ export const MODELS = {
|
|
|
2301
2284
|
contextWindow: 400000,
|
|
2302
2285
|
maxTokens: 128000,
|
|
2303
2286
|
},
|
|
2287
|
+
"gpt-5.3-chat-latest": {
|
|
2288
|
+
id: "gpt-5.3-chat-latest",
|
|
2289
|
+
name: "GPT-5.3 Chat (latest)",
|
|
2290
|
+
api: "azure-openai-responses",
|
|
2291
|
+
provider: "azure-openai-responses",
|
|
2292
|
+
baseUrl: "",
|
|
2293
|
+
reasoning: false,
|
|
2294
|
+
input: ["text", "image"],
|
|
2295
|
+
cost: {
|
|
2296
|
+
input: 1.75,
|
|
2297
|
+
output: 14,
|
|
2298
|
+
cacheRead: 0.175,
|
|
2299
|
+
cacheWrite: 0,
|
|
2300
|
+
},
|
|
2301
|
+
contextWindow: 128000,
|
|
2302
|
+
maxTokens: 16384,
|
|
2303
|
+
},
|
|
2304
2304
|
"gpt-5.3-codex": {
|
|
2305
2305
|
id: "gpt-5.3-codex",
|
|
2306
2306
|
name: "GPT-5.3 Codex",
|
|
@@ -5026,22 +5026,39 @@ export const MODELS = {
|
|
|
5026
5026
|
contextWindow: 128000,
|
|
5027
5027
|
maxTokens: 16384,
|
|
5028
5028
|
},
|
|
5029
|
+
"mistral-small-2603": {
|
|
5030
|
+
id: "mistral-small-2603",
|
|
5031
|
+
name: "Mistral Small 4",
|
|
5032
|
+
api: "mistral-conversations",
|
|
5033
|
+
provider: "mistral",
|
|
5034
|
+
baseUrl: "https://api.mistral.ai",
|
|
5035
|
+
reasoning: true,
|
|
5036
|
+
input: ["text", "image"],
|
|
5037
|
+
cost: {
|
|
5038
|
+
input: 0.15,
|
|
5039
|
+
output: 0.6,
|
|
5040
|
+
cacheRead: 0,
|
|
5041
|
+
cacheWrite: 0,
|
|
5042
|
+
},
|
|
5043
|
+
contextWindow: 256000,
|
|
5044
|
+
maxTokens: 256000,
|
|
5045
|
+
},
|
|
5029
5046
|
"mistral-small-latest": {
|
|
5030
5047
|
id: "mistral-small-latest",
|
|
5031
5048
|
name: "Mistral Small (latest)",
|
|
5032
5049
|
api: "mistral-conversations",
|
|
5033
5050
|
provider: "mistral",
|
|
5034
5051
|
baseUrl: "https://api.mistral.ai",
|
|
5035
|
-
reasoning:
|
|
5052
|
+
reasoning: true,
|
|
5036
5053
|
input: ["text", "image"],
|
|
5037
5054
|
cost: {
|
|
5038
|
-
input: 0.
|
|
5039
|
-
output: 0.
|
|
5055
|
+
input: 0.15,
|
|
5056
|
+
output: 0.6,
|
|
5040
5057
|
cacheRead: 0,
|
|
5041
5058
|
cacheWrite: 0,
|
|
5042
5059
|
},
|
|
5043
|
-
contextWindow:
|
|
5044
|
-
maxTokens:
|
|
5060
|
+
contextWindow: 256000,
|
|
5061
|
+
maxTokens: 256000,
|
|
5045
5062
|
},
|
|
5046
5063
|
"open-mistral-7b": {
|
|
5047
5064
|
id: "open-mistral-7b",
|
|
@@ -5572,6 +5589,23 @@ export const MODELS = {
|
|
|
5572
5589
|
contextWindow: 400000,
|
|
5573
5590
|
maxTokens: 128000,
|
|
5574
5591
|
},
|
|
5592
|
+
"gpt-5.3-chat-latest": {
|
|
5593
|
+
id: "gpt-5.3-chat-latest",
|
|
5594
|
+
name: "GPT-5.3 Chat (latest)",
|
|
5595
|
+
api: "openai-responses",
|
|
5596
|
+
provider: "openai",
|
|
5597
|
+
baseUrl: "https://api.openai.com/v1",
|
|
5598
|
+
reasoning: false,
|
|
5599
|
+
input: ["text", "image"],
|
|
5600
|
+
cost: {
|
|
5601
|
+
input: 1.75,
|
|
5602
|
+
output: 14,
|
|
5603
|
+
cacheRead: 0.175,
|
|
5604
|
+
cacheWrite: 0,
|
|
5605
|
+
},
|
|
5606
|
+
contextWindow: 128000,
|
|
5607
|
+
maxTokens: 16384,
|
|
5608
|
+
},
|
|
5575
5609
|
"gpt-5.3-codex": {
|
|
5576
5610
|
id: "gpt-5.3-codex",
|
|
5577
5611
|
name: "GPT-5.3 Codex",
|
|
@@ -6426,40 +6460,6 @@ export const MODELS = {
|
|
|
6426
6460
|
contextWindow: 262144,
|
|
6427
6461
|
maxTokens: 65536,
|
|
6428
6462
|
},
|
|
6429
|
-
"mimo-v2-omni-free": {
|
|
6430
|
-
id: "mimo-v2-omni-free",
|
|
6431
|
-
name: "MiMo V2 Omni Free",
|
|
6432
|
-
api: "openai-completions",
|
|
6433
|
-
provider: "opencode",
|
|
6434
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6435
|
-
reasoning: true,
|
|
6436
|
-
input: ["text", "image"],
|
|
6437
|
-
cost: {
|
|
6438
|
-
input: 0,
|
|
6439
|
-
output: 0,
|
|
6440
|
-
cacheRead: 0,
|
|
6441
|
-
cacheWrite: 0,
|
|
6442
|
-
},
|
|
6443
|
-
contextWindow: 262144,
|
|
6444
|
-
maxTokens: 64000,
|
|
6445
|
-
},
|
|
6446
|
-
"mimo-v2-pro-free": {
|
|
6447
|
-
id: "mimo-v2-pro-free",
|
|
6448
|
-
name: "MiMo V2 Pro Free",
|
|
6449
|
-
api: "openai-completions",
|
|
6450
|
-
provider: "opencode",
|
|
6451
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
6452
|
-
reasoning: true,
|
|
6453
|
-
input: ["text"],
|
|
6454
|
-
cost: {
|
|
6455
|
-
input: 0,
|
|
6456
|
-
output: 0,
|
|
6457
|
-
cacheRead: 0,
|
|
6458
|
-
cacheWrite: 0,
|
|
6459
|
-
},
|
|
6460
|
-
contextWindow: 1048576,
|
|
6461
|
-
maxTokens: 64000,
|
|
6462
|
-
},
|
|
6463
6463
|
"minimax-m2.5": {
|
|
6464
6464
|
id: "minimax-m2.5",
|
|
6465
6465
|
name: "MiniMax M2.5",
|
|
@@ -6508,9 +6508,26 @@ export const MODELS = {
|
|
|
6508
6508
|
cacheRead: 0,
|
|
6509
6509
|
cacheWrite: 0,
|
|
6510
6510
|
},
|
|
6511
|
-
contextWindow:
|
|
6511
|
+
contextWindow: 204800,
|
|
6512
6512
|
maxTokens: 128000,
|
|
6513
6513
|
},
|
|
6514
|
+
"qwen3.6-plus-free": {
|
|
6515
|
+
id: "qwen3.6-plus-free",
|
|
6516
|
+
name: "Qwen3.6 Plus Free",
|
|
6517
|
+
api: "openai-completions",
|
|
6518
|
+
provider: "opencode",
|
|
6519
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
6520
|
+
reasoning: true,
|
|
6521
|
+
input: ["text"],
|
|
6522
|
+
cost: {
|
|
6523
|
+
input: 0,
|
|
6524
|
+
output: 0,
|
|
6525
|
+
cacheRead: 0,
|
|
6526
|
+
cacheWrite: 0,
|
|
6527
|
+
},
|
|
6528
|
+
contextWindow: 1048576,
|
|
6529
|
+
maxTokens: 64000,
|
|
6530
|
+
},
|
|
6514
6531
|
},
|
|
6515
6532
|
"opencode-go": {
|
|
6516
6533
|
"glm-5": {
|
|
@@ -6547,12 +6564,46 @@ export const MODELS = {
|
|
|
6547
6564
|
contextWindow: 262144,
|
|
6548
6565
|
maxTokens: 65536,
|
|
6549
6566
|
},
|
|
6567
|
+
"mimo-v2-omni": {
|
|
6568
|
+
id: "mimo-v2-omni",
|
|
6569
|
+
name: "MiMo V2 Omni",
|
|
6570
|
+
api: "openai-completions",
|
|
6571
|
+
provider: "opencode-go",
|
|
6572
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6573
|
+
reasoning: true,
|
|
6574
|
+
input: ["text", "image"],
|
|
6575
|
+
cost: {
|
|
6576
|
+
input: 0.4,
|
|
6577
|
+
output: 2,
|
|
6578
|
+
cacheRead: 0.08,
|
|
6579
|
+
cacheWrite: 0,
|
|
6580
|
+
},
|
|
6581
|
+
contextWindow: 262144,
|
|
6582
|
+
maxTokens: 64000,
|
|
6583
|
+
},
|
|
6584
|
+
"mimo-v2-pro": {
|
|
6585
|
+
id: "mimo-v2-pro",
|
|
6586
|
+
name: "MiMo V2 Pro",
|
|
6587
|
+
api: "openai-completions",
|
|
6588
|
+
provider: "opencode-go",
|
|
6589
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6590
|
+
reasoning: true,
|
|
6591
|
+
input: ["text"],
|
|
6592
|
+
cost: {
|
|
6593
|
+
input: 1,
|
|
6594
|
+
output: 3,
|
|
6595
|
+
cacheRead: 0.2,
|
|
6596
|
+
cacheWrite: 0,
|
|
6597
|
+
},
|
|
6598
|
+
contextWindow: 1048576,
|
|
6599
|
+
maxTokens: 64000,
|
|
6600
|
+
},
|
|
6550
6601
|
"minimax-m2.5": {
|
|
6551
6602
|
id: "minimax-m2.5",
|
|
6552
6603
|
name: "MiniMax M2.5",
|
|
6553
|
-
api: "
|
|
6604
|
+
api: "openai-completions",
|
|
6554
6605
|
provider: "opencode-go",
|
|
6555
|
-
baseUrl: "https://opencode.ai/zen/go",
|
|
6606
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
6556
6607
|
reasoning: true,
|
|
6557
6608
|
input: ["text"],
|
|
6558
6609
|
cost: {
|
|
@@ -6957,6 +7008,23 @@ export const MODELS = {
|
|
|
6957
7008
|
contextWindow: 131000,
|
|
6958
7009
|
maxTokens: 4096,
|
|
6959
7010
|
},
|
|
7011
|
+
"arcee-ai/trinity-large-thinking": {
|
|
7012
|
+
id: "arcee-ai/trinity-large-thinking",
|
|
7013
|
+
name: "Arcee AI: Trinity Large Thinking",
|
|
7014
|
+
api: "openai-completions",
|
|
7015
|
+
provider: "openrouter",
|
|
7016
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7017
|
+
reasoning: true,
|
|
7018
|
+
input: ["text"],
|
|
7019
|
+
cost: {
|
|
7020
|
+
input: 0.22,
|
|
7021
|
+
output: 0.85,
|
|
7022
|
+
cacheRead: 0,
|
|
7023
|
+
cacheWrite: 0,
|
|
7024
|
+
},
|
|
7025
|
+
contextWindow: 262144,
|
|
7026
|
+
maxTokens: 262144,
|
|
7027
|
+
},
|
|
6960
7028
|
"arcee-ai/trinity-mini": {
|
|
6961
7029
|
id: "arcee-ai/trinity-mini",
|
|
6962
7030
|
name: "Arcee AI: Trinity Mini",
|
|
@@ -7397,7 +7465,7 @@ export const MODELS = {
|
|
|
7397
7465
|
cacheWrite: 0.08333333333333334,
|
|
7398
7466
|
},
|
|
7399
7467
|
contextWindow: 1048576,
|
|
7400
|
-
maxTokens:
|
|
7468
|
+
maxTokens: 65535,
|
|
7401
7469
|
},
|
|
7402
7470
|
"google/gemini-2.5-pro": {
|
|
7403
7471
|
id: "google/gemini-2.5-pro",
|
|
@@ -7518,6 +7586,40 @@ export const MODELS = {
|
|
|
7518
7586
|
contextWindow: 1048576,
|
|
7519
7587
|
maxTokens: 65536,
|
|
7520
7588
|
},
|
|
7589
|
+
"google/gemma-4-26b-a4b-it": {
|
|
7590
|
+
id: "google/gemma-4-26b-a4b-it",
|
|
7591
|
+
name: "Google: Gemma 4 26B A4B ",
|
|
7592
|
+
api: "openai-completions",
|
|
7593
|
+
provider: "openrouter",
|
|
7594
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7595
|
+
reasoning: true,
|
|
7596
|
+
input: ["text", "image"],
|
|
7597
|
+
cost: {
|
|
7598
|
+
input: 0.13,
|
|
7599
|
+
output: 0.39999999999999997,
|
|
7600
|
+
cacheRead: 0,
|
|
7601
|
+
cacheWrite: 0,
|
|
7602
|
+
},
|
|
7603
|
+
contextWindow: 262144,
|
|
7604
|
+
maxTokens: 262144,
|
|
7605
|
+
},
|
|
7606
|
+
"google/gemma-4-31b-it": {
|
|
7607
|
+
id: "google/gemma-4-31b-it",
|
|
7608
|
+
name: "Google: Gemma 4 31B",
|
|
7609
|
+
api: "openai-completions",
|
|
7610
|
+
provider: "openrouter",
|
|
7611
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
7612
|
+
reasoning: true,
|
|
7613
|
+
input: ["text", "image"],
|
|
7614
|
+
cost: {
|
|
7615
|
+
input: 0.14,
|
|
7616
|
+
output: 0.39999999999999997,
|
|
7617
|
+
cacheRead: 0,
|
|
7618
|
+
cacheWrite: 0,
|
|
7619
|
+
},
|
|
7620
|
+
contextWindow: 262144,
|
|
7621
|
+
maxTokens: 131072,
|
|
7622
|
+
},
|
|
7521
7623
|
"inception/mercury": {
|
|
7522
7624
|
id: "inception/mercury",
|
|
7523
7625
|
name: "Inception: Mercury",
|
|
@@ -7569,23 +7671,6 @@ export const MODELS = {
|
|
|
7569
7671
|
contextWindow: 128000,
|
|
7570
7672
|
maxTokens: 32000,
|
|
7571
7673
|
},
|
|
7572
|
-
"kwaipilot/kat-coder-pro": {
|
|
7573
|
-
id: "kwaipilot/kat-coder-pro",
|
|
7574
|
-
name: "Kwaipilot: KAT-Coder-Pro V1",
|
|
7575
|
-
api: "openai-completions",
|
|
7576
|
-
provider: "openrouter",
|
|
7577
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7578
|
-
reasoning: false,
|
|
7579
|
-
input: ["text"],
|
|
7580
|
-
cost: {
|
|
7581
|
-
input: 0.207,
|
|
7582
|
-
output: 0.828,
|
|
7583
|
-
cacheRead: 0.0414,
|
|
7584
|
-
cacheWrite: 0,
|
|
7585
|
-
},
|
|
7586
|
-
contextWindow: 256000,
|
|
7587
|
-
maxTokens: 128000,
|
|
7588
|
-
},
|
|
7589
7674
|
"kwaipilot/kat-coder-pro-v2": {
|
|
7590
7675
|
id: "kwaipilot/kat-coder-pro-v2",
|
|
7591
7676
|
name: "Kwaipilot: KAT-Coder-Pro V2",
|
|
@@ -7799,9 +7884,9 @@ export const MODELS = {
|
|
|
7799
7884
|
reasoning: true,
|
|
7800
7885
|
input: ["text"],
|
|
7801
7886
|
cost: {
|
|
7802
|
-
input: 0.
|
|
7803
|
-
output:
|
|
7804
|
-
cacheRead: 0.
|
|
7887
|
+
input: 0.118,
|
|
7888
|
+
output: 0.9900000000000001,
|
|
7889
|
+
cacheRead: 0.059,
|
|
7805
7890
|
cacheWrite: 0,
|
|
7806
7891
|
},
|
|
7807
7892
|
contextWindow: 196608,
|
|
@@ -8096,23 +8181,6 @@ export const MODELS = {
|
|
|
8096
8181
|
contextWindow: 32768,
|
|
8097
8182
|
maxTokens: 4096,
|
|
8098
8183
|
},
|
|
8099
|
-
"mistralai/mistral-small-24b-instruct-2501": {
|
|
8100
|
-
id: "mistralai/mistral-small-24b-instruct-2501",
|
|
8101
|
-
name: "Mistral: Mistral Small 3",
|
|
8102
|
-
api: "openai-completions",
|
|
8103
|
-
provider: "openrouter",
|
|
8104
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
8105
|
-
reasoning: false,
|
|
8106
|
-
input: ["text"],
|
|
8107
|
-
cost: {
|
|
8108
|
-
input: 0.049999999999999996,
|
|
8109
|
-
output: 0.08,
|
|
8110
|
-
cacheRead: 0,
|
|
8111
|
-
cacheWrite: 0,
|
|
8112
|
-
},
|
|
8113
|
-
contextWindow: 32768,
|
|
8114
|
-
maxTokens: 16384,
|
|
8115
|
-
},
|
|
8116
8184
|
"mistralai/mistral-small-2603": {
|
|
8117
8185
|
id: "mistralai/mistral-small-2603",
|
|
8118
8186
|
name: "Mistral: Mistral Small 4",
|
|
@@ -9167,6 +9235,40 @@ export const MODELS = {
|
|
|
9167
9235
|
contextWindow: 1050000,
|
|
9168
9236
|
maxTokens: 128000,
|
|
9169
9237
|
},
|
|
9238
|
+
"openai/gpt-audio": {
|
|
9239
|
+
id: "openai/gpt-audio",
|
|
9240
|
+
name: "OpenAI: GPT Audio",
|
|
9241
|
+
api: "openai-completions",
|
|
9242
|
+
provider: "openrouter",
|
|
9243
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9244
|
+
reasoning: false,
|
|
9245
|
+
input: ["text"],
|
|
9246
|
+
cost: {
|
|
9247
|
+
input: 2.5,
|
|
9248
|
+
output: 10,
|
|
9249
|
+
cacheRead: 0,
|
|
9250
|
+
cacheWrite: 0,
|
|
9251
|
+
},
|
|
9252
|
+
contextWindow: 128000,
|
|
9253
|
+
maxTokens: 16384,
|
|
9254
|
+
},
|
|
9255
|
+
"openai/gpt-audio-mini": {
|
|
9256
|
+
id: "openai/gpt-audio-mini",
|
|
9257
|
+
name: "OpenAI: GPT Audio Mini",
|
|
9258
|
+
api: "openai-completions",
|
|
9259
|
+
provider: "openrouter",
|
|
9260
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9261
|
+
reasoning: false,
|
|
9262
|
+
input: ["text"],
|
|
9263
|
+
cost: {
|
|
9264
|
+
input: 0.6,
|
|
9265
|
+
output: 2.4,
|
|
9266
|
+
cacheRead: 0,
|
|
9267
|
+
cacheWrite: 0,
|
|
9268
|
+
},
|
|
9269
|
+
contextWindow: 128000,
|
|
9270
|
+
maxTokens: 16384,
|
|
9271
|
+
},
|
|
9170
9272
|
"openai/gpt-oss-120b": {
|
|
9171
9273
|
id: "openai/gpt-oss-120b",
|
|
9172
9274
|
name: "OpenAI: gpt-oss-120b",
|
|
@@ -10134,7 +10236,7 @@ export const MODELS = {
|
|
|
10134
10236
|
cacheWrite: 0,
|
|
10135
10237
|
},
|
|
10136
10238
|
contextWindow: 256000,
|
|
10137
|
-
maxTokens:
|
|
10239
|
+
maxTokens: 32768,
|
|
10138
10240
|
},
|
|
10139
10241
|
"qwen/qwen3.5-flash-02-23": {
|
|
10140
10242
|
id: "qwen/qwen3.5-flash-02-23",
|
|
@@ -10170,6 +10272,23 @@ export const MODELS = {
|
|
|
10170
10272
|
contextWindow: 1000000,
|
|
10171
10273
|
maxTokens: 65536,
|
|
10172
10274
|
},
|
|
10275
|
+
"qwen/qwen3.6-plus:free": {
|
|
10276
|
+
id: "qwen/qwen3.6-plus:free",
|
|
10277
|
+
name: "Qwen: Qwen3.6 Plus (free)",
|
|
10278
|
+
api: "openai-completions",
|
|
10279
|
+
provider: "openrouter",
|
|
10280
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10281
|
+
reasoning: true,
|
|
10282
|
+
input: ["text", "image"],
|
|
10283
|
+
cost: {
|
|
10284
|
+
input: 0,
|
|
10285
|
+
output: 0,
|
|
10286
|
+
cacheRead: 0,
|
|
10287
|
+
cacheWrite: 0,
|
|
10288
|
+
},
|
|
10289
|
+
contextWindow: 1000000,
|
|
10290
|
+
maxTokens: 65536,
|
|
10291
|
+
},
|
|
10173
10292
|
"qwen/qwq-32b": {
|
|
10174
10293
|
id: "qwen/qwq-32b",
|
|
10175
10294
|
name: "Qwen: QwQ 32B",
|
|
@@ -10187,8 +10306,8 @@ export const MODELS = {
|
|
|
10187
10306
|
contextWindow: 131072,
|
|
10188
10307
|
maxTokens: 131072,
|
|
10189
10308
|
},
|
|
10190
|
-
"
|
|
10191
|
-
id: "
|
|
10309
|
+
"rekaai/reka-edge": {
|
|
10310
|
+
id: "rekaai/reka-edge",
|
|
10192
10311
|
name: "Reka Edge",
|
|
10193
10312
|
api: "openai-completions",
|
|
10194
10313
|
provider: "openrouter",
|
|
@@ -10266,11 +10385,11 @@ export const MODELS = {
|
|
|
10266
10385
|
cost: {
|
|
10267
10386
|
input: 0.09999999999999999,
|
|
10268
10387
|
output: 0.3,
|
|
10269
|
-
cacheRead: 0
|
|
10388
|
+
cacheRead: 0,
|
|
10270
10389
|
cacheWrite: 0,
|
|
10271
10390
|
},
|
|
10272
10391
|
contextWindow: 262144,
|
|
10273
|
-
maxTokens:
|
|
10392
|
+
maxTokens: 65536,
|
|
10274
10393
|
},
|
|
10275
10394
|
"stepfun/step-3.5-flash:free": {
|
|
10276
10395
|
id: "stepfun/step-3.5-flash:free",
|
|
@@ -10476,9 +10595,9 @@ export const MODELS = {
|
|
|
10476
10595
|
contextWindow: 2000000,
|
|
10477
10596
|
maxTokens: 30000,
|
|
10478
10597
|
},
|
|
10479
|
-
"x-ai/grok-4.20
|
|
10480
|
-
id: "x-ai/grok-4.20
|
|
10481
|
-
name: "xAI: Grok 4.20
|
|
10598
|
+
"x-ai/grok-4.20": {
|
|
10599
|
+
id: "x-ai/grok-4.20",
|
|
10600
|
+
name: "xAI: Grok 4.20",
|
|
10482
10601
|
api: "openai-completions",
|
|
10483
10602
|
provider: "openrouter",
|
|
10484
10603
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
@@ -10748,6 +10867,23 @@ export const MODELS = {
|
|
|
10748
10867
|
contextWindow: 202752,
|
|
10749
10868
|
maxTokens: 131072,
|
|
10750
10869
|
},
|
|
10870
|
+
"z-ai/glm-5v-turbo": {
|
|
10871
|
+
id: "z-ai/glm-5v-turbo",
|
|
10872
|
+
name: "Z.ai: GLM 5V Turbo",
|
|
10873
|
+
api: "openai-completions",
|
|
10874
|
+
provider: "openrouter",
|
|
10875
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10876
|
+
reasoning: true,
|
|
10877
|
+
input: ["text", "image"],
|
|
10878
|
+
cost: {
|
|
10879
|
+
input: 1.2,
|
|
10880
|
+
output: 4,
|
|
10881
|
+
cacheRead: 0.24,
|
|
10882
|
+
cacheWrite: 0,
|
|
10883
|
+
},
|
|
10884
|
+
contextWindow: 202752,
|
|
10885
|
+
maxTokens: 131072,
|
|
10886
|
+
},
|
|
10751
10887
|
},
|
|
10752
10888
|
"vercel-ai-gateway": {
|
|
10753
10889
|
"alibaba/qwen-3-14b": {
|
|
@@ -10776,12 +10912,12 @@ export const MODELS = {
|
|
|
10776
10912
|
reasoning: false,
|
|
10777
10913
|
input: ["text"],
|
|
10778
10914
|
cost: {
|
|
10779
|
-
input: 0.
|
|
10780
|
-
output: 0.
|
|
10781
|
-
cacheRead: 0,
|
|
10915
|
+
input: 0.22,
|
|
10916
|
+
output: 0.88,
|
|
10917
|
+
cacheRead: 0.11,
|
|
10782
10918
|
cacheWrite: 0,
|
|
10783
10919
|
},
|
|
10784
|
-
contextWindow:
|
|
10920
|
+
contextWindow: 32768,
|
|
10785
10921
|
maxTokens: 16384,
|
|
10786
10922
|
},
|
|
10787
10923
|
"alibaba/qwen-3-30b": {
|
|
@@ -10810,13 +10946,13 @@ export const MODELS = {
|
|
|
10810
10946
|
reasoning: true,
|
|
10811
10947
|
input: ["text"],
|
|
10812
10948
|
cost: {
|
|
10813
|
-
input: 0.
|
|
10814
|
-
output: 0.
|
|
10815
|
-
cacheRead: 0
|
|
10949
|
+
input: 0.16,
|
|
10950
|
+
output: 0.64,
|
|
10951
|
+
cacheRead: 0,
|
|
10816
10952
|
cacheWrite: 0,
|
|
10817
10953
|
},
|
|
10818
|
-
contextWindow:
|
|
10819
|
-
maxTokens:
|
|
10954
|
+
contextWindow: 128000,
|
|
10955
|
+
maxTokens: 8192,
|
|
10820
10956
|
},
|
|
10821
10957
|
"alibaba/qwen3-235b-a22b-thinking": {
|
|
10822
10958
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
@@ -10844,13 +10980,13 @@ export const MODELS = {
|
|
|
10844
10980
|
reasoning: false,
|
|
10845
10981
|
input: ["text"],
|
|
10846
10982
|
cost: {
|
|
10847
|
-
input:
|
|
10848
|
-
output:
|
|
10849
|
-
cacheRead: 0.
|
|
10983
|
+
input: 1.5,
|
|
10984
|
+
output: 7.5,
|
|
10985
|
+
cacheRead: 0.3,
|
|
10850
10986
|
cacheWrite: 0,
|
|
10851
10987
|
},
|
|
10852
10988
|
contextWindow: 262144,
|
|
10853
|
-
maxTokens:
|
|
10989
|
+
maxTokens: 65536,
|
|
10854
10990
|
},
|
|
10855
10991
|
"alibaba/qwen3-coder-30b-a3b": {
|
|
10856
10992
|
id: "alibaba/qwen3-coder-30b-a3b",
|
|
@@ -11005,6 +11141,23 @@ export const MODELS = {
|
|
|
11005
11141
|
contextWindow: 1000000,
|
|
11006
11142
|
maxTokens: 64000,
|
|
11007
11143
|
},
|
|
11144
|
+
"alibaba/qwen3.6-plus": {
|
|
11145
|
+
id: "alibaba/qwen3.6-plus",
|
|
11146
|
+
name: "Qwen 3.6 Plus",
|
|
11147
|
+
api: "anthropic-messages",
|
|
11148
|
+
provider: "vercel-ai-gateway",
|
|
11149
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11150
|
+
reasoning: true,
|
|
11151
|
+
input: ["text", "image"],
|
|
11152
|
+
cost: {
|
|
11153
|
+
input: 0.5,
|
|
11154
|
+
output: 3,
|
|
11155
|
+
cacheRead: 0.09999999999999999,
|
|
11156
|
+
cacheWrite: 0.625,
|
|
11157
|
+
},
|
|
11158
|
+
contextWindow: 1000000,
|
|
11159
|
+
maxTokens: 64000,
|
|
11160
|
+
},
|
|
11008
11161
|
"anthropic/claude-3-haiku": {
|
|
11009
11162
|
id: "anthropic/claude-3-haiku",
|
|
11010
11163
|
name: "Claude 3 Haiku",
|
|
@@ -11243,6 +11396,23 @@ export const MODELS = {
|
|
|
11243
11396
|
contextWindow: 131000,
|
|
11244
11397
|
maxTokens: 131000,
|
|
11245
11398
|
},
|
|
11399
|
+
"arcee-ai/trinity-large-thinking": {
|
|
11400
|
+
id: "arcee-ai/trinity-large-thinking",
|
|
11401
|
+
name: "Trinity Large Thinking",
|
|
11402
|
+
api: "anthropic-messages",
|
|
11403
|
+
provider: "vercel-ai-gateway",
|
|
11404
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11405
|
+
reasoning: true,
|
|
11406
|
+
input: ["text"],
|
|
11407
|
+
cost: {
|
|
11408
|
+
input: 0.25,
|
|
11409
|
+
output: 0.8999999999999999,
|
|
11410
|
+
cacheRead: 0,
|
|
11411
|
+
cacheWrite: 0,
|
|
11412
|
+
},
|
|
11413
|
+
contextWindow: 262100,
|
|
11414
|
+
maxTokens: 80000,
|
|
11415
|
+
},
|
|
11246
11416
|
"bytedance/seed-1.6": {
|
|
11247
11417
|
id: "bytedance/seed-1.6",
|
|
11248
11418
|
name: "Seed 1.6",
|
|
@@ -11320,13 +11490,13 @@ export const MODELS = {
|
|
|
11320
11490
|
reasoning: true,
|
|
11321
11491
|
input: ["text"],
|
|
11322
11492
|
cost: {
|
|
11323
|
-
input: 0.
|
|
11324
|
-
output: 1.
|
|
11325
|
-
cacheRead: 0,
|
|
11493
|
+
input: 0.56,
|
|
11494
|
+
output: 1.68,
|
|
11495
|
+
cacheRead: 0.28,
|
|
11326
11496
|
cacheWrite: 0,
|
|
11327
11497
|
},
|
|
11328
11498
|
contextWindow: 163840,
|
|
11329
|
-
maxTokens:
|
|
11499
|
+
maxTokens: 8192,
|
|
11330
11500
|
},
|
|
11331
11501
|
"deepseek/deepseek-v3.1-terminus": {
|
|
11332
11502
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -11532,6 +11702,40 @@ export const MODELS = {
|
|
|
11532
11702
|
contextWindow: 1000000,
|
|
11533
11703
|
maxTokens: 64000,
|
|
11534
11704
|
},
|
|
11705
|
+
"google/gemma-4-26b-a4b-it": {
|
|
11706
|
+
id: "google/gemma-4-26b-a4b-it",
|
|
11707
|
+
name: "Gemma 4 26B A4B IT",
|
|
11708
|
+
api: "anthropic-messages",
|
|
11709
|
+
provider: "vercel-ai-gateway",
|
|
11710
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11711
|
+
reasoning: true,
|
|
11712
|
+
input: ["text", "image"],
|
|
11713
|
+
cost: {
|
|
11714
|
+
input: 0.13,
|
|
11715
|
+
output: 0.39999999999999997,
|
|
11716
|
+
cacheRead: 0,
|
|
11717
|
+
cacheWrite: 0,
|
|
11718
|
+
},
|
|
11719
|
+
contextWindow: 262144,
|
|
11720
|
+
maxTokens: 131072,
|
|
11721
|
+
},
|
|
11722
|
+
"google/gemma-4-31b-it": {
|
|
11723
|
+
id: "google/gemma-4-31b-it",
|
|
11724
|
+
name: "Gemma 4 31B IT",
|
|
11725
|
+
api: "anthropic-messages",
|
|
11726
|
+
provider: "vercel-ai-gateway",
|
|
11727
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
11728
|
+
reasoning: true,
|
|
11729
|
+
input: ["text", "image"],
|
|
11730
|
+
cost: {
|
|
11731
|
+
input: 0.14,
|
|
11732
|
+
output: 0.39999999999999997,
|
|
11733
|
+
cacheRead: 0,
|
|
11734
|
+
cacheWrite: 0,
|
|
11735
|
+
},
|
|
11736
|
+
contextWindow: 262144,
|
|
11737
|
+
maxTokens: 131072,
|
|
11738
|
+
},
|
|
11535
11739
|
"inception/mercury-2": {
|
|
11536
11740
|
id: "inception/mercury-2",
|
|
11537
11741
|
name: "Mercury 2",
|
|
@@ -11598,7 +11802,7 @@ export const MODELS = {
|
|
|
11598
11802
|
cacheWrite: 0,
|
|
11599
11803
|
},
|
|
11600
11804
|
contextWindow: 128000,
|
|
11601
|
-
maxTokens:
|
|
11805
|
+
maxTokens: 8192,
|
|
11602
11806
|
},
|
|
11603
11807
|
"meituan/longcat-flash-thinking": {
|
|
11604
11808
|
id: "meituan/longcat-flash-thinking",
|
|
@@ -11643,13 +11847,13 @@ export const MODELS = {
|
|
|
11643
11847
|
reasoning: false,
|
|
11644
11848
|
input: ["text"],
|
|
11645
11849
|
cost: {
|
|
11646
|
-
input: 0.
|
|
11647
|
-
output: 0.
|
|
11648
|
-
cacheRead: 0
|
|
11850
|
+
input: 0.22,
|
|
11851
|
+
output: 0.22,
|
|
11852
|
+
cacheRead: 0,
|
|
11649
11853
|
cacheWrite: 0,
|
|
11650
11854
|
},
|
|
11651
11855
|
contextWindow: 128000,
|
|
11652
|
-
maxTokens:
|
|
11856
|
+
maxTokens: 8192,
|
|
11653
11857
|
},
|
|
11654
11858
|
"meta/llama-3.2-11b": {
|
|
11655
11859
|
id: "meta/llama-3.2-11b",
|
|
@@ -11711,12 +11915,12 @@ export const MODELS = {
|
|
|
11711
11915
|
reasoning: false,
|
|
11712
11916
|
input: ["text", "image"],
|
|
11713
11917
|
cost: {
|
|
11714
|
-
input: 0.
|
|
11715
|
-
output:
|
|
11918
|
+
input: 0.35,
|
|
11919
|
+
output: 1.15,
|
|
11716
11920
|
cacheRead: 0,
|
|
11717
11921
|
cacheWrite: 0,
|
|
11718
11922
|
},
|
|
11719
|
-
contextWindow:
|
|
11923
|
+
contextWindow: 524288,
|
|
11720
11924
|
maxTokens: 8192,
|
|
11721
11925
|
},
|
|
11722
11926
|
"meta/llama-4-scout": {
|
|
@@ -12036,7 +12240,7 @@ export const MODELS = {
|
|
|
12036
12240
|
cost: {
|
|
12037
12241
|
input: 0.6,
|
|
12038
12242
|
output: 2.5,
|
|
12039
|
-
cacheRead: 0
|
|
12243
|
+
cacheRead: 0,
|
|
12040
12244
|
cacheWrite: 0,
|
|
12041
12245
|
},
|
|
12042
12246
|
contextWindow: 131072,
|
|
@@ -12053,11 +12257,11 @@ export const MODELS = {
|
|
|
12053
12257
|
cost: {
|
|
12054
12258
|
input: 0.6,
|
|
12055
12259
|
output: 2.5,
|
|
12056
|
-
cacheRead: 0.
|
|
12260
|
+
cacheRead: 0.3,
|
|
12057
12261
|
cacheWrite: 0,
|
|
12058
12262
|
},
|
|
12059
12263
|
contextWindow: 256000,
|
|
12060
|
-
maxTokens:
|
|
12264
|
+
maxTokens: 128000,
|
|
12061
12265
|
},
|
|
12062
12266
|
"moonshotai/kimi-k2-thinking": {
|
|
12063
12267
|
id: "moonshotai/kimi-k2-thinking",
|
|
@@ -12620,6 +12824,23 @@ export const MODELS = {
|
|
|
12620
12824
|
contextWindow: 1050000,
|
|
12621
12825
|
maxTokens: 128000,
|
|
12622
12826
|
},
|
|
12827
|
+
"openai/gpt-oss-120b": {
|
|
12828
|
+
id: "openai/gpt-oss-120b",
|
|
12829
|
+
name: "gpt-oss-120b",
|
|
12830
|
+
api: "anthropic-messages",
|
|
12831
|
+
provider: "vercel-ai-gateway",
|
|
12832
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12833
|
+
reasoning: true,
|
|
12834
|
+
input: ["text"],
|
|
12835
|
+
cost: {
|
|
12836
|
+
input: 0.15,
|
|
12837
|
+
output: 0.6,
|
|
12838
|
+
cacheRead: 0,
|
|
12839
|
+
cacheWrite: 0,
|
|
12840
|
+
},
|
|
12841
|
+
contextWindow: 128000,
|
|
12842
|
+
maxTokens: 8192,
|
|
12843
|
+
},
|
|
12623
12844
|
"openai/gpt-oss-20b": {
|
|
12624
12845
|
id: "openai/gpt-oss-20b",
|
|
12625
12846
|
name: "gpt-oss-20b",
|
|
@@ -12629,12 +12850,12 @@ export const MODELS = {
|
|
|
12629
12850
|
reasoning: true,
|
|
12630
12851
|
input: ["text"],
|
|
12631
12852
|
cost: {
|
|
12632
|
-
input: 0.
|
|
12633
|
-
output: 0.
|
|
12853
|
+
input: 0.049999999999999996,
|
|
12854
|
+
output: 0.19999999999999998,
|
|
12634
12855
|
cacheRead: 0,
|
|
12635
12856
|
cacheWrite: 0,
|
|
12636
12857
|
},
|
|
12637
|
-
contextWindow:
|
|
12858
|
+
contextWindow: 131072,
|
|
12638
12859
|
maxTokens: 8192,
|
|
12639
12860
|
},
|
|
12640
12861
|
"openai/gpt-oss-safeguard-20b": {
|
|
@@ -13105,9 +13326,9 @@ export const MODELS = {
|
|
|
13105
13326
|
reasoning: true,
|
|
13106
13327
|
input: ["text"],
|
|
13107
13328
|
cost: {
|
|
13108
|
-
input: 0.
|
|
13109
|
-
output: 0.
|
|
13110
|
-
cacheRead: 0.
|
|
13329
|
+
input: 0.09,
|
|
13330
|
+
output: 0.29,
|
|
13331
|
+
cacheRead: 0.045,
|
|
13111
13332
|
cacheWrite: 0,
|
|
13112
13333
|
},
|
|
13113
13334
|
contextWindow: 262144,
|
|
@@ -13243,7 +13464,7 @@ export const MODELS = {
|
|
|
13243
13464
|
cost: {
|
|
13244
13465
|
input: 0.6,
|
|
13245
13466
|
output: 2.2,
|
|
13246
|
-
cacheRead: 0,
|
|
13467
|
+
cacheRead: 0.11,
|
|
13247
13468
|
cacheWrite: 0,
|
|
13248
13469
|
},
|
|
13249
13470
|
contextWindow: 200000,
|
|
@@ -13283,6 +13504,23 @@ export const MODELS = {
|
|
|
13283
13504
|
contextWindow: 200000,
|
|
13284
13505
|
maxTokens: 128000,
|
|
13285
13506
|
},
|
|
13507
|
+
"zai/glm-5": {
|
|
13508
|
+
id: "zai/glm-5",
|
|
13509
|
+
name: "GLM 5",
|
|
13510
|
+
api: "anthropic-messages",
|
|
13511
|
+
provider: "vercel-ai-gateway",
|
|
13512
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13513
|
+
reasoning: true,
|
|
13514
|
+
input: ["text"],
|
|
13515
|
+
cost: {
|
|
13516
|
+
input: 1,
|
|
13517
|
+
output: 3.1999999999999997,
|
|
13518
|
+
cacheRead: 0.19999999999999998,
|
|
13519
|
+
cacheWrite: 0,
|
|
13520
|
+
},
|
|
13521
|
+
contextWindow: 202800,
|
|
13522
|
+
maxTokens: 131100,
|
|
13523
|
+
},
|
|
13286
13524
|
"zai/glm-5-turbo": {
|
|
13287
13525
|
id: "zai/glm-5-turbo",
|
|
13288
13526
|
name: "GLM 5 Turbo",
|
|
@@ -13300,6 +13538,23 @@ export const MODELS = {
|
|
|
13300
13538
|
contextWindow: 202800,
|
|
13301
13539
|
maxTokens: 131100,
|
|
13302
13540
|
},
|
|
13541
|
+
"zai/glm-5v-turbo": {
|
|
13542
|
+
id: "zai/glm-5v-turbo",
|
|
13543
|
+
name: "GLM 5V Turbo",
|
|
13544
|
+
api: "anthropic-messages",
|
|
13545
|
+
provider: "vercel-ai-gateway",
|
|
13546
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13547
|
+
reasoning: true,
|
|
13548
|
+
input: ["text", "image"],
|
|
13549
|
+
cost: {
|
|
13550
|
+
input: 1.2,
|
|
13551
|
+
output: 4,
|
|
13552
|
+
cacheRead: 0.24,
|
|
13553
|
+
cacheWrite: 0,
|
|
13554
|
+
},
|
|
13555
|
+
contextWindow: 200000,
|
|
13556
|
+
maxTokens: 128000,
|
|
13557
|
+
},
|
|
13303
13558
|
},
|
|
13304
13559
|
"xai": {
|
|
13305
13560
|
"grok-2": {
|
|
@@ -13790,7 +14045,7 @@ export const MODELS = {
|
|
|
13790
14045
|
api: "openai-completions",
|
|
13791
14046
|
provider: "zai",
|
|
13792
14047
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13793
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14048
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13794
14049
|
reasoning: true,
|
|
13795
14050
|
input: ["text"],
|
|
13796
14051
|
cost: {
|
|
@@ -13808,7 +14063,7 @@ export const MODELS = {
|
|
|
13808
14063
|
api: "openai-completions",
|
|
13809
14064
|
provider: "zai",
|
|
13810
14065
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13811
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14066
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13812
14067
|
reasoning: true,
|
|
13813
14068
|
input: ["text", "image"],
|
|
13814
14069
|
cost: {
|
|
@@ -13826,7 +14081,7 @@ export const MODELS = {
|
|
|
13826
14081
|
api: "openai-completions",
|
|
13827
14082
|
provider: "zai",
|
|
13828
14083
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13829
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14084
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13830
14085
|
reasoning: true,
|
|
13831
14086
|
input: ["text"],
|
|
13832
14087
|
cost: {
|
|
@@ -13844,7 +14099,7 @@ export const MODELS = {
|
|
|
13844
14099
|
api: "openai-completions",
|
|
13845
14100
|
provider: "zai",
|
|
13846
14101
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13847
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14102
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13848
14103
|
reasoning: true,
|
|
13849
14104
|
input: ["text"],
|
|
13850
14105
|
cost: {
|
|
@@ -13862,7 +14117,7 @@ export const MODELS = {
|
|
|
13862
14117
|
api: "openai-completions",
|
|
13863
14118
|
provider: "zai",
|
|
13864
14119
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13865
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14120
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13866
14121
|
reasoning: true,
|
|
13867
14122
|
input: ["text"],
|
|
13868
14123
|
cost: {
|
|
@@ -13880,7 +14135,7 @@ export const MODELS = {
|
|
|
13880
14135
|
api: "openai-completions",
|
|
13881
14136
|
provider: "zai",
|
|
13882
14137
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13883
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14138
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13884
14139
|
reasoning: true,
|
|
13885
14140
|
input: ["text"],
|
|
13886
14141
|
cost: {
|
|
@@ -13898,7 +14153,7 @@ export const MODELS = {
|
|
|
13898
14153
|
api: "openai-completions",
|
|
13899
14154
|
provider: "zai",
|
|
13900
14155
|
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
13901
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14156
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
13902
14157
|
reasoning: true,
|
|
13903
14158
|
input: ["text"],
|
|
13904
14159
|
cost: {
|
|
@@ -13910,6 +14165,24 @@ export const MODELS = {
|
|
|
13910
14165
|
contextWindow: 200000,
|
|
13911
14166
|
maxTokens: 131072,
|
|
13912
14167
|
},
|
|
14168
|
+
"glm-5v-turbo": {
|
|
14169
|
+
id: "glm-5v-turbo",
|
|
14170
|
+
name: "glm-5v-turbo",
|
|
14171
|
+
api: "openai-completions",
|
|
14172
|
+
provider: "zai",
|
|
14173
|
+
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14174
|
+
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
14175
|
+
reasoning: true,
|
|
14176
|
+
input: ["text", "image"],
|
|
14177
|
+
cost: {
|
|
14178
|
+
input: 1.2,
|
|
14179
|
+
output: 4,
|
|
14180
|
+
cacheRead: 0.24,
|
|
14181
|
+
cacheWrite: 0,
|
|
14182
|
+
},
|
|
14183
|
+
contextWindow: 200000,
|
|
14184
|
+
maxTokens: 131072,
|
|
14185
|
+
},
|
|
13913
14186
|
},
|
|
13914
14187
|
};
|
|
13915
14188
|
//# sourceMappingURL=models.generated.js.map
|