@hyperspaceng/neural-ai 0.69.1 → 0.70.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +4 -1
- package/dist/env-api-keys.d.ts +9 -0
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +42 -31
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.d.ts +2 -1
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +298 -195
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +291 -200
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +7 -2
- package/dist/models.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +35 -15
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/azure-openai-responses.d.ts.map +1 -1
- package/dist/providers/azure-openai-responses.js +7 -4
- package/dist/providers/azure-openai-responses.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +34 -13
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +9 -8
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +97 -45
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +26 -23
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +2 -0
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/types.d.ts +35 -4
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -2536,8 +2536,8 @@ export const MODELS = {
|
|
|
2536
2536
|
cacheRead: 0.5,
|
|
2537
2537
|
cacheWrite: 0,
|
|
2538
2538
|
},
|
|
2539
|
-
contextWindow:
|
|
2540
|
-
maxTokens:
|
|
2539
|
+
contextWindow: 272000,
|
|
2540
|
+
maxTokens: 128000,
|
|
2541
2541
|
},
|
|
2542
2542
|
"o1": {
|
|
2543
2543
|
id: "o1",
|
|
@@ -2746,6 +2746,44 @@ export const MODELS = {
|
|
|
2746
2746
|
maxTokens: 40000,
|
|
2747
2747
|
},
|
|
2748
2748
|
},
|
|
2749
|
+
"deepseek": {
|
|
2750
|
+
"deepseek-v4-flash": {
|
|
2751
|
+
id: "deepseek-v4-flash",
|
|
2752
|
+
name: "DeepSeek V4 Flash",
|
|
2753
|
+
api: "openai-completions",
|
|
2754
|
+
provider: "deepseek",
|
|
2755
|
+
baseUrl: "https://api.deepseek.com",
|
|
2756
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek", "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
|
|
2757
|
+
reasoning: true,
|
|
2758
|
+
input: ["text"],
|
|
2759
|
+
cost: {
|
|
2760
|
+
input: 0.14,
|
|
2761
|
+
output: 0.28,
|
|
2762
|
+
cacheRead: 0.028,
|
|
2763
|
+
cacheWrite: 0,
|
|
2764
|
+
},
|
|
2765
|
+
contextWindow: 1000000,
|
|
2766
|
+
maxTokens: 384000,
|
|
2767
|
+
},
|
|
2768
|
+
"deepseek-v4-pro": {
|
|
2769
|
+
id: "deepseek-v4-pro",
|
|
2770
|
+
name: "DeepSeek V4 Pro",
|
|
2771
|
+
api: "openai-completions",
|
|
2772
|
+
provider: "deepseek",
|
|
2773
|
+
baseUrl: "https://api.deepseek.com",
|
|
2774
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek", "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
|
|
2775
|
+
reasoning: true,
|
|
2776
|
+
input: ["text"],
|
|
2777
|
+
cost: {
|
|
2778
|
+
input: 1.74,
|
|
2779
|
+
output: 3.48,
|
|
2780
|
+
cacheRead: 0.145,
|
|
2781
|
+
cacheWrite: 0,
|
|
2782
|
+
},
|
|
2783
|
+
contextWindow: 1000000,
|
|
2784
|
+
maxTokens: 384000,
|
|
2785
|
+
},
|
|
2786
|
+
},
|
|
2749
2787
|
"fireworks": {
|
|
2750
2788
|
"accounts/fireworks/models/deepseek-v3p1": {
|
|
2751
2789
|
id: "accounts/fireworks/models/deepseek-v3p1",
|
|
@@ -3062,6 +3100,7 @@ export const MODELS = {
|
|
|
3062
3100
|
provider: "github-copilot",
|
|
3063
3101
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3064
3102
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3103
|
+
compat: { "supportsEagerToolInputStreaming": false },
|
|
3065
3104
|
reasoning: true,
|
|
3066
3105
|
input: ["text", "image"],
|
|
3067
3106
|
cost: {
|
|
@@ -3134,6 +3173,7 @@ export const MODELS = {
|
|
|
3134
3173
|
provider: "github-copilot",
|
|
3135
3174
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3136
3175
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3176
|
+
compat: { "supportsEagerToolInputStreaming": false },
|
|
3137
3177
|
reasoning: true,
|
|
3138
3178
|
input: ["text", "image"],
|
|
3139
3179
|
cost: {
|
|
@@ -3152,6 +3192,7 @@ export const MODELS = {
|
|
|
3152
3192
|
provider: "github-copilot",
|
|
3153
3193
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3154
3194
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3195
|
+
compat: { "supportsEagerToolInputStreaming": false },
|
|
3155
3196
|
reasoning: true,
|
|
3156
3197
|
input: ["text", "image"],
|
|
3157
3198
|
cost: {
|
|
@@ -3493,6 +3534,24 @@ export const MODELS = {
|
|
|
3493
3534
|
contextWindow: 400000,
|
|
3494
3535
|
maxTokens: 128000,
|
|
3495
3536
|
},
|
|
3537
|
+
"gpt-5.5": {
|
|
3538
|
+
id: "gpt-5.5",
|
|
3539
|
+
name: "GPT-5.5",
|
|
3540
|
+
api: "openai-responses",
|
|
3541
|
+
provider: "github-copilot",
|
|
3542
|
+
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3543
|
+
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3544
|
+
reasoning: true,
|
|
3545
|
+
input: ["text", "image"],
|
|
3546
|
+
cost: {
|
|
3547
|
+
input: 0,
|
|
3548
|
+
output: 0,
|
|
3549
|
+
cacheRead: 0,
|
|
3550
|
+
cacheWrite: 0,
|
|
3551
|
+
},
|
|
3552
|
+
contextWindow: 400000,
|
|
3553
|
+
maxTokens: 128000,
|
|
3554
|
+
},
|
|
3496
3555
|
"grok-code-fast-1": {
|
|
3497
3556
|
id: "grok-code-fast-1",
|
|
3498
3557
|
name: "Grok Code Fast 1",
|
|
@@ -5168,6 +5227,7 @@ export const MODELS = {
|
|
|
5168
5227
|
api: "anthropic-messages",
|
|
5169
5228
|
provider: "kimi-coding",
|
|
5170
5229
|
baseUrl: "https://api.kimi.com/coding",
|
|
5230
|
+
headers: { "User-Agent": "KimiCLI/1.5" },
|
|
5171
5231
|
reasoning: true,
|
|
5172
5232
|
input: ["text", "image"],
|
|
5173
5233
|
cost: {
|
|
@@ -5185,6 +5245,7 @@ export const MODELS = {
|
|
|
5185
5245
|
api: "anthropic-messages",
|
|
5186
5246
|
provider: "kimi-coding",
|
|
5187
5247
|
baseUrl: "https://api.kimi.com/coding",
|
|
5248
|
+
headers: { "User-Agent": "KimiCLI/1.5" },
|
|
5188
5249
|
reasoning: true,
|
|
5189
5250
|
input: ["text", "image"],
|
|
5190
5251
|
cost: {
|
|
@@ -5202,6 +5263,7 @@ export const MODELS = {
|
|
|
5202
5263
|
api: "anthropic-messages",
|
|
5203
5264
|
provider: "kimi-coding",
|
|
5204
5265
|
baseUrl: "https://api.kimi.com/coding",
|
|
5266
|
+
headers: { "User-Agent": "KimiCLI/1.5" },
|
|
5205
5267
|
reasoning: true,
|
|
5206
5268
|
input: ["text"],
|
|
5207
5269
|
cost: {
|
|
@@ -6289,8 +6351,8 @@ export const MODELS = {
|
|
|
6289
6351
|
cacheRead: 0.5,
|
|
6290
6352
|
cacheWrite: 0,
|
|
6291
6353
|
},
|
|
6292
|
-
contextWindow:
|
|
6293
|
-
maxTokens:
|
|
6354
|
+
contextWindow: 272000,
|
|
6355
|
+
maxTokens: 128000,
|
|
6294
6356
|
},
|
|
6295
6357
|
"o1": {
|
|
6296
6358
|
id: "o1",
|
|
@@ -6583,6 +6645,23 @@ export const MODELS = {
|
|
|
6583
6645
|
contextWindow: 272000,
|
|
6584
6646
|
maxTokens: 128000,
|
|
6585
6647
|
},
|
|
6648
|
+
"gpt-5.5": {
|
|
6649
|
+
id: "gpt-5.5",
|
|
6650
|
+
name: "GPT-5.5",
|
|
6651
|
+
api: "openai-codex-responses",
|
|
6652
|
+
provider: "openai-codex",
|
|
6653
|
+
baseUrl: "https://chatgpt.com/backend-api",
|
|
6654
|
+
reasoning: true,
|
|
6655
|
+
input: ["text", "image"],
|
|
6656
|
+
cost: {
|
|
6657
|
+
input: 5,
|
|
6658
|
+
output: 30,
|
|
6659
|
+
cacheRead: 0.5,
|
|
6660
|
+
cacheWrite: 0,
|
|
6661
|
+
},
|
|
6662
|
+
contextWindow: 272000,
|
|
6663
|
+
maxTokens: 128000,
|
|
6664
|
+
},
|
|
6586
6665
|
},
|
|
6587
6666
|
"opencode": {
|
|
6588
6667
|
"big-pickle": {
|
|
@@ -7061,6 +7140,40 @@ export const MODELS = {
|
|
|
7061
7140
|
contextWindow: 1050000,
|
|
7062
7141
|
maxTokens: 128000,
|
|
7063
7142
|
},
|
|
7143
|
+
"gpt-5.5": {
|
|
7144
|
+
id: "gpt-5.5",
|
|
7145
|
+
name: "GPT-5.5",
|
|
7146
|
+
api: "openai-responses",
|
|
7147
|
+
provider: "opencode",
|
|
7148
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7149
|
+
reasoning: true,
|
|
7150
|
+
input: ["text", "image"],
|
|
7151
|
+
cost: {
|
|
7152
|
+
input: 5,
|
|
7153
|
+
output: 30,
|
|
7154
|
+
cacheRead: 0.5,
|
|
7155
|
+
cacheWrite: 0,
|
|
7156
|
+
},
|
|
7157
|
+
contextWindow: 1050000,
|
|
7158
|
+
maxTokens: 130000,
|
|
7159
|
+
},
|
|
7160
|
+
"gpt-5.5-pro": {
|
|
7161
|
+
id: "gpt-5.5-pro",
|
|
7162
|
+
name: "GPT-5.5 Pro",
|
|
7163
|
+
api: "openai-responses",
|
|
7164
|
+
provider: "opencode",
|
|
7165
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7166
|
+
reasoning: true,
|
|
7167
|
+
input: ["text", "image"],
|
|
7168
|
+
cost: {
|
|
7169
|
+
input: 30,
|
|
7170
|
+
output: 180,
|
|
7171
|
+
cacheRead: 30,
|
|
7172
|
+
cacheWrite: 0,
|
|
7173
|
+
},
|
|
7174
|
+
contextWindow: 1050000,
|
|
7175
|
+
maxTokens: 128000,
|
|
7176
|
+
},
|
|
7064
7177
|
"hy3-preview-free": {
|
|
7065
7178
|
id: "hy3-preview-free",
|
|
7066
7179
|
name: "Hy3 preview Free",
|
|
@@ -7233,6 +7346,42 @@ export const MODELS = {
|
|
|
7233
7346
|
},
|
|
7234
7347
|
},
|
|
7235
7348
|
"opencode-go": {
|
|
7349
|
+
"deepseek-v4-flash": {
|
|
7350
|
+
id: "deepseek-v4-flash",
|
|
7351
|
+
name: "DeepSeek V4 Flash",
|
|
7352
|
+
api: "openai-completions",
|
|
7353
|
+
provider: "opencode-go",
|
|
7354
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7355
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true },
|
|
7356
|
+
reasoning: true,
|
|
7357
|
+
input: ["text"],
|
|
7358
|
+
cost: {
|
|
7359
|
+
input: 0.14,
|
|
7360
|
+
output: 0.28,
|
|
7361
|
+
cacheRead: 0.028,
|
|
7362
|
+
cacheWrite: 0,
|
|
7363
|
+
},
|
|
7364
|
+
contextWindow: 1000000,
|
|
7365
|
+
maxTokens: 384000,
|
|
7366
|
+
},
|
|
7367
|
+
"deepseek-v4-pro": {
|
|
7368
|
+
id: "deepseek-v4-pro",
|
|
7369
|
+
name: "DeepSeek V4 Pro",
|
|
7370
|
+
api: "openai-completions",
|
|
7371
|
+
provider: "opencode-go",
|
|
7372
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7373
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true },
|
|
7374
|
+
reasoning: true,
|
|
7375
|
+
input: ["text"],
|
|
7376
|
+
cost: {
|
|
7377
|
+
input: 1.74,
|
|
7378
|
+
output: 3.48,
|
|
7379
|
+
cacheRead: 0.145,
|
|
7380
|
+
cacheWrite: 0,
|
|
7381
|
+
},
|
|
7382
|
+
contextWindow: 1000000,
|
|
7383
|
+
maxTokens: 384000,
|
|
7384
|
+
},
|
|
7236
7385
|
"glm-5": {
|
|
7237
7386
|
id: "glm-5",
|
|
7238
7387
|
name: "GLM-5",
|
|
@@ -7488,7 +7637,7 @@ export const MODELS = {
|
|
|
7488
7637
|
cacheWrite: 0,
|
|
7489
7638
|
},
|
|
7490
7639
|
contextWindow: 65536,
|
|
7491
|
-
maxTokens:
|
|
7640
|
+
maxTokens: 16384,
|
|
7492
7641
|
},
|
|
7493
7642
|
"amazon/nova-2-lite-v1": {
|
|
7494
7643
|
id: "amazon/nova-2-lite-v1",
|
|
@@ -8049,7 +8198,7 @@ export const MODELS = {
|
|
|
8049
8198
|
cacheWrite: 0,
|
|
8050
8199
|
},
|
|
8051
8200
|
contextWindow: 163840,
|
|
8052
|
-
maxTokens:
|
|
8201
|
+
maxTokens: 16384,
|
|
8053
8202
|
},
|
|
8054
8203
|
"deepseek/deepseek-chat-v3-0324": {
|
|
8055
8204
|
id: "deepseek/deepseek-chat-v3-0324",
|
|
@@ -8066,7 +8215,7 @@ export const MODELS = {
|
|
|
8066
8215
|
cacheWrite: 0,
|
|
8067
8216
|
},
|
|
8068
8217
|
contextWindow: 163840,
|
|
8069
|
-
maxTokens:
|
|
8218
|
+
maxTokens: 16384,
|
|
8070
8219
|
},
|
|
8071
8220
|
"deepseek/deepseek-chat-v3.1": {
|
|
8072
8221
|
id: "deepseek/deepseek-chat-v3.1",
|
|
@@ -8117,7 +8266,7 @@ export const MODELS = {
|
|
|
8117
8266
|
cacheWrite: 0,
|
|
8118
8267
|
},
|
|
8119
8268
|
contextWindow: 163840,
|
|
8120
|
-
maxTokens:
|
|
8269
|
+
maxTokens: 32768,
|
|
8121
8270
|
},
|
|
8122
8271
|
"deepseek/deepseek-v3.1-terminus": {
|
|
8123
8272
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -8130,11 +8279,11 @@ export const MODELS = {
|
|
|
8130
8279
|
cost: {
|
|
8131
8280
|
input: 0.21,
|
|
8132
8281
|
output: 0.7899999999999999,
|
|
8133
|
-
cacheRead: 0.
|
|
8282
|
+
cacheRead: 0.13,
|
|
8134
8283
|
cacheWrite: 0,
|
|
8135
8284
|
},
|
|
8136
8285
|
contextWindow: 163840,
|
|
8137
|
-
maxTokens:
|
|
8286
|
+
maxTokens: 32768,
|
|
8138
8287
|
},
|
|
8139
8288
|
"deepseek/deepseek-v3.2": {
|
|
8140
8289
|
id: "deepseek/deepseek-v3.2",
|
|
@@ -8176,6 +8325,7 @@ export const MODELS = {
|
|
|
8176
8325
|
api: "openai-completions",
|
|
8177
8326
|
provider: "openrouter",
|
|
8178
8327
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
8328
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true },
|
|
8179
8329
|
reasoning: true,
|
|
8180
8330
|
input: ["text"],
|
|
8181
8331
|
cost: {
|
|
@@ -8193,12 +8343,13 @@ export const MODELS = {
|
|
|
8193
8343
|
api: "openai-completions",
|
|
8194
8344
|
provider: "openrouter",
|
|
8195
8345
|
baseUrl: "https://openrouter.ai/api/v1",
|
|
8346
|
+
compat: { "requiresReasoningContentOnAssistantMessages": true },
|
|
8196
8347
|
reasoning: true,
|
|
8197
8348
|
input: ["text"],
|
|
8198
8349
|
cost: {
|
|
8199
|
-
input:
|
|
8200
|
-
output:
|
|
8201
|
-
cacheRead: 0.
|
|
8350
|
+
input: 0.435,
|
|
8351
|
+
output: 0.87,
|
|
8352
|
+
cacheRead: 0.03625,
|
|
8202
8353
|
cacheWrite: 0,
|
|
8203
8354
|
},
|
|
8204
8355
|
contextWindow: 1048576,
|
|
@@ -8235,7 +8386,7 @@ export const MODELS = {
|
|
|
8235
8386
|
cacheRead: 0.024999999999999998,
|
|
8236
8387
|
cacheWrite: 0.08333333333333334,
|
|
8237
8388
|
},
|
|
8238
|
-
contextWindow:
|
|
8389
|
+
contextWindow: 1000000,
|
|
8239
8390
|
maxTokens: 8192,
|
|
8240
8391
|
},
|
|
8241
8392
|
"google/gemini-2.0-flash-lite-001": {
|
|
@@ -8425,6 +8576,40 @@ export const MODELS = {
|
|
|
8425
8576
|
contextWindow: 1048576,
|
|
8426
8577
|
maxTokens: 65536,
|
|
8427
8578
|
},
|
|
8579
|
+
"google/gemma-3-12b-it": {
|
|
8580
|
+
id: "google/gemma-3-12b-it",
|
|
8581
|
+
name: "Google: Gemma 3 12B",
|
|
8582
|
+
api: "openai-completions",
|
|
8583
|
+
provider: "openrouter",
|
|
8584
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8585
|
+
reasoning: false,
|
|
8586
|
+
input: ["text", "image"],
|
|
8587
|
+
cost: {
|
|
8588
|
+
input: 0.04,
|
|
8589
|
+
output: 0.13,
|
|
8590
|
+
cacheRead: 0,
|
|
8591
|
+
cacheWrite: 0,
|
|
8592
|
+
},
|
|
8593
|
+
contextWindow: 131072,
|
|
8594
|
+
maxTokens: 16384,
|
|
8595
|
+
},
|
|
8596
|
+
"google/gemma-3-27b-it": {
|
|
8597
|
+
id: "google/gemma-3-27b-it",
|
|
8598
|
+
name: "Google: Gemma 3 27B",
|
|
8599
|
+
api: "openai-completions",
|
|
8600
|
+
provider: "openrouter",
|
|
8601
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8602
|
+
reasoning: false,
|
|
8603
|
+
input: ["text", "image"],
|
|
8604
|
+
cost: {
|
|
8605
|
+
input: 0.08,
|
|
8606
|
+
output: 0.16,
|
|
8607
|
+
cacheRead: 0,
|
|
8608
|
+
cacheWrite: 0,
|
|
8609
|
+
},
|
|
8610
|
+
contextWindow: 131072,
|
|
8611
|
+
maxTokens: 16384,
|
|
8612
|
+
},
|
|
8428
8613
|
"google/gemma-4-26b-a4b-it": {
|
|
8429
8614
|
id: "google/gemma-4-26b-a4b-it",
|
|
8430
8615
|
name: "Google: Gemma 4 26B A4B ",
|
|
@@ -8474,7 +8659,7 @@ export const MODELS = {
|
|
|
8474
8659
|
cacheWrite: 0,
|
|
8475
8660
|
},
|
|
8476
8661
|
contextWindow: 262144,
|
|
8477
|
-
maxTokens:
|
|
8662
|
+
maxTokens: 16384,
|
|
8478
8663
|
},
|
|
8479
8664
|
"google/gemma-4-31b-it:free": {
|
|
8480
8665
|
id: "google/gemma-4-31b-it:free",
|
|
@@ -8593,7 +8778,7 @@ export const MODELS = {
|
|
|
8593
8778
|
cacheWrite: 0,
|
|
8594
8779
|
},
|
|
8595
8780
|
contextWindow: 131072,
|
|
8596
|
-
maxTokens:
|
|
8781
|
+
maxTokens: 16384,
|
|
8597
8782
|
},
|
|
8598
8783
|
"meta-llama/llama-3.1-8b-instruct": {
|
|
8599
8784
|
id: "meta-llama/llama-3.1-8b-instruct",
|
|
@@ -8729,7 +8914,7 @@ export const MODELS = {
|
|
|
8729
8914
|
cacheWrite: 0,
|
|
8730
8915
|
},
|
|
8731
8916
|
contextWindow: 196608,
|
|
8732
|
-
maxTokens:
|
|
8917
|
+
maxTokens: 131072,
|
|
8733
8918
|
},
|
|
8734
8919
|
"minimax/minimax-m2.5:free": {
|
|
8735
8920
|
id: "minimax/minimax-m2.5:free",
|
|
@@ -9052,7 +9237,7 @@ export const MODELS = {
|
|
|
9052
9237
|
cacheWrite: 0,
|
|
9053
9238
|
},
|
|
9054
9239
|
contextWindow: 128000,
|
|
9055
|
-
maxTokens:
|
|
9240
|
+
maxTokens: 16384,
|
|
9056
9241
|
},
|
|
9057
9242
|
"mistralai/mistral-small-creative": {
|
|
9058
9243
|
id: "mistralai/mistral-small-creative",
|
|
@@ -9273,7 +9458,7 @@ export const MODELS = {
|
|
|
9273
9458
|
cacheWrite: 0,
|
|
9274
9459
|
},
|
|
9275
9460
|
contextWindow: 131072,
|
|
9276
|
-
maxTokens:
|
|
9461
|
+
maxTokens: 16384,
|
|
9277
9462
|
},
|
|
9278
9463
|
"nvidia/nemotron-3-nano-30b-a3b": {
|
|
9279
9464
|
id: "nvidia/nemotron-3-nano-30b-a3b",
|
|
@@ -9290,7 +9475,7 @@ export const MODELS = {
|
|
|
9290
9475
|
cacheWrite: 0,
|
|
9291
9476
|
},
|
|
9292
9477
|
contextWindow: 262144,
|
|
9293
|
-
maxTokens:
|
|
9478
|
+
maxTokens: 228000,
|
|
9294
9479
|
},
|
|
9295
9480
|
"nvidia/nemotron-3-nano-30b-a3b:free": {
|
|
9296
9481
|
id: "nvidia/nemotron-3-nano-30b-a3b:free",
|
|
@@ -9375,7 +9560,7 @@ export const MODELS = {
|
|
|
9375
9560
|
cacheWrite: 0,
|
|
9376
9561
|
},
|
|
9377
9562
|
contextWindow: 131072,
|
|
9378
|
-
maxTokens:
|
|
9563
|
+
maxTokens: 16384,
|
|
9379
9564
|
},
|
|
9380
9565
|
"nvidia/nemotron-nano-9b-v2:free": {
|
|
9381
9566
|
id: "nvidia/nemotron-nano-9b-v2:free",
|
|
@@ -10040,6 +10225,40 @@ export const MODELS = {
|
|
|
10040
10225
|
contextWindow: 1050000,
|
|
10041
10226
|
maxTokens: 128000,
|
|
10042
10227
|
},
|
|
10228
|
+
"openai/gpt-5.5": {
|
|
10229
|
+
id: "openai/gpt-5.5",
|
|
10230
|
+
name: "OpenAI: GPT-5.5",
|
|
10231
|
+
api: "openai-completions",
|
|
10232
|
+
provider: "openrouter",
|
|
10233
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10234
|
+
reasoning: true,
|
|
10235
|
+
input: ["text", "image"],
|
|
10236
|
+
cost: {
|
|
10237
|
+
input: 5,
|
|
10238
|
+
output: 30,
|
|
10239
|
+
cacheRead: 0.5,
|
|
10240
|
+
cacheWrite: 0,
|
|
10241
|
+
},
|
|
10242
|
+
contextWindow: 1050000,
|
|
10243
|
+
maxTokens: 128000,
|
|
10244
|
+
},
|
|
10245
|
+
"openai/gpt-5.5-pro": {
|
|
10246
|
+
id: "openai/gpt-5.5-pro",
|
|
10247
|
+
name: "OpenAI: GPT-5.5 Pro",
|
|
10248
|
+
api: "openai-completions",
|
|
10249
|
+
provider: "openrouter",
|
|
10250
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10251
|
+
reasoning: true,
|
|
10252
|
+
input: ["text", "image"],
|
|
10253
|
+
cost: {
|
|
10254
|
+
input: 30,
|
|
10255
|
+
output: 180,
|
|
10256
|
+
cacheRead: 0,
|
|
10257
|
+
cacheWrite: 0,
|
|
10258
|
+
},
|
|
10259
|
+
contextWindow: 1050000,
|
|
10260
|
+
maxTokens: 128000,
|
|
10261
|
+
},
|
|
10043
10262
|
"openai/gpt-audio": {
|
|
10044
10263
|
id: "openai/gpt-audio",
|
|
10045
10264
|
name: "OpenAI: GPT Audio",
|
|
@@ -10089,7 +10308,7 @@ export const MODELS = {
|
|
|
10089
10308
|
cacheWrite: 0,
|
|
10090
10309
|
},
|
|
10091
10310
|
contextWindow: 131072,
|
|
10092
|
-
maxTokens:
|
|
10311
|
+
maxTokens: 131072,
|
|
10093
10312
|
},
|
|
10094
10313
|
"openai/gpt-oss-120b:free": {
|
|
10095
10314
|
id: "openai/gpt-oss-120b:free",
|
|
@@ -10123,7 +10342,7 @@ export const MODELS = {
|
|
|
10123
10342
|
cacheWrite: 0,
|
|
10124
10343
|
},
|
|
10125
10344
|
contextWindow: 131072,
|
|
10126
|
-
maxTokens:
|
|
10345
|
+
maxTokens: 131072,
|
|
10127
10346
|
},
|
|
10128
10347
|
"openai/gpt-oss-20b:free": {
|
|
10129
10348
|
id: "openai/gpt-oss-20b:free",
|
|
@@ -10548,7 +10767,7 @@ export const MODELS = {
|
|
|
10548
10767
|
cacheWrite: 0,
|
|
10549
10768
|
},
|
|
10550
10769
|
contextWindow: 262144,
|
|
10551
|
-
maxTokens:
|
|
10770
|
+
maxTokens: 16384,
|
|
10552
10771
|
},
|
|
10553
10772
|
"qwen/qwen3-235b-a22b-thinking-2507": {
|
|
10554
10773
|
id: "qwen/qwen3-235b-a22b-thinking-2507",
|
|
@@ -10582,7 +10801,7 @@ export const MODELS = {
|
|
|
10582
10801
|
cacheWrite: 0,
|
|
10583
10802
|
},
|
|
10584
10803
|
contextWindow: 40960,
|
|
10585
|
-
maxTokens:
|
|
10804
|
+
maxTokens: 16384,
|
|
10586
10805
|
},
|
|
10587
10806
|
"qwen/qwen3-30b-a3b-instruct-2507": {
|
|
10588
10807
|
id: "qwen/qwen3-30b-a3b-instruct-2507",
|
|
@@ -10662,12 +10881,12 @@ export const MODELS = {
|
|
|
10662
10881
|
input: ["text"],
|
|
10663
10882
|
cost: {
|
|
10664
10883
|
input: 0.22,
|
|
10665
|
-
output: 1,
|
|
10666
|
-
cacheRead: 0
|
|
10884
|
+
output: 1.7999999999999998,
|
|
10885
|
+
cacheRead: 0,
|
|
10667
10886
|
cacheWrite: 0,
|
|
10668
10887
|
},
|
|
10669
10888
|
contextWindow: 262144,
|
|
10670
|
-
maxTokens:
|
|
10889
|
+
maxTokens: 65536,
|
|
10671
10890
|
},
|
|
10672
10891
|
"qwen/qwen3-coder-30b-a3b-instruct": {
|
|
10673
10892
|
id: "qwen/qwen3-coder-30b-a3b-instruct",
|
|
@@ -10712,9 +10931,9 @@ export const MODELS = {
|
|
|
10712
10931
|
reasoning: false,
|
|
10713
10932
|
input: ["text"],
|
|
10714
10933
|
cost: {
|
|
10715
|
-
input: 0.
|
|
10934
|
+
input: 0.14,
|
|
10716
10935
|
output: 0.7999999999999999,
|
|
10717
|
-
cacheRead: 0.
|
|
10936
|
+
cacheRead: 0.09,
|
|
10718
10937
|
cacheWrite: 0,
|
|
10719
10938
|
},
|
|
10720
10939
|
contextWindow: 262144,
|
|
@@ -10803,7 +11022,7 @@ export const MODELS = {
|
|
|
10803
11022
|
cacheWrite: 0,
|
|
10804
11023
|
},
|
|
10805
11024
|
contextWindow: 262144,
|
|
10806
|
-
maxTokens:
|
|
11025
|
+
maxTokens: 16384,
|
|
10807
11026
|
},
|
|
10808
11027
|
"qwen/qwen3-next-80b-a3b-instruct:free": {
|
|
10809
11028
|
id: "qwen/qwen3-next-80b-a3b-instruct:free",
|
|
@@ -10854,7 +11073,7 @@ export const MODELS = {
|
|
|
10854
11073
|
cacheWrite: 0,
|
|
10855
11074
|
},
|
|
10856
11075
|
contextWindow: 262144,
|
|
10857
|
-
maxTokens:
|
|
11076
|
+
maxTokens: 16384,
|
|
10858
11077
|
},
|
|
10859
11078
|
"qwen/qwen3-vl-235b-a22b-thinking": {
|
|
10860
11079
|
id: "qwen/qwen3-vl-235b-a22b-thinking",
|
|
@@ -11632,11 +11851,11 @@ export const MODELS = {
|
|
|
11632
11851
|
cost: {
|
|
11633
11852
|
input: 0.3,
|
|
11634
11853
|
output: 0.8999999999999999,
|
|
11635
|
-
cacheRead: 0,
|
|
11854
|
+
cacheRead: 0.049999999999999996,
|
|
11636
11855
|
cacheWrite: 0,
|
|
11637
11856
|
},
|
|
11638
11857
|
contextWindow: 131072,
|
|
11639
|
-
maxTokens:
|
|
11858
|
+
maxTokens: 24000,
|
|
11640
11859
|
},
|
|
11641
11860
|
"z-ai/glm-4.7": {
|
|
11642
11861
|
id: "z-ai/glm-4.7",
|
|
@@ -11666,11 +11885,11 @@ export const MODELS = {
|
|
|
11666
11885
|
cost: {
|
|
11667
11886
|
input: 0.06,
|
|
11668
11887
|
output: 0.39999999999999997,
|
|
11669
|
-
cacheRead: 0.
|
|
11888
|
+
cacheRead: 0.01,
|
|
11670
11889
|
cacheWrite: 0,
|
|
11671
11890
|
},
|
|
11672
11891
|
contextWindow: 202752,
|
|
11673
|
-
maxTokens:
|
|
11892
|
+
maxTokens: 16384,
|
|
11674
11893
|
},
|
|
11675
11894
|
"z-ai/glm-5": {
|
|
11676
11895
|
id: "z-ai/glm-5",
|
|
@@ -11687,7 +11906,7 @@ export const MODELS = {
|
|
|
11687
11906
|
cacheWrite: 0,
|
|
11688
11907
|
},
|
|
11689
11908
|
contextWindow: 202752,
|
|
11690
|
-
maxTokens:
|
|
11909
|
+
maxTokens: 16384,
|
|
11691
11910
|
},
|
|
11692
11911
|
"z-ai/glm-5-turbo": {
|
|
11693
11912
|
id: "z-ai/glm-5-turbo",
|
|
@@ -12433,7 +12652,7 @@ export const MODELS = {
|
|
|
12433
12652
|
cost: {
|
|
12434
12653
|
input: 0.14,
|
|
12435
12654
|
output: 0.28,
|
|
12436
|
-
cacheRead: 0.
|
|
12655
|
+
cacheRead: 0.028,
|
|
12437
12656
|
cacheWrite: 0,
|
|
12438
12657
|
},
|
|
12439
12658
|
contextWindow: 1000000,
|
|
@@ -13731,6 +13950,40 @@ export const MODELS = {
|
|
|
13731
13950
|
contextWindow: 1050000,
|
|
13732
13951
|
maxTokens: 128000,
|
|
13733
13952
|
},
|
|
13953
|
+
"openai/gpt-5.5": {
|
|
13954
|
+
id: "openai/gpt-5.5",
|
|
13955
|
+
name: "GPT 5.5",
|
|
13956
|
+
api: "anthropic-messages",
|
|
13957
|
+
provider: "vercel-ai-gateway",
|
|
13958
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13959
|
+
reasoning: true,
|
|
13960
|
+
input: ["text", "image"],
|
|
13961
|
+
cost: {
|
|
13962
|
+
input: 5,
|
|
13963
|
+
output: 30,
|
|
13964
|
+
cacheRead: 0.5,
|
|
13965
|
+
cacheWrite: 0,
|
|
13966
|
+
},
|
|
13967
|
+
contextWindow: 1000000,
|
|
13968
|
+
maxTokens: 128000,
|
|
13969
|
+
},
|
|
13970
|
+
"openai/gpt-5.5-pro": {
|
|
13971
|
+
id: "openai/gpt-5.5-pro",
|
|
13972
|
+
name: "GPT 5.5 Pro",
|
|
13973
|
+
api: "anthropic-messages",
|
|
13974
|
+
provider: "vercel-ai-gateway",
|
|
13975
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13976
|
+
reasoning: true,
|
|
13977
|
+
input: ["text", "image"],
|
|
13978
|
+
cost: {
|
|
13979
|
+
input: 30,
|
|
13980
|
+
output: 180,
|
|
13981
|
+
cacheRead: 0,
|
|
13982
|
+
cacheWrite: 0,
|
|
13983
|
+
},
|
|
13984
|
+
contextWindow: 1000000,
|
|
13985
|
+
maxTokens: 128000,
|
|
13986
|
+
},
|
|
13734
13987
|
"openai/gpt-oss-20b": {
|
|
13735
13988
|
id: "openai/gpt-oss-20b",
|
|
13736
13989
|
name: "GPT OSS 120B",
|
|
@@ -14857,24 +15110,6 @@ export const MODELS = {
|
|
|
14857
15110
|
},
|
|
14858
15111
|
},
|
|
14859
15112
|
"zai": {
|
|
14860
|
-
"glm-4.5": {
|
|
14861
|
-
id: "glm-4.5",
|
|
14862
|
-
name: "GLM-4.5",
|
|
14863
|
-
api: "openai-completions",
|
|
14864
|
-
provider: "zai",
|
|
14865
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14866
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14867
|
-
reasoning: true,
|
|
14868
|
-
input: ["text"],
|
|
14869
|
-
cost: {
|
|
14870
|
-
input: 0,
|
|
14871
|
-
output: 0,
|
|
14872
|
-
cacheRead: 0,
|
|
14873
|
-
cacheWrite: 0,
|
|
14874
|
-
},
|
|
14875
|
-
contextWindow: 131072,
|
|
14876
|
-
maxTokens: 98304,
|
|
14877
|
-
},
|
|
14878
15113
|
"glm-4.5-air": {
|
|
14879
15114
|
id: "glm-4.5-air",
|
|
14880
15115
|
name: "GLM-4.5-Air",
|
|
@@ -14893,78 +15128,6 @@ export const MODELS = {
|
|
|
14893
15128
|
contextWindow: 131072,
|
|
14894
15129
|
maxTokens: 98304,
|
|
14895
15130
|
},
|
|
14896
|
-
"glm-4.5-flash": {
|
|
14897
|
-
id: "glm-4.5-flash",
|
|
14898
|
-
name: "GLM-4.5-Flash",
|
|
14899
|
-
api: "openai-completions",
|
|
14900
|
-
provider: "zai",
|
|
14901
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14902
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14903
|
-
reasoning: true,
|
|
14904
|
-
input: ["text"],
|
|
14905
|
-
cost: {
|
|
14906
|
-
input: 0,
|
|
14907
|
-
output: 0,
|
|
14908
|
-
cacheRead: 0,
|
|
14909
|
-
cacheWrite: 0,
|
|
14910
|
-
},
|
|
14911
|
-
contextWindow: 131072,
|
|
14912
|
-
maxTokens: 98304,
|
|
14913
|
-
},
|
|
14914
|
-
"glm-4.5v": {
|
|
14915
|
-
id: "glm-4.5v",
|
|
14916
|
-
name: "GLM-4.5V",
|
|
14917
|
-
api: "openai-completions",
|
|
14918
|
-
provider: "zai",
|
|
14919
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14920
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
|
|
14921
|
-
reasoning: true,
|
|
14922
|
-
input: ["text", "image"],
|
|
14923
|
-
cost: {
|
|
14924
|
-
input: 0,
|
|
14925
|
-
output: 0,
|
|
14926
|
-
cacheRead: 0,
|
|
14927
|
-
cacheWrite: 0,
|
|
14928
|
-
},
|
|
14929
|
-
contextWindow: 64000,
|
|
14930
|
-
maxTokens: 16384,
|
|
14931
|
-
},
|
|
14932
|
-
"glm-4.6": {
|
|
14933
|
-
id: "glm-4.6",
|
|
14934
|
-
name: "GLM-4.6",
|
|
14935
|
-
api: "openai-completions",
|
|
14936
|
-
provider: "zai",
|
|
14937
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14938
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
14939
|
-
reasoning: true,
|
|
14940
|
-
input: ["text"],
|
|
14941
|
-
cost: {
|
|
14942
|
-
input: 0,
|
|
14943
|
-
output: 0,
|
|
14944
|
-
cacheRead: 0,
|
|
14945
|
-
cacheWrite: 0,
|
|
14946
|
-
},
|
|
14947
|
-
contextWindow: 204800,
|
|
14948
|
-
maxTokens: 131072,
|
|
14949
|
-
},
|
|
14950
|
-
"glm-4.6v": {
|
|
14951
|
-
id: "glm-4.6v",
|
|
14952
|
-
name: "GLM-4.6V",
|
|
14953
|
-
api: "openai-completions",
|
|
14954
|
-
provider: "zai",
|
|
14955
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14956
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
14957
|
-
reasoning: true,
|
|
14958
|
-
input: ["text", "image"],
|
|
14959
|
-
cost: {
|
|
14960
|
-
input: 0,
|
|
14961
|
-
output: 0,
|
|
14962
|
-
cacheRead: 0,
|
|
14963
|
-
cacheWrite: 0,
|
|
14964
|
-
},
|
|
14965
|
-
contextWindow: 128000,
|
|
14966
|
-
maxTokens: 32768,
|
|
14967
|
-
},
|
|
14968
15131
|
"glm-4.7": {
|
|
14969
15132
|
id: "glm-4.7",
|
|
14970
15133
|
name: "GLM-4.7",
|
|
@@ -14983,60 +15146,6 @@ export const MODELS = {
|
|
|
14983
15146
|
contextWindow: 204800,
|
|
14984
15147
|
maxTokens: 131072,
|
|
14985
15148
|
},
|
|
14986
|
-
"glm-4.7-flash": {
|
|
14987
|
-
id: "glm-4.7-flash",
|
|
14988
|
-
name: "GLM-4.7-Flash",
|
|
14989
|
-
api: "openai-completions",
|
|
14990
|
-
provider: "zai",
|
|
14991
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
14992
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
14993
|
-
reasoning: true,
|
|
14994
|
-
input: ["text"],
|
|
14995
|
-
cost: {
|
|
14996
|
-
input: 0,
|
|
14997
|
-
output: 0,
|
|
14998
|
-
cacheRead: 0,
|
|
14999
|
-
cacheWrite: 0,
|
|
15000
|
-
},
|
|
15001
|
-
contextWindow: 200000,
|
|
15002
|
-
maxTokens: 131072,
|
|
15003
|
-
},
|
|
15004
|
-
"glm-4.7-flashx": {
|
|
15005
|
-
id: "glm-4.7-flashx",
|
|
15006
|
-
name: "GLM-4.7-FlashX",
|
|
15007
|
-
api: "openai-completions",
|
|
15008
|
-
provider: "zai",
|
|
15009
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15010
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
15011
|
-
reasoning: true,
|
|
15012
|
-
input: ["text"],
|
|
15013
|
-
cost: {
|
|
15014
|
-
input: 0.07,
|
|
15015
|
-
output: 0.4,
|
|
15016
|
-
cacheRead: 0.01,
|
|
15017
|
-
cacheWrite: 0,
|
|
15018
|
-
},
|
|
15019
|
-
contextWindow: 200000,
|
|
15020
|
-
maxTokens: 131072,
|
|
15021
|
-
},
|
|
15022
|
-
"glm-5": {
|
|
15023
|
-
id: "glm-5",
|
|
15024
|
-
name: "GLM-5",
|
|
15025
|
-
api: "openai-completions",
|
|
15026
|
-
provider: "zai",
|
|
15027
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15028
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
15029
|
-
reasoning: true,
|
|
15030
|
-
input: ["text"],
|
|
15031
|
-
cost: {
|
|
15032
|
-
input: 0,
|
|
15033
|
-
output: 0,
|
|
15034
|
-
cacheRead: 0,
|
|
15035
|
-
cacheWrite: 0,
|
|
15036
|
-
},
|
|
15037
|
-
contextWindow: 204800,
|
|
15038
|
-
maxTokens: 131072,
|
|
15039
|
-
},
|
|
15040
15149
|
"glm-5-turbo": {
|
|
15041
15150
|
id: "glm-5-turbo",
|
|
15042
15151
|
name: "GLM-5-Turbo",
|
|
@@ -15073,24 +15182,6 @@ export const MODELS = {
|
|
|
15073
15182
|
contextWindow: 200000,
|
|
15074
15183
|
maxTokens: 131072,
|
|
15075
15184
|
},
|
|
15076
|
-
"glm-5v-turbo": {
|
|
15077
|
-
id: "glm-5v-turbo",
|
|
15078
|
-
name: "glm-5v-turbo",
|
|
15079
|
-
api: "openai-completions",
|
|
15080
|
-
provider: "zai",
|
|
15081
|
-
baseUrl: "https://api.z.ai/api/coding/paas/v4",
|
|
15082
|
-
compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
|
|
15083
|
-
reasoning: true,
|
|
15084
|
-
input: ["text", "image"],
|
|
15085
|
-
cost: {
|
|
15086
|
-
input: 0,
|
|
15087
|
-
output: 0,
|
|
15088
|
-
cacheRead: 0,
|
|
15089
|
-
cacheWrite: 0,
|
|
15090
|
-
},
|
|
15091
|
-
contextWindow: 200000,
|
|
15092
|
-
maxTokens: 131072,
|
|
15093
|
-
},
|
|
15094
15185
|
},
|
|
15095
15186
|
};
|
|
15096
15187
|
//# sourceMappingURL=models.generated.js.map
|