@hyperspaceng/neural-ai 0.69.1 → 0.70.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/env-api-keys.d.ts +9 -0
- package/dist/env-api-keys.d.ts.map +1 -1
- package/dist/env-api-keys.js +41 -31
- package/dist/env-api-keys.js.map +1 -1
- package/dist/models.d.ts +1 -1
- package/dist/models.d.ts.map +1 -1
- package/dist/models.generated.d.ts +205 -0
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +222 -29
- package/dist/models.generated.js.map +1 -1
- package/dist/models.js +5 -2
- package/dist/models.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +28 -11
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-vertex.d.ts.map +1 -1
- package/dist/providers/google-vertex.js +34 -13
- package/dist/providers/google-vertex.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +8 -7
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +53 -31
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses.d.ts.map +1 -1
- package/dist/providers/openai-responses.js +19 -19
- package/dist/providers/openai-responses.js.map +1 -1
- package/dist/types.d.ts +20 -1
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -3062,6 +3062,7 @@ export const MODELS = {
|
|
|
3062
3062
|
provider: "github-copilot",
|
|
3063
3063
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3064
3064
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3065
|
+
compat: { "supportsEagerToolInputStreaming": false },
|
|
3065
3066
|
reasoning: true,
|
|
3066
3067
|
input: ["text", "image"],
|
|
3067
3068
|
cost: {
|
|
@@ -3134,6 +3135,7 @@ export const MODELS = {
|
|
|
3134
3135
|
provider: "github-copilot",
|
|
3135
3136
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3136
3137
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3138
|
+
compat: { "supportsEagerToolInputStreaming": false },
|
|
3137
3139
|
reasoning: true,
|
|
3138
3140
|
input: ["text", "image"],
|
|
3139
3141
|
cost: {
|
|
@@ -3152,6 +3154,7 @@ export const MODELS = {
|
|
|
3152
3154
|
provider: "github-copilot",
|
|
3153
3155
|
baseUrl: "https://api.individual.githubcopilot.com",
|
|
3154
3156
|
headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
|
|
3157
|
+
compat: { "supportsEagerToolInputStreaming": false },
|
|
3155
3158
|
reasoning: true,
|
|
3156
3159
|
input: ["text", "image"],
|
|
3157
3160
|
cost: {
|
|
@@ -5168,6 +5171,7 @@ export const MODELS = {
|
|
|
5168
5171
|
api: "anthropic-messages",
|
|
5169
5172
|
provider: "kimi-coding",
|
|
5170
5173
|
baseUrl: "https://api.kimi.com/coding",
|
|
5174
|
+
headers: { "User-Agent": "KimiCLI/1.5" },
|
|
5171
5175
|
reasoning: true,
|
|
5172
5176
|
input: ["text", "image"],
|
|
5173
5177
|
cost: {
|
|
@@ -5185,6 +5189,7 @@ export const MODELS = {
|
|
|
5185
5189
|
api: "anthropic-messages",
|
|
5186
5190
|
provider: "kimi-coding",
|
|
5187
5191
|
baseUrl: "https://api.kimi.com/coding",
|
|
5192
|
+
headers: { "User-Agent": "KimiCLI/1.5" },
|
|
5188
5193
|
reasoning: true,
|
|
5189
5194
|
input: ["text", "image"],
|
|
5190
5195
|
cost: {
|
|
@@ -5202,6 +5207,7 @@ export const MODELS = {
|
|
|
5202
5207
|
api: "anthropic-messages",
|
|
5203
5208
|
provider: "kimi-coding",
|
|
5204
5209
|
baseUrl: "https://api.kimi.com/coding",
|
|
5210
|
+
headers: { "User-Agent": "KimiCLI/1.5" },
|
|
5205
5211
|
reasoning: true,
|
|
5206
5212
|
input: ["text"],
|
|
5207
5213
|
cost: {
|
|
@@ -6583,6 +6589,23 @@ export const MODELS = {
|
|
|
6583
6589
|
contextWindow: 272000,
|
|
6584
6590
|
maxTokens: 128000,
|
|
6585
6591
|
},
|
|
6592
|
+
"gpt-5.5": {
|
|
6593
|
+
id: "gpt-5.5",
|
|
6594
|
+
name: "GPT-5.5",
|
|
6595
|
+
api: "openai-codex-responses",
|
|
6596
|
+
provider: "openai-codex",
|
|
6597
|
+
baseUrl: "https://chatgpt.com/backend-api",
|
|
6598
|
+
reasoning: true,
|
|
6599
|
+
input: ["text", "image"],
|
|
6600
|
+
cost: {
|
|
6601
|
+
input: 5,
|
|
6602
|
+
output: 30,
|
|
6603
|
+
cacheRead: 0.5,
|
|
6604
|
+
cacheWrite: 0,
|
|
6605
|
+
},
|
|
6606
|
+
contextWindow: 400000,
|
|
6607
|
+
maxTokens: 128000,
|
|
6608
|
+
},
|
|
6586
6609
|
},
|
|
6587
6610
|
"opencode": {
|
|
6588
6611
|
"big-pickle": {
|
|
@@ -7061,6 +7084,40 @@ export const MODELS = {
|
|
|
7061
7084
|
contextWindow: 1050000,
|
|
7062
7085
|
maxTokens: 128000,
|
|
7063
7086
|
},
|
|
7087
|
+
"gpt-5.5": {
|
|
7088
|
+
id: "gpt-5.5",
|
|
7089
|
+
name: "GPT-5.5",
|
|
7090
|
+
api: "openai-responses",
|
|
7091
|
+
provider: "opencode",
|
|
7092
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7093
|
+
reasoning: true,
|
|
7094
|
+
input: ["text", "image"],
|
|
7095
|
+
cost: {
|
|
7096
|
+
input: 5,
|
|
7097
|
+
output: 30,
|
|
7098
|
+
cacheRead: 0.5,
|
|
7099
|
+
cacheWrite: 0,
|
|
7100
|
+
},
|
|
7101
|
+
contextWindow: 1050000,
|
|
7102
|
+
maxTokens: 130000,
|
|
7103
|
+
},
|
|
7104
|
+
"gpt-5.5-pro": {
|
|
7105
|
+
id: "gpt-5.5-pro",
|
|
7106
|
+
name: "GPT-5.5 Pro",
|
|
7107
|
+
api: "openai-responses",
|
|
7108
|
+
provider: "opencode",
|
|
7109
|
+
baseUrl: "https://opencode.ai/zen/v1",
|
|
7110
|
+
reasoning: true,
|
|
7111
|
+
input: ["text", "image"],
|
|
7112
|
+
cost: {
|
|
7113
|
+
input: 30,
|
|
7114
|
+
output: 180,
|
|
7115
|
+
cacheRead: 30,
|
|
7116
|
+
cacheWrite: 0,
|
|
7117
|
+
},
|
|
7118
|
+
contextWindow: 1050000,
|
|
7119
|
+
maxTokens: 128000,
|
|
7120
|
+
},
|
|
7064
7121
|
"hy3-preview-free": {
|
|
7065
7122
|
id: "hy3-preview-free",
|
|
7066
7123
|
name: "Hy3 preview Free",
|
|
@@ -7233,6 +7290,40 @@ export const MODELS = {
|
|
|
7233
7290
|
},
|
|
7234
7291
|
},
|
|
7235
7292
|
"opencode-go": {
|
|
7293
|
+
"deepseek-v4-flash": {
|
|
7294
|
+
id: "deepseek-v4-flash",
|
|
7295
|
+
name: "DeepSeek V4 Flash",
|
|
7296
|
+
api: "openai-completions",
|
|
7297
|
+
provider: "opencode-go",
|
|
7298
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7299
|
+
reasoning: true,
|
|
7300
|
+
input: ["text"],
|
|
7301
|
+
cost: {
|
|
7302
|
+
input: 0.14,
|
|
7303
|
+
output: 0.28,
|
|
7304
|
+
cacheRead: 0.028,
|
|
7305
|
+
cacheWrite: 0,
|
|
7306
|
+
},
|
|
7307
|
+
contextWindow: 1000000,
|
|
7308
|
+
maxTokens: 384000,
|
|
7309
|
+
},
|
|
7310
|
+
"deepseek-v4-pro": {
|
|
7311
|
+
id: "deepseek-v4-pro",
|
|
7312
|
+
name: "DeepSeek V4 Pro",
|
|
7313
|
+
api: "openai-completions",
|
|
7314
|
+
provider: "opencode-go",
|
|
7315
|
+
baseUrl: "https://opencode.ai/zen/go/v1",
|
|
7316
|
+
reasoning: true,
|
|
7317
|
+
input: ["text"],
|
|
7318
|
+
cost: {
|
|
7319
|
+
input: 1.74,
|
|
7320
|
+
output: 3.48,
|
|
7321
|
+
cacheRead: 0.145,
|
|
7322
|
+
cacheWrite: 0,
|
|
7323
|
+
},
|
|
7324
|
+
contextWindow: 1000000,
|
|
7325
|
+
maxTokens: 384000,
|
|
7326
|
+
},
|
|
7236
7327
|
"glm-5": {
|
|
7237
7328
|
id: "glm-5",
|
|
7238
7329
|
name: "GLM-5",
|
|
@@ -7488,7 +7579,7 @@ export const MODELS = {
|
|
|
7488
7579
|
cacheWrite: 0,
|
|
7489
7580
|
},
|
|
7490
7581
|
contextWindow: 65536,
|
|
7491
|
-
maxTokens:
|
|
7582
|
+
maxTokens: 16384,
|
|
7492
7583
|
},
|
|
7493
7584
|
"amazon/nova-2-lite-v1": {
|
|
7494
7585
|
id: "amazon/nova-2-lite-v1",
|
|
@@ -8049,7 +8140,7 @@ export const MODELS = {
|
|
|
8049
8140
|
cacheWrite: 0,
|
|
8050
8141
|
},
|
|
8051
8142
|
contextWindow: 163840,
|
|
8052
|
-
maxTokens:
|
|
8143
|
+
maxTokens: 16384,
|
|
8053
8144
|
},
|
|
8054
8145
|
"deepseek/deepseek-chat-v3-0324": {
|
|
8055
8146
|
id: "deepseek/deepseek-chat-v3-0324",
|
|
@@ -8066,7 +8157,7 @@ export const MODELS = {
|
|
|
8066
8157
|
cacheWrite: 0,
|
|
8067
8158
|
},
|
|
8068
8159
|
contextWindow: 163840,
|
|
8069
|
-
maxTokens:
|
|
8160
|
+
maxTokens: 16384,
|
|
8070
8161
|
},
|
|
8071
8162
|
"deepseek/deepseek-chat-v3.1": {
|
|
8072
8163
|
id: "deepseek/deepseek-chat-v3.1",
|
|
@@ -8117,7 +8208,7 @@ export const MODELS = {
|
|
|
8117
8208
|
cacheWrite: 0,
|
|
8118
8209
|
},
|
|
8119
8210
|
contextWindow: 163840,
|
|
8120
|
-
maxTokens:
|
|
8211
|
+
maxTokens: 32768,
|
|
8121
8212
|
},
|
|
8122
8213
|
"deepseek/deepseek-v3.1-terminus": {
|
|
8123
8214
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -8130,11 +8221,11 @@ export const MODELS = {
|
|
|
8130
8221
|
cost: {
|
|
8131
8222
|
input: 0.21,
|
|
8132
8223
|
output: 0.7899999999999999,
|
|
8133
|
-
cacheRead: 0.
|
|
8224
|
+
cacheRead: 0.13,
|
|
8134
8225
|
cacheWrite: 0,
|
|
8135
8226
|
},
|
|
8136
8227
|
contextWindow: 163840,
|
|
8137
|
-
maxTokens:
|
|
8228
|
+
maxTokens: 32768,
|
|
8138
8229
|
},
|
|
8139
8230
|
"deepseek/deepseek-v3.2": {
|
|
8140
8231
|
id: "deepseek/deepseek-v3.2",
|
|
@@ -8235,7 +8326,7 @@ export const MODELS = {
|
|
|
8235
8326
|
cacheRead: 0.024999999999999998,
|
|
8236
8327
|
cacheWrite: 0.08333333333333334,
|
|
8237
8328
|
},
|
|
8238
|
-
contextWindow:
|
|
8329
|
+
contextWindow: 1000000,
|
|
8239
8330
|
maxTokens: 8192,
|
|
8240
8331
|
},
|
|
8241
8332
|
"google/gemini-2.0-flash-lite-001": {
|
|
@@ -8425,6 +8516,40 @@ export const MODELS = {
|
|
|
8425
8516
|
contextWindow: 1048576,
|
|
8426
8517
|
maxTokens: 65536,
|
|
8427
8518
|
},
|
|
8519
|
+
"google/gemma-3-12b-it": {
|
|
8520
|
+
id: "google/gemma-3-12b-it",
|
|
8521
|
+
name: "Google: Gemma 3 12B",
|
|
8522
|
+
api: "openai-completions",
|
|
8523
|
+
provider: "openrouter",
|
|
8524
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8525
|
+
reasoning: false,
|
|
8526
|
+
input: ["text", "image"],
|
|
8527
|
+
cost: {
|
|
8528
|
+
input: 0.04,
|
|
8529
|
+
output: 0.13,
|
|
8530
|
+
cacheRead: 0,
|
|
8531
|
+
cacheWrite: 0,
|
|
8532
|
+
},
|
|
8533
|
+
contextWindow: 131072,
|
|
8534
|
+
maxTokens: 16384,
|
|
8535
|
+
},
|
|
8536
|
+
"google/gemma-3-27b-it": {
|
|
8537
|
+
id: "google/gemma-3-27b-it",
|
|
8538
|
+
name: "Google: Gemma 3 27B",
|
|
8539
|
+
api: "openai-completions",
|
|
8540
|
+
provider: "openrouter",
|
|
8541
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
8542
|
+
reasoning: false,
|
|
8543
|
+
input: ["text", "image"],
|
|
8544
|
+
cost: {
|
|
8545
|
+
input: 0.08,
|
|
8546
|
+
output: 0.16,
|
|
8547
|
+
cacheRead: 0,
|
|
8548
|
+
cacheWrite: 0,
|
|
8549
|
+
},
|
|
8550
|
+
contextWindow: 131072,
|
|
8551
|
+
maxTokens: 16384,
|
|
8552
|
+
},
|
|
8428
8553
|
"google/gemma-4-26b-a4b-it": {
|
|
8429
8554
|
id: "google/gemma-4-26b-a4b-it",
|
|
8430
8555
|
name: "Google: Gemma 4 26B A4B ",
|
|
@@ -8474,7 +8599,7 @@ export const MODELS = {
|
|
|
8474
8599
|
cacheWrite: 0,
|
|
8475
8600
|
},
|
|
8476
8601
|
contextWindow: 262144,
|
|
8477
|
-
maxTokens:
|
|
8602
|
+
maxTokens: 16384,
|
|
8478
8603
|
},
|
|
8479
8604
|
"google/gemma-4-31b-it:free": {
|
|
8480
8605
|
id: "google/gemma-4-31b-it:free",
|
|
@@ -8593,7 +8718,7 @@ export const MODELS = {
|
|
|
8593
8718
|
cacheWrite: 0,
|
|
8594
8719
|
},
|
|
8595
8720
|
contextWindow: 131072,
|
|
8596
|
-
maxTokens:
|
|
8721
|
+
maxTokens: 16384,
|
|
8597
8722
|
},
|
|
8598
8723
|
"meta-llama/llama-3.1-8b-instruct": {
|
|
8599
8724
|
id: "meta-llama/llama-3.1-8b-instruct",
|
|
@@ -8729,7 +8854,7 @@ export const MODELS = {
|
|
|
8729
8854
|
cacheWrite: 0,
|
|
8730
8855
|
},
|
|
8731
8856
|
contextWindow: 196608,
|
|
8732
|
-
maxTokens:
|
|
8857
|
+
maxTokens: 131072,
|
|
8733
8858
|
},
|
|
8734
8859
|
"minimax/minimax-m2.5:free": {
|
|
8735
8860
|
id: "minimax/minimax-m2.5:free",
|
|
@@ -9052,7 +9177,7 @@ export const MODELS = {
|
|
|
9052
9177
|
cacheWrite: 0,
|
|
9053
9178
|
},
|
|
9054
9179
|
contextWindow: 128000,
|
|
9055
|
-
maxTokens:
|
|
9180
|
+
maxTokens: 16384,
|
|
9056
9181
|
},
|
|
9057
9182
|
"mistralai/mistral-small-creative": {
|
|
9058
9183
|
id: "mistralai/mistral-small-creative",
|
|
@@ -9273,7 +9398,7 @@ export const MODELS = {
|
|
|
9273
9398
|
cacheWrite: 0,
|
|
9274
9399
|
},
|
|
9275
9400
|
contextWindow: 131072,
|
|
9276
|
-
maxTokens:
|
|
9401
|
+
maxTokens: 16384,
|
|
9277
9402
|
},
|
|
9278
9403
|
"nvidia/nemotron-3-nano-30b-a3b": {
|
|
9279
9404
|
id: "nvidia/nemotron-3-nano-30b-a3b",
|
|
@@ -9290,7 +9415,7 @@ export const MODELS = {
|
|
|
9290
9415
|
cacheWrite: 0,
|
|
9291
9416
|
},
|
|
9292
9417
|
contextWindow: 262144,
|
|
9293
|
-
maxTokens:
|
|
9418
|
+
maxTokens: 228000,
|
|
9294
9419
|
},
|
|
9295
9420
|
"nvidia/nemotron-3-nano-30b-a3b:free": {
|
|
9296
9421
|
id: "nvidia/nemotron-3-nano-30b-a3b:free",
|
|
@@ -9375,7 +9500,7 @@ export const MODELS = {
|
|
|
9375
9500
|
cacheWrite: 0,
|
|
9376
9501
|
},
|
|
9377
9502
|
contextWindow: 131072,
|
|
9378
|
-
maxTokens:
|
|
9503
|
+
maxTokens: 16384,
|
|
9379
9504
|
},
|
|
9380
9505
|
"nvidia/nemotron-nano-9b-v2:free": {
|
|
9381
9506
|
id: "nvidia/nemotron-nano-9b-v2:free",
|
|
@@ -10040,6 +10165,40 @@ export const MODELS = {
|
|
|
10040
10165
|
contextWindow: 1050000,
|
|
10041
10166
|
maxTokens: 128000,
|
|
10042
10167
|
},
|
|
10168
|
+
"openai/gpt-5.5": {
|
|
10169
|
+
id: "openai/gpt-5.5",
|
|
10170
|
+
name: "OpenAI: GPT-5.5",
|
|
10171
|
+
api: "openai-completions",
|
|
10172
|
+
provider: "openrouter",
|
|
10173
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10174
|
+
reasoning: true,
|
|
10175
|
+
input: ["text", "image"],
|
|
10176
|
+
cost: {
|
|
10177
|
+
input: 5,
|
|
10178
|
+
output: 30,
|
|
10179
|
+
cacheRead: 0.5,
|
|
10180
|
+
cacheWrite: 0,
|
|
10181
|
+
},
|
|
10182
|
+
contextWindow: 1050000,
|
|
10183
|
+
maxTokens: 128000,
|
|
10184
|
+
},
|
|
10185
|
+
"openai/gpt-5.5-pro": {
|
|
10186
|
+
id: "openai/gpt-5.5-pro",
|
|
10187
|
+
name: "OpenAI: GPT-5.5 Pro",
|
|
10188
|
+
api: "openai-completions",
|
|
10189
|
+
provider: "openrouter",
|
|
10190
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
10191
|
+
reasoning: true,
|
|
10192
|
+
input: ["text", "image"],
|
|
10193
|
+
cost: {
|
|
10194
|
+
input: 30,
|
|
10195
|
+
output: 180,
|
|
10196
|
+
cacheRead: 0,
|
|
10197
|
+
cacheWrite: 0,
|
|
10198
|
+
},
|
|
10199
|
+
contextWindow: 1050000,
|
|
10200
|
+
maxTokens: 128000,
|
|
10201
|
+
},
|
|
10043
10202
|
"openai/gpt-audio": {
|
|
10044
10203
|
id: "openai/gpt-audio",
|
|
10045
10204
|
name: "OpenAI: GPT Audio",
|
|
@@ -10089,7 +10248,7 @@ export const MODELS = {
|
|
|
10089
10248
|
cacheWrite: 0,
|
|
10090
10249
|
},
|
|
10091
10250
|
contextWindow: 131072,
|
|
10092
|
-
maxTokens:
|
|
10251
|
+
maxTokens: 131072,
|
|
10093
10252
|
},
|
|
10094
10253
|
"openai/gpt-oss-120b:free": {
|
|
10095
10254
|
id: "openai/gpt-oss-120b:free",
|
|
@@ -10123,7 +10282,7 @@ export const MODELS = {
|
|
|
10123
10282
|
cacheWrite: 0,
|
|
10124
10283
|
},
|
|
10125
10284
|
contextWindow: 131072,
|
|
10126
|
-
maxTokens:
|
|
10285
|
+
maxTokens: 131072,
|
|
10127
10286
|
},
|
|
10128
10287
|
"openai/gpt-oss-20b:free": {
|
|
10129
10288
|
id: "openai/gpt-oss-20b:free",
|
|
@@ -10548,7 +10707,7 @@ export const MODELS = {
|
|
|
10548
10707
|
cacheWrite: 0,
|
|
10549
10708
|
},
|
|
10550
10709
|
contextWindow: 262144,
|
|
10551
|
-
maxTokens:
|
|
10710
|
+
maxTokens: 16384,
|
|
10552
10711
|
},
|
|
10553
10712
|
"qwen/qwen3-235b-a22b-thinking-2507": {
|
|
10554
10713
|
id: "qwen/qwen3-235b-a22b-thinking-2507",
|
|
@@ -10582,7 +10741,7 @@ export const MODELS = {
|
|
|
10582
10741
|
cacheWrite: 0,
|
|
10583
10742
|
},
|
|
10584
10743
|
contextWindow: 40960,
|
|
10585
|
-
maxTokens:
|
|
10744
|
+
maxTokens: 16384,
|
|
10586
10745
|
},
|
|
10587
10746
|
"qwen/qwen3-30b-a3b-instruct-2507": {
|
|
10588
10747
|
id: "qwen/qwen3-30b-a3b-instruct-2507",
|
|
@@ -10667,7 +10826,7 @@ export const MODELS = {
|
|
|
10667
10826
|
cacheWrite: 0,
|
|
10668
10827
|
},
|
|
10669
10828
|
contextWindow: 262144,
|
|
10670
|
-
maxTokens:
|
|
10829
|
+
maxTokens: 65536,
|
|
10671
10830
|
},
|
|
10672
10831
|
"qwen/qwen3-coder-30b-a3b-instruct": {
|
|
10673
10832
|
id: "qwen/qwen3-coder-30b-a3b-instruct",
|
|
@@ -10712,9 +10871,9 @@ export const MODELS = {
|
|
|
10712
10871
|
reasoning: false,
|
|
10713
10872
|
input: ["text"],
|
|
10714
10873
|
cost: {
|
|
10715
|
-
input: 0.
|
|
10874
|
+
input: 0.14,
|
|
10716
10875
|
output: 0.7999999999999999,
|
|
10717
|
-
cacheRead: 0.
|
|
10876
|
+
cacheRead: 0.09,
|
|
10718
10877
|
cacheWrite: 0,
|
|
10719
10878
|
},
|
|
10720
10879
|
contextWindow: 262144,
|
|
@@ -10803,7 +10962,7 @@ export const MODELS = {
|
|
|
10803
10962
|
cacheWrite: 0,
|
|
10804
10963
|
},
|
|
10805
10964
|
contextWindow: 262144,
|
|
10806
|
-
maxTokens:
|
|
10965
|
+
maxTokens: 16384,
|
|
10807
10966
|
},
|
|
10808
10967
|
"qwen/qwen3-next-80b-a3b-instruct:free": {
|
|
10809
10968
|
id: "qwen/qwen3-next-80b-a3b-instruct:free",
|
|
@@ -10854,7 +11013,7 @@ export const MODELS = {
|
|
|
10854
11013
|
cacheWrite: 0,
|
|
10855
11014
|
},
|
|
10856
11015
|
contextWindow: 262144,
|
|
10857
|
-
maxTokens:
|
|
11016
|
+
maxTokens: 16384,
|
|
10858
11017
|
},
|
|
10859
11018
|
"qwen/qwen3-vl-235b-a22b-thinking": {
|
|
10860
11019
|
id: "qwen/qwen3-vl-235b-a22b-thinking",
|
|
@@ -11632,11 +11791,11 @@ export const MODELS = {
|
|
|
11632
11791
|
cost: {
|
|
11633
11792
|
input: 0.3,
|
|
11634
11793
|
output: 0.8999999999999999,
|
|
11635
|
-
cacheRead: 0,
|
|
11794
|
+
cacheRead: 0.049999999999999996,
|
|
11636
11795
|
cacheWrite: 0,
|
|
11637
11796
|
},
|
|
11638
11797
|
contextWindow: 131072,
|
|
11639
|
-
maxTokens:
|
|
11798
|
+
maxTokens: 24000,
|
|
11640
11799
|
},
|
|
11641
11800
|
"z-ai/glm-4.7": {
|
|
11642
11801
|
id: "z-ai/glm-4.7",
|
|
@@ -11666,11 +11825,11 @@ export const MODELS = {
|
|
|
11666
11825
|
cost: {
|
|
11667
11826
|
input: 0.06,
|
|
11668
11827
|
output: 0.39999999999999997,
|
|
11669
|
-
cacheRead: 0.
|
|
11828
|
+
cacheRead: 0.01,
|
|
11670
11829
|
cacheWrite: 0,
|
|
11671
11830
|
},
|
|
11672
11831
|
contextWindow: 202752,
|
|
11673
|
-
maxTokens:
|
|
11832
|
+
maxTokens: 16384,
|
|
11674
11833
|
},
|
|
11675
11834
|
"z-ai/glm-5": {
|
|
11676
11835
|
id: "z-ai/glm-5",
|
|
@@ -11687,7 +11846,7 @@ export const MODELS = {
|
|
|
11687
11846
|
cacheWrite: 0,
|
|
11688
11847
|
},
|
|
11689
11848
|
contextWindow: 202752,
|
|
11690
|
-
maxTokens:
|
|
11849
|
+
maxTokens: 16384,
|
|
11691
11850
|
},
|
|
11692
11851
|
"z-ai/glm-5-turbo": {
|
|
11693
11852
|
id: "z-ai/glm-5-turbo",
|
|
@@ -12433,7 +12592,7 @@ export const MODELS = {
|
|
|
12433
12592
|
cost: {
|
|
12434
12593
|
input: 0.14,
|
|
12435
12594
|
output: 0.28,
|
|
12436
|
-
cacheRead: 0.
|
|
12595
|
+
cacheRead: 0.028,
|
|
12437
12596
|
cacheWrite: 0,
|
|
12438
12597
|
},
|
|
12439
12598
|
contextWindow: 1000000,
|
|
@@ -13731,6 +13890,40 @@ export const MODELS = {
|
|
|
13731
13890
|
contextWindow: 1050000,
|
|
13732
13891
|
maxTokens: 128000,
|
|
13733
13892
|
},
|
|
13893
|
+
"openai/gpt-5.5": {
|
|
13894
|
+
id: "openai/gpt-5.5",
|
|
13895
|
+
name: "GPT 5.5",
|
|
13896
|
+
api: "anthropic-messages",
|
|
13897
|
+
provider: "vercel-ai-gateway",
|
|
13898
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13899
|
+
reasoning: true,
|
|
13900
|
+
input: ["text", "image"],
|
|
13901
|
+
cost: {
|
|
13902
|
+
input: 5,
|
|
13903
|
+
output: 30,
|
|
13904
|
+
cacheRead: 0.5,
|
|
13905
|
+
cacheWrite: 0,
|
|
13906
|
+
},
|
|
13907
|
+
contextWindow: 1000000,
|
|
13908
|
+
maxTokens: 128000,
|
|
13909
|
+
},
|
|
13910
|
+
"openai/gpt-5.5-pro": {
|
|
13911
|
+
id: "openai/gpt-5.5-pro",
|
|
13912
|
+
name: "GPT 5.5 Pro",
|
|
13913
|
+
api: "anthropic-messages",
|
|
13914
|
+
provider: "vercel-ai-gateway",
|
|
13915
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13916
|
+
reasoning: true,
|
|
13917
|
+
input: ["text", "image"],
|
|
13918
|
+
cost: {
|
|
13919
|
+
input: 30,
|
|
13920
|
+
output: 180,
|
|
13921
|
+
cacheRead: 0,
|
|
13922
|
+
cacheWrite: 0,
|
|
13923
|
+
},
|
|
13924
|
+
contextWindow: 1000000,
|
|
13925
|
+
maxTokens: 128000,
|
|
13926
|
+
},
|
|
13734
13927
|
"openai/gpt-oss-20b": {
|
|
13735
13928
|
id: "openai/gpt-oss-20b",
|
|
13736
13929
|
name: "GPT OSS 120B",
|