@hyperspaceng/neural-ai 0.69.1 → 0.70.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (38) hide show
  1. package/README.md +4 -1
  2. package/dist/env-api-keys.d.ts +9 -0
  3. package/dist/env-api-keys.d.ts.map +1 -1
  4. package/dist/env-api-keys.js +42 -31
  5. package/dist/env-api-keys.js.map +1 -1
  6. package/dist/models.d.ts +2 -1
  7. package/dist/models.d.ts.map +1 -1
  8. package/dist/models.generated.d.ts +298 -195
  9. package/dist/models.generated.d.ts.map +1 -1
  10. package/dist/models.generated.js +291 -200
  11. package/dist/models.generated.js.map +1 -1
  12. package/dist/models.js +7 -2
  13. package/dist/models.js.map +1 -1
  14. package/dist/providers/anthropic.d.ts.map +1 -1
  15. package/dist/providers/anthropic.js +35 -15
  16. package/dist/providers/anthropic.js.map +1 -1
  17. package/dist/providers/azure-openai-responses.d.ts.map +1 -1
  18. package/dist/providers/azure-openai-responses.js +7 -4
  19. package/dist/providers/azure-openai-responses.js.map +1 -1
  20. package/dist/providers/google-vertex.d.ts.map +1 -1
  21. package/dist/providers/google-vertex.js +34 -13
  22. package/dist/providers/google-vertex.js.map +1 -1
  23. package/dist/providers/openai-codex-responses.d.ts.map +1 -1
  24. package/dist/providers/openai-codex-responses.js +9 -8
  25. package/dist/providers/openai-codex-responses.js.map +1 -1
  26. package/dist/providers/openai-completions.d.ts.map +1 -1
  27. package/dist/providers/openai-completions.js +97 -45
  28. package/dist/providers/openai-completions.js.map +1 -1
  29. package/dist/providers/openai-responses.d.ts.map +1 -1
  30. package/dist/providers/openai-responses.js +26 -23
  31. package/dist/providers/openai-responses.js.map +1 -1
  32. package/dist/providers/simple-options.d.ts.map +1 -1
  33. package/dist/providers/simple-options.js +2 -0
  34. package/dist/providers/simple-options.js.map +1 -1
  35. package/dist/types.d.ts +35 -4
  36. package/dist/types.d.ts.map +1 -1
  37. package/dist/types.js.map +1 -1
  38. package/package.json +1 -1
@@ -2536,8 +2536,8 @@ export const MODELS = {
2536
2536
  cacheRead: 0.5,
2537
2537
  cacheWrite: 0,
2538
2538
  },
2539
- contextWindow: 1050000,
2540
- maxTokens: 130000,
2539
+ contextWindow: 272000,
2540
+ maxTokens: 128000,
2541
2541
  },
2542
2542
  "o1": {
2543
2543
  id: "o1",
@@ -2746,6 +2746,44 @@ export const MODELS = {
2746
2746
  maxTokens: 40000,
2747
2747
  },
2748
2748
  },
2749
+ "deepseek": {
2750
+ "deepseek-v4-flash": {
2751
+ id: "deepseek-v4-flash",
2752
+ name: "DeepSeek V4 Flash",
2753
+ api: "openai-completions",
2754
+ provider: "deepseek",
2755
+ baseUrl: "https://api.deepseek.com",
2756
+ compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek", "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
2757
+ reasoning: true,
2758
+ input: ["text"],
2759
+ cost: {
2760
+ input: 0.14,
2761
+ output: 0.28,
2762
+ cacheRead: 0.028,
2763
+ cacheWrite: 0,
2764
+ },
2765
+ contextWindow: 1000000,
2766
+ maxTokens: 384000,
2767
+ },
2768
+ "deepseek-v4-pro": {
2769
+ id: "deepseek-v4-pro",
2770
+ name: "DeepSeek V4 Pro",
2771
+ api: "openai-completions",
2772
+ provider: "deepseek",
2773
+ baseUrl: "https://api.deepseek.com",
2774
+ compat: { "requiresReasoningContentOnAssistantMessages": true, "thinkingFormat": "deepseek", "reasoningEffortMap": { "minimal": "high", "low": "high", "medium": "high", "high": "high", "xhigh": "max" } },
2775
+ reasoning: true,
2776
+ input: ["text"],
2777
+ cost: {
2778
+ input: 1.74,
2779
+ output: 3.48,
2780
+ cacheRead: 0.145,
2781
+ cacheWrite: 0,
2782
+ },
2783
+ contextWindow: 1000000,
2784
+ maxTokens: 384000,
2785
+ },
2786
+ },
2749
2787
  "fireworks": {
2750
2788
  "accounts/fireworks/models/deepseek-v3p1": {
2751
2789
  id: "accounts/fireworks/models/deepseek-v3p1",
@@ -3062,6 +3100,7 @@ export const MODELS = {
3062
3100
  provider: "github-copilot",
3063
3101
  baseUrl: "https://api.individual.githubcopilot.com",
3064
3102
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3103
+ compat: { "supportsEagerToolInputStreaming": false },
3065
3104
  reasoning: true,
3066
3105
  input: ["text", "image"],
3067
3106
  cost: {
@@ -3134,6 +3173,7 @@ export const MODELS = {
3134
3173
  provider: "github-copilot",
3135
3174
  baseUrl: "https://api.individual.githubcopilot.com",
3136
3175
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3176
+ compat: { "supportsEagerToolInputStreaming": false },
3137
3177
  reasoning: true,
3138
3178
  input: ["text", "image"],
3139
3179
  cost: {
@@ -3152,6 +3192,7 @@ export const MODELS = {
3152
3192
  provider: "github-copilot",
3153
3193
  baseUrl: "https://api.individual.githubcopilot.com",
3154
3194
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3195
+ compat: { "supportsEagerToolInputStreaming": false },
3155
3196
  reasoning: true,
3156
3197
  input: ["text", "image"],
3157
3198
  cost: {
@@ -3493,6 +3534,24 @@ export const MODELS = {
3493
3534
  contextWindow: 400000,
3494
3535
  maxTokens: 128000,
3495
3536
  },
3537
+ "gpt-5.5": {
3538
+ id: "gpt-5.5",
3539
+ name: "GPT-5.5",
3540
+ api: "openai-responses",
3541
+ provider: "github-copilot",
3542
+ baseUrl: "https://api.individual.githubcopilot.com",
3543
+ headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3544
+ reasoning: true,
3545
+ input: ["text", "image"],
3546
+ cost: {
3547
+ input: 0,
3548
+ output: 0,
3549
+ cacheRead: 0,
3550
+ cacheWrite: 0,
3551
+ },
3552
+ contextWindow: 400000,
3553
+ maxTokens: 128000,
3554
+ },
3496
3555
  "grok-code-fast-1": {
3497
3556
  id: "grok-code-fast-1",
3498
3557
  name: "Grok Code Fast 1",
@@ -5168,6 +5227,7 @@ export const MODELS = {
5168
5227
  api: "anthropic-messages",
5169
5228
  provider: "kimi-coding",
5170
5229
  baseUrl: "https://api.kimi.com/coding",
5230
+ headers: { "User-Agent": "KimiCLI/1.5" },
5171
5231
  reasoning: true,
5172
5232
  input: ["text", "image"],
5173
5233
  cost: {
@@ -5185,6 +5245,7 @@ export const MODELS = {
5185
5245
  api: "anthropic-messages",
5186
5246
  provider: "kimi-coding",
5187
5247
  baseUrl: "https://api.kimi.com/coding",
5248
+ headers: { "User-Agent": "KimiCLI/1.5" },
5188
5249
  reasoning: true,
5189
5250
  input: ["text", "image"],
5190
5251
  cost: {
@@ -5202,6 +5263,7 @@ export const MODELS = {
5202
5263
  api: "anthropic-messages",
5203
5264
  provider: "kimi-coding",
5204
5265
  baseUrl: "https://api.kimi.com/coding",
5266
+ headers: { "User-Agent": "KimiCLI/1.5" },
5205
5267
  reasoning: true,
5206
5268
  input: ["text"],
5207
5269
  cost: {
@@ -6289,8 +6351,8 @@ export const MODELS = {
6289
6351
  cacheRead: 0.5,
6290
6352
  cacheWrite: 0,
6291
6353
  },
6292
- contextWindow: 1050000,
6293
- maxTokens: 130000,
6354
+ contextWindow: 272000,
6355
+ maxTokens: 128000,
6294
6356
  },
6295
6357
  "o1": {
6296
6358
  id: "o1",
@@ -6583,6 +6645,23 @@ export const MODELS = {
6583
6645
  contextWindow: 272000,
6584
6646
  maxTokens: 128000,
6585
6647
  },
6648
+ "gpt-5.5": {
6649
+ id: "gpt-5.5",
6650
+ name: "GPT-5.5",
6651
+ api: "openai-codex-responses",
6652
+ provider: "openai-codex",
6653
+ baseUrl: "https://chatgpt.com/backend-api",
6654
+ reasoning: true,
6655
+ input: ["text", "image"],
6656
+ cost: {
6657
+ input: 5,
6658
+ output: 30,
6659
+ cacheRead: 0.5,
6660
+ cacheWrite: 0,
6661
+ },
6662
+ contextWindow: 272000,
6663
+ maxTokens: 128000,
6664
+ },
6586
6665
  },
6587
6666
  "opencode": {
6588
6667
  "big-pickle": {
@@ -7061,6 +7140,40 @@ export const MODELS = {
7061
7140
  contextWindow: 1050000,
7062
7141
  maxTokens: 128000,
7063
7142
  },
7143
+ "gpt-5.5": {
7144
+ id: "gpt-5.5",
7145
+ name: "GPT-5.5",
7146
+ api: "openai-responses",
7147
+ provider: "opencode",
7148
+ baseUrl: "https://opencode.ai/zen/v1",
7149
+ reasoning: true,
7150
+ input: ["text", "image"],
7151
+ cost: {
7152
+ input: 5,
7153
+ output: 30,
7154
+ cacheRead: 0.5,
7155
+ cacheWrite: 0,
7156
+ },
7157
+ contextWindow: 1050000,
7158
+ maxTokens: 130000,
7159
+ },
7160
+ "gpt-5.5-pro": {
7161
+ id: "gpt-5.5-pro",
7162
+ name: "GPT-5.5 Pro",
7163
+ api: "openai-responses",
7164
+ provider: "opencode",
7165
+ baseUrl: "https://opencode.ai/zen/v1",
7166
+ reasoning: true,
7167
+ input: ["text", "image"],
7168
+ cost: {
7169
+ input: 30,
7170
+ output: 180,
7171
+ cacheRead: 30,
7172
+ cacheWrite: 0,
7173
+ },
7174
+ contextWindow: 1050000,
7175
+ maxTokens: 128000,
7176
+ },
7064
7177
  "hy3-preview-free": {
7065
7178
  id: "hy3-preview-free",
7066
7179
  name: "Hy3 preview Free",
@@ -7233,6 +7346,42 @@ export const MODELS = {
7233
7346
  },
7234
7347
  },
7235
7348
  "opencode-go": {
7349
+ "deepseek-v4-flash": {
7350
+ id: "deepseek-v4-flash",
7351
+ name: "DeepSeek V4 Flash",
7352
+ api: "openai-completions",
7353
+ provider: "opencode-go",
7354
+ baseUrl: "https://opencode.ai/zen/go/v1",
7355
+ compat: { "requiresReasoningContentOnAssistantMessages": true },
7356
+ reasoning: true,
7357
+ input: ["text"],
7358
+ cost: {
7359
+ input: 0.14,
7360
+ output: 0.28,
7361
+ cacheRead: 0.028,
7362
+ cacheWrite: 0,
7363
+ },
7364
+ contextWindow: 1000000,
7365
+ maxTokens: 384000,
7366
+ },
7367
+ "deepseek-v4-pro": {
7368
+ id: "deepseek-v4-pro",
7369
+ name: "DeepSeek V4 Pro",
7370
+ api: "openai-completions",
7371
+ provider: "opencode-go",
7372
+ baseUrl: "https://opencode.ai/zen/go/v1",
7373
+ compat: { "requiresReasoningContentOnAssistantMessages": true },
7374
+ reasoning: true,
7375
+ input: ["text"],
7376
+ cost: {
7377
+ input: 1.74,
7378
+ output: 3.48,
7379
+ cacheRead: 0.145,
7380
+ cacheWrite: 0,
7381
+ },
7382
+ contextWindow: 1000000,
7383
+ maxTokens: 384000,
7384
+ },
7236
7385
  "glm-5": {
7237
7386
  id: "glm-5",
7238
7387
  name: "GLM-5",
@@ -7488,7 +7637,7 @@ export const MODELS = {
7488
7637
  cacheWrite: 0,
7489
7638
  },
7490
7639
  contextWindow: 65536,
7491
- maxTokens: 4096,
7640
+ maxTokens: 16384,
7492
7641
  },
7493
7642
  "amazon/nova-2-lite-v1": {
7494
7643
  id: "amazon/nova-2-lite-v1",
@@ -8049,7 +8198,7 @@ export const MODELS = {
8049
8198
  cacheWrite: 0,
8050
8199
  },
8051
8200
  contextWindow: 163840,
8052
- maxTokens: 163840,
8201
+ maxTokens: 16384,
8053
8202
  },
8054
8203
  "deepseek/deepseek-chat-v3-0324": {
8055
8204
  id: "deepseek/deepseek-chat-v3-0324",
@@ -8066,7 +8215,7 @@ export const MODELS = {
8066
8215
  cacheWrite: 0,
8067
8216
  },
8068
8217
  contextWindow: 163840,
8069
- maxTokens: 4096,
8218
+ maxTokens: 16384,
8070
8219
  },
8071
8220
  "deepseek/deepseek-chat-v3.1": {
8072
8221
  id: "deepseek/deepseek-chat-v3.1",
@@ -8117,7 +8266,7 @@ export const MODELS = {
8117
8266
  cacheWrite: 0,
8118
8267
  },
8119
8268
  contextWindow: 163840,
8120
- maxTokens: 4096,
8269
+ maxTokens: 32768,
8121
8270
  },
8122
8271
  "deepseek/deepseek-v3.1-terminus": {
8123
8272
  id: "deepseek/deepseek-v3.1-terminus",
@@ -8130,11 +8279,11 @@ export const MODELS = {
8130
8279
  cost: {
8131
8280
  input: 0.21,
8132
8281
  output: 0.7899999999999999,
8133
- cacheRead: 0.1300000002,
8282
+ cacheRead: 0.13,
8134
8283
  cacheWrite: 0,
8135
8284
  },
8136
8285
  contextWindow: 163840,
8137
- maxTokens: 4096,
8286
+ maxTokens: 32768,
8138
8287
  },
8139
8288
  "deepseek/deepseek-v3.2": {
8140
8289
  id: "deepseek/deepseek-v3.2",
@@ -8176,6 +8325,7 @@ export const MODELS = {
8176
8325
  api: "openai-completions",
8177
8326
  provider: "openrouter",
8178
8327
  baseUrl: "https://openrouter.ai/api/v1",
8328
+ compat: { "requiresReasoningContentOnAssistantMessages": true },
8179
8329
  reasoning: true,
8180
8330
  input: ["text"],
8181
8331
  cost: {
@@ -8193,12 +8343,13 @@ export const MODELS = {
8193
8343
  api: "openai-completions",
8194
8344
  provider: "openrouter",
8195
8345
  baseUrl: "https://openrouter.ai/api/v1",
8346
+ compat: { "requiresReasoningContentOnAssistantMessages": true },
8196
8347
  reasoning: true,
8197
8348
  input: ["text"],
8198
8349
  cost: {
8199
- input: 1.74,
8200
- output: 3.48,
8201
- cacheRead: 0.145,
8350
+ input: 0.435,
8351
+ output: 0.87,
8352
+ cacheRead: 0.03625,
8202
8353
  cacheWrite: 0,
8203
8354
  },
8204
8355
  contextWindow: 1048576,
@@ -8235,7 +8386,7 @@ export const MODELS = {
8235
8386
  cacheRead: 0.024999999999999998,
8236
8387
  cacheWrite: 0.08333333333333334,
8237
8388
  },
8238
- contextWindow: 1048576,
8389
+ contextWindow: 1000000,
8239
8390
  maxTokens: 8192,
8240
8391
  },
8241
8392
  "google/gemini-2.0-flash-lite-001": {
@@ -8425,6 +8576,40 @@ export const MODELS = {
8425
8576
  contextWindow: 1048576,
8426
8577
  maxTokens: 65536,
8427
8578
  },
8579
+ "google/gemma-3-12b-it": {
8580
+ id: "google/gemma-3-12b-it",
8581
+ name: "Google: Gemma 3 12B",
8582
+ api: "openai-completions",
8583
+ provider: "openrouter",
8584
+ baseUrl: "https://openrouter.ai/api/v1",
8585
+ reasoning: false,
8586
+ input: ["text", "image"],
8587
+ cost: {
8588
+ input: 0.04,
8589
+ output: 0.13,
8590
+ cacheRead: 0,
8591
+ cacheWrite: 0,
8592
+ },
8593
+ contextWindow: 131072,
8594
+ maxTokens: 16384,
8595
+ },
8596
+ "google/gemma-3-27b-it": {
8597
+ id: "google/gemma-3-27b-it",
8598
+ name: "Google: Gemma 3 27B",
8599
+ api: "openai-completions",
8600
+ provider: "openrouter",
8601
+ baseUrl: "https://openrouter.ai/api/v1",
8602
+ reasoning: false,
8603
+ input: ["text", "image"],
8604
+ cost: {
8605
+ input: 0.08,
8606
+ output: 0.16,
8607
+ cacheRead: 0,
8608
+ cacheWrite: 0,
8609
+ },
8610
+ contextWindow: 131072,
8611
+ maxTokens: 16384,
8612
+ },
8428
8613
  "google/gemma-4-26b-a4b-it": {
8429
8614
  id: "google/gemma-4-26b-a4b-it",
8430
8615
  name: "Google: Gemma 4 26B A4B ",
@@ -8474,7 +8659,7 @@ export const MODELS = {
8474
8659
  cacheWrite: 0,
8475
8660
  },
8476
8661
  contextWindow: 262144,
8477
- maxTokens: 4096,
8662
+ maxTokens: 16384,
8478
8663
  },
8479
8664
  "google/gemma-4-31b-it:free": {
8480
8665
  id: "google/gemma-4-31b-it:free",
@@ -8593,7 +8778,7 @@ export const MODELS = {
8593
8778
  cacheWrite: 0,
8594
8779
  },
8595
8780
  contextWindow: 131072,
8596
- maxTokens: 4096,
8781
+ maxTokens: 16384,
8597
8782
  },
8598
8783
  "meta-llama/llama-3.1-8b-instruct": {
8599
8784
  id: "meta-llama/llama-3.1-8b-instruct",
@@ -8729,7 +8914,7 @@ export const MODELS = {
8729
8914
  cacheWrite: 0,
8730
8915
  },
8731
8916
  contextWindow: 196608,
8732
- maxTokens: 4096,
8917
+ maxTokens: 131072,
8733
8918
  },
8734
8919
  "minimax/minimax-m2.5:free": {
8735
8920
  id: "minimax/minimax-m2.5:free",
@@ -9052,7 +9237,7 @@ export const MODELS = {
9052
9237
  cacheWrite: 0,
9053
9238
  },
9054
9239
  contextWindow: 128000,
9055
- maxTokens: 4096,
9240
+ maxTokens: 16384,
9056
9241
  },
9057
9242
  "mistralai/mistral-small-creative": {
9058
9243
  id: "mistralai/mistral-small-creative",
@@ -9273,7 +9458,7 @@ export const MODELS = {
9273
9458
  cacheWrite: 0,
9274
9459
  },
9275
9460
  contextWindow: 131072,
9276
- maxTokens: 4096,
9461
+ maxTokens: 16384,
9277
9462
  },
9278
9463
  "nvidia/nemotron-3-nano-30b-a3b": {
9279
9464
  id: "nvidia/nemotron-3-nano-30b-a3b",
@@ -9290,7 +9475,7 @@ export const MODELS = {
9290
9475
  cacheWrite: 0,
9291
9476
  },
9292
9477
  contextWindow: 262144,
9293
- maxTokens: 4096,
9478
+ maxTokens: 228000,
9294
9479
  },
9295
9480
  "nvidia/nemotron-3-nano-30b-a3b:free": {
9296
9481
  id: "nvidia/nemotron-3-nano-30b-a3b:free",
@@ -9375,7 +9560,7 @@ export const MODELS = {
9375
9560
  cacheWrite: 0,
9376
9561
  },
9377
9562
  contextWindow: 131072,
9378
- maxTokens: 4096,
9563
+ maxTokens: 16384,
9379
9564
  },
9380
9565
  "nvidia/nemotron-nano-9b-v2:free": {
9381
9566
  id: "nvidia/nemotron-nano-9b-v2:free",
@@ -10040,6 +10225,40 @@ export const MODELS = {
10040
10225
  contextWindow: 1050000,
10041
10226
  maxTokens: 128000,
10042
10227
  },
10228
+ "openai/gpt-5.5": {
10229
+ id: "openai/gpt-5.5",
10230
+ name: "OpenAI: GPT-5.5",
10231
+ api: "openai-completions",
10232
+ provider: "openrouter",
10233
+ baseUrl: "https://openrouter.ai/api/v1",
10234
+ reasoning: true,
10235
+ input: ["text", "image"],
10236
+ cost: {
10237
+ input: 5,
10238
+ output: 30,
10239
+ cacheRead: 0.5,
10240
+ cacheWrite: 0,
10241
+ },
10242
+ contextWindow: 1050000,
10243
+ maxTokens: 128000,
10244
+ },
10245
+ "openai/gpt-5.5-pro": {
10246
+ id: "openai/gpt-5.5-pro",
10247
+ name: "OpenAI: GPT-5.5 Pro",
10248
+ api: "openai-completions",
10249
+ provider: "openrouter",
10250
+ baseUrl: "https://openrouter.ai/api/v1",
10251
+ reasoning: true,
10252
+ input: ["text", "image"],
10253
+ cost: {
10254
+ input: 30,
10255
+ output: 180,
10256
+ cacheRead: 0,
10257
+ cacheWrite: 0,
10258
+ },
10259
+ contextWindow: 1050000,
10260
+ maxTokens: 128000,
10261
+ },
10043
10262
  "openai/gpt-audio": {
10044
10263
  id: "openai/gpt-audio",
10045
10264
  name: "OpenAI: GPT Audio",
@@ -10089,7 +10308,7 @@ export const MODELS = {
10089
10308
  cacheWrite: 0,
10090
10309
  },
10091
10310
  contextWindow: 131072,
10092
- maxTokens: 4096,
10311
+ maxTokens: 131072,
10093
10312
  },
10094
10313
  "openai/gpt-oss-120b:free": {
10095
10314
  id: "openai/gpt-oss-120b:free",
@@ -10123,7 +10342,7 @@ export const MODELS = {
10123
10342
  cacheWrite: 0,
10124
10343
  },
10125
10344
  contextWindow: 131072,
10126
- maxTokens: 4096,
10345
+ maxTokens: 131072,
10127
10346
  },
10128
10347
  "openai/gpt-oss-20b:free": {
10129
10348
  id: "openai/gpt-oss-20b:free",
@@ -10548,7 +10767,7 @@ export const MODELS = {
10548
10767
  cacheWrite: 0,
10549
10768
  },
10550
10769
  contextWindow: 262144,
10551
- maxTokens: 4096,
10770
+ maxTokens: 16384,
10552
10771
  },
10553
10772
  "qwen/qwen3-235b-a22b-thinking-2507": {
10554
10773
  id: "qwen/qwen3-235b-a22b-thinking-2507",
@@ -10582,7 +10801,7 @@ export const MODELS = {
10582
10801
  cacheWrite: 0,
10583
10802
  },
10584
10803
  contextWindow: 40960,
10585
- maxTokens: 40960,
10804
+ maxTokens: 16384,
10586
10805
  },
10587
10806
  "qwen/qwen3-30b-a3b-instruct-2507": {
10588
10807
  id: "qwen/qwen3-30b-a3b-instruct-2507",
@@ -10662,12 +10881,12 @@ export const MODELS = {
10662
10881
  input: ["text"],
10663
10882
  cost: {
10664
10883
  input: 0.22,
10665
- output: 1,
10666
- cacheRead: 0.022,
10884
+ output: 1.7999999999999998,
10885
+ cacheRead: 0,
10667
10886
  cacheWrite: 0,
10668
10887
  },
10669
10888
  contextWindow: 262144,
10670
- maxTokens: 4096,
10889
+ maxTokens: 65536,
10671
10890
  },
10672
10891
  "qwen/qwen3-coder-30b-a3b-instruct": {
10673
10892
  id: "qwen/qwen3-coder-30b-a3b-instruct",
@@ -10712,9 +10931,9 @@ export const MODELS = {
10712
10931
  reasoning: false,
10713
10932
  input: ["text"],
10714
10933
  cost: {
10715
- input: 0.15,
10934
+ input: 0.14,
10716
10935
  output: 0.7999999999999999,
10717
- cacheRead: 0.11,
10936
+ cacheRead: 0.09,
10718
10937
  cacheWrite: 0,
10719
10938
  },
10720
10939
  contextWindow: 262144,
@@ -10803,7 +11022,7 @@ export const MODELS = {
10803
11022
  cacheWrite: 0,
10804
11023
  },
10805
11024
  contextWindow: 262144,
10806
- maxTokens: 4096,
11025
+ maxTokens: 16384,
10807
11026
  },
10808
11027
  "qwen/qwen3-next-80b-a3b-instruct:free": {
10809
11028
  id: "qwen/qwen3-next-80b-a3b-instruct:free",
@@ -10854,7 +11073,7 @@ export const MODELS = {
10854
11073
  cacheWrite: 0,
10855
11074
  },
10856
11075
  contextWindow: 262144,
10857
- maxTokens: 4096,
11076
+ maxTokens: 16384,
10858
11077
  },
10859
11078
  "qwen/qwen3-vl-235b-a22b-thinking": {
10860
11079
  id: "qwen/qwen3-vl-235b-a22b-thinking",
@@ -11632,11 +11851,11 @@ export const MODELS = {
11632
11851
  cost: {
11633
11852
  input: 0.3,
11634
11853
  output: 0.8999999999999999,
11635
- cacheRead: 0,
11854
+ cacheRead: 0.049999999999999996,
11636
11855
  cacheWrite: 0,
11637
11856
  },
11638
11857
  contextWindow: 131072,
11639
- maxTokens: 131072,
11858
+ maxTokens: 24000,
11640
11859
  },
11641
11860
  "z-ai/glm-4.7": {
11642
11861
  id: "z-ai/glm-4.7",
@@ -11666,11 +11885,11 @@ export const MODELS = {
11666
11885
  cost: {
11667
11886
  input: 0.06,
11668
11887
  output: 0.39999999999999997,
11669
- cacheRead: 0.0100000002,
11888
+ cacheRead: 0.01,
11670
11889
  cacheWrite: 0,
11671
11890
  },
11672
11891
  contextWindow: 202752,
11673
- maxTokens: 4096,
11892
+ maxTokens: 16384,
11674
11893
  },
11675
11894
  "z-ai/glm-5": {
11676
11895
  id: "z-ai/glm-5",
@@ -11687,7 +11906,7 @@ export const MODELS = {
11687
11906
  cacheWrite: 0,
11688
11907
  },
11689
11908
  contextWindow: 202752,
11690
- maxTokens: 4096,
11909
+ maxTokens: 16384,
11691
11910
  },
11692
11911
  "z-ai/glm-5-turbo": {
11693
11912
  id: "z-ai/glm-5-turbo",
@@ -12433,7 +12652,7 @@ export const MODELS = {
12433
12652
  cost: {
12434
12653
  input: 0.14,
12435
12654
  output: 0.28,
12436
- cacheRead: 0.014,
12655
+ cacheRead: 0.028,
12437
12656
  cacheWrite: 0,
12438
12657
  },
12439
12658
  contextWindow: 1000000,
@@ -13731,6 +13950,40 @@ export const MODELS = {
13731
13950
  contextWindow: 1050000,
13732
13951
  maxTokens: 128000,
13733
13952
  },
13953
+ "openai/gpt-5.5": {
13954
+ id: "openai/gpt-5.5",
13955
+ name: "GPT 5.5",
13956
+ api: "anthropic-messages",
13957
+ provider: "vercel-ai-gateway",
13958
+ baseUrl: "https://ai-gateway.vercel.sh",
13959
+ reasoning: true,
13960
+ input: ["text", "image"],
13961
+ cost: {
13962
+ input: 5,
13963
+ output: 30,
13964
+ cacheRead: 0.5,
13965
+ cacheWrite: 0,
13966
+ },
13967
+ contextWindow: 1000000,
13968
+ maxTokens: 128000,
13969
+ },
13970
+ "openai/gpt-5.5-pro": {
13971
+ id: "openai/gpt-5.5-pro",
13972
+ name: "GPT 5.5 Pro",
13973
+ api: "anthropic-messages",
13974
+ provider: "vercel-ai-gateway",
13975
+ baseUrl: "https://ai-gateway.vercel.sh",
13976
+ reasoning: true,
13977
+ input: ["text", "image"],
13978
+ cost: {
13979
+ input: 30,
13980
+ output: 180,
13981
+ cacheRead: 0,
13982
+ cacheWrite: 0,
13983
+ },
13984
+ contextWindow: 1000000,
13985
+ maxTokens: 128000,
13986
+ },
13734
13987
  "openai/gpt-oss-20b": {
13735
13988
  id: "openai/gpt-oss-20b",
13736
13989
  name: "GPT OSS 120B",
@@ -14857,24 +15110,6 @@ export const MODELS = {
14857
15110
  },
14858
15111
  },
14859
15112
  "zai": {
14860
- "glm-4.5": {
14861
- id: "glm-4.5",
14862
- name: "GLM-4.5",
14863
- api: "openai-completions",
14864
- provider: "zai",
14865
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
14866
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14867
- reasoning: true,
14868
- input: ["text"],
14869
- cost: {
14870
- input: 0,
14871
- output: 0,
14872
- cacheRead: 0,
14873
- cacheWrite: 0,
14874
- },
14875
- contextWindow: 131072,
14876
- maxTokens: 98304,
14877
- },
14878
15113
  "glm-4.5-air": {
14879
15114
  id: "glm-4.5-air",
14880
15115
  name: "GLM-4.5-Air",
@@ -14893,78 +15128,6 @@ export const MODELS = {
14893
15128
  contextWindow: 131072,
14894
15129
  maxTokens: 98304,
14895
15130
  },
14896
- "glm-4.5-flash": {
14897
- id: "glm-4.5-flash",
14898
- name: "GLM-4.5-Flash",
14899
- api: "openai-completions",
14900
- provider: "zai",
14901
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
14902
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14903
- reasoning: true,
14904
- input: ["text"],
14905
- cost: {
14906
- input: 0,
14907
- output: 0,
14908
- cacheRead: 0,
14909
- cacheWrite: 0,
14910
- },
14911
- contextWindow: 131072,
14912
- maxTokens: 98304,
14913
- },
14914
- "glm-4.5v": {
14915
- id: "glm-4.5v",
14916
- name: "GLM-4.5V",
14917
- api: "openai-completions",
14918
- provider: "zai",
14919
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
14920
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai" },
14921
- reasoning: true,
14922
- input: ["text", "image"],
14923
- cost: {
14924
- input: 0,
14925
- output: 0,
14926
- cacheRead: 0,
14927
- cacheWrite: 0,
14928
- },
14929
- contextWindow: 64000,
14930
- maxTokens: 16384,
14931
- },
14932
- "glm-4.6": {
14933
- id: "glm-4.6",
14934
- name: "GLM-4.6",
14935
- api: "openai-completions",
14936
- provider: "zai",
14937
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
14938
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14939
- reasoning: true,
14940
- input: ["text"],
14941
- cost: {
14942
- input: 0,
14943
- output: 0,
14944
- cacheRead: 0,
14945
- cacheWrite: 0,
14946
- },
14947
- contextWindow: 204800,
14948
- maxTokens: 131072,
14949
- },
14950
- "glm-4.6v": {
14951
- id: "glm-4.6v",
14952
- name: "GLM-4.6V",
14953
- api: "openai-completions",
14954
- provider: "zai",
14955
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
14956
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14957
- reasoning: true,
14958
- input: ["text", "image"],
14959
- cost: {
14960
- input: 0,
14961
- output: 0,
14962
- cacheRead: 0,
14963
- cacheWrite: 0,
14964
- },
14965
- contextWindow: 128000,
14966
- maxTokens: 32768,
14967
- },
14968
15131
  "glm-4.7": {
14969
15132
  id: "glm-4.7",
14970
15133
  name: "GLM-4.7",
@@ -14983,60 +15146,6 @@ export const MODELS = {
14983
15146
  contextWindow: 204800,
14984
15147
  maxTokens: 131072,
14985
15148
  },
14986
- "glm-4.7-flash": {
14987
- id: "glm-4.7-flash",
14988
- name: "GLM-4.7-Flash",
14989
- api: "openai-completions",
14990
- provider: "zai",
14991
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
14992
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
14993
- reasoning: true,
14994
- input: ["text"],
14995
- cost: {
14996
- input: 0,
14997
- output: 0,
14998
- cacheRead: 0,
14999
- cacheWrite: 0,
15000
- },
15001
- contextWindow: 200000,
15002
- maxTokens: 131072,
15003
- },
15004
- "glm-4.7-flashx": {
15005
- id: "glm-4.7-flashx",
15006
- name: "GLM-4.7-FlashX",
15007
- api: "openai-completions",
15008
- provider: "zai",
15009
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15010
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
15011
- reasoning: true,
15012
- input: ["text"],
15013
- cost: {
15014
- input: 0.07,
15015
- output: 0.4,
15016
- cacheRead: 0.01,
15017
- cacheWrite: 0,
15018
- },
15019
- contextWindow: 200000,
15020
- maxTokens: 131072,
15021
- },
15022
- "glm-5": {
15023
- id: "glm-5",
15024
- name: "GLM-5",
15025
- api: "openai-completions",
15026
- provider: "zai",
15027
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15028
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
15029
- reasoning: true,
15030
- input: ["text"],
15031
- cost: {
15032
- input: 0,
15033
- output: 0,
15034
- cacheRead: 0,
15035
- cacheWrite: 0,
15036
- },
15037
- contextWindow: 204800,
15038
- maxTokens: 131072,
15039
- },
15040
15149
  "glm-5-turbo": {
15041
15150
  id: "glm-5-turbo",
15042
15151
  name: "GLM-5-Turbo",
@@ -15073,24 +15182,6 @@ export const MODELS = {
15073
15182
  contextWindow: 200000,
15074
15183
  maxTokens: 131072,
15075
15184
  },
15076
- "glm-5v-turbo": {
15077
- id: "glm-5v-turbo",
15078
- name: "glm-5v-turbo",
15079
- api: "openai-completions",
15080
- provider: "zai",
15081
- baseUrl: "https://api.z.ai/api/coding/paas/v4",
15082
- compat: { "supportsDeveloperRole": false, "thinkingFormat": "zai", "zaiToolStream": true },
15083
- reasoning: true,
15084
- input: ["text", "image"],
15085
- cost: {
15086
- input: 0,
15087
- output: 0,
15088
- cacheRead: 0,
15089
- cacheWrite: 0,
15090
- },
15091
- contextWindow: 200000,
15092
- maxTokens: 131072,
15093
- },
15094
15185
  },
15095
15186
  };
15096
15187
  //# sourceMappingURL=models.generated.js.map