@hyperspaceng/neural-ai 0.69.1 → 0.70.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3062,6 +3062,7 @@ export const MODELS = {
3062
3062
  provider: "github-copilot",
3063
3063
  baseUrl: "https://api.individual.githubcopilot.com",
3064
3064
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3065
+ compat: { "supportsEagerToolInputStreaming": false },
3065
3066
  reasoning: true,
3066
3067
  input: ["text", "image"],
3067
3068
  cost: {
@@ -3134,6 +3135,7 @@ export const MODELS = {
3134
3135
  provider: "github-copilot",
3135
3136
  baseUrl: "https://api.individual.githubcopilot.com",
3136
3137
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3138
+ compat: { "supportsEagerToolInputStreaming": false },
3137
3139
  reasoning: true,
3138
3140
  input: ["text", "image"],
3139
3141
  cost: {
@@ -3152,6 +3154,7 @@ export const MODELS = {
3152
3154
  provider: "github-copilot",
3153
3155
  baseUrl: "https://api.individual.githubcopilot.com",
3154
3156
  headers: { "User-Agent": "GitHubCopilotChat/0.35.0", "Editor-Version": "vscode/1.107.0", "Editor-Plugin-Version": "copilot-chat/0.35.0", "Copilot-Integration-Id": "vscode-chat" },
3157
+ compat: { "supportsEagerToolInputStreaming": false },
3155
3158
  reasoning: true,
3156
3159
  input: ["text", "image"],
3157
3160
  cost: {
@@ -5168,6 +5171,7 @@ export const MODELS = {
5168
5171
  api: "anthropic-messages",
5169
5172
  provider: "kimi-coding",
5170
5173
  baseUrl: "https://api.kimi.com/coding",
5174
+ headers: { "User-Agent": "KimiCLI/1.5" },
5171
5175
  reasoning: true,
5172
5176
  input: ["text", "image"],
5173
5177
  cost: {
@@ -5185,6 +5189,7 @@ export const MODELS = {
5185
5189
  api: "anthropic-messages",
5186
5190
  provider: "kimi-coding",
5187
5191
  baseUrl: "https://api.kimi.com/coding",
5192
+ headers: { "User-Agent": "KimiCLI/1.5" },
5188
5193
  reasoning: true,
5189
5194
  input: ["text", "image"],
5190
5195
  cost: {
@@ -5202,6 +5207,7 @@ export const MODELS = {
5202
5207
  api: "anthropic-messages",
5203
5208
  provider: "kimi-coding",
5204
5209
  baseUrl: "https://api.kimi.com/coding",
5210
+ headers: { "User-Agent": "KimiCLI/1.5" },
5205
5211
  reasoning: true,
5206
5212
  input: ["text"],
5207
5213
  cost: {
@@ -6583,6 +6589,23 @@ export const MODELS = {
6583
6589
  contextWindow: 272000,
6584
6590
  maxTokens: 128000,
6585
6591
  },
6592
+ "gpt-5.5": {
6593
+ id: "gpt-5.5",
6594
+ name: "GPT-5.5",
6595
+ api: "openai-codex-responses",
6596
+ provider: "openai-codex",
6597
+ baseUrl: "https://chatgpt.com/backend-api",
6598
+ reasoning: true,
6599
+ input: ["text", "image"],
6600
+ cost: {
6601
+ input: 5,
6602
+ output: 30,
6603
+ cacheRead: 0.5,
6604
+ cacheWrite: 0,
6605
+ },
6606
+ contextWindow: 400000,
6607
+ maxTokens: 128000,
6608
+ },
6586
6609
  },
6587
6610
  "opencode": {
6588
6611
  "big-pickle": {
@@ -7061,6 +7084,40 @@ export const MODELS = {
7061
7084
  contextWindow: 1050000,
7062
7085
  maxTokens: 128000,
7063
7086
  },
7087
+ "gpt-5.5": {
7088
+ id: "gpt-5.5",
7089
+ name: "GPT-5.5",
7090
+ api: "openai-responses",
7091
+ provider: "opencode",
7092
+ baseUrl: "https://opencode.ai/zen/v1",
7093
+ reasoning: true,
7094
+ input: ["text", "image"],
7095
+ cost: {
7096
+ input: 5,
7097
+ output: 30,
7098
+ cacheRead: 0.5,
7099
+ cacheWrite: 0,
7100
+ },
7101
+ contextWindow: 1050000,
7102
+ maxTokens: 130000,
7103
+ },
7104
+ "gpt-5.5-pro": {
7105
+ id: "gpt-5.5-pro",
7106
+ name: "GPT-5.5 Pro",
7107
+ api: "openai-responses",
7108
+ provider: "opencode",
7109
+ baseUrl: "https://opencode.ai/zen/v1",
7110
+ reasoning: true,
7111
+ input: ["text", "image"],
7112
+ cost: {
7113
+ input: 30,
7114
+ output: 180,
7115
+ cacheRead: 30,
7116
+ cacheWrite: 0,
7117
+ },
7118
+ contextWindow: 1050000,
7119
+ maxTokens: 128000,
7120
+ },
7064
7121
  "hy3-preview-free": {
7065
7122
  id: "hy3-preview-free",
7066
7123
  name: "Hy3 preview Free",
@@ -7233,6 +7290,40 @@ export const MODELS = {
7233
7290
  },
7234
7291
  },
7235
7292
  "opencode-go": {
7293
+ "deepseek-v4-flash": {
7294
+ id: "deepseek-v4-flash",
7295
+ name: "DeepSeek V4 Flash",
7296
+ api: "openai-completions",
7297
+ provider: "opencode-go",
7298
+ baseUrl: "https://opencode.ai/zen/go/v1",
7299
+ reasoning: true,
7300
+ input: ["text"],
7301
+ cost: {
7302
+ input: 0.14,
7303
+ output: 0.28,
7304
+ cacheRead: 0.028,
7305
+ cacheWrite: 0,
7306
+ },
7307
+ contextWindow: 1000000,
7308
+ maxTokens: 384000,
7309
+ },
7310
+ "deepseek-v4-pro": {
7311
+ id: "deepseek-v4-pro",
7312
+ name: "DeepSeek V4 Pro",
7313
+ api: "openai-completions",
7314
+ provider: "opencode-go",
7315
+ baseUrl: "https://opencode.ai/zen/go/v1",
7316
+ reasoning: true,
7317
+ input: ["text"],
7318
+ cost: {
7319
+ input: 1.74,
7320
+ output: 3.48,
7321
+ cacheRead: 0.145,
7322
+ cacheWrite: 0,
7323
+ },
7324
+ contextWindow: 1000000,
7325
+ maxTokens: 384000,
7326
+ },
7236
7327
  "glm-5": {
7237
7328
  id: "glm-5",
7238
7329
  name: "GLM-5",
@@ -7488,7 +7579,7 @@ export const MODELS = {
7488
7579
  cacheWrite: 0,
7489
7580
  },
7490
7581
  contextWindow: 65536,
7491
- maxTokens: 4096,
7582
+ maxTokens: 16384,
7492
7583
  },
7493
7584
  "amazon/nova-2-lite-v1": {
7494
7585
  id: "amazon/nova-2-lite-v1",
@@ -8049,7 +8140,7 @@ export const MODELS = {
8049
8140
  cacheWrite: 0,
8050
8141
  },
8051
8142
  contextWindow: 163840,
8052
- maxTokens: 163840,
8143
+ maxTokens: 16384,
8053
8144
  },
8054
8145
  "deepseek/deepseek-chat-v3-0324": {
8055
8146
  id: "deepseek/deepseek-chat-v3-0324",
@@ -8066,7 +8157,7 @@ export const MODELS = {
8066
8157
  cacheWrite: 0,
8067
8158
  },
8068
8159
  contextWindow: 163840,
8069
- maxTokens: 4096,
8160
+ maxTokens: 16384,
8070
8161
  },
8071
8162
  "deepseek/deepseek-chat-v3.1": {
8072
8163
  id: "deepseek/deepseek-chat-v3.1",
@@ -8117,7 +8208,7 @@ export const MODELS = {
8117
8208
  cacheWrite: 0,
8118
8209
  },
8119
8210
  contextWindow: 163840,
8120
- maxTokens: 4096,
8211
+ maxTokens: 32768,
8121
8212
  },
8122
8213
  "deepseek/deepseek-v3.1-terminus": {
8123
8214
  id: "deepseek/deepseek-v3.1-terminus",
@@ -8130,11 +8221,11 @@ export const MODELS = {
8130
8221
  cost: {
8131
8222
  input: 0.21,
8132
8223
  output: 0.7899999999999999,
8133
- cacheRead: 0.1300000002,
8224
+ cacheRead: 0.13,
8134
8225
  cacheWrite: 0,
8135
8226
  },
8136
8227
  contextWindow: 163840,
8137
- maxTokens: 4096,
8228
+ maxTokens: 32768,
8138
8229
  },
8139
8230
  "deepseek/deepseek-v3.2": {
8140
8231
  id: "deepseek/deepseek-v3.2",
@@ -8235,7 +8326,7 @@ export const MODELS = {
8235
8326
  cacheRead: 0.024999999999999998,
8236
8327
  cacheWrite: 0.08333333333333334,
8237
8328
  },
8238
- contextWindow: 1048576,
8329
+ contextWindow: 1000000,
8239
8330
  maxTokens: 8192,
8240
8331
  },
8241
8332
  "google/gemini-2.0-flash-lite-001": {
@@ -8425,6 +8516,40 @@ export const MODELS = {
8425
8516
  contextWindow: 1048576,
8426
8517
  maxTokens: 65536,
8427
8518
  },
8519
+ "google/gemma-3-12b-it": {
8520
+ id: "google/gemma-3-12b-it",
8521
+ name: "Google: Gemma 3 12B",
8522
+ api: "openai-completions",
8523
+ provider: "openrouter",
8524
+ baseUrl: "https://openrouter.ai/api/v1",
8525
+ reasoning: false,
8526
+ input: ["text", "image"],
8527
+ cost: {
8528
+ input: 0.04,
8529
+ output: 0.13,
8530
+ cacheRead: 0,
8531
+ cacheWrite: 0,
8532
+ },
8533
+ contextWindow: 131072,
8534
+ maxTokens: 16384,
8535
+ },
8536
+ "google/gemma-3-27b-it": {
8537
+ id: "google/gemma-3-27b-it",
8538
+ name: "Google: Gemma 3 27B",
8539
+ api: "openai-completions",
8540
+ provider: "openrouter",
8541
+ baseUrl: "https://openrouter.ai/api/v1",
8542
+ reasoning: false,
8543
+ input: ["text", "image"],
8544
+ cost: {
8545
+ input: 0.08,
8546
+ output: 0.16,
8547
+ cacheRead: 0,
8548
+ cacheWrite: 0,
8549
+ },
8550
+ contextWindow: 131072,
8551
+ maxTokens: 16384,
8552
+ },
8428
8553
  "google/gemma-4-26b-a4b-it": {
8429
8554
  id: "google/gemma-4-26b-a4b-it",
8430
8555
  name: "Google: Gemma 4 26B A4B ",
@@ -8474,7 +8599,7 @@ export const MODELS = {
8474
8599
  cacheWrite: 0,
8475
8600
  },
8476
8601
  contextWindow: 262144,
8477
- maxTokens: 4096,
8602
+ maxTokens: 16384,
8478
8603
  },
8479
8604
  "google/gemma-4-31b-it:free": {
8480
8605
  id: "google/gemma-4-31b-it:free",
@@ -8593,7 +8718,7 @@ export const MODELS = {
8593
8718
  cacheWrite: 0,
8594
8719
  },
8595
8720
  contextWindow: 131072,
8596
- maxTokens: 4096,
8721
+ maxTokens: 16384,
8597
8722
  },
8598
8723
  "meta-llama/llama-3.1-8b-instruct": {
8599
8724
  id: "meta-llama/llama-3.1-8b-instruct",
@@ -8729,7 +8854,7 @@ export const MODELS = {
8729
8854
  cacheWrite: 0,
8730
8855
  },
8731
8856
  contextWindow: 196608,
8732
- maxTokens: 4096,
8857
+ maxTokens: 131072,
8733
8858
  },
8734
8859
  "minimax/minimax-m2.5:free": {
8735
8860
  id: "minimax/minimax-m2.5:free",
@@ -9052,7 +9177,7 @@ export const MODELS = {
9052
9177
  cacheWrite: 0,
9053
9178
  },
9054
9179
  contextWindow: 128000,
9055
- maxTokens: 4096,
9180
+ maxTokens: 16384,
9056
9181
  },
9057
9182
  "mistralai/mistral-small-creative": {
9058
9183
  id: "mistralai/mistral-small-creative",
@@ -9273,7 +9398,7 @@ export const MODELS = {
9273
9398
  cacheWrite: 0,
9274
9399
  },
9275
9400
  contextWindow: 131072,
9276
- maxTokens: 4096,
9401
+ maxTokens: 16384,
9277
9402
  },
9278
9403
  "nvidia/nemotron-3-nano-30b-a3b": {
9279
9404
  id: "nvidia/nemotron-3-nano-30b-a3b",
@@ -9290,7 +9415,7 @@ export const MODELS = {
9290
9415
  cacheWrite: 0,
9291
9416
  },
9292
9417
  contextWindow: 262144,
9293
- maxTokens: 4096,
9418
+ maxTokens: 228000,
9294
9419
  },
9295
9420
  "nvidia/nemotron-3-nano-30b-a3b:free": {
9296
9421
  id: "nvidia/nemotron-3-nano-30b-a3b:free",
@@ -9375,7 +9500,7 @@ export const MODELS = {
9375
9500
  cacheWrite: 0,
9376
9501
  },
9377
9502
  contextWindow: 131072,
9378
- maxTokens: 4096,
9503
+ maxTokens: 16384,
9379
9504
  },
9380
9505
  "nvidia/nemotron-nano-9b-v2:free": {
9381
9506
  id: "nvidia/nemotron-nano-9b-v2:free",
@@ -10040,6 +10165,40 @@ export const MODELS = {
10040
10165
  contextWindow: 1050000,
10041
10166
  maxTokens: 128000,
10042
10167
  },
10168
+ "openai/gpt-5.5": {
10169
+ id: "openai/gpt-5.5",
10170
+ name: "OpenAI: GPT-5.5",
10171
+ api: "openai-completions",
10172
+ provider: "openrouter",
10173
+ baseUrl: "https://openrouter.ai/api/v1",
10174
+ reasoning: true,
10175
+ input: ["text", "image"],
10176
+ cost: {
10177
+ input: 5,
10178
+ output: 30,
10179
+ cacheRead: 0.5,
10180
+ cacheWrite: 0,
10181
+ },
10182
+ contextWindow: 1050000,
10183
+ maxTokens: 128000,
10184
+ },
10185
+ "openai/gpt-5.5-pro": {
10186
+ id: "openai/gpt-5.5-pro",
10187
+ name: "OpenAI: GPT-5.5 Pro",
10188
+ api: "openai-completions",
10189
+ provider: "openrouter",
10190
+ baseUrl: "https://openrouter.ai/api/v1",
10191
+ reasoning: true,
10192
+ input: ["text", "image"],
10193
+ cost: {
10194
+ input: 30,
10195
+ output: 180,
10196
+ cacheRead: 0,
10197
+ cacheWrite: 0,
10198
+ },
10199
+ contextWindow: 1050000,
10200
+ maxTokens: 128000,
10201
+ },
10043
10202
  "openai/gpt-audio": {
10044
10203
  id: "openai/gpt-audio",
10045
10204
  name: "OpenAI: GPT Audio",
@@ -10089,7 +10248,7 @@ export const MODELS = {
10089
10248
  cacheWrite: 0,
10090
10249
  },
10091
10250
  contextWindow: 131072,
10092
- maxTokens: 4096,
10251
+ maxTokens: 131072,
10093
10252
  },
10094
10253
  "openai/gpt-oss-120b:free": {
10095
10254
  id: "openai/gpt-oss-120b:free",
@@ -10123,7 +10282,7 @@ export const MODELS = {
10123
10282
  cacheWrite: 0,
10124
10283
  },
10125
10284
  contextWindow: 131072,
10126
- maxTokens: 4096,
10285
+ maxTokens: 131072,
10127
10286
  },
10128
10287
  "openai/gpt-oss-20b:free": {
10129
10288
  id: "openai/gpt-oss-20b:free",
@@ -10548,7 +10707,7 @@ export const MODELS = {
10548
10707
  cacheWrite: 0,
10549
10708
  },
10550
10709
  contextWindow: 262144,
10551
- maxTokens: 4096,
10710
+ maxTokens: 16384,
10552
10711
  },
10553
10712
  "qwen/qwen3-235b-a22b-thinking-2507": {
10554
10713
  id: "qwen/qwen3-235b-a22b-thinking-2507",
@@ -10582,7 +10741,7 @@ export const MODELS = {
10582
10741
  cacheWrite: 0,
10583
10742
  },
10584
10743
  contextWindow: 40960,
10585
- maxTokens: 40960,
10744
+ maxTokens: 16384,
10586
10745
  },
10587
10746
  "qwen/qwen3-30b-a3b-instruct-2507": {
10588
10747
  id: "qwen/qwen3-30b-a3b-instruct-2507",
@@ -10667,7 +10826,7 @@ export const MODELS = {
10667
10826
  cacheWrite: 0,
10668
10827
  },
10669
10828
  contextWindow: 262144,
10670
- maxTokens: 4096,
10829
+ maxTokens: 65536,
10671
10830
  },
10672
10831
  "qwen/qwen3-coder-30b-a3b-instruct": {
10673
10832
  id: "qwen/qwen3-coder-30b-a3b-instruct",
@@ -10712,9 +10871,9 @@ export const MODELS = {
10712
10871
  reasoning: false,
10713
10872
  input: ["text"],
10714
10873
  cost: {
10715
- input: 0.15,
10874
+ input: 0.14,
10716
10875
  output: 0.7999999999999999,
10717
- cacheRead: 0.11,
10876
+ cacheRead: 0.09,
10718
10877
  cacheWrite: 0,
10719
10878
  },
10720
10879
  contextWindow: 262144,
@@ -10803,7 +10962,7 @@ export const MODELS = {
10803
10962
  cacheWrite: 0,
10804
10963
  },
10805
10964
  contextWindow: 262144,
10806
- maxTokens: 4096,
10965
+ maxTokens: 16384,
10807
10966
  },
10808
10967
  "qwen/qwen3-next-80b-a3b-instruct:free": {
10809
10968
  id: "qwen/qwen3-next-80b-a3b-instruct:free",
@@ -10854,7 +11013,7 @@ export const MODELS = {
10854
11013
  cacheWrite: 0,
10855
11014
  },
10856
11015
  contextWindow: 262144,
10857
- maxTokens: 4096,
11016
+ maxTokens: 16384,
10858
11017
  },
10859
11018
  "qwen/qwen3-vl-235b-a22b-thinking": {
10860
11019
  id: "qwen/qwen3-vl-235b-a22b-thinking",
@@ -11632,11 +11791,11 @@ export const MODELS = {
11632
11791
  cost: {
11633
11792
  input: 0.3,
11634
11793
  output: 0.8999999999999999,
11635
- cacheRead: 0,
11794
+ cacheRead: 0.049999999999999996,
11636
11795
  cacheWrite: 0,
11637
11796
  },
11638
11797
  contextWindow: 131072,
11639
- maxTokens: 131072,
11798
+ maxTokens: 24000,
11640
11799
  },
11641
11800
  "z-ai/glm-4.7": {
11642
11801
  id: "z-ai/glm-4.7",
@@ -11666,11 +11825,11 @@ export const MODELS = {
11666
11825
  cost: {
11667
11826
  input: 0.06,
11668
11827
  output: 0.39999999999999997,
11669
- cacheRead: 0.0100000002,
11828
+ cacheRead: 0.01,
11670
11829
  cacheWrite: 0,
11671
11830
  },
11672
11831
  contextWindow: 202752,
11673
- maxTokens: 4096,
11832
+ maxTokens: 16384,
11674
11833
  },
11675
11834
  "z-ai/glm-5": {
11676
11835
  id: "z-ai/glm-5",
@@ -11687,7 +11846,7 @@ export const MODELS = {
11687
11846
  cacheWrite: 0,
11688
11847
  },
11689
11848
  contextWindow: 202752,
11690
- maxTokens: 4096,
11849
+ maxTokens: 16384,
11691
11850
  },
11692
11851
  "z-ai/glm-5-turbo": {
11693
11852
  id: "z-ai/glm-5-turbo",
@@ -12433,7 +12592,7 @@ export const MODELS = {
12433
12592
  cost: {
12434
12593
  input: 0.14,
12435
12594
  output: 0.28,
12436
- cacheRead: 0.014,
12595
+ cacheRead: 0.028,
12437
12596
  cacheWrite: 0,
12438
12597
  },
12439
12598
  contextWindow: 1000000,
@@ -13731,6 +13890,40 @@ export const MODELS = {
13731
13890
  contextWindow: 1050000,
13732
13891
  maxTokens: 128000,
13733
13892
  },
13893
+ "openai/gpt-5.5": {
13894
+ id: "openai/gpt-5.5",
13895
+ name: "GPT 5.5",
13896
+ api: "anthropic-messages",
13897
+ provider: "vercel-ai-gateway",
13898
+ baseUrl: "https://ai-gateway.vercel.sh",
13899
+ reasoning: true,
13900
+ input: ["text", "image"],
13901
+ cost: {
13902
+ input: 5,
13903
+ output: 30,
13904
+ cacheRead: 0.5,
13905
+ cacheWrite: 0,
13906
+ },
13907
+ contextWindow: 1000000,
13908
+ maxTokens: 128000,
13909
+ },
13910
+ "openai/gpt-5.5-pro": {
13911
+ id: "openai/gpt-5.5-pro",
13912
+ name: "GPT 5.5 Pro",
13913
+ api: "anthropic-messages",
13914
+ provider: "vercel-ai-gateway",
13915
+ baseUrl: "https://ai-gateway.vercel.sh",
13916
+ reasoning: true,
13917
+ input: ["text", "image"],
13918
+ cost: {
13919
+ input: 30,
13920
+ output: 180,
13921
+ cacheRead: 0,
13922
+ cacheWrite: 0,
13923
+ },
13924
+ contextWindow: 1000000,
13925
+ maxTokens: 128000,
13926
+ },
13734
13927
  "openai/gpt-oss-20b": {
13735
13928
  id: "openai/gpt-oss-20b",
13736
13929
  name: "GPT OSS 120B",