@hyperspaceng/neural-ai 0.68.2 → 0.69.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -325,6 +325,40 @@ export const MODELS = {
325
325
  contextWindow: 1000000,
326
326
  maxTokens: 64000,
327
327
  },
328
+ "au.anthropic.claude-opus-4-6-v1": {
329
+ id: "au.anthropic.claude-opus-4-6-v1",
330
+ name: "AU Anthropic Claude Opus 4.6",
331
+ api: "bedrock-converse-stream",
332
+ provider: "amazon-bedrock",
333
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
334
+ reasoning: true,
335
+ input: ["text", "image"],
336
+ cost: {
337
+ input: 16.5,
338
+ output: 82.5,
339
+ cacheRead: 0.5,
340
+ cacheWrite: 6.25,
341
+ },
342
+ contextWindow: 1000000,
343
+ maxTokens: 128000,
344
+ },
345
+ "au.anthropic.claude-sonnet-4-6": {
346
+ id: "au.anthropic.claude-sonnet-4-6",
347
+ name: "AU Anthropic Claude Sonnet 4.6",
348
+ api: "bedrock-converse-stream",
349
+ provider: "amazon-bedrock",
350
+ baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
351
+ reasoning: true,
352
+ input: ["text", "image"],
353
+ cost: {
354
+ input: 3.3,
355
+ output: 16.5,
356
+ cacheRead: 0.33,
357
+ cacheWrite: 4.125,
358
+ },
359
+ contextWindow: 1000000,
360
+ maxTokens: 128000,
361
+ },
328
362
  "deepseek.r1-v1:0": {
329
363
  id: "deepseek.r1-v1:0",
330
364
  name: "DeepSeek-R1",
@@ -2488,6 +2522,23 @@ export const MODELS = {
2488
2522
  contextWindow: 1050000,
2489
2523
  maxTokens: 128000,
2490
2524
  },
2525
+ "gpt-5.5": {
2526
+ id: "gpt-5.5",
2527
+ name: "GPT-5.5",
2528
+ api: "azure-openai-responses",
2529
+ provider: "azure-openai-responses",
2530
+ baseUrl: "",
2531
+ reasoning: true,
2532
+ input: ["text", "image"],
2533
+ cost: {
2534
+ input: 5,
2535
+ output: 30,
2536
+ cacheRead: 0.5,
2537
+ cacheWrite: 0,
2538
+ },
2539
+ contextWindow: 1050000,
2540
+ maxTokens: 130000,
2541
+ },
2491
2542
  "o1": {
2492
2543
  id: "o1",
2493
2544
  name: "o1",
@@ -4164,6 +4215,23 @@ export const MODELS = {
4164
4215
  contextWindow: 1048576,
4165
4216
  maxTokens: 65535,
4166
4217
  },
4218
+ "gemini-3.1-flash-lite-preview": {
4219
+ id: "gemini-3.1-flash-lite-preview",
4220
+ name: "Gemini 3.1 Flash Lite Preview (Cloud Code Assist)",
4221
+ api: "google-gemini-cli",
4222
+ provider: "google-gemini-cli",
4223
+ baseUrl: "https://cloudcode-pa.googleapis.com",
4224
+ reasoning: true,
4225
+ input: ["text", "image"],
4226
+ cost: {
4227
+ input: 0,
4228
+ output: 0,
4229
+ cacheRead: 0,
4230
+ cacheWrite: 0,
4231
+ },
4232
+ contextWindow: 1048576,
4233
+ maxTokens: 65535,
4234
+ },
4167
4235
  "gemini-3.1-pro-preview": {
4168
4236
  id: "gemini-3.1-pro-preview",
4169
4237
  name: "Gemini 3.1 Pro Preview (Cloud Code Assist)",
@@ -6207,6 +6275,23 @@ export const MODELS = {
6207
6275
  contextWindow: 1050000,
6208
6276
  maxTokens: 128000,
6209
6277
  },
6278
+ "gpt-5.5": {
6279
+ id: "gpt-5.5",
6280
+ name: "GPT-5.5",
6281
+ api: "openai-responses",
6282
+ provider: "openai",
6283
+ baseUrl: "https://api.openai.com/v1",
6284
+ reasoning: true,
6285
+ input: ["text", "image"],
6286
+ cost: {
6287
+ input: 5,
6288
+ output: 30,
6289
+ cacheRead: 0.5,
6290
+ cacheWrite: 0,
6291
+ },
6292
+ contextWindow: 1050000,
6293
+ maxTokens: 130000,
6294
+ },
6210
6295
  "o1": {
6211
6296
  id: "o1",
6212
6297
  name: "o1",
@@ -6976,6 +7061,23 @@ export const MODELS = {
6976
7061
  contextWindow: 1050000,
6977
7062
  maxTokens: 128000,
6978
7063
  },
7064
+ "hy3-preview-free": {
7065
+ id: "hy3-preview-free",
7066
+ name: "Hy3 preview Free",
7067
+ api: "openai-completions",
7068
+ provider: "opencode",
7069
+ baseUrl: "https://opencode.ai/zen/v1",
7070
+ reasoning: true,
7071
+ input: ["text"],
7072
+ cost: {
7073
+ input: 0,
7074
+ output: 0,
7075
+ cacheRead: 0,
7076
+ cacheWrite: 0,
7077
+ },
7078
+ contextWindow: 256000,
7079
+ maxTokens: 64000,
7080
+ },
6979
7081
  "kimi-k2.5": {
6980
7082
  id: "kimi-k2.5",
6981
7083
  name: "Kimi K2.5",
@@ -8068,6 +8170,40 @@ export const MODELS = {
8068
8170
  contextWindow: 163840,
8069
8171
  maxTokens: 65536,
8070
8172
  },
8173
+ "deepseek/deepseek-v4-flash": {
8174
+ id: "deepseek/deepseek-v4-flash",
8175
+ name: "DeepSeek: DeepSeek V4 Flash",
8176
+ api: "openai-completions",
8177
+ provider: "openrouter",
8178
+ baseUrl: "https://openrouter.ai/api/v1",
8179
+ reasoning: true,
8180
+ input: ["text"],
8181
+ cost: {
8182
+ input: 0.14,
8183
+ output: 0.28,
8184
+ cacheRead: 0.028,
8185
+ cacheWrite: 0,
8186
+ },
8187
+ contextWindow: 1048576,
8188
+ maxTokens: 384000,
8189
+ },
8190
+ "deepseek/deepseek-v4-pro": {
8191
+ id: "deepseek/deepseek-v4-pro",
8192
+ name: "DeepSeek: DeepSeek V4 Pro",
8193
+ api: "openai-completions",
8194
+ provider: "openrouter",
8195
+ baseUrl: "https://openrouter.ai/api/v1",
8196
+ reasoning: true,
8197
+ input: ["text"],
8198
+ cost: {
8199
+ input: 1.74,
8200
+ output: 3.48,
8201
+ cacheRead: 0.145,
8202
+ cacheWrite: 0,
8203
+ },
8204
+ contextWindow: 1048576,
8205
+ maxTokens: 384000,
8206
+ },
8071
8207
  "essentialai/rnj-1-instruct": {
8072
8208
  id: "essentialai/rnj-1-instruct",
8073
8209
  name: "EssentialAI: Rnj 1 Instruct",
@@ -8099,7 +8235,7 @@ export const MODELS = {
8099
8235
  cacheRead: 0.024999999999999998,
8100
8236
  cacheWrite: 0.08333333333333334,
8101
8237
  },
8102
- contextWindow: 1000000,
8238
+ contextWindow: 1048576,
8103
8239
  maxTokens: 8192,
8104
8240
  },
8105
8241
  "google/gemini-2.0-flash-lite-001": {
@@ -8298,8 +8434,8 @@ export const MODELS = {
8298
8434
  reasoning: true,
8299
8435
  input: ["text", "image"],
8300
8436
  cost: {
8301
- input: 0.07,
8302
- output: 0.33999999999999997,
8437
+ input: 0.06,
8438
+ output: 0.33,
8303
8439
  cacheRead: 0,
8304
8440
  cacheWrite: 0,
8305
8441
  },
@@ -8374,6 +8510,23 @@ export const MODELS = {
8374
8510
  contextWindow: 128000,
8375
8511
  maxTokens: 50000,
8376
8512
  },
8513
+ "inclusionai/ling-2.6-1t:free": {
8514
+ id: "inclusionai/ling-2.6-1t:free",
8515
+ name: "inclusionAI: Ling-2.6-1T (free)",
8516
+ api: "openai-completions",
8517
+ provider: "openrouter",
8518
+ baseUrl: "https://openrouter.ai/api/v1",
8519
+ reasoning: false,
8520
+ input: ["text"],
8521
+ cost: {
8522
+ input: 0,
8523
+ output: 0,
8524
+ cacheRead: 0,
8525
+ cacheWrite: 0,
8526
+ },
8527
+ contextWindow: 262144,
8528
+ maxTokens: 32768,
8529
+ },
8377
8530
  "inclusionai/ling-2.6-flash:free": {
8378
8531
  id: "inclusionai/ling-2.6-flash:free",
8379
8532
  name: "inclusionAI: Ling-2.6-flash (free)",
@@ -8571,12 +8724,12 @@ export const MODELS = {
8571
8724
  input: ["text"],
8572
8725
  cost: {
8573
8726
  input: 0.15,
8574
- output: 1.2,
8575
- cacheRead: 0.075,
8727
+ output: 1.15,
8728
+ cacheRead: 0.03,
8576
8729
  cacheWrite: 0,
8577
8730
  },
8578
8731
  contextWindow: 196608,
8579
- maxTokens: 65536,
8732
+ maxTokens: 4096,
8580
8733
  },
8581
8734
  "minimax/minimax-m2.5:free": {
8582
8735
  id: "minimax/minimax-m2.5:free",
@@ -8842,13 +8995,13 @@ export const MODELS = {
8842
8995
  reasoning: false,
8843
8996
  input: ["text"],
8844
8997
  cost: {
8845
- input: 0.02,
8846
- output: 0.04,
8998
+ input: 0.01,
8999
+ output: 0.03,
8847
9000
  cacheRead: 0,
8848
9001
  cacheWrite: 0,
8849
9002
  },
8850
9003
  contextWindow: 131072,
8851
- maxTokens: 16384,
9004
+ maxTokens: 4096,
8852
9005
  },
8853
9006
  "mistralai/mistral-saba": {
8854
9007
  id: "mistralai/mistral-saba",
@@ -9063,13 +9216,13 @@ export const MODELS = {
9063
9216
  reasoning: true,
9064
9217
  input: ["text", "image"],
9065
9218
  cost: {
9066
- input: 0.75,
9067
- output: 3.5,
9068
- cacheRead: 0.15,
9219
+ input: 0.7448,
9220
+ output: 4.655,
9221
+ cacheRead: 0.1463,
9069
9222
  cacheWrite: 0,
9070
9223
  },
9071
- contextWindow: 262144,
9072
- maxTokens: 4096,
9224
+ contextWindow: 256000,
9225
+ maxTokens: 65536,
9073
9226
  },
9074
9227
  "nex-agi/deepseek-v3.1-nex-n1": {
9075
9228
  id: "nex-agi/deepseek-v3.1-nex-n1",
@@ -11043,6 +11196,23 @@ export const MODELS = {
11043
11196
  contextWindow: 262144,
11044
11197
  maxTokens: 65536,
11045
11198
  },
11199
+ "tencent/hy3-preview:free": {
11200
+ id: "tencent/hy3-preview:free",
11201
+ name: "Tencent: Hy3 preview (free)",
11202
+ api: "openai-completions",
11203
+ provider: "openrouter",
11204
+ baseUrl: "https://openrouter.ai/api/v1",
11205
+ reasoning: true,
11206
+ input: ["text"],
11207
+ cost: {
11208
+ input: 0,
11209
+ output: 0,
11210
+ cacheRead: 0,
11211
+ cacheWrite: 0,
11212
+ },
11213
+ contextWindow: 262144,
11214
+ maxTokens: 262144,
11215
+ },
11046
11216
  "thedrummer/rocinante-12b": {
11047
11217
  id: "thedrummer/rocinante-12b",
11048
11218
  name: "TheDrummer: Rocinante 12B",
@@ -12252,6 +12422,40 @@ export const MODELS = {
12252
12422
  contextWindow: 128000,
12253
12423
  maxTokens: 64000,
12254
12424
  },
12425
+ "deepseek/deepseek-v4-flash": {
12426
+ id: "deepseek/deepseek-v4-flash",
12427
+ name: "DeepSeek V4 Flash",
12428
+ api: "anthropic-messages",
12429
+ provider: "vercel-ai-gateway",
12430
+ baseUrl: "https://ai-gateway.vercel.sh",
12431
+ reasoning: true,
12432
+ input: ["text"],
12433
+ cost: {
12434
+ input: 0.14,
12435
+ output: 0.28,
12436
+ cacheRead: 0.014,
12437
+ cacheWrite: 0,
12438
+ },
12439
+ contextWindow: 1000000,
12440
+ maxTokens: 384000,
12441
+ },
12442
+ "deepseek/deepseek-v4-pro": {
12443
+ id: "deepseek/deepseek-v4-pro",
12444
+ name: "DeepSeek V4 Pro",
12445
+ api: "anthropic-messages",
12446
+ provider: "vercel-ai-gateway",
12447
+ baseUrl: "https://ai-gateway.vercel.sh",
12448
+ reasoning: true,
12449
+ input: ["text"],
12450
+ cost: {
12451
+ input: 1.74,
12452
+ output: 3.48,
12453
+ cacheRead: 0.145,
12454
+ cacheWrite: 0,
12455
+ },
12456
+ contextWindow: 1000000,
12457
+ maxTokens: 384000,
12458
+ },
12255
12459
  "google/gemini-2.0-flash": {
12256
12460
  id: "google/gemini-2.0-flash",
12257
12461
  name: "Gemini 2.0 Flash",