@hyperspaceng/neural-ai 0.70.6 → 0.70.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -7423,23 +7423,6 @@ export const MODELS = {
7423
7423
  contextWindow: 262144,
7424
7424
  maxTokens: 65536,
7425
7425
  },
7426
- "ling-2.6-flash-free": {
7427
- id: "ling-2.6-flash-free",
7428
- name: "Ling 2.6 Flash Free",
7429
- api: "openai-completions",
7430
- provider: "opencode",
7431
- baseUrl: "https://opencode.ai/zen/v1",
7432
- reasoning: false,
7433
- input: ["text"],
7434
- cost: {
7435
- input: 0,
7436
- output: 0,
7437
- cacheRead: 0,
7438
- cacheWrite: 0,
7439
- },
7440
- contextWindow: 262100,
7441
- maxTokens: 32800,
7442
- },
7443
7426
  "minimax-m2.5": {
7444
7427
  id: "minimax-m2.5",
7445
7428
  name: "MiniMax M2.5",
@@ -9437,23 +9420,6 @@ export const MODELS = {
9437
9420
  contextWindow: 128000,
9438
9421
  maxTokens: 16384,
9439
9422
  },
9440
- "mistralai/mistral-small-creative": {
9441
- id: "mistralai/mistral-small-creative",
9442
- name: "Mistral: Mistral Small Creative",
9443
- api: "openai-completions",
9444
- provider: "openrouter",
9445
- baseUrl: "https://openrouter.ai/api/v1",
9446
- reasoning: false,
9447
- input: ["text"],
9448
- cost: {
9449
- input: 0.09999999999999999,
9450
- output: 0.3,
9451
- cacheRead: 0.01,
9452
- cacheWrite: 0,
9453
- },
9454
- contextWindow: 32768,
9455
- maxTokens: 4096,
9456
- },
9457
9423
  "mistralai/mixtral-8x22b-instruct": {
9458
9424
  id: "mistralai/mixtral-8x22b-instruct",
9459
9425
  name: "Mistral: Mixtral 8x22B Instruct",
@@ -9599,13 +9565,13 @@ export const MODELS = {
9599
9565
  reasoning: true,
9600
9566
  input: ["text", "image"],
9601
9567
  cost: {
9602
- input: 0.7448,
9603
- output: 4.655,
9604
- cacheRead: 0.1463,
9568
+ input: 0.75,
9569
+ output: 3.5,
9570
+ cacheRead: 0.15,
9605
9571
  cacheWrite: 0,
9606
9572
  },
9607
- contextWindow: 256000,
9608
- maxTokens: 65536,
9573
+ contextWindow: 262144,
9574
+ maxTokens: 16384,
9609
9575
  },
9610
9576
  "nex-agi/deepseek-v3.1-nex-n1": {
9611
9577
  id: "nex-agi/deepseek-v3.1-nex-n1",
@@ -12353,13 +12319,13 @@ export const MODELS = {
12353
12319
  reasoning: true,
12354
12320
  input: ["text", "image"],
12355
12321
  cost: {
12356
- input: 0.7448,
12357
- output: 4.655,
12358
- cacheRead: 0.1463,
12322
+ input: 0.75,
12323
+ output: 3.5,
12324
+ cacheRead: 0.15,
12359
12325
  cacheWrite: 0,
12360
12326
  },
12361
- contextWindow: 256000,
12362
- maxTokens: 65536,
12327
+ contextWindow: 262144,
12328
+ maxTokens: 16384,
12363
12329
  },
12364
12330
  "~openai/gpt-latest": {
12365
12331
  id: "~openai/gpt-latest",
@@ -12484,20 +12450,20 @@ export const MODELS = {
12484
12450
  },
12485
12451
  "alibaba/qwen3-235b-a22b-thinking": {
12486
12452
  id: "alibaba/qwen3-235b-a22b-thinking",
12487
- name: "Qwen3 235B A22B Thinking 2507",
12453
+ name: "Qwen3 VL 235B A22B Thinking",
12488
12454
  api: "anthropic-messages",
12489
12455
  provider: "vercel-ai-gateway",
12490
12456
  baseUrl: "https://ai-gateway.vercel.sh",
12491
12457
  reasoning: true,
12492
12458
  input: ["text", "image"],
12493
12459
  cost: {
12494
- input: 0.22999999999999998,
12495
- output: 2.3,
12496
- cacheRead: 0.19999999999999998,
12460
+ input: 0.39999999999999997,
12461
+ output: 4,
12462
+ cacheRead: 0,
12497
12463
  cacheWrite: 0,
12498
12464
  },
12499
- contextWindow: 262114,
12500
- maxTokens: 262114,
12465
+ contextWindow: 131072,
12466
+ maxTokens: 32768,
12501
12467
  },
12502
12468
  "alibaba/qwen3-coder": {
12503
12469
  id: "alibaba/qwen3-coder",
@@ -12669,6 +12635,23 @@ export const MODELS = {
12669
12635
  contextWindow: 1000000,
12670
12636
  maxTokens: 64000,
12671
12637
  },
12638
+ "alibaba/qwen3.6-27b": {
12639
+ id: "alibaba/qwen3.6-27b",
12640
+ name: "Qwen 3.6 27B",
12641
+ api: "anthropic-messages",
12642
+ provider: "vercel-ai-gateway",
12643
+ baseUrl: "https://ai-gateway.vercel.sh",
12644
+ reasoning: true,
12645
+ input: ["text", "image"],
12646
+ cost: {
12647
+ input: 0.6,
12648
+ output: 3.5999999999999996,
12649
+ cacheRead: 0,
12650
+ cacheWrite: 0,
12651
+ },
12652
+ contextWindow: 256000,
12653
+ maxTokens: 256000,
12654
+ },
12672
12655
  "alibaba/qwen3.6-plus": {
12673
12656
  id: "alibaba/qwen3.6-plus",
12674
12657
  name: "Qwen 3.6 Plus",
@@ -13049,16 +13032,16 @@ export const MODELS = {
13049
13032
  api: "anthropic-messages",
13050
13033
  provider: "vercel-ai-gateway",
13051
13034
  baseUrl: "https://ai-gateway.vercel.sh",
13052
- reasoning: true,
13035
+ reasoning: false,
13053
13036
  input: ["text"],
13054
13037
  cost: {
13055
- input: 0.28,
13056
- output: 0.42,
13057
- cacheRead: 0.028,
13038
+ input: 0.62,
13039
+ output: 1.85,
13040
+ cacheRead: 0,
13058
13041
  cacheWrite: 0,
13059
13042
  },
13060
13043
  contextWindow: 128000,
13061
- maxTokens: 64000,
13044
+ maxTokens: 8000,
13062
13045
  },
13063
13046
  "deepseek/deepseek-v4-flash": {
13064
13047
  id: "deepseek/deepseek-v4-flash",
@@ -14896,6 +14879,40 @@ export const MODELS = {
14896
14879
  contextWindow: 1000000,
14897
14880
  maxTokens: 128000,
14898
14881
  },
14882
+ "xiaomi/mimo-v2.5": {
14883
+ id: "xiaomi/mimo-v2.5",
14884
+ name: "MiMo M2.5",
14885
+ api: "anthropic-messages",
14886
+ provider: "vercel-ai-gateway",
14887
+ baseUrl: "https://ai-gateway.vercel.sh",
14888
+ reasoning: true,
14889
+ input: ["text", "image"],
14890
+ cost: {
14891
+ input: 0.39999999999999997,
14892
+ output: 2,
14893
+ cacheRead: 0.08,
14894
+ cacheWrite: 0,
14895
+ },
14896
+ contextWindow: 1050000,
14897
+ maxTokens: 131100,
14898
+ },
14899
+ "xiaomi/mimo-v2.5-pro": {
14900
+ id: "xiaomi/mimo-v2.5-pro",
14901
+ name: "MiMo V2.5 Pro",
14902
+ api: "anthropic-messages",
14903
+ provider: "vercel-ai-gateway",
14904
+ baseUrl: "https://ai-gateway.vercel.sh",
14905
+ reasoning: true,
14906
+ input: ["text", "image"],
14907
+ cost: {
14908
+ input: 1,
14909
+ output: 3,
14910
+ cacheRead: 0.19999999999999998,
14911
+ cacheWrite: 0,
14912
+ },
14913
+ contextWindow: 1050000,
14914
+ maxTokens: 131000,
14915
+ },
14899
14916
  "zai/glm-4.5": {
14900
14917
  id: "zai/glm-4.5",
14901
14918
  name: "GLM-4.5",