@hyperspaceng/neural-ai 0.70.6 → 0.70.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +52 -35
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +72 -55
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +14 -15
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -7423,23 +7423,6 @@ export const MODELS = {
|
|
|
7423
7423
|
contextWindow: 262144,
|
|
7424
7424
|
maxTokens: 65536,
|
|
7425
7425
|
},
|
|
7426
|
-
"ling-2.6-flash-free": {
|
|
7427
|
-
id: "ling-2.6-flash-free",
|
|
7428
|
-
name: "Ling 2.6 Flash Free",
|
|
7429
|
-
api: "openai-completions",
|
|
7430
|
-
provider: "opencode",
|
|
7431
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
7432
|
-
reasoning: false,
|
|
7433
|
-
input: ["text"],
|
|
7434
|
-
cost: {
|
|
7435
|
-
input: 0,
|
|
7436
|
-
output: 0,
|
|
7437
|
-
cacheRead: 0,
|
|
7438
|
-
cacheWrite: 0,
|
|
7439
|
-
},
|
|
7440
|
-
contextWindow: 262100,
|
|
7441
|
-
maxTokens: 32800,
|
|
7442
|
-
},
|
|
7443
7426
|
"minimax-m2.5": {
|
|
7444
7427
|
id: "minimax-m2.5",
|
|
7445
7428
|
name: "MiniMax M2.5",
|
|
@@ -9437,23 +9420,6 @@ export const MODELS = {
|
|
|
9437
9420
|
contextWindow: 128000,
|
|
9438
9421
|
maxTokens: 16384,
|
|
9439
9422
|
},
|
|
9440
|
-
"mistralai/mistral-small-creative": {
|
|
9441
|
-
id: "mistralai/mistral-small-creative",
|
|
9442
|
-
name: "Mistral: Mistral Small Creative",
|
|
9443
|
-
api: "openai-completions",
|
|
9444
|
-
provider: "openrouter",
|
|
9445
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
9446
|
-
reasoning: false,
|
|
9447
|
-
input: ["text"],
|
|
9448
|
-
cost: {
|
|
9449
|
-
input: 0.09999999999999999,
|
|
9450
|
-
output: 0.3,
|
|
9451
|
-
cacheRead: 0.01,
|
|
9452
|
-
cacheWrite: 0,
|
|
9453
|
-
},
|
|
9454
|
-
contextWindow: 32768,
|
|
9455
|
-
maxTokens: 4096,
|
|
9456
|
-
},
|
|
9457
9423
|
"mistralai/mixtral-8x22b-instruct": {
|
|
9458
9424
|
id: "mistralai/mixtral-8x22b-instruct",
|
|
9459
9425
|
name: "Mistral: Mixtral 8x22B Instruct",
|
|
@@ -9599,13 +9565,13 @@ export const MODELS = {
|
|
|
9599
9565
|
reasoning: true,
|
|
9600
9566
|
input: ["text", "image"],
|
|
9601
9567
|
cost: {
|
|
9602
|
-
input: 0.
|
|
9603
|
-
output:
|
|
9604
|
-
cacheRead: 0.
|
|
9568
|
+
input: 0.75,
|
|
9569
|
+
output: 3.5,
|
|
9570
|
+
cacheRead: 0.15,
|
|
9605
9571
|
cacheWrite: 0,
|
|
9606
9572
|
},
|
|
9607
|
-
contextWindow:
|
|
9608
|
-
maxTokens:
|
|
9573
|
+
contextWindow: 262144,
|
|
9574
|
+
maxTokens: 16384,
|
|
9609
9575
|
},
|
|
9610
9576
|
"nex-agi/deepseek-v3.1-nex-n1": {
|
|
9611
9577
|
id: "nex-agi/deepseek-v3.1-nex-n1",
|
|
@@ -12353,13 +12319,13 @@ export const MODELS = {
|
|
|
12353
12319
|
reasoning: true,
|
|
12354
12320
|
input: ["text", "image"],
|
|
12355
12321
|
cost: {
|
|
12356
|
-
input: 0.
|
|
12357
|
-
output:
|
|
12358
|
-
cacheRead: 0.
|
|
12322
|
+
input: 0.75,
|
|
12323
|
+
output: 3.5,
|
|
12324
|
+
cacheRead: 0.15,
|
|
12359
12325
|
cacheWrite: 0,
|
|
12360
12326
|
},
|
|
12361
|
-
contextWindow:
|
|
12362
|
-
maxTokens:
|
|
12327
|
+
contextWindow: 262144,
|
|
12328
|
+
maxTokens: 16384,
|
|
12363
12329
|
},
|
|
12364
12330
|
"~openai/gpt-latest": {
|
|
12365
12331
|
id: "~openai/gpt-latest",
|
|
@@ -12484,20 +12450,20 @@ export const MODELS = {
|
|
|
12484
12450
|
},
|
|
12485
12451
|
"alibaba/qwen3-235b-a22b-thinking": {
|
|
12486
12452
|
id: "alibaba/qwen3-235b-a22b-thinking",
|
|
12487
|
-
name: "Qwen3 235B A22B Thinking
|
|
12453
|
+
name: "Qwen3 VL 235B A22B Thinking",
|
|
12488
12454
|
api: "anthropic-messages",
|
|
12489
12455
|
provider: "vercel-ai-gateway",
|
|
12490
12456
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12491
12457
|
reasoning: true,
|
|
12492
12458
|
input: ["text", "image"],
|
|
12493
12459
|
cost: {
|
|
12494
|
-
input: 0.
|
|
12495
|
-
output:
|
|
12496
|
-
cacheRead: 0
|
|
12460
|
+
input: 0.39999999999999997,
|
|
12461
|
+
output: 4,
|
|
12462
|
+
cacheRead: 0,
|
|
12497
12463
|
cacheWrite: 0,
|
|
12498
12464
|
},
|
|
12499
|
-
contextWindow:
|
|
12500
|
-
maxTokens:
|
|
12465
|
+
contextWindow: 131072,
|
|
12466
|
+
maxTokens: 32768,
|
|
12501
12467
|
},
|
|
12502
12468
|
"alibaba/qwen3-coder": {
|
|
12503
12469
|
id: "alibaba/qwen3-coder",
|
|
@@ -12669,6 +12635,23 @@ export const MODELS = {
|
|
|
12669
12635
|
contextWindow: 1000000,
|
|
12670
12636
|
maxTokens: 64000,
|
|
12671
12637
|
},
|
|
12638
|
+
"alibaba/qwen3.6-27b": {
|
|
12639
|
+
id: "alibaba/qwen3.6-27b",
|
|
12640
|
+
name: "Qwen 3.6 27B",
|
|
12641
|
+
api: "anthropic-messages",
|
|
12642
|
+
provider: "vercel-ai-gateway",
|
|
12643
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
12644
|
+
reasoning: true,
|
|
12645
|
+
input: ["text", "image"],
|
|
12646
|
+
cost: {
|
|
12647
|
+
input: 0.6,
|
|
12648
|
+
output: 3.5999999999999996,
|
|
12649
|
+
cacheRead: 0,
|
|
12650
|
+
cacheWrite: 0,
|
|
12651
|
+
},
|
|
12652
|
+
contextWindow: 256000,
|
|
12653
|
+
maxTokens: 256000,
|
|
12654
|
+
},
|
|
12672
12655
|
"alibaba/qwen3.6-plus": {
|
|
12673
12656
|
id: "alibaba/qwen3.6-plus",
|
|
12674
12657
|
name: "Qwen 3.6 Plus",
|
|
@@ -13049,16 +13032,16 @@ export const MODELS = {
|
|
|
13049
13032
|
api: "anthropic-messages",
|
|
13050
13033
|
provider: "vercel-ai-gateway",
|
|
13051
13034
|
baseUrl: "https://ai-gateway.vercel.sh",
|
|
13052
|
-
reasoning:
|
|
13035
|
+
reasoning: false,
|
|
13053
13036
|
input: ["text"],
|
|
13054
13037
|
cost: {
|
|
13055
|
-
input: 0.
|
|
13056
|
-
output:
|
|
13057
|
-
cacheRead: 0
|
|
13038
|
+
input: 0.62,
|
|
13039
|
+
output: 1.85,
|
|
13040
|
+
cacheRead: 0,
|
|
13058
13041
|
cacheWrite: 0,
|
|
13059
13042
|
},
|
|
13060
13043
|
contextWindow: 128000,
|
|
13061
|
-
maxTokens:
|
|
13044
|
+
maxTokens: 8000,
|
|
13062
13045
|
},
|
|
13063
13046
|
"deepseek/deepseek-v4-flash": {
|
|
13064
13047
|
id: "deepseek/deepseek-v4-flash",
|
|
@@ -14896,6 +14879,40 @@ export const MODELS = {
|
|
|
14896
14879
|
contextWindow: 1000000,
|
|
14897
14880
|
maxTokens: 128000,
|
|
14898
14881
|
},
|
|
14882
|
+
"xiaomi/mimo-v2.5": {
|
|
14883
|
+
id: "xiaomi/mimo-v2.5",
|
|
14884
|
+
name: "MiMo M2.5",
|
|
14885
|
+
api: "anthropic-messages",
|
|
14886
|
+
provider: "vercel-ai-gateway",
|
|
14887
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
14888
|
+
reasoning: true,
|
|
14889
|
+
input: ["text", "image"],
|
|
14890
|
+
cost: {
|
|
14891
|
+
input: 0.39999999999999997,
|
|
14892
|
+
output: 2,
|
|
14893
|
+
cacheRead: 0.08,
|
|
14894
|
+
cacheWrite: 0,
|
|
14895
|
+
},
|
|
14896
|
+
contextWindow: 1050000,
|
|
14897
|
+
maxTokens: 131100,
|
|
14898
|
+
},
|
|
14899
|
+
"xiaomi/mimo-v2.5-pro": {
|
|
14900
|
+
id: "xiaomi/mimo-v2.5-pro",
|
|
14901
|
+
name: "MiMo V2.5 Pro",
|
|
14902
|
+
api: "anthropic-messages",
|
|
14903
|
+
provider: "vercel-ai-gateway",
|
|
14904
|
+
baseUrl: "https://ai-gateway.vercel.sh",
|
|
14905
|
+
reasoning: true,
|
|
14906
|
+
input: ["text", "image"],
|
|
14907
|
+
cost: {
|
|
14908
|
+
input: 1,
|
|
14909
|
+
output: 3,
|
|
14910
|
+
cacheRead: 0.19999999999999998,
|
|
14911
|
+
cacheWrite: 0,
|
|
14912
|
+
},
|
|
14913
|
+
contextWindow: 1050000,
|
|
14914
|
+
maxTokens: 131000,
|
|
14915
|
+
},
|
|
14899
14916
|
"zai/glm-4.5": {
|
|
14900
14917
|
id: "zai/glm-4.5",
|
|
14901
14918
|
name: "GLM-4.5",
|