@hyperspaceng/neural-ai 0.66.2 → 0.67.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +37 -51
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +58 -74
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts +2 -0
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +33 -4
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/package.json +1 -1
package/dist/models.generated.js
CHANGED
|
@@ -4356,6 +4356,24 @@ export const MODELS = {
|
|
|
4356
4356
|
contextWindow: 204800,
|
|
4357
4357
|
maxTokens: 131072,
|
|
4358
4358
|
},
|
|
4359
|
+
"MiniMaxAI/MiniMax-M2.7": {
|
|
4360
|
+
id: "MiniMaxAI/MiniMax-M2.7",
|
|
4361
|
+
name: "MiniMax-M2.7",
|
|
4362
|
+
api: "openai-completions",
|
|
4363
|
+
provider: "huggingface",
|
|
4364
|
+
baseUrl: "https://router.huggingface.co/v1",
|
|
4365
|
+
compat: { "supportsDeveloperRole": false },
|
|
4366
|
+
reasoning: true,
|
|
4367
|
+
input: ["text"],
|
|
4368
|
+
cost: {
|
|
4369
|
+
input: 0.3,
|
|
4370
|
+
output: 1.2,
|
|
4371
|
+
cacheRead: 0.06,
|
|
4372
|
+
cacheWrite: 0,
|
|
4373
|
+
},
|
|
4374
|
+
contextWindow: 204800,
|
|
4375
|
+
maxTokens: 131072,
|
|
4376
|
+
},
|
|
4359
4377
|
"Qwen/Qwen3-235B-A22B-Thinking-2507": {
|
|
4360
4378
|
id: "Qwen/Qwen3-235B-A22B-Thinking-2507",
|
|
4361
4379
|
name: "Qwen3-235B-A22B-Thinking-2507",
|
|
@@ -7105,7 +7123,7 @@ export const MODELS = {
|
|
|
7105
7123
|
cost: {
|
|
7106
7124
|
input: 0.22,
|
|
7107
7125
|
output: 0.85,
|
|
7108
|
-
cacheRead: 0,
|
|
7126
|
+
cacheRead: 0.06,
|
|
7109
7127
|
cacheWrite: 0,
|
|
7110
7128
|
},
|
|
7111
7129
|
contextWindow: 262144,
|
|
@@ -7375,13 +7393,13 @@ export const MODELS = {
|
|
|
7375
7393
|
reasoning: true,
|
|
7376
7394
|
input: ["text"],
|
|
7377
7395
|
cost: {
|
|
7378
|
-
input: 0.
|
|
7396
|
+
input: 0.5,
|
|
7379
7397
|
output: 2.1500000000000004,
|
|
7380
|
-
cacheRead: 0.
|
|
7398
|
+
cacheRead: 0.35,
|
|
7381
7399
|
cacheWrite: 0,
|
|
7382
7400
|
},
|
|
7383
7401
|
contextWindow: 163840,
|
|
7384
|
-
maxTokens:
|
|
7402
|
+
maxTokens: 4096,
|
|
7385
7403
|
},
|
|
7386
7404
|
"deepseek/deepseek-v3.1-terminus": {
|
|
7387
7405
|
id: "deepseek/deepseek-v3.1-terminus",
|
|
@@ -7664,13 +7682,13 @@ export const MODELS = {
|
|
|
7664
7682
|
reasoning: true,
|
|
7665
7683
|
input: ["text", "image"],
|
|
7666
7684
|
cost: {
|
|
7667
|
-
input: 0.
|
|
7668
|
-
output: 0.
|
|
7669
|
-
cacheRead: 0,
|
|
7685
|
+
input: 0.08,
|
|
7686
|
+
output: 0.35,
|
|
7687
|
+
cacheRead: 0.01,
|
|
7670
7688
|
cacheWrite: 0,
|
|
7671
7689
|
},
|
|
7672
7690
|
contextWindow: 262144,
|
|
7673
|
-
maxTokens:
|
|
7691
|
+
maxTokens: 4096,
|
|
7674
7692
|
},
|
|
7675
7693
|
"google/gemma-4-26b-a4b-it:free": {
|
|
7676
7694
|
id: "google/gemma-4-26b-a4b-it:free",
|
|
@@ -7698,13 +7716,13 @@ export const MODELS = {
|
|
|
7698
7716
|
reasoning: true,
|
|
7699
7717
|
input: ["text", "image"],
|
|
7700
7718
|
cost: {
|
|
7701
|
-
input: 0.
|
|
7702
|
-
output: 0.
|
|
7703
|
-
cacheRead: 0,
|
|
7719
|
+
input: 0.13,
|
|
7720
|
+
output: 0.38,
|
|
7721
|
+
cacheRead: 0.019999999499999997,
|
|
7704
7722
|
cacheWrite: 0,
|
|
7705
7723
|
},
|
|
7706
7724
|
contextWindow: 262144,
|
|
7707
|
-
maxTokens:
|
|
7725
|
+
maxTokens: 4096,
|
|
7708
7726
|
},
|
|
7709
7727
|
"google/gemma-4-31b-it:free": {
|
|
7710
7728
|
id: "google/gemma-4-31b-it:free",
|
|
@@ -7723,23 +7741,6 @@ export const MODELS = {
|
|
|
7723
7741
|
contextWindow: 262144,
|
|
7724
7742
|
maxTokens: 32768,
|
|
7725
7743
|
},
|
|
7726
|
-
"inception/mercury": {
|
|
7727
|
-
id: "inception/mercury",
|
|
7728
|
-
name: "Inception: Mercury",
|
|
7729
|
-
api: "openai-completions",
|
|
7730
|
-
provider: "openrouter",
|
|
7731
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7732
|
-
reasoning: false,
|
|
7733
|
-
input: ["text"],
|
|
7734
|
-
cost: {
|
|
7735
|
-
input: 0.25,
|
|
7736
|
-
output: 0.75,
|
|
7737
|
-
cacheRead: 0.024999999999999998,
|
|
7738
|
-
cacheWrite: 0,
|
|
7739
|
-
},
|
|
7740
|
-
contextWindow: 128000,
|
|
7741
|
-
maxTokens: 32000,
|
|
7742
|
-
},
|
|
7743
7744
|
"inception/mercury-2": {
|
|
7744
7745
|
id: "inception/mercury-2",
|
|
7745
7746
|
name: "Inception: Mercury 2",
|
|
@@ -7757,23 +7758,6 @@ export const MODELS = {
|
|
|
7757
7758
|
contextWindow: 128000,
|
|
7758
7759
|
maxTokens: 50000,
|
|
7759
7760
|
},
|
|
7760
|
-
"inception/mercury-coder": {
|
|
7761
|
-
id: "inception/mercury-coder",
|
|
7762
|
-
name: "Inception: Mercury Coder",
|
|
7763
|
-
api: "openai-completions",
|
|
7764
|
-
provider: "openrouter",
|
|
7765
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7766
|
-
reasoning: false,
|
|
7767
|
-
input: ["text"],
|
|
7768
|
-
cost: {
|
|
7769
|
-
input: 0.25,
|
|
7770
|
-
output: 0.75,
|
|
7771
|
-
cacheRead: 0.024999999999999998,
|
|
7772
|
-
cacheWrite: 0,
|
|
7773
|
-
},
|
|
7774
|
-
contextWindow: 128000,
|
|
7775
|
-
maxTokens: 32000,
|
|
7776
|
-
},
|
|
7777
7761
|
"kwaipilot/kat-coder-pro-v2": {
|
|
7778
7762
|
id: "kwaipilot/kat-coder-pro-v2",
|
|
7779
7763
|
name: "Kwaipilot: KAT-Coder-Pro V2",
|
|
@@ -7791,23 +7775,6 @@ export const MODELS = {
|
|
|
7791
7775
|
contextWindow: 256000,
|
|
7792
7776
|
maxTokens: 80000,
|
|
7793
7777
|
},
|
|
7794
|
-
"meituan/longcat-flash-chat": {
|
|
7795
|
-
id: "meituan/longcat-flash-chat",
|
|
7796
|
-
name: "Meituan: LongCat Flash Chat",
|
|
7797
|
-
api: "openai-completions",
|
|
7798
|
-
provider: "openrouter",
|
|
7799
|
-
baseUrl: "https://openrouter.ai/api/v1",
|
|
7800
|
-
reasoning: false,
|
|
7801
|
-
input: ["text"],
|
|
7802
|
-
cost: {
|
|
7803
|
-
input: 0.19999999999999998,
|
|
7804
|
-
output: 0.7999999999999999,
|
|
7805
|
-
cacheRead: 0.19999999999999998,
|
|
7806
|
-
cacheWrite: 0,
|
|
7807
|
-
},
|
|
7808
|
-
contextWindow: 131072,
|
|
7809
|
-
maxTokens: 131072,
|
|
7810
|
-
},
|
|
7811
7778
|
"meta-llama/llama-3-8b-instruct": {
|
|
7812
7779
|
id: "meta-llama/llama-3-8b-instruct",
|
|
7813
7780
|
name: "Meta: Llama 3 8B Instruct",
|
|
@@ -8010,7 +7977,7 @@ export const MODELS = {
|
|
|
8010
7977
|
cacheWrite: 0,
|
|
8011
7978
|
},
|
|
8012
7979
|
contextWindow: 196608,
|
|
8013
|
-
maxTokens:
|
|
7980
|
+
maxTokens: 8192,
|
|
8014
7981
|
},
|
|
8015
7982
|
"minimax/minimax-m2.7": {
|
|
8016
7983
|
id: "minimax/minimax-m2.7",
|
|
@@ -8023,11 +7990,11 @@ export const MODELS = {
|
|
|
8023
7990
|
cost: {
|
|
8024
7991
|
input: 0.3,
|
|
8025
7992
|
output: 1.2,
|
|
8026
|
-
cacheRead: 0.
|
|
7993
|
+
cacheRead: 0.059,
|
|
8027
7994
|
cacheWrite: 0,
|
|
8028
7995
|
},
|
|
8029
|
-
contextWindow:
|
|
8030
|
-
maxTokens:
|
|
7996
|
+
contextWindow: 196608,
|
|
7997
|
+
maxTokens: 4096,
|
|
8031
7998
|
},
|
|
8032
7999
|
"mistralai/codestral-2508": {
|
|
8033
8000
|
id: "mistralai/codestral-2508",
|
|
@@ -8448,11 +8415,11 @@ export const MODELS = {
|
|
|
8448
8415
|
cost: {
|
|
8449
8416
|
input: 0.6,
|
|
8450
8417
|
output: 2.5,
|
|
8451
|
-
cacheRead: 0,
|
|
8418
|
+
cacheRead: 0.15,
|
|
8452
8419
|
cacheWrite: 0,
|
|
8453
8420
|
},
|
|
8454
8421
|
contextWindow: 262144,
|
|
8455
|
-
maxTokens:
|
|
8422
|
+
maxTokens: 262144,
|
|
8456
8423
|
},
|
|
8457
8424
|
"moonshotai/kimi-k2.5": {
|
|
8458
8425
|
id: "moonshotai/kimi-k2.5",
|
|
@@ -9438,7 +9405,7 @@ export const MODELS = {
|
|
|
9438
9405
|
cacheWrite: 0,
|
|
9439
9406
|
},
|
|
9440
9407
|
contextWindow: 131072,
|
|
9441
|
-
maxTokens:
|
|
9408
|
+
maxTokens: 8192,
|
|
9442
9409
|
},
|
|
9443
9410
|
"openai/gpt-oss-safeguard-20b": {
|
|
9444
9411
|
id: "openai/gpt-oss-safeguard-20b",
|
|
@@ -9627,6 +9594,23 @@ export const MODELS = {
|
|
|
9627
9594
|
contextWindow: 2000000,
|
|
9628
9595
|
maxTokens: 4096,
|
|
9629
9596
|
},
|
|
9597
|
+
"openrouter/elephant-alpha": {
|
|
9598
|
+
id: "openrouter/elephant-alpha",
|
|
9599
|
+
name: "Elephant",
|
|
9600
|
+
api: "openai-completions",
|
|
9601
|
+
provider: "openrouter",
|
|
9602
|
+
baseUrl: "https://openrouter.ai/api/v1",
|
|
9603
|
+
reasoning: false,
|
|
9604
|
+
input: ["text"],
|
|
9605
|
+
cost: {
|
|
9606
|
+
input: 0,
|
|
9607
|
+
output: 0,
|
|
9608
|
+
cacheRead: 0,
|
|
9609
|
+
cacheWrite: 0,
|
|
9610
|
+
},
|
|
9611
|
+
contextWindow: 262144,
|
|
9612
|
+
maxTokens: 32768,
|
|
9613
|
+
},
|
|
9630
9614
|
"openrouter/free": {
|
|
9631
9615
|
id: "openrouter/free",
|
|
9632
9616
|
name: "Free Models Router",
|
|
@@ -10010,13 +9994,13 @@ export const MODELS = {
|
|
|
10010
9994
|
reasoning: false,
|
|
10011
9995
|
input: ["text"],
|
|
10012
9996
|
cost: {
|
|
10013
|
-
input: 0.
|
|
10014
|
-
output: 0.
|
|
10015
|
-
cacheRead: 0.
|
|
9997
|
+
input: 0.15,
|
|
9998
|
+
output: 0.7999999999999999,
|
|
9999
|
+
cacheRead: 0.12,
|
|
10016
10000
|
cacheWrite: 0,
|
|
10017
10001
|
},
|
|
10018
10002
|
contextWindow: 262144,
|
|
10019
|
-
maxTokens:
|
|
10003
|
+
maxTokens: 262144,
|
|
10020
10004
|
},
|
|
10021
10005
|
"qwen/qwen3-coder-plus": {
|
|
10022
10006
|
id: "qwen/qwen3-coder-plus",
|