@hyperspaceng/neural-ai 0.66.2 → 0.67.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -4356,6 +4356,24 @@ export const MODELS = {
4356
4356
  contextWindow: 204800,
4357
4357
  maxTokens: 131072,
4358
4358
  },
4359
+ "MiniMaxAI/MiniMax-M2.7": {
4360
+ id: "MiniMaxAI/MiniMax-M2.7",
4361
+ name: "MiniMax-M2.7",
4362
+ api: "openai-completions",
4363
+ provider: "huggingface",
4364
+ baseUrl: "https://router.huggingface.co/v1",
4365
+ compat: { "supportsDeveloperRole": false },
4366
+ reasoning: true,
4367
+ input: ["text"],
4368
+ cost: {
4369
+ input: 0.3,
4370
+ output: 1.2,
4371
+ cacheRead: 0.06,
4372
+ cacheWrite: 0,
4373
+ },
4374
+ contextWindow: 204800,
4375
+ maxTokens: 131072,
4376
+ },
4359
4377
  "Qwen/Qwen3-235B-A22B-Thinking-2507": {
4360
4378
  id: "Qwen/Qwen3-235B-A22B-Thinking-2507",
4361
4379
  name: "Qwen3-235B-A22B-Thinking-2507",
@@ -7105,7 +7123,7 @@ export const MODELS = {
7105
7123
  cost: {
7106
7124
  input: 0.22,
7107
7125
  output: 0.85,
7108
- cacheRead: 0,
7126
+ cacheRead: 0.06,
7109
7127
  cacheWrite: 0,
7110
7128
  },
7111
7129
  contextWindow: 262144,
@@ -7375,13 +7393,13 @@ export const MODELS = {
7375
7393
  reasoning: true,
7376
7394
  input: ["text"],
7377
7395
  cost: {
7378
- input: 0.44999999999999996,
7396
+ input: 0.5,
7379
7397
  output: 2.1500000000000004,
7380
- cacheRead: 0.22499999999999998,
7398
+ cacheRead: 0.35,
7381
7399
  cacheWrite: 0,
7382
7400
  },
7383
7401
  contextWindow: 163840,
7384
- maxTokens: 65536,
7402
+ maxTokens: 4096,
7385
7403
  },
7386
7404
  "deepseek/deepseek-v3.1-terminus": {
7387
7405
  id: "deepseek/deepseek-v3.1-terminus",
@@ -7664,13 +7682,13 @@ export const MODELS = {
7664
7682
  reasoning: true,
7665
7683
  input: ["text", "image"],
7666
7684
  cost: {
7667
- input: 0.12,
7668
- output: 0.39999999999999997,
7669
- cacheRead: 0,
7685
+ input: 0.08,
7686
+ output: 0.35,
7687
+ cacheRead: 0.01,
7670
7688
  cacheWrite: 0,
7671
7689
  },
7672
7690
  contextWindow: 262144,
7673
- maxTokens: 262144,
7691
+ maxTokens: 4096,
7674
7692
  },
7675
7693
  "google/gemma-4-26b-a4b-it:free": {
7676
7694
  id: "google/gemma-4-26b-a4b-it:free",
@@ -7698,13 +7716,13 @@ export const MODELS = {
7698
7716
  reasoning: true,
7699
7717
  input: ["text", "image"],
7700
7718
  cost: {
7701
- input: 0.14,
7702
- output: 0.39999999999999997,
7703
- cacheRead: 0,
7719
+ input: 0.13,
7720
+ output: 0.38,
7721
+ cacheRead: 0.019999999499999997,
7704
7722
  cacheWrite: 0,
7705
7723
  },
7706
7724
  contextWindow: 262144,
7707
- maxTokens: 131072,
7725
+ maxTokens: 4096,
7708
7726
  },
7709
7727
  "google/gemma-4-31b-it:free": {
7710
7728
  id: "google/gemma-4-31b-it:free",
@@ -7723,23 +7741,6 @@ export const MODELS = {
7723
7741
  contextWindow: 262144,
7724
7742
  maxTokens: 32768,
7725
7743
  },
7726
- "inception/mercury": {
7727
- id: "inception/mercury",
7728
- name: "Inception: Mercury",
7729
- api: "openai-completions",
7730
- provider: "openrouter",
7731
- baseUrl: "https://openrouter.ai/api/v1",
7732
- reasoning: false,
7733
- input: ["text"],
7734
- cost: {
7735
- input: 0.25,
7736
- output: 0.75,
7737
- cacheRead: 0.024999999999999998,
7738
- cacheWrite: 0,
7739
- },
7740
- contextWindow: 128000,
7741
- maxTokens: 32000,
7742
- },
7743
7744
  "inception/mercury-2": {
7744
7745
  id: "inception/mercury-2",
7745
7746
  name: "Inception: Mercury 2",
@@ -7757,23 +7758,6 @@ export const MODELS = {
7757
7758
  contextWindow: 128000,
7758
7759
  maxTokens: 50000,
7759
7760
  },
7760
- "inception/mercury-coder": {
7761
- id: "inception/mercury-coder",
7762
- name: "Inception: Mercury Coder",
7763
- api: "openai-completions",
7764
- provider: "openrouter",
7765
- baseUrl: "https://openrouter.ai/api/v1",
7766
- reasoning: false,
7767
- input: ["text"],
7768
- cost: {
7769
- input: 0.25,
7770
- output: 0.75,
7771
- cacheRead: 0.024999999999999998,
7772
- cacheWrite: 0,
7773
- },
7774
- contextWindow: 128000,
7775
- maxTokens: 32000,
7776
- },
7777
7761
  "kwaipilot/kat-coder-pro-v2": {
7778
7762
  id: "kwaipilot/kat-coder-pro-v2",
7779
7763
  name: "Kwaipilot: KAT-Coder-Pro V2",
@@ -7791,23 +7775,6 @@ export const MODELS = {
7791
7775
  contextWindow: 256000,
7792
7776
  maxTokens: 80000,
7793
7777
  },
7794
- "meituan/longcat-flash-chat": {
7795
- id: "meituan/longcat-flash-chat",
7796
- name: "Meituan: LongCat Flash Chat",
7797
- api: "openai-completions",
7798
- provider: "openrouter",
7799
- baseUrl: "https://openrouter.ai/api/v1",
7800
- reasoning: false,
7801
- input: ["text"],
7802
- cost: {
7803
- input: 0.19999999999999998,
7804
- output: 0.7999999999999999,
7805
- cacheRead: 0.19999999999999998,
7806
- cacheWrite: 0,
7807
- },
7808
- contextWindow: 131072,
7809
- maxTokens: 131072,
7810
- },
7811
7778
  "meta-llama/llama-3-8b-instruct": {
7812
7779
  id: "meta-llama/llama-3-8b-instruct",
7813
7780
  name: "Meta: Llama 3 8B Instruct",
@@ -8010,7 +7977,7 @@ export const MODELS = {
8010
7977
  cacheWrite: 0,
8011
7978
  },
8012
7979
  contextWindow: 196608,
8013
- maxTokens: 196608,
7980
+ maxTokens: 8192,
8014
7981
  },
8015
7982
  "minimax/minimax-m2.7": {
8016
7983
  id: "minimax/minimax-m2.7",
@@ -8023,11 +7990,11 @@ export const MODELS = {
8023
7990
  cost: {
8024
7991
  input: 0.3,
8025
7992
  output: 1.2,
8026
- cacheRead: 0.06,
7993
+ cacheRead: 0.059,
8027
7994
  cacheWrite: 0,
8028
7995
  },
8029
- contextWindow: 204800,
8030
- maxTokens: 131072,
7996
+ contextWindow: 196608,
7997
+ maxTokens: 4096,
8031
7998
  },
8032
7999
  "mistralai/codestral-2508": {
8033
8000
  id: "mistralai/codestral-2508",
@@ -8448,11 +8415,11 @@ export const MODELS = {
8448
8415
  cost: {
8449
8416
  input: 0.6,
8450
8417
  output: 2.5,
8451
- cacheRead: 0,
8418
+ cacheRead: 0.15,
8452
8419
  cacheWrite: 0,
8453
8420
  },
8454
8421
  contextWindow: 262144,
8455
- maxTokens: 4096,
8422
+ maxTokens: 262144,
8456
8423
  },
8457
8424
  "moonshotai/kimi-k2.5": {
8458
8425
  id: "moonshotai/kimi-k2.5",
@@ -9438,7 +9405,7 @@ export const MODELS = {
9438
9405
  cacheWrite: 0,
9439
9406
  },
9440
9407
  contextWindow: 131072,
9441
- maxTokens: 131072,
9408
+ maxTokens: 8192,
9442
9409
  },
9443
9410
  "openai/gpt-oss-safeguard-20b": {
9444
9411
  id: "openai/gpt-oss-safeguard-20b",
@@ -9627,6 +9594,23 @@ export const MODELS = {
9627
9594
  contextWindow: 2000000,
9628
9595
  maxTokens: 4096,
9629
9596
  },
9597
+ "openrouter/elephant-alpha": {
9598
+ id: "openrouter/elephant-alpha",
9599
+ name: "Elephant",
9600
+ api: "openai-completions",
9601
+ provider: "openrouter",
9602
+ baseUrl: "https://openrouter.ai/api/v1",
9603
+ reasoning: false,
9604
+ input: ["text"],
9605
+ cost: {
9606
+ input: 0,
9607
+ output: 0,
9608
+ cacheRead: 0,
9609
+ cacheWrite: 0,
9610
+ },
9611
+ contextWindow: 262144,
9612
+ maxTokens: 32768,
9613
+ },
9630
9614
  "openrouter/free": {
9631
9615
  id: "openrouter/free",
9632
9616
  name: "Free Models Router",
@@ -10010,13 +9994,13 @@ export const MODELS = {
10010
9994
  reasoning: false,
10011
9995
  input: ["text"],
10012
9996
  cost: {
10013
- input: 0.12,
10014
- output: 0.75,
10015
- cacheRead: 0.06,
9997
+ input: 0.15,
9998
+ output: 0.7999999999999999,
9999
+ cacheRead: 0.12,
10016
10000
  cacheWrite: 0,
10017
10001
  },
10018
10002
  contextWindow: 262144,
10019
- maxTokens: 65536,
10003
+ maxTokens: 262144,
10020
10004
  },
10021
10005
  "qwen/qwen3-coder-plus": {
10022
10006
  id: "qwen/qwen3-coder-plus",