@hyperspaceng/neural-ai 0.66.2 → 0.67.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -4506,6 +4506,26 @@ export declare const MODELS: {
4506
4506
  contextWindow: number;
4507
4507
  maxTokens: number;
4508
4508
  };
4509
+ readonly "MiniMaxAI/MiniMax-M2.7": {
4510
+ id: string;
4511
+ name: string;
4512
+ api: "openai-completions";
4513
+ provider: string;
4514
+ baseUrl: string;
4515
+ compat: {
4516
+ supportsDeveloperRole: false;
4517
+ };
4518
+ reasoning: true;
4519
+ input: "text"[];
4520
+ cost: {
4521
+ input: number;
4522
+ output: number;
4523
+ cacheRead: number;
4524
+ cacheWrite: number;
4525
+ };
4526
+ contextWindow: number;
4527
+ maxTokens: number;
4528
+ };
4509
4529
  readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
4510
4530
  id: string;
4511
4531
  name: string;
@@ -7907,23 +7927,6 @@ export declare const MODELS: {
7907
7927
  contextWindow: number;
7908
7928
  maxTokens: number;
7909
7929
  };
7910
- readonly "inception/mercury": {
7911
- id: string;
7912
- name: string;
7913
- api: "openai-completions";
7914
- provider: string;
7915
- baseUrl: string;
7916
- reasoning: false;
7917
- input: "text"[];
7918
- cost: {
7919
- input: number;
7920
- output: number;
7921
- cacheRead: number;
7922
- cacheWrite: number;
7923
- };
7924
- contextWindow: number;
7925
- maxTokens: number;
7926
- };
7927
7930
  readonly "inception/mercury-2": {
7928
7931
  id: string;
7929
7932
  name: string;
@@ -7941,23 +7944,6 @@ export declare const MODELS: {
7941
7944
  contextWindow: number;
7942
7945
  maxTokens: number;
7943
7946
  };
7944
- readonly "inception/mercury-coder": {
7945
- id: string;
7946
- name: string;
7947
- api: "openai-completions";
7948
- provider: string;
7949
- baseUrl: string;
7950
- reasoning: false;
7951
- input: "text"[];
7952
- cost: {
7953
- input: number;
7954
- output: number;
7955
- cacheRead: number;
7956
- cacheWrite: number;
7957
- };
7958
- contextWindow: number;
7959
- maxTokens: number;
7960
- };
7961
7947
  readonly "kwaipilot/kat-coder-pro-v2": {
7962
7948
  id: string;
7963
7949
  name: string;
@@ -7975,23 +7961,6 @@ export declare const MODELS: {
7975
7961
  contextWindow: number;
7976
7962
  maxTokens: number;
7977
7963
  };
7978
- readonly "meituan/longcat-flash-chat": {
7979
- id: string;
7980
- name: string;
7981
- api: "openai-completions";
7982
- provider: string;
7983
- baseUrl: string;
7984
- reasoning: false;
7985
- input: "text"[];
7986
- cost: {
7987
- input: number;
7988
- output: number;
7989
- cacheRead: number;
7990
- cacheWrite: number;
7991
- };
7992
- contextWindow: number;
7993
- maxTokens: number;
7994
- };
7995
7964
  readonly "meta-llama/llama-3-8b-instruct": {
7996
7965
  id: string;
7997
7966
  name: string;
@@ -9811,6 +9780,23 @@ export declare const MODELS: {
9811
9780
  contextWindow: number;
9812
9781
  maxTokens: number;
9813
9782
  };
9783
+ readonly "openrouter/elephant-alpha": {
9784
+ id: string;
9785
+ name: string;
9786
+ api: "openai-completions";
9787
+ provider: string;
9788
+ baseUrl: string;
9789
+ reasoning: false;
9790
+ input: "text"[];
9791
+ cost: {
9792
+ input: number;
9793
+ output: number;
9794
+ cacheRead: number;
9795
+ cacheWrite: number;
9796
+ };
9797
+ contextWindow: number;
9798
+ maxTokens: number;
9799
+ };
9814
9800
  readonly "openrouter/free": {
9815
9801
  id: string;
9816
9802
  name: string;