@hyperspaceng/neural-ai 0.67.69 → 0.68.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (43) hide show
  1. package/README.md +3 -1
  2. package/dist/models.generated.d.ts +194 -55
  3. package/dist/models.generated.d.ts.map +1 -1
  4. package/dist/models.generated.js +222 -85
  5. package/dist/models.generated.js.map +1 -1
  6. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  7. package/dist/providers/amazon-bedrock.js +9 -1
  8. package/dist/providers/amazon-bedrock.js.map +1 -1
  9. package/dist/providers/anthropic.d.ts.map +1 -1
  10. package/dist/providers/anthropic.js +1 -2
  11. package/dist/providers/anthropic.js.map +1 -1
  12. package/dist/providers/google-shared.d.ts.map +1 -1
  13. package/dist/providers/google-shared.js +30 -4
  14. package/dist/providers/google-shared.js.map +1 -1
  15. package/dist/providers/openai-completions.d.ts +5 -1
  16. package/dist/providers/openai-completions.d.ts.map +1 -1
  17. package/dist/providers/openai-completions.js +149 -60
  18. package/dist/providers/openai-completions.js.map +1 -1
  19. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  20. package/dist/providers/openai-responses-shared.js +2 -5
  21. package/dist/providers/openai-responses-shared.js.map +1 -1
  22. package/dist/providers/simple-options.d.ts.map +1 -1
  23. package/dist/providers/simple-options.js +1 -1
  24. package/dist/providers/simple-options.js.map +1 -1
  25. package/dist/providers/transform-messages.d.ts.map +1 -1
  26. package/dist/providers/transform-messages.js +41 -2
  27. package/dist/providers/transform-messages.js.map +1 -1
  28. package/dist/types.d.ts +4 -0
  29. package/dist/types.d.ts.map +1 -1
  30. package/dist/types.js.map +1 -1
  31. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  32. package/dist/utils/oauth/anthropic.js +1 -1
  33. package/dist/utils/oauth/anthropic.js.map +1 -1
  34. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  35. package/dist/utils/oauth/google-antigravity.js +2 -1
  36. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  37. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  38. package/dist/utils/oauth/google-gemini-cli.js +2 -1
  39. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  40. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  41. package/dist/utils/oauth/openai-codex.js +3 -2
  42. package/dist/utils/oauth/openai-codex.js.map +1 -1
  43. package/package.json +1 -1
@@ -3251,7 +3251,7 @@ export const MODELS = {
3251
3251
  cost: {
3252
3252
  input: 0.3,
3253
3253
  output: 2.5,
3254
- cacheRead: 0.075,
3254
+ cacheRead: 0.03,
3255
3255
  cacheWrite: 0,
3256
3256
  },
3257
3257
  contextWindow: 1048576,
@@ -3370,7 +3370,7 @@ export const MODELS = {
3370
3370
  cost: {
3371
3371
  input: 1.25,
3372
3372
  output: 10,
3373
- cacheRead: 0.31,
3373
+ cacheRead: 0.125,
3374
3374
  cacheWrite: 0,
3375
3375
  },
3376
3376
  contextWindow: 1048576,
@@ -4694,6 +4694,24 @@ export const MODELS = {
4694
4694
  contextWindow: 262144,
4695
4695
  maxTokens: 262144,
4696
4696
  },
4697
+ "moonshotai/Kimi-K2.6": {
4698
+ id: "moonshotai/Kimi-K2.6",
4699
+ name: "Kimi-K2.6",
4700
+ api: "openai-completions",
4701
+ provider: "huggingface",
4702
+ baseUrl: "https://router.huggingface.co/v1",
4703
+ compat: { "supportsDeveloperRole": false },
4704
+ reasoning: true,
4705
+ input: ["text", "image"],
4706
+ cost: {
4707
+ input: 0.95,
4708
+ output: 4,
4709
+ cacheRead: 0.16,
4710
+ cacheWrite: 0,
4711
+ },
4712
+ contextWindow: 262144,
4713
+ maxTokens: 262144,
4714
+ },
4697
4715
  "zai-org/GLM-4.7": {
4698
4716
  id: "zai-org/GLM-4.7",
4699
4717
  name: "GLM-4.7",
@@ -4768,6 +4786,23 @@ export const MODELS = {
4768
4786
  },
4769
4787
  },
4770
4788
  "kimi-coding": {
4789
+ "k2p6": {
4790
+ id: "k2p6",
4791
+ name: "Kimi K2.6",
4792
+ api: "anthropic-messages",
4793
+ provider: "kimi-coding",
4794
+ baseUrl: "https://api.kimi.com/coding",
4795
+ reasoning: true,
4796
+ input: ["text", "image"],
4797
+ cost: {
4798
+ input: 0,
4799
+ output: 0,
4800
+ cacheRead: 0,
4801
+ cacheWrite: 0,
4802
+ },
4803
+ contextWindow: 262144,
4804
+ maxTokens: 32768,
4805
+ },
4771
4806
  "kimi-for-coding": {
4772
4807
  id: "kimi-for-coding",
4773
4808
  name: "Kimi For Coding",
@@ -6650,6 +6685,40 @@ export const MODELS = {
6650
6685
  contextWindow: 262144,
6651
6686
  maxTokens: 65536,
6652
6687
  },
6688
+ "kimi-k2.6": {
6689
+ id: "kimi-k2.6",
6690
+ name: "Kimi K2.6",
6691
+ api: "openai-completions",
6692
+ provider: "opencode",
6693
+ baseUrl: "https://opencode.ai/zen/v1",
6694
+ reasoning: true,
6695
+ input: ["text", "image"],
6696
+ cost: {
6697
+ input: 0.95,
6698
+ output: 4,
6699
+ cacheRead: 0.16,
6700
+ cacheWrite: 0,
6701
+ },
6702
+ contextWindow: 262144,
6703
+ maxTokens: 65536,
6704
+ },
6705
+ "ling-2.6-flash-free": {
6706
+ id: "ling-2.6-flash-free",
6707
+ name: "Ling 2.6 Flash Free",
6708
+ api: "openai-completions",
6709
+ provider: "opencode",
6710
+ baseUrl: "https://opencode.ai/zen/v1",
6711
+ reasoning: false,
6712
+ input: ["text"],
6713
+ cost: {
6714
+ input: 0,
6715
+ output: 0,
6716
+ cacheRead: 0,
6717
+ cacheWrite: 0,
6718
+ },
6719
+ contextWindow: 262100,
6720
+ maxTokens: 32800,
6721
+ },
6653
6722
  "minimax-m2.5": {
6654
6723
  id: "minimax-m2.5",
6655
6724
  name: "MiniMax M2.5",
@@ -6684,6 +6753,23 @@ export const MODELS = {
6684
6753
  contextWindow: 204800,
6685
6754
  maxTokens: 131072,
6686
6755
  },
6756
+ "minimax-m2.7": {
6757
+ id: "minimax-m2.7",
6758
+ name: "MiniMax M2.7",
6759
+ api: "openai-completions",
6760
+ provider: "opencode",
6761
+ baseUrl: "https://opencode.ai/zen/v1",
6762
+ reasoning: true,
6763
+ input: ["text"],
6764
+ cost: {
6765
+ input: 0.3,
6766
+ output: 1.2,
6767
+ cacheRead: 0.06,
6768
+ cacheWrite: 0,
6769
+ },
6770
+ contextWindow: 204800,
6771
+ maxTokens: 131072,
6772
+ },
6687
6773
  "nemotron-3-super-free": {
6688
6774
  id: "nemotron-3-super-free",
6689
6775
  name: "Nemotron 3 Super Free",
@@ -6704,9 +6790,9 @@ export const MODELS = {
6704
6790
  "qwen3.5-plus": {
6705
6791
  id: "qwen3.5-plus",
6706
6792
  name: "Qwen3.5 Plus",
6707
- api: "openai-completions",
6793
+ api: "anthropic-messages",
6708
6794
  provider: "opencode",
6709
- baseUrl: "https://opencode.ai/zen/v1",
6795
+ baseUrl: "https://opencode.ai/zen",
6710
6796
  reasoning: true,
6711
6797
  input: ["text", "image"],
6712
6798
  cost: {
@@ -6721,9 +6807,9 @@ export const MODELS = {
6721
6807
  "qwen3.6-plus": {
6722
6808
  id: "qwen3.6-plus",
6723
6809
  name: "Qwen3.6 Plus",
6724
- api: "openai-completions",
6810
+ api: "anthropic-messages",
6725
6811
  provider: "opencode",
6726
- baseUrl: "https://opencode.ai/zen/v1",
6812
+ baseUrl: "https://opencode.ai/zen",
6727
6813
  reasoning: true,
6728
6814
  input: ["text", "image"],
6729
6815
  cost: {
@@ -6788,6 +6874,23 @@ export const MODELS = {
6788
6874
  contextWindow: 262144,
6789
6875
  maxTokens: 65536,
6790
6876
  },
6877
+ "kimi-k2.6": {
6878
+ id: "kimi-k2.6",
6879
+ name: "Kimi K2.6 (3x limits)",
6880
+ api: "openai-completions",
6881
+ provider: "opencode-go",
6882
+ baseUrl: "https://opencode.ai/zen/go/v1",
6883
+ reasoning: true,
6884
+ input: ["text", "image"],
6885
+ cost: {
6886
+ input: 0.32,
6887
+ output: 1.34,
6888
+ cacheRead: 0.054,
6889
+ cacheWrite: 0,
6890
+ },
6891
+ contextWindow: 262144,
6892
+ maxTokens: 65536,
6893
+ },
6791
6894
  "mimo-v2-omni": {
6792
6895
  id: "mimo-v2-omni",
6793
6896
  name: "MiMo V2 Omni",
@@ -6859,9 +6962,9 @@ export const MODELS = {
6859
6962
  "qwen3.5-plus": {
6860
6963
  id: "qwen3.5-plus",
6861
6964
  name: "Qwen3.5 Plus",
6862
- api: "openai-completions",
6965
+ api: "anthropic-messages",
6863
6966
  provider: "opencode-go",
6864
- baseUrl: "https://opencode.ai/zen/go/v1",
6967
+ baseUrl: "https://opencode.ai/zen/go",
6865
6968
  reasoning: true,
6866
6969
  input: ["text", "image"],
6867
6970
  cost: {
@@ -6876,9 +6979,9 @@ export const MODELS = {
6876
6979
  "qwen3.6-plus": {
6877
6980
  id: "qwen3.6-plus",
6878
6981
  name: "Qwen3.6 Plus",
6879
- api: "openai-completions",
6982
+ api: "anthropic-messages",
6880
6983
  provider: "opencode-go",
6881
- baseUrl: "https://opencode.ai/zen/go/v1",
6984
+ baseUrl: "https://opencode.ai/zen/go",
6882
6985
  reasoning: true,
6883
6986
  input: ["text", "image"],
6884
6987
  cost: {
@@ -7598,13 +7701,13 @@ export const MODELS = {
7598
7701
  reasoning: true,
7599
7702
  input: ["text"],
7600
7703
  cost: {
7601
- input: 0.25899999999999995,
7602
- output: 0.42,
7603
- cacheRead: 0.135,
7704
+ input: 0.252,
7705
+ output: 0.378,
7706
+ cacheRead: 0.0252,
7604
7707
  cacheWrite: 0,
7605
7708
  },
7606
- contextWindow: 163840,
7607
- maxTokens: 163840,
7709
+ contextWindow: 131072,
7710
+ maxTokens: 32768,
7608
7711
  },
7609
7712
  "deepseek/deepseek-v3.2-exp": {
7610
7713
  id: "deepseek/deepseek-v3.2-exp",
@@ -7853,13 +7956,13 @@ export const MODELS = {
7853
7956
  reasoning: true,
7854
7957
  input: ["text", "image"],
7855
7958
  cost: {
7856
- input: 0.08,
7959
+ input: 0.07,
7857
7960
  output: 0.35,
7858
- cacheRead: 0.01,
7961
+ cacheRead: 0.04,
7859
7962
  cacheWrite: 0,
7860
7963
  },
7861
7964
  contextWindow: 262144,
7862
- maxTokens: 4096,
7965
+ maxTokens: 262144,
7863
7966
  },
7864
7967
  "google/gemma-4-26b-a4b-it:free": {
7865
7968
  id: "google/gemma-4-26b-a4b-it:free",
@@ -7929,6 +8032,23 @@ export const MODELS = {
7929
8032
  contextWindow: 128000,
7930
8033
  maxTokens: 50000,
7931
8034
  },
8035
+ "inclusionai/ling-2.6-flash:free": {
8036
+ id: "inclusionai/ling-2.6-flash:free",
8037
+ name: "inclusionAI: Ling-2.6-flash (free)",
8038
+ api: "openai-completions",
8039
+ provider: "openrouter",
8040
+ baseUrl: "https://openrouter.ai/api/v1",
8041
+ reasoning: false,
8042
+ input: ["text"],
8043
+ cost: {
8044
+ input: 0,
8045
+ output: 0,
8046
+ cacheRead: 0,
8047
+ cacheWrite: 0,
8048
+ },
8049
+ contextWindow: 262144,
8050
+ maxTokens: 32768,
8051
+ },
7932
8052
  "kwaipilot/kat-coder-pro-v2": {
7933
8053
  id: "kwaipilot/kat-coder-pro-v2",
7934
8054
  name: "Kwaipilot: KAT-Coder-Pro V2",
@@ -8006,13 +8126,13 @@ export const MODELS = {
8006
8126
  reasoning: false,
8007
8127
  input: ["text"],
8008
8128
  cost: {
8009
- input: 0.12,
8010
- output: 0.38,
8129
+ input: 0.09999999999999999,
8130
+ output: 0.32,
8011
8131
  cacheRead: 0,
8012
8132
  cacheWrite: 0,
8013
8133
  },
8014
8134
  contextWindow: 131072,
8015
- maxTokens: 131072,
8135
+ maxTokens: 16384,
8016
8136
  },
8017
8137
  "meta-llama/llama-3.3-70b-instruct:free": {
8018
8138
  id: "meta-llama/llama-3.3-70b-instruct:free",
@@ -8108,9 +8228,9 @@ export const MODELS = {
8108
8228
  reasoning: true,
8109
8229
  input: ["text"],
8110
8230
  cost: {
8111
- input: 0.118,
8112
- output: 0.9900000000000001,
8113
- cacheRead: 0.059,
8231
+ input: 0.15,
8232
+ output: 1.2,
8233
+ cacheRead: 0.075,
8114
8234
  cacheWrite: 0,
8115
8235
  },
8116
8236
  contextWindow: 196608,
@@ -8592,6 +8712,23 @@ export const MODELS = {
8592
8712
  contextWindow: 262144,
8593
8713
  maxTokens: 4096,
8594
8714
  },
8715
+ "moonshotai/kimi-k2.6": {
8716
+ id: "moonshotai/kimi-k2.6",
8717
+ name: "MoonshotAI: Kimi K2.6",
8718
+ api: "openai-completions",
8719
+ provider: "openrouter",
8720
+ baseUrl: "https://openrouter.ai/api/v1",
8721
+ reasoning: true,
8722
+ input: ["text", "image"],
8723
+ cost: {
8724
+ input: 0.7999999999999999,
8725
+ output: 3.5,
8726
+ cacheRead: 0.19999999999999998,
8727
+ cacheWrite: 0,
8728
+ },
8729
+ contextWindow: 262144,
8730
+ maxTokens: 262144,
8731
+ },
8595
8732
  "nex-agi/deepseek-v3.1-nex-n1": {
8596
8733
  id: "nex-agi/deepseek-v3.1-nex-n1",
8597
8734
  name: "Nex AGI: DeepSeek V3.1 Nex N1",
@@ -9102,40 +9239,6 @@ export const MODELS = {
9102
9239
  contextWindow: 400000,
9103
9240
  maxTokens: 128000,
9104
9241
  },
9105
- "openai/gpt-5-image": {
9106
- id: "openai/gpt-5-image",
9107
- name: "OpenAI: GPT-5 Image",
9108
- api: "openai-completions",
9109
- provider: "openrouter",
9110
- baseUrl: "https://openrouter.ai/api/v1",
9111
- reasoning: true,
9112
- input: ["text", "image"],
9113
- cost: {
9114
- input: 10,
9115
- output: 10,
9116
- cacheRead: 1.25,
9117
- cacheWrite: 0,
9118
- },
9119
- contextWindow: 400000,
9120
- maxTokens: 128000,
9121
- },
9122
- "openai/gpt-5-image-mini": {
9123
- id: "openai/gpt-5-image-mini",
9124
- name: "OpenAI: GPT-5 Image Mini",
9125
- api: "openai-completions",
9126
- provider: "openrouter",
9127
- baseUrl: "https://openrouter.ai/api/v1",
9128
- reasoning: true,
9129
- input: ["text", "image"],
9130
- cost: {
9131
- input: 2.5,
9132
- output: 2,
9133
- cacheRead: 0.25,
9134
- cacheWrite: 0,
9135
- },
9136
- contextWindow: 400000,
9137
- maxTokens: 128000,
9138
- },
9139
9242
  "openai/gpt-5-mini": {
9140
9243
  id: "openai/gpt-5-mini",
9141
9244
  name: "OpenAI: GPT-5 Mini",
@@ -9731,23 +9834,6 @@ export const MODELS = {
9731
9834
  contextWindow: 2000000,
9732
9835
  maxTokens: 4096,
9733
9836
  },
9734
- "openrouter/elephant-alpha": {
9735
- id: "openrouter/elephant-alpha",
9736
- name: "Elephant",
9737
- api: "openai-completions",
9738
- provider: "openrouter",
9739
- baseUrl: "https://openrouter.ai/api/v1",
9740
- reasoning: false,
9741
- input: ["text"],
9742
- cost: {
9743
- input: 0,
9744
- output: 0,
9745
- cacheRead: 0,
9746
- cacheWrite: 0,
9747
- },
9748
- contextWindow: 262144,
9749
- maxTokens: 32768,
9750
- },
9751
9837
  "openrouter/free": {
9752
9838
  id: "openrouter/free",
9753
9839
  name: "Free Models Router",
@@ -10133,7 +10219,7 @@ export const MODELS = {
10133
10219
  cost: {
10134
10220
  input: 0.15,
10135
10221
  output: 0.7999999999999999,
10136
- cacheRead: 0.12,
10222
+ cacheRead: 0.11,
10137
10223
  cacheWrite: 0,
10138
10224
  },
10139
10225
  contextWindow: 262144,
@@ -11015,13 +11101,13 @@ export const MODELS = {
11015
11101
  reasoning: true,
11016
11102
  input: ["text"],
11017
11103
  cost: {
11018
- input: 0.39,
11019
- output: 1.75,
11020
- cacheRead: 0.195,
11104
+ input: 0.38,
11105
+ output: 1.74,
11106
+ cacheRead: 0,
11021
11107
  cacheWrite: 0,
11022
11108
  },
11023
11109
  contextWindow: 202752,
11024
- maxTokens: 65535,
11110
+ maxTokens: 4096,
11025
11111
  },
11026
11112
  "z-ai/glm-4.7-flash": {
11027
11113
  id: "z-ai/glm-4.7-flash",
@@ -11054,8 +11140,8 @@ export const MODELS = {
11054
11140
  cacheRead: 0.119,
11055
11141
  cacheWrite: 0,
11056
11142
  },
11057
- contextWindow: 80000,
11058
- maxTokens: 131072,
11143
+ contextWindow: 202752,
11144
+ maxTokens: 4096,
11059
11145
  },
11060
11146
  "z-ai/glm-5-turbo": {
11061
11147
  id: "z-ai/glm-5-turbo",
@@ -11083,9 +11169,9 @@ export const MODELS = {
11083
11169
  reasoning: true,
11084
11170
  input: ["text"],
11085
11171
  cost: {
11086
- input: 0.95,
11087
- output: 3.15,
11088
- cacheRead: 0.475,
11172
+ input: 1.0499999999999998,
11173
+ output: 3.5,
11174
+ cacheRead: 0.5249999999999999,
11089
11175
  cacheWrite: 0,
11090
11176
  },
11091
11177
  contextWindow: 202752,
@@ -11108,6 +11194,23 @@ export const MODELS = {
11108
11194
  contextWindow: 202752,
11109
11195
  maxTokens: 131072,
11110
11196
  },
11197
+ "~anthropic/claude-opus-latest": {
11198
+ id: "~anthropic/claude-opus-latest",
11199
+ name: "Anthropic: Claude Opus Latest",
11200
+ api: "openai-completions",
11201
+ provider: "openrouter",
11202
+ baseUrl: "https://openrouter.ai/api/v1",
11203
+ reasoning: true,
11204
+ input: ["text", "image"],
11205
+ cost: {
11206
+ input: 5,
11207
+ output: 25,
11208
+ cacheRead: 0.5,
11209
+ cacheWrite: 6.25,
11210
+ },
11211
+ contextWindow: 1000000,
11212
+ maxTokens: 128000,
11213
+ },
11111
11214
  },
11112
11215
  "vercel-ai-gateway": {
11113
11216
  "alibaba/qwen-3-14b": {
@@ -11178,6 +11281,23 @@ export const MODELS = {
11178
11281
  contextWindow: 128000,
11179
11282
  maxTokens: 8192,
11180
11283
  },
11284
+ "alibaba/qwen-3.6-max-preview": {
11285
+ id: "alibaba/qwen-3.6-max-preview",
11286
+ name: "Qwen 3.6 Max Preview",
11287
+ api: "anthropic-messages",
11288
+ provider: "vercel-ai-gateway",
11289
+ baseUrl: "https://ai-gateway.vercel.sh",
11290
+ reasoning: true,
11291
+ input: ["text", "image"],
11292
+ cost: {
11293
+ input: 1.3,
11294
+ output: 7.8,
11295
+ cacheRead: 0.26,
11296
+ cacheWrite: 1.625,
11297
+ },
11298
+ contextWindow: 240000,
11299
+ maxTokens: 64000,
11300
+ },
11181
11301
  "alibaba/qwen3-235b-a22b-thinking": {
11182
11302
  id: "alibaba/qwen3-235b-a22b-thinking",
11183
11303
  name: "Qwen3 235B A22B Thinking 2507",
@@ -12521,6 +12641,23 @@ export const MODELS = {
12521
12641
  contextWindow: 262114,
12522
12642
  maxTokens: 262114,
12523
12643
  },
12644
+ "moonshotai/kimi-k2.6": {
12645
+ id: "moonshotai/kimi-k2.6",
12646
+ name: "Kimi K2.6",
12647
+ api: "anthropic-messages",
12648
+ provider: "vercel-ai-gateway",
12649
+ baseUrl: "https://ai-gateway.vercel.sh",
12650
+ reasoning: true,
12651
+ input: ["text", "image"],
12652
+ cost: {
12653
+ input: 0.95,
12654
+ output: 4,
12655
+ cacheRead: 0.16,
12656
+ cacheWrite: 0,
12657
+ },
12658
+ contextWindow: 262000,
12659
+ maxTokens: 262000,
12660
+ },
12524
12661
  "nvidia/nemotron-nano-12b-v2-vl": {
12525
12662
  id: "nvidia/nemotron-nano-12b-v2-vl",
12526
12663
  name: "Nvidia Nemotron Nano 12B V2 VL",