@hyperspaceng/neural-ai 0.67.7 → 0.68.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (47) hide show
  1. package/README.md +3 -1
  2. package/dist/models.generated.d.ts +186 -81
  3. package/dist/models.generated.d.ts.map +1 -1
  4. package/dist/models.generated.js +222 -119
  5. package/dist/models.generated.js.map +1 -1
  6. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  7. package/dist/providers/amazon-bedrock.js +28 -38
  8. package/dist/providers/amazon-bedrock.js.map +1 -1
  9. package/dist/providers/anthropic.d.ts.map +1 -1
  10. package/dist/providers/anthropic.js +1 -2
  11. package/dist/providers/anthropic.js.map +1 -1
  12. package/dist/providers/google-shared.d.ts.map +1 -1
  13. package/dist/providers/google-shared.js +30 -4
  14. package/dist/providers/google-shared.js.map +1 -1
  15. package/dist/providers/mistral.d.ts +3 -0
  16. package/dist/providers/mistral.d.ts.map +1 -1
  17. package/dist/providers/mistral.js +31 -4
  18. package/dist/providers/mistral.js.map +1 -1
  19. package/dist/providers/openai-completions.d.ts +5 -1
  20. package/dist/providers/openai-completions.d.ts.map +1 -1
  21. package/dist/providers/openai-completions.js +153 -61
  22. package/dist/providers/openai-completions.js.map +1 -1
  23. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  24. package/dist/providers/openai-responses-shared.js +2 -5
  25. package/dist/providers/openai-responses-shared.js.map +1 -1
  26. package/dist/providers/simple-options.d.ts.map +1 -1
  27. package/dist/providers/simple-options.js +1 -1
  28. package/dist/providers/simple-options.js.map +1 -1
  29. package/dist/providers/transform-messages.d.ts.map +1 -1
  30. package/dist/providers/transform-messages.js +41 -2
  31. package/dist/providers/transform-messages.js.map +1 -1
  32. package/dist/types.d.ts +4 -0
  33. package/dist/types.d.ts.map +1 -1
  34. package/dist/types.js.map +1 -1
  35. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  36. package/dist/utils/oauth/anthropic.js +1 -1
  37. package/dist/utils/oauth/anthropic.js.map +1 -1
  38. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  39. package/dist/utils/oauth/google-antigravity.js +2 -1
  40. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  41. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  42. package/dist/utils/oauth/google-gemini-cli.js +2 -1
  43. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  44. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  45. package/dist/utils/oauth/openai-codex.js +3 -2
  46. package/dist/utils/oauth/openai-codex.js.map +1 -1
  47. package/package.json +2 -2
package/README.md CHANGED
@@ -851,11 +851,13 @@ interface OpenAICompletionsCompat {
851
851
  supportsReasoningEffort?: boolean; // Whether provider supports `reasoning_effort` (default: true)
852
852
  supportsUsageInStreaming?: boolean; // Whether provider supports `stream_options: { include_usage: true }` (default: true)
853
853
  supportsStrictMode?: boolean; // Whether provider supports `strict` in tool definitions (default: true)
854
+ sendSessionAffinityHeaders?: boolean; // Whether to send `session_id`, `x-client-request-id`, and `x-session-affinity` from `sessionId` when caching is enabled (default: false)
854
855
  maxTokensField?: 'max_completion_tokens' | 'max_tokens'; // Which field name to use (default: max_completion_tokens)
855
856
  requiresToolResultName?: boolean; // Whether tool results require the `name` field (default: false)
856
857
  requiresAssistantAfterToolResult?: boolean; // Whether tool results must be followed by an assistant message (default: false)
857
858
  requiresThinkingAsText?: boolean; // Whether thinking blocks must be converted to text (default: false)
858
- thinkingFormat?: 'openai' | 'zai' | 'qwen'; // Format for reasoning param: 'openai' uses reasoning_effort, 'zai' uses thinking: { type: "enabled" }, 'qwen' uses enable_thinking: boolean (default: openai)
859
+ thinkingFormat?: 'openai' | 'zai' | 'qwen' | 'qwen-chat-template'; // Format for reasoning param: 'openai' uses reasoning_effort, 'zai' uses thinking: { type: "enabled" }, 'qwen' uses enable_thinking: boolean, 'qwen-chat-template' uses chat_template_kwargs.enable_thinking (default: openai)
860
+ cacheControlFormat?: 'anthropic'; // Anthropic-style cache_control on system prompt, last tool, and last user/assistant text content
859
861
  openRouterRouting?: OpenRouterRouting; // OpenRouter routing preferences (default: {})
860
862
  vercelGatewayRouting?: VercelGatewayRouting; // Vercel AI Gateway routing preferences (default: {})
861
863
  }
@@ -4877,6 +4877,26 @@ export declare const MODELS: {
4877
4877
  contextWindow: number;
4878
4878
  maxTokens: number;
4879
4879
  };
4880
+ readonly "moonshotai/Kimi-K2.6": {
4881
+ id: string;
4882
+ name: string;
4883
+ api: "openai-completions";
4884
+ provider: string;
4885
+ baseUrl: string;
4886
+ compat: {
4887
+ supportsDeveloperRole: false;
4888
+ };
4889
+ reasoning: true;
4890
+ input: ("image" | "text")[];
4891
+ cost: {
4892
+ input: number;
4893
+ output: number;
4894
+ cacheRead: number;
4895
+ cacheWrite: number;
4896
+ };
4897
+ contextWindow: number;
4898
+ maxTokens: number;
4899
+ };
4880
4900
  readonly "zai-org/GLM-4.7": {
4881
4901
  id: string;
4882
4902
  name: string;
@@ -4959,6 +4979,23 @@ export declare const MODELS: {
4959
4979
  };
4960
4980
  };
4961
4981
  readonly "kimi-coding": {
4982
+ readonly k2p6: {
4983
+ id: string;
4984
+ name: string;
4985
+ api: "anthropic-messages";
4986
+ provider: string;
4987
+ baseUrl: string;
4988
+ reasoning: true;
4989
+ input: ("image" | "text")[];
4990
+ cost: {
4991
+ input: number;
4992
+ output: number;
4993
+ cacheRead: number;
4994
+ cacheWrite: number;
4995
+ };
4996
+ contextWindow: number;
4997
+ maxTokens: number;
4998
+ };
4962
4999
  readonly "kimi-for-coding": {
4963
5000
  id: string;
4964
5001
  name: string;
@@ -6841,6 +6878,40 @@ export declare const MODELS: {
6841
6878
  contextWindow: number;
6842
6879
  maxTokens: number;
6843
6880
  };
6881
+ readonly "kimi-k2.6": {
6882
+ id: string;
6883
+ name: string;
6884
+ api: "openai-completions";
6885
+ provider: string;
6886
+ baseUrl: string;
6887
+ reasoning: true;
6888
+ input: ("image" | "text")[];
6889
+ cost: {
6890
+ input: number;
6891
+ output: number;
6892
+ cacheRead: number;
6893
+ cacheWrite: number;
6894
+ };
6895
+ contextWindow: number;
6896
+ maxTokens: number;
6897
+ };
6898
+ readonly "ling-2.6-flash-free": {
6899
+ id: string;
6900
+ name: string;
6901
+ api: "openai-completions";
6902
+ provider: string;
6903
+ baseUrl: string;
6904
+ reasoning: false;
6905
+ input: "text"[];
6906
+ cost: {
6907
+ input: number;
6908
+ output: number;
6909
+ cacheRead: number;
6910
+ cacheWrite: number;
6911
+ };
6912
+ contextWindow: number;
6913
+ maxTokens: number;
6914
+ };
6844
6915
  readonly "minimax-m2.5": {
6845
6916
  id: string;
6846
6917
  name: string;
@@ -6875,6 +6946,23 @@ export declare const MODELS: {
6875
6946
  contextWindow: number;
6876
6947
  maxTokens: number;
6877
6948
  };
6949
+ readonly "minimax-m2.7": {
6950
+ id: string;
6951
+ name: string;
6952
+ api: "openai-completions";
6953
+ provider: string;
6954
+ baseUrl: string;
6955
+ reasoning: true;
6956
+ input: "text"[];
6957
+ cost: {
6958
+ input: number;
6959
+ output: number;
6960
+ cacheRead: number;
6961
+ cacheWrite: number;
6962
+ };
6963
+ contextWindow: number;
6964
+ maxTokens: number;
6965
+ };
6878
6966
  readonly "nemotron-3-super-free": {
6879
6967
  id: string;
6880
6968
  name: string;
@@ -6895,7 +6983,7 @@ export declare const MODELS: {
6895
6983
  readonly "qwen3.5-plus": {
6896
6984
  id: string;
6897
6985
  name: string;
6898
- api: "openai-completions";
6986
+ api: "anthropic-messages";
6899
6987
  provider: string;
6900
6988
  baseUrl: string;
6901
6989
  reasoning: true;
@@ -6912,7 +7000,7 @@ export declare const MODELS: {
6912
7000
  readonly "qwen3.6-plus": {
6913
7001
  id: string;
6914
7002
  name: string;
6915
- api: "openai-completions";
7003
+ api: "anthropic-messages";
6916
7004
  provider: string;
6917
7005
  baseUrl: string;
6918
7006
  reasoning: true;
@@ -6979,6 +7067,23 @@ export declare const MODELS: {
6979
7067
  contextWindow: number;
6980
7068
  maxTokens: number;
6981
7069
  };
7070
+ readonly "kimi-k2.6": {
7071
+ id: string;
7072
+ name: string;
7073
+ api: "openai-completions";
7074
+ provider: string;
7075
+ baseUrl: string;
7076
+ reasoning: true;
7077
+ input: ("image" | "text")[];
7078
+ cost: {
7079
+ input: number;
7080
+ output: number;
7081
+ cacheRead: number;
7082
+ cacheWrite: number;
7083
+ };
7084
+ contextWindow: number;
7085
+ maxTokens: number;
7086
+ };
6982
7087
  readonly "mimo-v2-omni": {
6983
7088
  id: string;
6984
7089
  name: string;
@@ -7050,7 +7155,7 @@ export declare const MODELS: {
7050
7155
  readonly "qwen3.5-plus": {
7051
7156
  id: string;
7052
7157
  name: string;
7053
- api: "openai-completions";
7158
+ api: "anthropic-messages";
7054
7159
  provider: string;
7055
7160
  baseUrl: string;
7056
7161
  reasoning: true;
@@ -7067,7 +7172,7 @@ export declare const MODELS: {
7067
7172
  readonly "qwen3.6-plus": {
7068
7173
  id: string;
7069
7174
  name: string;
7070
- api: "openai-completions";
7175
+ api: "anthropic-messages";
7071
7176
  provider: string;
7072
7177
  baseUrl: string;
7073
7178
  reasoning: true;
@@ -8120,7 +8225,7 @@ export declare const MODELS: {
8120
8225
  contextWindow: number;
8121
8226
  maxTokens: number;
8122
8227
  };
8123
- readonly "kwaipilot/kat-coder-pro-v2": {
8228
+ readonly "inclusionai/ling-2.6-flash:free": {
8124
8229
  id: string;
8125
8230
  name: string;
8126
8231
  api: "openai-completions";
@@ -8137,7 +8242,7 @@ export declare const MODELS: {
8137
8242
  contextWindow: number;
8138
8243
  maxTokens: number;
8139
8244
  };
8140
- readonly "meta-llama/llama-3-8b-instruct": {
8245
+ readonly "kwaipilot/kat-coder-pro-v2": {
8141
8246
  id: string;
8142
8247
  name: string;
8143
8248
  api: "openai-completions";
@@ -8154,7 +8259,7 @@ export declare const MODELS: {
8154
8259
  contextWindow: number;
8155
8260
  maxTokens: number;
8156
8261
  };
8157
- readonly "meta-llama/llama-3.1-70b-instruct": {
8262
+ readonly "meta-llama/llama-3-8b-instruct": {
8158
8263
  id: string;
8159
8264
  name: string;
8160
8265
  api: "openai-completions";
@@ -8171,7 +8276,7 @@ export declare const MODELS: {
8171
8276
  contextWindow: number;
8172
8277
  maxTokens: number;
8173
8278
  };
8174
- readonly "meta-llama/llama-3.1-8b-instruct": {
8279
+ readonly "meta-llama/llama-3.1-70b-instruct": {
8175
8280
  id: string;
8176
8281
  name: string;
8177
8282
  api: "openai-completions";
@@ -8188,7 +8293,7 @@ export declare const MODELS: {
8188
8293
  contextWindow: number;
8189
8294
  maxTokens: number;
8190
8295
  };
8191
- readonly "meta-llama/llama-3.3-70b-instruct": {
8296
+ readonly "meta-llama/llama-3.1-8b-instruct": {
8192
8297
  id: string;
8193
8298
  name: string;
8194
8299
  api: "openai-completions";
@@ -8205,7 +8310,7 @@ export declare const MODELS: {
8205
8310
  contextWindow: number;
8206
8311
  maxTokens: number;
8207
8312
  };
8208
- readonly "meta-llama/llama-3.3-70b-instruct:free": {
8313
+ readonly "meta-llama/llama-3.3-70b-instruct": {
8209
8314
  id: string;
8210
8315
  name: string;
8211
8316
  api: "openai-completions";
@@ -8222,14 +8327,14 @@ export declare const MODELS: {
8222
8327
  contextWindow: number;
8223
8328
  maxTokens: number;
8224
8329
  };
8225
- readonly "meta-llama/llama-4-maverick": {
8330
+ readonly "meta-llama/llama-3.3-70b-instruct:free": {
8226
8331
  id: string;
8227
8332
  name: string;
8228
8333
  api: "openai-completions";
8229
8334
  provider: string;
8230
8335
  baseUrl: string;
8231
8336
  reasoning: false;
8232
- input: ("image" | "text")[];
8337
+ input: "text"[];
8233
8338
  cost: {
8234
8339
  input: number;
8235
8340
  output: number;
@@ -8800,6 +8905,23 @@ export declare const MODELS: {
8800
8905
  contextWindow: number;
8801
8906
  maxTokens: number;
8802
8907
  };
8908
+ readonly "moonshotai/kimi-k2.6": {
8909
+ id: string;
8910
+ name: string;
8911
+ api: "openai-completions";
8912
+ provider: string;
8913
+ baseUrl: string;
8914
+ reasoning: true;
8915
+ input: ("image" | "text")[];
8916
+ cost: {
8917
+ input: number;
8918
+ output: number;
8919
+ cacheRead: number;
8920
+ cacheWrite: number;
8921
+ };
8922
+ contextWindow: number;
8923
+ maxTokens: number;
8924
+ };
8803
8925
  readonly "nex-agi/deepseek-v3.1-nex-n1": {
8804
8926
  id: string;
8805
8927
  name: string;
@@ -9276,23 +9398,6 @@ export declare const MODELS: {
9276
9398
  contextWindow: number;
9277
9399
  maxTokens: number;
9278
9400
  };
9279
- readonly "openai/gpt-4o:extended": {
9280
- id: string;
9281
- name: string;
9282
- api: "openai-completions";
9283
- provider: string;
9284
- baseUrl: string;
9285
- reasoning: false;
9286
- input: ("image" | "text")[];
9287
- cost: {
9288
- input: number;
9289
- output: number;
9290
- cacheRead: number;
9291
- cacheWrite: number;
9292
- };
9293
- contextWindow: number;
9294
- maxTokens: number;
9295
- };
9296
9401
  readonly "openai/gpt-5": {
9297
9402
  id: string;
9298
9403
  name: string;
@@ -9327,40 +9432,6 @@ export declare const MODELS: {
9327
9432
  contextWindow: number;
9328
9433
  maxTokens: number;
9329
9434
  };
9330
- readonly "openai/gpt-5-image": {
9331
- id: string;
9332
- name: string;
9333
- api: "openai-completions";
9334
- provider: string;
9335
- baseUrl: string;
9336
- reasoning: true;
9337
- input: ("image" | "text")[];
9338
- cost: {
9339
- input: number;
9340
- output: number;
9341
- cacheRead: number;
9342
- cacheWrite: number;
9343
- };
9344
- contextWindow: number;
9345
- maxTokens: number;
9346
- };
9347
- readonly "openai/gpt-5-image-mini": {
9348
- id: string;
9349
- name: string;
9350
- api: "openai-completions";
9351
- provider: string;
9352
- baseUrl: string;
9353
- reasoning: true;
9354
- input: ("image" | "text")[];
9355
- cost: {
9356
- input: number;
9357
- output: number;
9358
- cacheRead: number;
9359
- cacheWrite: number;
9360
- };
9361
- contextWindow: number;
9362
- maxTokens: number;
9363
- };
9364
9435
  readonly "openai/gpt-5-mini": {
9365
9436
  id: string;
9366
9437
  name: string;
@@ -9956,23 +10027,6 @@ export declare const MODELS: {
9956
10027
  contextWindow: number;
9957
10028
  maxTokens: number;
9958
10029
  };
9959
- readonly "openrouter/elephant-alpha": {
9960
- id: string;
9961
- name: string;
9962
- api: "openai-completions";
9963
- provider: string;
9964
- baseUrl: string;
9965
- reasoning: false;
9966
- input: "text"[];
9967
- cost: {
9968
- input: number;
9969
- output: number;
9970
- cacheRead: number;
9971
- cacheWrite: number;
9972
- };
9973
- contextWindow: number;
9974
- maxTokens: number;
9975
- };
9976
10030
  readonly "openrouter/free": {
9977
10031
  id: string;
9978
10032
  name: string;
@@ -10183,7 +10237,7 @@ export declare const MODELS: {
10183
10237
  api: "openai-completions";
10184
10238
  provider: string;
10185
10239
  baseUrl: string;
10186
- reasoning: true;
10240
+ reasoning: false;
10187
10241
  input: "text"[];
10188
10242
  cost: {
10189
10243
  input: number;
@@ -11333,6 +11387,23 @@ export declare const MODELS: {
11333
11387
  contextWindow: number;
11334
11388
  maxTokens: number;
11335
11389
  };
11390
+ readonly "~anthropic/claude-opus-latest": {
11391
+ id: string;
11392
+ name: string;
11393
+ api: "openai-completions";
11394
+ provider: string;
11395
+ baseUrl: string;
11396
+ reasoning: true;
11397
+ input: ("image" | "text")[];
11398
+ cost: {
11399
+ input: number;
11400
+ output: number;
11401
+ cacheRead: number;
11402
+ cacheWrite: number;
11403
+ };
11404
+ contextWindow: number;
11405
+ maxTokens: number;
11406
+ };
11336
11407
  };
11337
11408
  readonly "vercel-ai-gateway": {
11338
11409
  readonly "alibaba/qwen-3-14b": {
@@ -11403,6 +11474,23 @@ export declare const MODELS: {
11403
11474
  contextWindow: number;
11404
11475
  maxTokens: number;
11405
11476
  };
11477
+ readonly "alibaba/qwen-3.6-max-preview": {
11478
+ id: string;
11479
+ name: string;
11480
+ api: "anthropic-messages";
11481
+ provider: string;
11482
+ baseUrl: string;
11483
+ reasoning: true;
11484
+ input: ("image" | "text")[];
11485
+ cost: {
11486
+ input: number;
11487
+ output: number;
11488
+ cacheRead: number;
11489
+ cacheWrite: number;
11490
+ };
11491
+ contextWindow: number;
11492
+ maxTokens: number;
11493
+ };
11406
11494
  readonly "alibaba/qwen3-235b-a22b-thinking": {
11407
11495
  id: string;
11408
11496
  name: string;
@@ -12746,6 +12834,23 @@ export declare const MODELS: {
12746
12834
  contextWindow: number;
12747
12835
  maxTokens: number;
12748
12836
  };
12837
+ readonly "moonshotai/kimi-k2.6": {
12838
+ id: string;
12839
+ name: string;
12840
+ api: "anthropic-messages";
12841
+ provider: string;
12842
+ baseUrl: string;
12843
+ reasoning: true;
12844
+ input: ("image" | "text")[];
12845
+ cost: {
12846
+ input: number;
12847
+ output: number;
12848
+ cacheRead: number;
12849
+ cacheWrite: number;
12850
+ };
12851
+ contextWindow: number;
12852
+ maxTokens: number;
12853
+ };
12749
12854
  readonly "nvidia/nemotron-nano-12b-v2-vl": {
12750
12855
  id: string;
12751
12856
  name: string;