@hyperspaceng/neural-ai 0.67.7 → 0.68.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +3 -1
- package/dist/models.generated.d.ts +186 -81
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +222 -119
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/amazon-bedrock.d.ts.map +1 -1
- package/dist/providers/amazon-bedrock.js +28 -38
- package/dist/providers/amazon-bedrock.js.map +1 -1
- package/dist/providers/anthropic.d.ts.map +1 -1
- package/dist/providers/anthropic.js +1 -2
- package/dist/providers/anthropic.js.map +1 -1
- package/dist/providers/google-shared.d.ts.map +1 -1
- package/dist/providers/google-shared.js +30 -4
- package/dist/providers/google-shared.js.map +1 -1
- package/dist/providers/mistral.d.ts +3 -0
- package/dist/providers/mistral.d.ts.map +1 -1
- package/dist/providers/mistral.js +31 -4
- package/dist/providers/mistral.js.map +1 -1
- package/dist/providers/openai-completions.d.ts +5 -1
- package/dist/providers/openai-completions.d.ts.map +1 -1
- package/dist/providers/openai-completions.js +153 -61
- package/dist/providers/openai-completions.js.map +1 -1
- package/dist/providers/openai-responses-shared.d.ts.map +1 -1
- package/dist/providers/openai-responses-shared.js +2 -5
- package/dist/providers/openai-responses-shared.js.map +1 -1
- package/dist/providers/simple-options.d.ts.map +1 -1
- package/dist/providers/simple-options.js +1 -1
- package/dist/providers/simple-options.js.map +1 -1
- package/dist/providers/transform-messages.d.ts.map +1 -1
- package/dist/providers/transform-messages.js +41 -2
- package/dist/providers/transform-messages.js.map +1 -1
- package/dist/types.d.ts +4 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/dist/utils/oauth/anthropic.d.ts.map +1 -1
- package/dist/utils/oauth/anthropic.js +1 -1
- package/dist/utils/oauth/anthropic.js.map +1 -1
- package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
- package/dist/utils/oauth/google-antigravity.js +2 -1
- package/dist/utils/oauth/google-antigravity.js.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
- package/dist/utils/oauth/google-gemini-cli.js +2 -1
- package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
- package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
- package/dist/utils/oauth/openai-codex.js +3 -2
- package/dist/utils/oauth/openai-codex.js.map +1 -1
- package/package.json +2 -2
package/README.md
CHANGED
|
@@ -851,11 +851,13 @@ interface OpenAICompletionsCompat {
|
|
|
851
851
|
supportsReasoningEffort?: boolean; // Whether provider supports `reasoning_effort` (default: true)
|
|
852
852
|
supportsUsageInStreaming?: boolean; // Whether provider supports `stream_options: { include_usage: true }` (default: true)
|
|
853
853
|
supportsStrictMode?: boolean; // Whether provider supports `strict` in tool definitions (default: true)
|
|
854
|
+
sendSessionAffinityHeaders?: boolean; // Whether to send `session_id`, `x-client-request-id`, and `x-session-affinity` from `sessionId` when caching is enabled (default: false)
|
|
854
855
|
maxTokensField?: 'max_completion_tokens' | 'max_tokens'; // Which field name to use (default: max_completion_tokens)
|
|
855
856
|
requiresToolResultName?: boolean; // Whether tool results require the `name` field (default: false)
|
|
856
857
|
requiresAssistantAfterToolResult?: boolean; // Whether tool results must be followed by an assistant message (default: false)
|
|
857
858
|
requiresThinkingAsText?: boolean; // Whether thinking blocks must be converted to text (default: false)
|
|
858
|
-
thinkingFormat?: 'openai' | 'zai' | 'qwen'; // Format for reasoning param: 'openai' uses reasoning_effort, 'zai' uses thinking: { type: "enabled" }, 'qwen' uses enable_thinking: boolean (default: openai)
|
|
859
|
+
thinkingFormat?: 'openai' | 'zai' | 'qwen' | 'qwen-chat-template'; // Format for reasoning param: 'openai' uses reasoning_effort, 'zai' uses thinking: { type: "enabled" }, 'qwen' uses enable_thinking: boolean, 'qwen-chat-template' uses chat_template_kwargs.enable_thinking (default: openai)
|
|
860
|
+
cacheControlFormat?: 'anthropic'; // Anthropic-style cache_control on system prompt, last tool, and last user/assistant text content
|
|
859
861
|
openRouterRouting?: OpenRouterRouting; // OpenRouter routing preferences (default: {})
|
|
860
862
|
vercelGatewayRouting?: VercelGatewayRouting; // Vercel AI Gateway routing preferences (default: {})
|
|
861
863
|
}
|
|
@@ -4877,6 +4877,26 @@ export declare const MODELS: {
|
|
|
4877
4877
|
contextWindow: number;
|
|
4878
4878
|
maxTokens: number;
|
|
4879
4879
|
};
|
|
4880
|
+
readonly "moonshotai/Kimi-K2.6": {
|
|
4881
|
+
id: string;
|
|
4882
|
+
name: string;
|
|
4883
|
+
api: "openai-completions";
|
|
4884
|
+
provider: string;
|
|
4885
|
+
baseUrl: string;
|
|
4886
|
+
compat: {
|
|
4887
|
+
supportsDeveloperRole: false;
|
|
4888
|
+
};
|
|
4889
|
+
reasoning: true;
|
|
4890
|
+
input: ("image" | "text")[];
|
|
4891
|
+
cost: {
|
|
4892
|
+
input: number;
|
|
4893
|
+
output: number;
|
|
4894
|
+
cacheRead: number;
|
|
4895
|
+
cacheWrite: number;
|
|
4896
|
+
};
|
|
4897
|
+
contextWindow: number;
|
|
4898
|
+
maxTokens: number;
|
|
4899
|
+
};
|
|
4880
4900
|
readonly "zai-org/GLM-4.7": {
|
|
4881
4901
|
id: string;
|
|
4882
4902
|
name: string;
|
|
@@ -4959,6 +4979,23 @@ export declare const MODELS: {
|
|
|
4959
4979
|
};
|
|
4960
4980
|
};
|
|
4961
4981
|
readonly "kimi-coding": {
|
|
4982
|
+
readonly k2p6: {
|
|
4983
|
+
id: string;
|
|
4984
|
+
name: string;
|
|
4985
|
+
api: "anthropic-messages";
|
|
4986
|
+
provider: string;
|
|
4987
|
+
baseUrl: string;
|
|
4988
|
+
reasoning: true;
|
|
4989
|
+
input: ("image" | "text")[];
|
|
4990
|
+
cost: {
|
|
4991
|
+
input: number;
|
|
4992
|
+
output: number;
|
|
4993
|
+
cacheRead: number;
|
|
4994
|
+
cacheWrite: number;
|
|
4995
|
+
};
|
|
4996
|
+
contextWindow: number;
|
|
4997
|
+
maxTokens: number;
|
|
4998
|
+
};
|
|
4962
4999
|
readonly "kimi-for-coding": {
|
|
4963
5000
|
id: string;
|
|
4964
5001
|
name: string;
|
|
@@ -6841,6 +6878,40 @@ export declare const MODELS: {
|
|
|
6841
6878
|
contextWindow: number;
|
|
6842
6879
|
maxTokens: number;
|
|
6843
6880
|
};
|
|
6881
|
+
readonly "kimi-k2.6": {
|
|
6882
|
+
id: string;
|
|
6883
|
+
name: string;
|
|
6884
|
+
api: "openai-completions";
|
|
6885
|
+
provider: string;
|
|
6886
|
+
baseUrl: string;
|
|
6887
|
+
reasoning: true;
|
|
6888
|
+
input: ("image" | "text")[];
|
|
6889
|
+
cost: {
|
|
6890
|
+
input: number;
|
|
6891
|
+
output: number;
|
|
6892
|
+
cacheRead: number;
|
|
6893
|
+
cacheWrite: number;
|
|
6894
|
+
};
|
|
6895
|
+
contextWindow: number;
|
|
6896
|
+
maxTokens: number;
|
|
6897
|
+
};
|
|
6898
|
+
readonly "ling-2.6-flash-free": {
|
|
6899
|
+
id: string;
|
|
6900
|
+
name: string;
|
|
6901
|
+
api: "openai-completions";
|
|
6902
|
+
provider: string;
|
|
6903
|
+
baseUrl: string;
|
|
6904
|
+
reasoning: false;
|
|
6905
|
+
input: "text"[];
|
|
6906
|
+
cost: {
|
|
6907
|
+
input: number;
|
|
6908
|
+
output: number;
|
|
6909
|
+
cacheRead: number;
|
|
6910
|
+
cacheWrite: number;
|
|
6911
|
+
};
|
|
6912
|
+
contextWindow: number;
|
|
6913
|
+
maxTokens: number;
|
|
6914
|
+
};
|
|
6844
6915
|
readonly "minimax-m2.5": {
|
|
6845
6916
|
id: string;
|
|
6846
6917
|
name: string;
|
|
@@ -6875,6 +6946,23 @@ export declare const MODELS: {
|
|
|
6875
6946
|
contextWindow: number;
|
|
6876
6947
|
maxTokens: number;
|
|
6877
6948
|
};
|
|
6949
|
+
readonly "minimax-m2.7": {
|
|
6950
|
+
id: string;
|
|
6951
|
+
name: string;
|
|
6952
|
+
api: "openai-completions";
|
|
6953
|
+
provider: string;
|
|
6954
|
+
baseUrl: string;
|
|
6955
|
+
reasoning: true;
|
|
6956
|
+
input: "text"[];
|
|
6957
|
+
cost: {
|
|
6958
|
+
input: number;
|
|
6959
|
+
output: number;
|
|
6960
|
+
cacheRead: number;
|
|
6961
|
+
cacheWrite: number;
|
|
6962
|
+
};
|
|
6963
|
+
contextWindow: number;
|
|
6964
|
+
maxTokens: number;
|
|
6965
|
+
};
|
|
6878
6966
|
readonly "nemotron-3-super-free": {
|
|
6879
6967
|
id: string;
|
|
6880
6968
|
name: string;
|
|
@@ -6895,7 +6983,7 @@ export declare const MODELS: {
|
|
|
6895
6983
|
readonly "qwen3.5-plus": {
|
|
6896
6984
|
id: string;
|
|
6897
6985
|
name: string;
|
|
6898
|
-
api: "
|
|
6986
|
+
api: "anthropic-messages";
|
|
6899
6987
|
provider: string;
|
|
6900
6988
|
baseUrl: string;
|
|
6901
6989
|
reasoning: true;
|
|
@@ -6912,7 +7000,7 @@ export declare const MODELS: {
|
|
|
6912
7000
|
readonly "qwen3.6-plus": {
|
|
6913
7001
|
id: string;
|
|
6914
7002
|
name: string;
|
|
6915
|
-
api: "
|
|
7003
|
+
api: "anthropic-messages";
|
|
6916
7004
|
provider: string;
|
|
6917
7005
|
baseUrl: string;
|
|
6918
7006
|
reasoning: true;
|
|
@@ -6979,6 +7067,23 @@ export declare const MODELS: {
|
|
|
6979
7067
|
contextWindow: number;
|
|
6980
7068
|
maxTokens: number;
|
|
6981
7069
|
};
|
|
7070
|
+
readonly "kimi-k2.6": {
|
|
7071
|
+
id: string;
|
|
7072
|
+
name: string;
|
|
7073
|
+
api: "openai-completions";
|
|
7074
|
+
provider: string;
|
|
7075
|
+
baseUrl: string;
|
|
7076
|
+
reasoning: true;
|
|
7077
|
+
input: ("image" | "text")[];
|
|
7078
|
+
cost: {
|
|
7079
|
+
input: number;
|
|
7080
|
+
output: number;
|
|
7081
|
+
cacheRead: number;
|
|
7082
|
+
cacheWrite: number;
|
|
7083
|
+
};
|
|
7084
|
+
contextWindow: number;
|
|
7085
|
+
maxTokens: number;
|
|
7086
|
+
};
|
|
6982
7087
|
readonly "mimo-v2-omni": {
|
|
6983
7088
|
id: string;
|
|
6984
7089
|
name: string;
|
|
@@ -7050,7 +7155,7 @@ export declare const MODELS: {
|
|
|
7050
7155
|
readonly "qwen3.5-plus": {
|
|
7051
7156
|
id: string;
|
|
7052
7157
|
name: string;
|
|
7053
|
-
api: "
|
|
7158
|
+
api: "anthropic-messages";
|
|
7054
7159
|
provider: string;
|
|
7055
7160
|
baseUrl: string;
|
|
7056
7161
|
reasoning: true;
|
|
@@ -7067,7 +7172,7 @@ export declare const MODELS: {
|
|
|
7067
7172
|
readonly "qwen3.6-plus": {
|
|
7068
7173
|
id: string;
|
|
7069
7174
|
name: string;
|
|
7070
|
-
api: "
|
|
7175
|
+
api: "anthropic-messages";
|
|
7071
7176
|
provider: string;
|
|
7072
7177
|
baseUrl: string;
|
|
7073
7178
|
reasoning: true;
|
|
@@ -8120,7 +8225,7 @@ export declare const MODELS: {
|
|
|
8120
8225
|
contextWindow: number;
|
|
8121
8226
|
maxTokens: number;
|
|
8122
8227
|
};
|
|
8123
|
-
readonly "
|
|
8228
|
+
readonly "inclusionai/ling-2.6-flash:free": {
|
|
8124
8229
|
id: string;
|
|
8125
8230
|
name: string;
|
|
8126
8231
|
api: "openai-completions";
|
|
@@ -8137,7 +8242,7 @@ export declare const MODELS: {
|
|
|
8137
8242
|
contextWindow: number;
|
|
8138
8243
|
maxTokens: number;
|
|
8139
8244
|
};
|
|
8140
|
-
readonly "
|
|
8245
|
+
readonly "kwaipilot/kat-coder-pro-v2": {
|
|
8141
8246
|
id: string;
|
|
8142
8247
|
name: string;
|
|
8143
8248
|
api: "openai-completions";
|
|
@@ -8154,7 +8259,7 @@ export declare const MODELS: {
|
|
|
8154
8259
|
contextWindow: number;
|
|
8155
8260
|
maxTokens: number;
|
|
8156
8261
|
};
|
|
8157
|
-
readonly "meta-llama/llama-3
|
|
8262
|
+
readonly "meta-llama/llama-3-8b-instruct": {
|
|
8158
8263
|
id: string;
|
|
8159
8264
|
name: string;
|
|
8160
8265
|
api: "openai-completions";
|
|
@@ -8171,7 +8276,7 @@ export declare const MODELS: {
|
|
|
8171
8276
|
contextWindow: number;
|
|
8172
8277
|
maxTokens: number;
|
|
8173
8278
|
};
|
|
8174
|
-
readonly "meta-llama/llama-3.1-
|
|
8279
|
+
readonly "meta-llama/llama-3.1-70b-instruct": {
|
|
8175
8280
|
id: string;
|
|
8176
8281
|
name: string;
|
|
8177
8282
|
api: "openai-completions";
|
|
@@ -8188,7 +8293,7 @@ export declare const MODELS: {
|
|
|
8188
8293
|
contextWindow: number;
|
|
8189
8294
|
maxTokens: number;
|
|
8190
8295
|
};
|
|
8191
|
-
readonly "meta-llama/llama-3.
|
|
8296
|
+
readonly "meta-llama/llama-3.1-8b-instruct": {
|
|
8192
8297
|
id: string;
|
|
8193
8298
|
name: string;
|
|
8194
8299
|
api: "openai-completions";
|
|
@@ -8205,7 +8310,7 @@ export declare const MODELS: {
|
|
|
8205
8310
|
contextWindow: number;
|
|
8206
8311
|
maxTokens: number;
|
|
8207
8312
|
};
|
|
8208
|
-
readonly "meta-llama/llama-3.3-70b-instruct
|
|
8313
|
+
readonly "meta-llama/llama-3.3-70b-instruct": {
|
|
8209
8314
|
id: string;
|
|
8210
8315
|
name: string;
|
|
8211
8316
|
api: "openai-completions";
|
|
@@ -8222,14 +8327,14 @@ export declare const MODELS: {
|
|
|
8222
8327
|
contextWindow: number;
|
|
8223
8328
|
maxTokens: number;
|
|
8224
8329
|
};
|
|
8225
|
-
readonly "meta-llama/llama-
|
|
8330
|
+
readonly "meta-llama/llama-3.3-70b-instruct:free": {
|
|
8226
8331
|
id: string;
|
|
8227
8332
|
name: string;
|
|
8228
8333
|
api: "openai-completions";
|
|
8229
8334
|
provider: string;
|
|
8230
8335
|
baseUrl: string;
|
|
8231
8336
|
reasoning: false;
|
|
8232
|
-
input:
|
|
8337
|
+
input: "text"[];
|
|
8233
8338
|
cost: {
|
|
8234
8339
|
input: number;
|
|
8235
8340
|
output: number;
|
|
@@ -8800,6 +8905,23 @@ export declare const MODELS: {
|
|
|
8800
8905
|
contextWindow: number;
|
|
8801
8906
|
maxTokens: number;
|
|
8802
8907
|
};
|
|
8908
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
8909
|
+
id: string;
|
|
8910
|
+
name: string;
|
|
8911
|
+
api: "openai-completions";
|
|
8912
|
+
provider: string;
|
|
8913
|
+
baseUrl: string;
|
|
8914
|
+
reasoning: true;
|
|
8915
|
+
input: ("image" | "text")[];
|
|
8916
|
+
cost: {
|
|
8917
|
+
input: number;
|
|
8918
|
+
output: number;
|
|
8919
|
+
cacheRead: number;
|
|
8920
|
+
cacheWrite: number;
|
|
8921
|
+
};
|
|
8922
|
+
contextWindow: number;
|
|
8923
|
+
maxTokens: number;
|
|
8924
|
+
};
|
|
8803
8925
|
readonly "nex-agi/deepseek-v3.1-nex-n1": {
|
|
8804
8926
|
id: string;
|
|
8805
8927
|
name: string;
|
|
@@ -9276,23 +9398,6 @@ export declare const MODELS: {
|
|
|
9276
9398
|
contextWindow: number;
|
|
9277
9399
|
maxTokens: number;
|
|
9278
9400
|
};
|
|
9279
|
-
readonly "openai/gpt-4o:extended": {
|
|
9280
|
-
id: string;
|
|
9281
|
-
name: string;
|
|
9282
|
-
api: "openai-completions";
|
|
9283
|
-
provider: string;
|
|
9284
|
-
baseUrl: string;
|
|
9285
|
-
reasoning: false;
|
|
9286
|
-
input: ("image" | "text")[];
|
|
9287
|
-
cost: {
|
|
9288
|
-
input: number;
|
|
9289
|
-
output: number;
|
|
9290
|
-
cacheRead: number;
|
|
9291
|
-
cacheWrite: number;
|
|
9292
|
-
};
|
|
9293
|
-
contextWindow: number;
|
|
9294
|
-
maxTokens: number;
|
|
9295
|
-
};
|
|
9296
9401
|
readonly "openai/gpt-5": {
|
|
9297
9402
|
id: string;
|
|
9298
9403
|
name: string;
|
|
@@ -9327,40 +9432,6 @@ export declare const MODELS: {
|
|
|
9327
9432
|
contextWindow: number;
|
|
9328
9433
|
maxTokens: number;
|
|
9329
9434
|
};
|
|
9330
|
-
readonly "openai/gpt-5-image": {
|
|
9331
|
-
id: string;
|
|
9332
|
-
name: string;
|
|
9333
|
-
api: "openai-completions";
|
|
9334
|
-
provider: string;
|
|
9335
|
-
baseUrl: string;
|
|
9336
|
-
reasoning: true;
|
|
9337
|
-
input: ("image" | "text")[];
|
|
9338
|
-
cost: {
|
|
9339
|
-
input: number;
|
|
9340
|
-
output: number;
|
|
9341
|
-
cacheRead: number;
|
|
9342
|
-
cacheWrite: number;
|
|
9343
|
-
};
|
|
9344
|
-
contextWindow: number;
|
|
9345
|
-
maxTokens: number;
|
|
9346
|
-
};
|
|
9347
|
-
readonly "openai/gpt-5-image-mini": {
|
|
9348
|
-
id: string;
|
|
9349
|
-
name: string;
|
|
9350
|
-
api: "openai-completions";
|
|
9351
|
-
provider: string;
|
|
9352
|
-
baseUrl: string;
|
|
9353
|
-
reasoning: true;
|
|
9354
|
-
input: ("image" | "text")[];
|
|
9355
|
-
cost: {
|
|
9356
|
-
input: number;
|
|
9357
|
-
output: number;
|
|
9358
|
-
cacheRead: number;
|
|
9359
|
-
cacheWrite: number;
|
|
9360
|
-
};
|
|
9361
|
-
contextWindow: number;
|
|
9362
|
-
maxTokens: number;
|
|
9363
|
-
};
|
|
9364
9435
|
readonly "openai/gpt-5-mini": {
|
|
9365
9436
|
id: string;
|
|
9366
9437
|
name: string;
|
|
@@ -9956,23 +10027,6 @@ export declare const MODELS: {
|
|
|
9956
10027
|
contextWindow: number;
|
|
9957
10028
|
maxTokens: number;
|
|
9958
10029
|
};
|
|
9959
|
-
readonly "openrouter/elephant-alpha": {
|
|
9960
|
-
id: string;
|
|
9961
|
-
name: string;
|
|
9962
|
-
api: "openai-completions";
|
|
9963
|
-
provider: string;
|
|
9964
|
-
baseUrl: string;
|
|
9965
|
-
reasoning: false;
|
|
9966
|
-
input: "text"[];
|
|
9967
|
-
cost: {
|
|
9968
|
-
input: number;
|
|
9969
|
-
output: number;
|
|
9970
|
-
cacheRead: number;
|
|
9971
|
-
cacheWrite: number;
|
|
9972
|
-
};
|
|
9973
|
-
contextWindow: number;
|
|
9974
|
-
maxTokens: number;
|
|
9975
|
-
};
|
|
9976
10030
|
readonly "openrouter/free": {
|
|
9977
10031
|
id: string;
|
|
9978
10032
|
name: string;
|
|
@@ -10183,7 +10237,7 @@ export declare const MODELS: {
|
|
|
10183
10237
|
api: "openai-completions";
|
|
10184
10238
|
provider: string;
|
|
10185
10239
|
baseUrl: string;
|
|
10186
|
-
reasoning:
|
|
10240
|
+
reasoning: false;
|
|
10187
10241
|
input: "text"[];
|
|
10188
10242
|
cost: {
|
|
10189
10243
|
input: number;
|
|
@@ -11333,6 +11387,23 @@ export declare const MODELS: {
|
|
|
11333
11387
|
contextWindow: number;
|
|
11334
11388
|
maxTokens: number;
|
|
11335
11389
|
};
|
|
11390
|
+
readonly "~anthropic/claude-opus-latest": {
|
|
11391
|
+
id: string;
|
|
11392
|
+
name: string;
|
|
11393
|
+
api: "openai-completions";
|
|
11394
|
+
provider: string;
|
|
11395
|
+
baseUrl: string;
|
|
11396
|
+
reasoning: true;
|
|
11397
|
+
input: ("image" | "text")[];
|
|
11398
|
+
cost: {
|
|
11399
|
+
input: number;
|
|
11400
|
+
output: number;
|
|
11401
|
+
cacheRead: number;
|
|
11402
|
+
cacheWrite: number;
|
|
11403
|
+
};
|
|
11404
|
+
contextWindow: number;
|
|
11405
|
+
maxTokens: number;
|
|
11406
|
+
};
|
|
11336
11407
|
};
|
|
11337
11408
|
readonly "vercel-ai-gateway": {
|
|
11338
11409
|
readonly "alibaba/qwen-3-14b": {
|
|
@@ -11403,6 +11474,23 @@ export declare const MODELS: {
|
|
|
11403
11474
|
contextWindow: number;
|
|
11404
11475
|
maxTokens: number;
|
|
11405
11476
|
};
|
|
11477
|
+
readonly "alibaba/qwen-3.6-max-preview": {
|
|
11478
|
+
id: string;
|
|
11479
|
+
name: string;
|
|
11480
|
+
api: "anthropic-messages";
|
|
11481
|
+
provider: string;
|
|
11482
|
+
baseUrl: string;
|
|
11483
|
+
reasoning: true;
|
|
11484
|
+
input: ("image" | "text")[];
|
|
11485
|
+
cost: {
|
|
11486
|
+
input: number;
|
|
11487
|
+
output: number;
|
|
11488
|
+
cacheRead: number;
|
|
11489
|
+
cacheWrite: number;
|
|
11490
|
+
};
|
|
11491
|
+
contextWindow: number;
|
|
11492
|
+
maxTokens: number;
|
|
11493
|
+
};
|
|
11406
11494
|
readonly "alibaba/qwen3-235b-a22b-thinking": {
|
|
11407
11495
|
id: string;
|
|
11408
11496
|
name: string;
|
|
@@ -12746,6 +12834,23 @@ export declare const MODELS: {
|
|
|
12746
12834
|
contextWindow: number;
|
|
12747
12835
|
maxTokens: number;
|
|
12748
12836
|
};
|
|
12837
|
+
readonly "moonshotai/kimi-k2.6": {
|
|
12838
|
+
id: string;
|
|
12839
|
+
name: string;
|
|
12840
|
+
api: "anthropic-messages";
|
|
12841
|
+
provider: string;
|
|
12842
|
+
baseUrl: string;
|
|
12843
|
+
reasoning: true;
|
|
12844
|
+
input: ("image" | "text")[];
|
|
12845
|
+
cost: {
|
|
12846
|
+
input: number;
|
|
12847
|
+
output: number;
|
|
12848
|
+
cacheRead: number;
|
|
12849
|
+
cacheWrite: number;
|
|
12850
|
+
};
|
|
12851
|
+
contextWindow: number;
|
|
12852
|
+
maxTokens: number;
|
|
12853
|
+
};
|
|
12749
12854
|
readonly "nvidia/nemotron-nano-12b-v2-vl": {
|
|
12750
12855
|
id: string;
|
|
12751
12856
|
name: string;
|