@xsai-ext/providers 0.4.3 → 0.5.0-beta.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.ts CHANGED
@@ -1,6 +1,6 @@
1
1
  import { C as ChatProvider, M as ModelProvider, E as EmbedProvider, I as ImageProvider, S as SpeechProvider, T as TranscriptionProvider, a as ChatProviderWithExtraOptions } from './types-DW1hvH0W.js';
2
- import { A as AnthropicOptions, O as OpenRouterOptions } from './anthropic-C6HzSBof.js';
3
- import { A as AnthropicModels, O as OpenrouterModels, T as TogetheraiModels } from './types-B2HZyaCH.js';
2
+ import { A as AnthropicOptions, O as OpenRouterOptions } from './anthropic-DF1J2SNf.js';
3
+ import { A as AnthropicModels, O as OpenrouterModels, T as TogetheraiModels } from './types-JlPM4HgL.js';
4
4
  import '@xsai/shared';
5
5
 
6
6
  /**
@@ -10,7 +10,7 @@ import '@xsai/shared';
10
10
  * - baseURL - `https://routellm.abacus.ai/v1`
11
11
  * - apiKey - `ABACUS_API_KEY`
12
12
  */
13
- declare const abacus: ChatProvider<"gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "gemini-2.0-flash-001" | "gemini-3-flash-preview" | "route-llm" | "grok-code-fast-1" | "kimi-k2-turbo-preview" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "gpt-4.1-mini" | "claude-opus-4-5-20251101" | "qwen-2.5-coder-32b" | "claude-sonnet-4-5-20250929" | "grok-4-0709" | "o3-mini" | "gpt-5.2-chat-latest" | "gemini-2.0-pro-exp-02-05" | "gpt-5.1" | "gpt-5-nano" | "claude-sonnet-4-20250514" | "gpt-4.1" | "o4-mini" | "claude-opus-4-20250514" | "gpt-5-mini" | "o3-pro" | "claude-3-7-sonnet-20250219" | "gemini-2.5-pro" | "gpt-4o-2024-11-20" | "o3" | "gpt-4o-mini" | "qwen3-max" | "gpt-5" | "grok-4-1-fast-non-reasoning" | "llama-3.3-70b-versatile" | "claude-opus-4-1-20250805" | "gpt-5.2" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "deepseek/deepseek-v3.1" | "openai/gpt-oss-120b" | "meta-llama/Meta-Llama-3.1-8B-Instruct" | "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "meta-llama/Meta-Llama-3.1-70B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/QwQ-32B" | "Qwen/Qwen3-32B" | "Qwen/qwen3-coder-480b-a35b-instruct" | "zai-org/glm-4.7" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2"> & ModelProvider;
13
+ declare const abacus: ChatProvider<"gpt-5.2-codex" | "llama-3.3-70b-versatile" | "claude-opus-4-5-20251101" | "o3-mini" | "gpt-5.2-chat-latest" | "gpt-5" | "claude-opus-4-20250514" | "gemini-3.1-pro-preview" | "claude-3-7-sonnet-20250219" | "kimi-k2.5" | "grok-4-fast-non-reasoning" | "grok-code-fast-1" | "gpt-5.3-codex" | "gpt-5-mini" | "claude-opus-4-6" | "claude-sonnet-4-5-20250929" | "gpt-4o-mini" | "gpt-5.1-codex-max" | "claude-sonnet-4-6" | "gpt-4.1" | "gpt-5.1-chat-latest" | "gpt-5.3-codex-xhigh" | "gpt-5.4" | "o3" | "grok-4-1-fast-non-reasoning" | "gpt-5.3-chat-latest" | "claude-sonnet-4-20250514" | "grok-4-0709" | "gemini-3-flash-preview" | "gemini-2.5-pro" | "claude-opus-4-1-20250805" | "kimi-k2-turbo-preview" | "gemini-2.5-flash" | "gpt-4o-2024-11-20" | "gpt-5.2" | "gpt-5.1" | "gpt-4.1-mini" | "qwen-2.5-coder-32b" | "gpt-5-nano" | "gemini-3.1-flash-lite-preview" | "claude-haiku-4-5-20251001" | "qwen3-max" | "o3-pro" | "gpt-4.1-nano" | "gpt-5-codex" | "o4-mini" | "route-llm" | "gpt-5.1-codex" | "openai/gpt-oss-120b" | "deepseek/deepseek-v3.1" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "zai-org/glm-5" | "zai-org/glm-4.7" | "meta-llama/Meta-Llama-3.1-8B-Instruct" | "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2" | "Qwen/Qwen3-32B" | "Qwen/qwen3-coder-480b-a35b-instruct" | "Qwen/QwQ-32B" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507"> & ModelProvider;
14
14
  /**
15
15
  * Alibaba Provider
16
16
  * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
@@ -18,7 +18,7 @@ declare const abacus: ChatProvider<"gpt-4.1-nano" | "grok-4-fast-non-reasoning"
18
18
  * - baseURL - `https://dashscope-intl.aliyuncs.com/compatible-mode/v1`
19
19
  * - apiKey - `DASHSCOPE_API_KEY`
20
20
  */
21
- declare const alibaba: ModelProvider & ChatProvider<"qwen3-max" | "qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus">;
21
+ declare const alibaba: ModelProvider & ChatProvider<"qwen3-max" | "qwen2-5-72b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-8b" | "qwen-mt-plus" | "qwen3.5-plus" | "qwen2-5-omni-7b" | "qwen-turbo" | "qwen-vl-max" | "qwen-omni-turbo-realtime" | "qwen-vl-plus" | "qwen-max" | "qvq-max" | "qwen-plus-character-ja" | "qwq-plus" | "qwen3-omni-flash" | "qwen3-14b" | "qwen-plus" | "qwen2-5-7b-instruct" | "qwen2-5-32b-instruct" | "qwen3-omni-flash-realtime" | "qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-vl-30b-a3b" | "qwen3-next-80b-a3b-instruct" | "qwen-mt-turbo" | "qwen3-vl-plus" | "qwen3-235b-a22b" | "qwen2-5-vl-7b-instruct" | "qwen-vl-ocr" | "qwen-omni-turbo" | "qwen3.5-397b-a17b" | "qwen3-livetranslate-flash-realtime" | "qwen3-coder-plus" | "qwen-flash" | "qwen2-5-vl-72b-instruct" | "qwen3-32b" | "qwen3-asr-flash" | "qwen3-coder-flash" | "qwen2-5-14b-instruct" | "qwen3-vl-235b-a22b">;
22
22
  /**
23
23
  * Alibaba (China) Provider
24
24
  * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
@@ -26,7 +26,23 @@ declare const alibaba: ModelProvider & ChatProvider<"qwen3-max" | "qwen3-livetra
26
26
  * - baseURL - `https://dashscope.aliyuncs.com/compatible-mode/v1`
27
27
  * - apiKey - `DASHSCOPE_API_KEY`
28
28
  */
29
- declare const alibabaCn: ModelProvider & ChatProvider<"qwen3-max" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
29
+ declare const alibabaCn: ModelProvider & ChatProvider<"kimi-k2.5" | "qwen3-max" | "qwen2-5-72b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-8b" | "qwen-mt-plus" | "qwen3.5-plus" | "qwen2-5-omni-7b" | "qwen-turbo" | "qwen-vl-max" | "qwen-omni-turbo-realtime" | "qwen-vl-plus" | "qwen-max" | "qvq-max" | "qwq-plus" | "qwen3-omni-flash" | "qwen3-14b" | "qwen-plus" | "qwen2-5-7b-instruct" | "qwen2-5-32b-instruct" | "qwen3-omni-flash-realtime" | "qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-vl-30b-a3b" | "qwen3-next-80b-a3b-instruct" | "qwen-mt-turbo" | "qwen3-vl-plus" | "qwen3-235b-a22b" | "qwen2-5-vl-7b-instruct" | "qwen-vl-ocr" | "qwen-omni-turbo" | "qwen3.5-397b-a17b" | "qwen3-coder-plus" | "qwen-flash" | "qwen2-5-vl-72b-instruct" | "qwen3-32b" | "qwen3-asr-flash" | "qwen3-coder-flash" | "qwen2-5-14b-instruct" | "qwen3-vl-235b-a22b" | "qwen-math-plus" | "deepseek-v3-2-exp" | "deepseek-r1-distill-llama-70b" | "qwen-plus-character" | "qwen-doc-turbo" | "qwq-32b" | "kimi-k2-thinking" | "deepseek-r1" | "qwen-long" | "qwen-deep-research" | "moonshot-kimi-k2-instruct" | "qwen2-5-coder-32b-instruct" | "deepseek-v3" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-1-5b" | "deepseek-r1-distill-qwen-14b" | "deepseek-r1-distill-qwen-7b" | "qwen-math-turbo" | "tongyi-intent-detect-v3" | "MiniMax-M2.5" | "deepseek-r1-distill-qwen-32b" | "deepseek-v3-1" | "glm-5" | "qwen2-5-math-7b-instruct" | "deepseek-r1-0528" | "qwen2-5-math-72b-instruct" | "deepseek-r1-distill-llama-8b" | "qwen3.5-flash" | "MiniMax/MiniMax-M2.5" | "kimi/kimi-k2.5" | "siliconflow/deepseek-v3.1-terminus" | "siliconflow/deepseek-v3-0324" | "siliconflow/deepseek-r1-0528" | "siliconflow/deepseek-v3.2">;
30
+ /**
31
+ * Alibaba Coding Plan Provider
32
+ * @see {@link https://www.alibabacloud.com/help/en/model-studio/coding-plan}
33
+ * @remarks
34
+ * - baseURL - `https://coding-intl.dashscope.aliyuncs.com/v1`
35
+ * - apiKey - `ALIBABA_CODING_PLAN_API_KEY`
36
+ */
37
+ declare const alibabaCodingPlan: ModelProvider & ChatProvider<"kimi-k2.5" | "qwen3.5-plus" | "qwen3-coder-plus" | "MiniMax-M2.5" | "glm-5" | "qwen3-coder-next" | "glm-4.7" | "qwen3-max-2026-01-23">;
38
+ /**
39
+ * Alibaba Coding Plan (China) Provider
40
+ * @see {@link https://help.aliyun.com/zh/model-studio/coding-plan}
41
+ * @remarks
42
+ * - baseURL - `https://coding.dashscope.aliyuncs.com/v1`
43
+ * - apiKey - `ALIBABA_CODING_PLAN_API_KEY`
44
+ */
45
+ declare const alibabaCodingPlanCn: ModelProvider & ChatProvider<"kimi-k2.5" | "qwen3.5-plus" | "qwen3-coder-plus" | "MiniMax-M2.5" | "glm-5" | "qwen3-coder-next" | "glm-4.7" | "qwen3-max-2026-01-23">;
30
46
  /**
31
47
  * Bailing Provider
32
48
  * @see {@link https://alipaytbox.yuque.com/sxs0ba/ling/intro}
@@ -42,7 +58,15 @@ declare const bailing: ModelProvider & ChatProvider<"Ling-1T" | "Ring-1T">;
42
58
  * - baseURL - `https://inference.baseten.co/v1`
43
59
  * - apiKey - `BASETEN_API_KEY`
44
60
  */
45
- declare const baseten: ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "zai-org/GLM-4.6">;
61
+ declare const baseten: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "deepseek-ai/DeepSeek-V3.2" | "zai-org/GLM-4.7" | "zai-org/GLM-5" | "zai-org/GLM-4.6" | "nvidia/Nemotron-120B-A12B" | "MiniMaxAI/MiniMax-M2.5" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2.5" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking">;
62
+ /**
63
+ * Berget.AI Provider
64
+ * @see {@link https://api.berget.ai}
65
+ * @remarks
66
+ * - baseURL - `https://api.berget.ai/v1`
67
+ * - apiKey - `BERGET_API_KEY`
68
+ */
69
+ declare const berget: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "zai-org/GLM-4.7" | "meta-llama/Llama-3.3-70B-Instruct" | "intfloat/multilingual-e5-large-instruct" | "intfloat/multilingual-e5-large" | "KBLab/kb-whisper-large" | "BAAI/bge-reranker-v2-m3" | "mistralai/Mistral-Small-3.2-24B-Instruct-2506">;
46
70
  /**
47
71
  * Cerebras Provider
48
72
  * @see {@link https://inference-docs.cerebras.ai/models/overview}
@@ -50,7 +74,7 @@ declare const baseten: ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-V3.2"
50
74
  * - baseURL - `https://api.cerebras.ai/v1/`
51
75
  * - apiKey - `CEREBRAS_API_KEY`
52
76
  */
53
- declare const cerebras: ModelProvider & ChatProvider<"zai-glm-4.7" | "qwen-3-235b-a22b-instruct-2507" | "gpt-oss-120b"> & EmbedProvider<string>;
77
+ declare const cerebras: ModelProvider & ChatProvider<"gpt-oss-120b" | "llama3.1-8b" | "zai-glm-4.7" | "qwen-3-235b-a22b-instruct-2507"> & EmbedProvider<string>;
54
78
  /**
55
79
  * Chutes Provider
56
80
  * @see {@link https://llm.chutes.ai/v1/models}
@@ -58,7 +82,15 @@ declare const cerebras: ModelProvider & ChatProvider<"zai-glm-4.7" | "qwen-3-235
58
82
  * - baseURL - `https://llm.chutes.ai/v1`
59
83
  * - apiKey - `CHUTES_API_KEY`
60
84
  */
61
- declare const chutes: ModelProvider & ChatProvider<"Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-32B" | "moonshotai/Kimi-K2-Instruct-0905" | "NousResearch/Hermes-4.3-36B" | "NousResearch/Hermes-4-70B" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4-405B-FP8-TEE" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "rednote-hilab/dots.ocr" | "moonshotai/Kimi-K2-Thinking-TEE" | "MiniMaxAI/MiniMax-M2.1-TEE" | "nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "tngtech/TNG-R1T-Chimera-TEE" | "XiaomiMiMo/MiMo-V2-Flash" | "OpenGVLab/InternVL3-78B-TEE" | "openai/gpt-oss-120b-TEE" | "openai/gpt-oss-20b" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "mistralai/Devstral-2-123B-Instruct-2512" | "mistralai/Devstral-2-123B-Instruct-2512-TEE" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "unsloth/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3Guard-Gen-0.6B" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-VL-72B-Instruct-TEE" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-235B-A22B-Instruct-2507-TEE" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8-TEE" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "miromind-ai/MiroThinker-v1.5-235B" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.5-TEE" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.7-TEE" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-V3-0324-TEE" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-V3.1-Terminus-TEE" | "deepseek-ai/DeepSeek-V3" | "deepseek-ai/DeepSeek-R1-TEE" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-R1-0528-TEE" | "deepseek-ai/DeepSeek-V3.2-TEE" | "deepseek-ai/DeepSeek-V3.1-TEE">;
85
+ declare const chutes: ModelProvider & ChatProvider<"Qwen/Qwen3-32B" | "Qwen/Qwen2.5-72B-Instruct" | "moonshotai/Kimi-K2-Instruct-0905" | "unsloth/gemma-3-27b-it" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/Llama-3.2-3B-Instruct" | "unsloth/Llama-3.2-1B-Instruct" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "openai/gpt-oss-120b-TEE" | "openai/gpt-oss-20b" | "NousResearch/Hermes-4-405B-FP8-TEE" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4.3-36B" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "NousResearch/Hermes-4-70B" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.5-Air" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.7-TEE" | "zai-org/GLM-4.6-FP8" | "zai-org/GLM-4.7-Flash" | "zai-org/GLM-4.5-TEE" | "zai-org/GLM-4.5-FP8" | "zai-org/GLM-5-TEE" | "zai-org/GLM-5-Turbo" | "zai-org/GLM-4.7-FP8" | "nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16" | "rednote-hilab/dots.ocr" | "miromind-ai/MiroThinker-v1.5-235B" | "MiniMaxAI/MiniMax-M2.5-TEE" | "MiniMaxAI/MiniMax-M2.1-TEE" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-V3.1-Terminus-TEE" | "deepseek-ai/DeepSeek-R1-0528-TEE" | "deepseek-ai/DeepSeek-V3-0324-TEE" | "deepseek-ai/DeepSeek-V3.2-TEE" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-R1-TEE" | "deepseek-ai/DeepSeek-V3.1-TEE" | "deepseek-ai/DeepSeek-V3" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-VL-72B-Instruct-TEE" | "Qwen/Qwen3Guard-Gen-0.6B" | "Qwen/Qwen3-14B" | "Qwen/Qwen3-Coder-Next" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "Qwen/Qwen3.5-397B-A17B-TEE" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8-TEE" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Instruct-2507-TEE" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "moonshotai/Kimi-K2-Thinking-TEE" | "moonshotai/Kimi-K2.5-TEE" | "OpenGVLab/InternVL3-78B-TEE" | "XiaomiMiMo/MiMo-V2-Flash" | "tngtech/TNG-R1T-Chimera-TEE" | "tngtech/TNG-R1T-Chimera-Turbo" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "mistralai/Devstral-2-123B-Instruct-2512-TEE">;
86
+ /**
87
+ * CloudFerro Sherlock Provider
88
+ * @see {@link https://docs.sherlock.cloudferro.com/}
89
+ * @remarks
90
+ * - baseURL - `https://api-sherlock.cloudferro.com/openai/v1/`
91
+ * - apiKey - `CLOUDFERRO_SHERLOCK_API_KEY`
92
+ */
93
+ declare const cloudferroSherlock: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "MiniMaxAI/MiniMax-M2.5" | "meta-llama/Llama-3.3-70B-Instruct" | "speakleash/Bielik-11B-v3.0-Instruct" | "speakleash/Bielik-11B-v2.6-Instruct">;
62
94
  /**
63
95
  * Cohere Provider
64
96
  * @see {@link https://docs.cohere.com/docs/models}
@@ -66,7 +98,7 @@ declare const chutes: ModelProvider & ChatProvider<"Qwen/Qwen2.5-72B-Instruct" |
66
98
  * - baseURL - `https://api.cohere.ai/compatibility/v1/`
67
99
  * - apiKey - `COHERE_API_KEY`
68
100
  */
69
- declare const cohere: ModelProvider & EmbedProvider<string> & ChatProvider<"command-a-translate-08-2025" | "command-a-03-2025" | "command-r-08-2024" | "command-r-plus-08-2024" | "command-r7b-12-2024" | "command-a-reasoning-08-2025" | "command-a-vision-07-2025">;
101
+ declare const cohere: ModelProvider & EmbedProvider<string> & ChatProvider<"command-a-reasoning-08-2025" | "command-r-08-2024" | "command-a-translate-08-2025" | "command-a-03-2025" | "command-r-plus-08-2024" | "c4ai-aya-expanse-32b" | "command-a-vision-07-2025" | "command-r7b-arabic-02-2025" | "c4ai-aya-vision-8b" | "command-r7b-12-2024" | "c4ai-aya-expanse-8b" | "c4ai-aya-vision-32b">;
70
102
  /**
71
103
  * Cortecs Provider
72
104
  * @see {@link https://api.cortecs.ai/v1/models}
@@ -74,7 +106,7 @@ declare const cohere: ModelProvider & EmbedProvider<string> & ChatProvider<"comm
74
106
  * - baseURL - `https://api.cortecs.ai/v1`
75
107
  * - apiKey - `CORTECS_API_KEY`
76
108
  */
77
- declare const cortecs: ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro" | "qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "devstral-small-2512" | "claude-sonnet-4" | "llama-3.1-405b-instruct">;
109
+ declare const cortecs: ModelProvider & ChatProvider<"kimi-k2.5" | "gpt-4.1" | "gemini-2.5-pro" | "qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "kimi-k2-thinking" | "glm-4.7" | "gpt-oss-120b" | "claude-4-6-sonnet" | "deepseek-v3-0324" | "devstral-2512" | "llama-3.1-405b-instruct" | "glm-4.7-flash" | "minimax-m2" | "glm-4.5" | "minimax-m2.1" | "claude-opus4-5" | "nova-pro-v1" | "claude-sonnet-4" | "intellect-3" | "glm-4.5-air" | "devstral-small-2512" | "claude-opus4-6" | "claude-haiku-4-5" | "minimax-m2.5" | "kimi-k2-instruct" | "claude-4-5-sonnet">;
78
110
  /**
79
111
  * Deep Infra Provider
80
112
  * @see {@link https://deepinfra.com/models}
@@ -82,15 +114,39 @@ declare const cortecs: ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro"
82
114
  * - baseURL - `https://api.deepinfra.com/v1/openai/`
83
115
  * - apiKey - `DEEPINFRA_API_KEY`
84
116
  */
85
- declare const deepinfra: ModelProvider & EmbedProvider<string> & ChatProvider<"openai/gpt-oss-120b" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "openai/gpt-oss-20b" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2" | "MiniMaxAI/MiniMax-M2.1" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo" | "zai-org/GLM-4.5">;
117
+ declare const deepinfra: ModelProvider & EmbedProvider<string> & ChatProvider<"openai/gpt-oss-120b" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "deepseek-ai/DeepSeek-V3.2" | "zai-org/GLM-4.7" | "zai-org/GLM-5" | "zai-org/GLM-4.6" | "MiniMaxAI/MiniMax-M2.5" | "moonshotai/Kimi-K2.5" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "openai/gpt-oss-20b" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.7-Flash" | "anthropic/claude-4-opus" | "anthropic/claude-3-7-sonnet-latest" | "zai-org/GLM-4.5" | "meta-llama/Llama-4-Scout-17B-16E-Instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-3.1-8B-Instruct-Turbo" | "meta-llama/Llama-3.1-70B-Instruct-Turbo" | "meta-llama/Llama-3.1-70B-Instruct" | "meta-llama/Llama-3.3-70B-Instruct-Turbo" | "MiniMaxAI/MiniMax-M2.1" | "MiniMaxAI/MiniMax-M2" | "deepseek-ai/DeepSeek-R1-0528" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "moonshotai/Kimi-K2-Instruct">;
86
118
  /**
87
119
  * DeepSeek Provider
88
- * @see {@link https://platform.deepseek.com/api-docs/pricing}
120
+ * @see {@link https://api-docs.deepseek.com/quick_start/pricing}
89
121
  * @remarks
90
122
  * - baseURL - `https://api.deepseek.com`
91
123
  * - apiKey - `DEEPSEEK_API_KEY`
92
124
  */
93
125
  declare const deepseek: ModelProvider & ChatProvider<"deepseek-chat" | "deepseek-reasoner">;
126
+ /**
127
+ * DInference Provider
128
+ * @see {@link https://dinference.com}
129
+ * @remarks
130
+ * - baseURL - `https://api.dinference.com/v1`
131
+ * - apiKey - `DINFERENCE_API_KEY`
132
+ */
133
+ declare const dinference: ModelProvider & ChatProvider<"glm-5" | "glm-4.7" | "gpt-oss-120b">;
134
+ /**
135
+ * D.Run (China) Provider
136
+ * @see {@link https://www.d.run}
137
+ * @remarks
138
+ * - baseURL - `https://chat.d.run/v1`
139
+ * - apiKey - `DRUN_API_KEY`
140
+ */
141
+ declare const drun: ModelProvider & ChatProvider<"public/deepseek-r1" | "public/deepseek-v3" | "public/minimax-m25">;
142
+ /**
143
+ * evroc Provider
144
+ * @see {@link https://docs.evroc.com/products/think/overview.html}
145
+ * @remarks
146
+ * - baseURL - `https://models.think.evroc.com/v1`
147
+ * - apiKey - `EVROC_API_KEY`
148
+ */
149
+ declare const evroc: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "moonshotai/Kimi-K2.5" | "intfloat/multilingual-e5-large-instruct" | "KBLab/kb-whisper-large" | "openai/whisper-large-v3" | "microsoft/Phi-4-multimodal-instruct" | "nvidia/Llama-3.3-70B-Instruct-FP8" | "Qwen/Qwen3-VL-30B-A3B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507-FP8" | "Qwen/Qwen3-Embedding-8B" | "mistralai/devstral-small-2-24b-instruct-2512" | "mistralai/Magistral-Small-2509" | "mistralai/Voxtral-Small-24B-2507">;
94
150
  /**
95
151
  * FastRouter Provider
96
152
  * @see {@link https://fastrouter.ai/models}
@@ -98,7 +154,7 @@ declare const deepseek: ModelProvider & ChatProvider<"deepseek-chat" | "deepseek
98
154
  * - baseURL - `https://go.fastrouter.ai/api/v1`
99
155
  * - apiKey - `FASTROUTER_API_KEY`
100
156
  */
101
- declare const fastrouter: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "openai/gpt-oss-20b" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
157
+ declare const fastrouter: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "openai/gpt-oss-20b" | "openai/gpt-5" | "openai/gpt-5-mini" | "openai/gpt-4.1" | "openai/gpt-5-nano" | "x-ai/grok-4" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "z-ai/glm-5" | "google/gemini-2.5-pro" | "google/gemini-2.5-flash" | "deepseek-ai/deepseek-r1-distill-llama-70b" | "qwen/qwen3-coder" | "moonshotai/kimi-k2">;
102
158
  /**
103
159
  * Fireworks AI Provider
104
160
  * @see {@link https://fireworks.ai/docs/}
@@ -106,15 +162,15 @@ declare const fastrouter: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "
106
162
  * - baseURL - `https://api.fireworks.ai/inference/v1/`
107
163
  * - apiKey - `FIREWORKS_API_KEY`
108
164
  */
109
- declare const fireworks: ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/deepseek-v3p2" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/minimax-m2p1" | "accounts/fireworks/models/glm-4p7" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/glm-4p6" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
165
+ declare const fireworks: ModelProvider & ChatProvider<"accounts/fireworks/routers/kimi-k2p5-turbo" | "accounts/fireworks/models/kimi-k2p5" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2p1" | "accounts/fireworks/models/minimax-m2p5" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p7" | "accounts/fireworks/models/deepseek-v3p2" | "accounts/fireworks/models/glm-4p5" | "accounts/fireworks/models/glm-5" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/kimi-k2-instruct">;
110
166
  /**
111
167
  * Firmware Provider
112
- * @see {@link https://docs.firmware.ai}
168
+ * @see {@link https://docs.frogbot.ai}
113
169
  * @remarks
114
- * - baseURL - `https://app.firmware.ai/api/v1`
170
+ * - baseURL - `https://app.frogbot.ai/api/v1`
115
171
  * - apiKey - `FIRMWARE_API_KEY`
116
172
  */
117
- declare const firmware: ModelProvider & ChatProvider<"grok-4-fast-non-reasoning" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "claude-sonnet-4-5-20250929" | "gpt-5-nano" | "gpt-5-mini" | "gemini-2.5-pro" | "gpt-5" | "gpt-5.2" | "claude-haiku-4-5-20251001" | "deepseek-chat" | "deepseek-reasoner" | "grok-4-fast-reasoning" | "deepseek-coder" | "claude-opus-4-5" | "gpt-4o">;
173
+ declare const firmware: ModelProvider & ChatProvider<"kimi-k2.5" | "grok-code-fast-1" | "gpt-5-mini" | "claude-opus-4-6" | "claude-sonnet-4-6" | "grok-4-1-fast-non-reasoning" | "gemini-3-flash-preview" | "gemini-2.5-pro" | "gemini-2.5-flash" | "gpt-5-nano" | "gpt-oss-120b" | "claude-haiku-4-5" | "claude-opus-4-5" | "gemini-3-pro-preview" | "gpt-4o" | "zai-glm-5" | "deepseek-v3-2" | "minimax-m2-5" | "gpt-5-4" | "gemini-3-1-pro-preview" | "grok-4-1-fast-reasoning" | "gpt-5-3-codex" | "gpt-oss-20b" | "claude-sonnet-4-5">;
118
174
  /**
119
175
  * Friendli Provider
120
176
  * @see {@link https://friendli.ai/docs/guides/serverless_endpoints/introduction}
@@ -122,7 +178,7 @@ declare const firmware: ModelProvider & ChatProvider<"grok-4-fast-non-reasoning"
122
178
  * - baseURL - `https://api.friendli.ai/serverless/v1`
123
179
  * - apiKey - `FRIENDLI_TOKEN`
124
180
  */
125
- declare const friendli: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-32B" | "zai-org/GLM-4.6" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "meta-llama-3.3-70b-instruct" | "meta-llama-3.1-8b-instruct" | "LGAI-EXAONE/K-EXAONE-236B-A23B" | "LGAI-EXAONE/EXAONE-4.0.1-32B" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
181
+ declare const friendli: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "zai-org/GLM-4.7" | "zai-org/GLM-5" | "MiniMaxAI/MiniMax-M2.5" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "MiniMaxAI/MiniMax-M2.1">;
126
182
  /**
127
183
  * GitHub Copilot Provider
128
184
  * @see {@link https://docs.github.com/en/copilot}
@@ -130,7 +186,7 @@ declare const friendli: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instr
130
186
  * - baseURL - `https://api.githubcopilot.com`
131
187
  * - apiKey - `GITHUB_TOKEN`
132
188
  */
133
- declare const githubCopilot: ModelProvider & ChatProvider<"gemini-2.0-flash-001" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gemini-3-pro-preview" | "o3-mini" | "gpt-5.1" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "gemini-2.5-pro" | "o3" | "gpt-5" | "gpt-5.2" | "claude-sonnet-4" | "gpt-4o" | "claude-opus-4" | "gpt-5.2-codex" | "gpt-5.1-codex" | "claude-haiku-4.5" | "claude-3.5-sonnet" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "claude-opus-41" | "claude-3.7-sonnet" | "gpt-5.1-codex-max" | "claude-3.7-sonnet-thought" | "claude-opus-4.5" | "claude-sonnet-4.5">;
189
+ declare const githubCopilot: ModelProvider & ChatProvider<"gpt-5.2-codex" | "gpt-5" | "gemini-3.1-pro-preview" | "grok-code-fast-1" | "gpt-5.3-codex" | "gpt-5-mini" | "gpt-5.1-codex-max" | "gpt-4.1" | "gpt-5.4" | "gemini-3-flash-preview" | "gemini-2.5-pro" | "gpt-5.2" | "gpt-5.1" | "gpt-5.1-codex" | "claude-sonnet-4" | "gemini-3-pro-preview" | "gpt-4o" | "gpt-5.1-codex-mini" | "claude-opus-4.6" | "gpt-5.4-mini" | "claude-haiku-4.5" | "claude-sonnet-4.6" | "claude-opus-4.5" | "claude-opus-41" | "claude-sonnet-4.5">;
134
190
  /**
135
191
  * GitHub Models Provider
136
192
  * @see {@link https://docs.github.com/en/github-models}
@@ -138,7 +194,7 @@ declare const githubCopilot: ModelProvider & ChatProvider<"gemini-2.0-flash-001"
138
194
  * - baseURL - `https://models.github.ai/inference`
139
195
  * - apiKey - `GITHUB_TOKEN`
140
196
  */
141
- declare const githubModels: ModelProvider & ChatProvider<"openai/gpt-4.1" | "core42/jais-30b-chat" | "xai/grok-3" | "xai/grok-3-mini" | "cohere/cohere-command-r-08-2024" | "cohere/cohere-command-a" | "cohere/cohere-command-r-plus-08-2024" | "cohere/cohere-command-r" | "cohere/cohere-command-r-plus" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-r1" | "deepseek/deepseek-v3-0324" | "mistral-ai/mistral-medium-2505" | "mistral-ai/ministral-3b" | "mistral-ai/mistral-nemo" | "mistral-ai/mistral-large-2411" | "mistral-ai/codestral-2501" | "mistral-ai/mistral-small-2503" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-mini-4k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-4" | "microsoft/phi-4-mini-reasoning" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-mini-instruct" | "microsoft/phi-4-multimodal-instruct" | "microsoft/phi-3-mini-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-mini-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-4-reasoning" | "microsoft/mai-ds-r1" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o1-preview" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o1-mini" | "openai/o3" | "openai/gpt-4o-mini" | "meta/llama-3.2-11b-vision-instruct" | "meta/meta-llama-3.1-405b-instruct" | "meta/llama-4-maverick-17b-128e-instruct-fp8" | "meta/meta-llama-3-70b-instruct" | "meta/meta-llama-3.1-70b-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-3.2-90b-vision-instruct" | "meta/meta-llama-3-8b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/meta-llama-3.1-8b-instruct" | "ai21-labs/ai21-jamba-1.5-large" | "ai21-labs/ai21-jamba-1.5-mini">;
197
+ declare const githubModels: ModelProvider & ChatProvider<"openai/gpt-4.1" | "mistral-ai/codestral-2501" | "mistral-ai/mistral-large-2411" | "mistral-ai/mistral-small-2503" | "mistral-ai/mistral-medium-2505" | "mistral-ai/ministral-3b" | "mistral-ai/mistral-nemo" | "ai21-labs/ai21-jamba-1.5-mini" | "ai21-labs/ai21-jamba-1.5-large" | "openai/o3-mini" | "openai/gpt-4o" | "openai/gpt-4o-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4.1-mini" | "openai/gpt-4.1-nano" | "openai/o1-preview" | "openai/o4-mini" | "openai/o1-mini" | "microsoft/phi-3-mini-128k-instruct" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-4-reasoning" | "microsoft/phi-4-mini-reasoning" | "microsoft/phi-3-mini-4k-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/mai-ds-r1" | "microsoft/phi-3.5-mini-instruct" | "microsoft/phi-4" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-multimodal-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-4-mini-instruct" | "cohere/cohere-command-r-plus-08-2024" | "cohere/cohere-command-r" | "cohere/cohere-command-r-08-2024" | "cohere/cohere-command-r-plus" | "cohere/cohere-command-a" | "deepseek/deepseek-v3-0324" | "deepseek/deepseek-r1" | "deepseek/deepseek-r1-0528" | "xai/grok-3-mini" | "xai/grok-3" | "core42/jais-30b-chat" | "meta/meta-llama-3.1-8b-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/llama-3.2-11b-vision-instruct" | "meta/meta-llama-3-70b-instruct" | "meta/meta-llama-3-8b-instruct" | "meta/llama-3.2-90b-vision-instruct" | "meta/meta-llama-3.1-405b-instruct" | "meta/llama-4-maverick-17b-128e-instruct-fp8" | "meta/meta-llama-3.1-70b-instruct">;
142
198
  /**
143
199
  * Google Provider
144
200
  * @see {@link https://ai.google.dev/gemini-api/docs/pricing}
@@ -146,7 +202,7 @@ declare const githubModels: ModelProvider & ChatProvider<"openai/gpt-4.1" | "cor
146
202
  * - baseURL - `https://generativelanguage.googleapis.com/v1beta/openai/`
147
203
  * - apiKey - `GOOGLE_GENERATIVE_AI_API_KEY or GEMINI_API_KEY`
148
204
  */
149
- declare const google: ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-3-flash-preview" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "gemini-2.5-pro" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
205
+ declare const google: ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-3.1-pro-preview" | "gemini-3-flash-preview" | "gemini-2.5-pro" | "gemini-2.5-flash" | "gemini-3.1-flash-lite-preview" | "gemini-3-pro-preview" | "gemini-2.5-flash-lite" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemma-3n-e2b-it" | "gemini-2.5-flash-preview-05-20" | "gemma-3-27b-it" | "gemma-3-4b-it" | "gemma-3n-e4b-it" | "gemini-2.5-pro-preview-06-05" | "gemini-2.5-pro-preview-05-06" | "gemini-2.0-flash-lite" | "gemini-1.5-flash-8b" | "gemini-1.5-flash" | "gemini-2.5-pro-preview-tts" | "gemini-3.1-pro-preview-customtools" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.0-flash" | "gemini-1.5-pro" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-preview-tts" | "gemini-flash-lite-latest" | "gemini-3.1-flash-image-preview" | "gemini-2.5-flash-image" | "gemini-flash-latest" | "gemma-3-12b-it" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-embedding-001" | "gemini-live-2.5-flash" | "gemini-2.5-flash-image-preview">;
150
206
  /**
151
207
  * Groq Provider
152
208
  * @see {@link https://console.groq.com/docs/models}
@@ -154,7 +210,7 @@ declare const google: ModelProvider & EmbedProvider<string> & ChatProvider<"gemi
154
210
  * - baseURL - `https://api.groq.com/openai/v1/`
155
211
  * - apiKey - `GROQ_API_KEY`
156
212
  */
157
- declare const groq: ModelProvider & EmbedProvider<string> & ChatProvider<"llama-3.3-70b-versatile" | "openai/gpt-oss-120b" | "deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b">;
213
+ declare const groq: ModelProvider & EmbedProvider<string> & ChatProvider<"llama-3.3-70b-versatile" | "openai/gpt-oss-120b" | "deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "llama3-70b-8192" | "gemma2-9b-it" | "llama-3.1-8b-instant" | "qwen-qwq-32b" | "llama-guard-3-8b" | "mistral-saba-24b" | "whisper-large-v3" | "whisper-large-v3-turbo" | "llama3-8b-8192" | "allam-2-7b" | "openai/gpt-oss-safeguard-20b" | "canopylabs/orpheus-arabic-saudi" | "canopylabs/orpheus-v1-english" | "meta-llama/llama-guard-4-12b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-prompt-guard-2-22m" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-prompt-guard-2-86m" | "qwen/qwen3-32b" | "groq/compound-mini" | "groq/compound" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct">;
158
214
  /**
159
215
  * Helicone Provider
160
216
  * @see {@link https://helicone.ai/models}
@@ -162,7 +218,7 @@ declare const groq: ModelProvider & EmbedProvider<string> & ChatProvider<"llama-
162
218
  * - baseURL - `https://ai-gateway.helicone.ai/v1`
163
219
  * - apiKey - `HELICONE_API_KEY`
164
220
  */
165
- declare const helicone: ModelProvider & ChatProvider<"gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "grok-code-fast-1" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "gpt-4.1-mini" | "claude-sonnet-4-5-20250929" | "o3-mini" | "gpt-5.1" | "gpt-5-nano" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "o3-pro" | "gemini-2.5-pro" | "o3" | "gpt-4o-mini" | "gpt-5" | "grok-4-1-fast-non-reasoning" | "llama-3.3-70b-versatile" | "claude-opus-4-1-20250805" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "qwen3-next-80b-a3b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-thinking" | "claude-sonnet-4" | "deepseek-reasoner" | "grok-4-fast-reasoning" | "gpt-4o" | "claude-opus-4" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "claude-3.7-sonnet" | "gemini-2.5-flash-lite" | "llama-3.1-8b-instant" | "gemma2-9b-it" | "qwen3-coder" | "claude-opus-4-1" | "grok-4" | "llama-4-maverick" | "llama-prompt-guard-2-86m" | "grok-4-1-fast-reasoning" | "claude-4.5-haiku" | "llama-3.1-8b-instruct-turbo" | "gpt-4.1-mini-2025-04-14" | "llama-guard-4" | "llama-3.1-8b-instruct" | "deepseek-v3.1-terminus" | "llama-prompt-guard-2-22m" | "claude-3.5-sonnet-v2" | "sonar-deep-research" | "grok-3" | "mistral-small" | "kimi-k2-0711" | "chatgpt-4o-latest" | "kimi-k2-0905" | "sonar-reasoning" | "llama-3.3-70b-instruct" | "claude-4.5-sonnet" | "codex-mini-latest" | "deepseek-tng-r1t2-chimera" | "claude-4.5-opus" | "sonar" | "glm-4.6" | "qwen3-235b-a22b-thinking" | "hermes-2-pro-llama-3-8b" | "o1" | "grok-3-mini" | "sonar-pro" | "o1-mini" | "claude-3-haiku-20240307" | "qwen2.5-coder-7b-fast" | "gemma-3-12b-it" | "mistral-nemo" | "gpt-oss-20b" | "claude-3.5-haiku" | "gpt-5-chat-latest" | "sonar-reasoning-pro" | "qwen3-vl-235b-a22b-instruct" | "qwen3-30b-a3b" | "deepseek-v3.2" | "gpt-5-pro" | "mistral-large-2411" | "ernie-4.5-21b-a3b-thinking" | "llama-4-scout">;
221
+ declare const helicone: ModelProvider & ChatProvider<"llama-3.3-70b-versatile" | "o3-mini" | "gpt-5" | "grok-4-fast-non-reasoning" | "grok-code-fast-1" | "gpt-5-mini" | "claude-sonnet-4-5-20250929" | "gpt-4o-mini" | "gpt-4.1" | "gpt-5.1-chat-latest" | "o3" | "grok-4-1-fast-non-reasoning" | "gemini-2.5-pro" | "claude-opus-4-1-20250805" | "gemini-2.5-flash" | "gpt-5.1" | "gpt-4.1-mini" | "gpt-5-nano" | "claude-haiku-4-5-20251001" | "o3-pro" | "gpt-4.1-nano" | "gpt-5-codex" | "o4-mini" | "gpt-5.1-codex" | "qwen3-coder-30b-a3b-instruct" | "qwen3-next-80b-a3b-instruct" | "qwen3-32b" | "deepseek-r1-distill-llama-70b" | "kimi-k2-thinking" | "deepseek-v3" | "gpt-oss-120b" | "claude-sonnet-4" | "deepseek-reasoner" | "gemini-3-pro-preview" | "gpt-4o" | "grok-4-1-fast-reasoning" | "gpt-oss-20b" | "gpt-5.1-codex-mini" | "gemini-2.5-flash-lite" | "gemma-3-12b-it" | "gemma2-9b-it" | "llama-3.1-8b-instant" | "claude-4.5-opus" | "hermes-2-pro-llama-3-8b" | "deepseek-v3.1-terminus" | "sonar-reasoning" | "gpt-5-pro" | "qwen3-vl-235b-a22b-instruct" | "kimi-k2-0711" | "llama-3.3-70b-instruct" | "chatgpt-4o-latest" | "grok-4" | "kimi-k2-0905" | "mistral-large-2411" | "ernie-4.5-21b-a3b-thinking" | "llama-guard-4" | "qwen2.5-coder-7b-fast" | "qwen3-30b-a3b" | "llama-3.1-8b-instruct" | "sonar-deep-research" | "sonar" | "llama-4-scout" | "llama-prompt-guard-2-22m" | "qwen3-coder" | "o1" | "codex-mini-latest" | "sonar-reasoning-pro" | "claude-3-haiku-20240307" | "claude-3.5-sonnet-v2" | "llama-4-maverick" | "grok-3-mini" | "gpt-4.1-mini-2025-04-14" | "claude-opus-4-1" | "glm-4.6" | "llama-3.1-8b-instruct-turbo" | "claude-3.7-sonnet" | "qwen3-235b-a22b-thinking" | "sonar-pro" | "mistral-nemo" | "claude-3.5-haiku" | "mistral-small" | "gpt-5-chat-latest" | "grok-4-fast-reasoning" | "grok-3" | "deepseek-tng-r1t2-chimera" | "claude-4.5-sonnet" | "claude-4.5-haiku" | "deepseek-v3.2" | "llama-prompt-guard-2-86m" | "o1-mini" | "claude-opus-4">;
166
222
  /**
167
223
  * Hugging Face Provider
168
224
  * @see {@link https://huggingface.co/docs/inference-providers}
@@ -170,7 +226,7 @@ declare const helicone: ModelProvider & ChatProvider<"gpt-4.1-nano" | "grok-4-fa
170
226
  * - baseURL - `https://router.huggingface.co/v1`
171
227
  * - apiKey - `HF_TOKEN`
172
228
  */
173
- declare const huggingface: ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "XiaomiMiMo/MiMo-V2-Flash" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2.1" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/DeepSeek-R1-0528">;
229
+ declare const huggingface: ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-V3.2" | "zai-org/GLM-4.7" | "zai-org/GLM-5" | "MiniMaxAI/MiniMax-M2.5" | "moonshotai/Kimi-K2.5" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.7-Flash" | "Qwen/Qwen3-Coder-Next" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "XiaomiMiMo/MiMo-V2-Flash" | "MiniMaxAI/MiniMax-M2.1" | "deepseek-ai/DeepSeek-R1-0528" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "Qwen/Qwen3.5-397B-A17B" | "Qwen/Qwen3-Embedding-4B">;
174
230
  /**
175
231
  * iFlow Provider
176
232
  * @see {@link https://platform.iflow.cn/en/docs}
@@ -178,7 +234,7 @@ declare const huggingface: ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-V3
178
234
  * - baseURL - `https://apis.iflow.cn/v1`
179
235
  * - apiKey - `IFLOW_API_KEY`
180
236
  */
181
- declare const iflowcn: ModelProvider & ChatProvider<"qwen3-max" | "qwen3-vl-plus" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "kimi-k2" | "qwen3-235b" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct" | "qwen3-max-preview">;
237
+ declare const iflowcn: ModelProvider & ChatProvider<"qwen3-max" | "qwen3-vl-plus" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-r1" | "deepseek-v3" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "qwen3-max-preview" | "kimi-k2" | "qwen3-235b" | "qwen3-235b-a22b-instruct" | "qwen3-235b-a22b-thinking-2507">;
182
238
  /**
183
239
  * Inception Provider
184
240
  * @see {@link https://platform.inceptionlabs.ai/docs}
@@ -186,7 +242,7 @@ declare const iflowcn: ModelProvider & ChatProvider<"qwen3-max" | "qwen3-vl-plus
186
242
  * - baseURL - `https://api.inceptionlabs.ai/v1/`
187
243
  * - apiKey - `INCEPTION_API_KEY`
188
244
  */
189
- declare const inception: ModelProvider & ChatProvider<"mercury-coder" | "mercury">;
245
+ declare const inception: ModelProvider & ChatProvider<"mercury-edit" | "mercury-2" | "mercury-coder" | "mercury">;
190
246
  /**
191
247
  * Inference Provider
192
248
  * @see {@link https://inference.net/models}
@@ -194,7 +250,7 @@ declare const inception: ModelProvider & ChatProvider<"mercury-coder" | "mercury
194
250
  * - baseURL - `https://inference.net/v1`
195
251
  * - apiKey - `INFERENCE_API_KEY`
196
252
  */
197
- declare const inference: ModelProvider & ChatProvider<"meta/llama-3.2-11b-vision-instruct" | "mistral/mistral-nemo-12b-instruct" | "google/gemma-3" | "osmosis/osmosis-structure-0.6b" | "qwen/qwen3-embedding-4b" | "qwen/qwen-2.5-7b-vision-instruct" | "meta/llama-3.1-8b-instruct" | "meta/llama-3.2-3b-instruct" | "meta/llama-3.2-1b-instruct">;
253
+ declare const inference: ModelProvider & ChatProvider<"meta/llama-3.2-11b-vision-instruct" | "google/gemma-3" | "qwen/qwen3-embedding-4b" | "qwen/qwen-2.5-7b-vision-instruct" | "mistral/mistral-nemo-12b-instruct" | "meta/llama-3.2-1b-instruct" | "meta/llama-3.1-8b-instruct" | "meta/llama-3.2-3b-instruct" | "osmosis/osmosis-structure-0.6b">;
198
254
  /**
199
255
  * IO.NET Provider
200
256
  * @see {@link https://io.net/docs/guides/intelligence/io-intelligence}
@@ -202,7 +258,23 @@ declare const inference: ModelProvider & ChatProvider<"meta/llama-3.2-11b-vision
202
258
  * - baseURL - `https://api.intelligence.io.solutions/api/v1`
203
259
  * - apiKey - `IOINTELLIGENCE_API_KEY`
204
260
  */
205
- declare const ioNet: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506" | "mistralai/Mistral-Large-Instruct-2411" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar">;
261
+ declare const ioNet: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "zai-org/GLM-4.6" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "meta-llama/Llama-3.3-70B-Instruct" | "openai/gpt-oss-20b" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "Qwen/Qwen2.5-VL-32B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Large-Instruct-2411" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506">;
262
+ /**
263
+ * Jiekou.AI Provider
264
+ * @see {@link https://docs.jiekou.ai/docs/support/quickstart?utm_source=github_models.dev}
265
+ * @remarks
266
+ * - baseURL - `https://api.jiekou.ai/openai`
267
+ * - apiKey - `JIEKOU_API_KEY`
268
+ */
269
+ declare const jiekou: ModelProvider & ChatProvider<"gpt-5.2-codex" | "claude-opus-4-5-20251101" | "o3-mini" | "claude-opus-4-20250514" | "grok-4-fast-non-reasoning" | "grok-code-fast-1" | "gpt-5-mini" | "claude-opus-4-6" | "claude-sonnet-4-5-20250929" | "gpt-5.1-codex-max" | "o3" | "grok-4-1-fast-non-reasoning" | "claude-sonnet-4-20250514" | "grok-4-0709" | "gemini-3-flash-preview" | "gemini-2.5-pro" | "claude-opus-4-1-20250805" | "gemini-2.5-flash" | "gpt-5.2" | "gpt-5.1" | "gpt-5-nano" | "claude-haiku-4-5-20251001" | "gpt-5-codex" | "o4-mini" | "gpt-5.1-codex" | "deepseek/deepseek-v3.1" | "zai-org/glm-4.5" | "zai-org/glm-4.7" | "gemini-3-pro-preview" | "grok-4-1-fast-reasoning" | "gpt-5.1-codex-mini" | "deepseek/deepseek-v3-0324" | "deepseek/deepseek-r1-0528" | "gemini-2.5-flash-lite" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-2.5-flash-preview-05-20" | "gemini-2.5-pro-preview-06-05" | "gemini-2.5-flash-lite-preview-06-17" | "moonshotai/kimi-k2-instruct" | "gpt-5-pro" | "gpt-5-chat-latest" | "grok-4-fast-reasoning" | "gpt-5.2-pro" | "zai-org/glm-4.7-flash" | "zai-org/glm-4.5v" | "baidu/ernie-4.5-300b-a47b-paddle" | "baidu/ernie-4.5-vl-424b-a47b" | "minimaxai/minimax-m1-80k" | "minimax/minimax-m2.1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-next" | "qwen/qwen3-30b-a3b-fp8" | "qwen/qwen3-235b-a22b-fp8" | "qwen/qwen3-coder-480b-a35b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-32b-fp8" | "moonshotai/kimi-k2.5" | "moonshotai/kimi-k2-0905" | "xiaomimimo/mimo-v2-flash">;
270
+ /**
271
+ * Kilo Gateway Provider
272
+ * @see {@link https://kilo.ai}
273
+ * @remarks
274
+ * - baseURL - `https://api.kilo.ai/api/gateway`
275
+ * - apiKey - `KILO_API_KEY`
276
+ */
277
+ declare const kilo: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "openai/gpt-oss-20b" | "openai/gpt-5" | "openai/gpt-5-mini" | "openai/gpt-4.1" | "openai/gpt-5-nano" | "x-ai/grok-4" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "z-ai/glm-5" | "google/gemini-2.5-pro" | "google/gemini-2.5-flash" | "qwen/qwen3-coder" | "moonshotai/kimi-k2" | "openai/o3-mini" | "openai/gpt-4o" | "openai/gpt-4o-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4.1-mini" | "openai/gpt-4.1-nano" | "openai/o4-mini" | "microsoft/phi-4" | "deepseek/deepseek-r1" | "deepseek/deepseek-r1-0528" | "openai/gpt-oss-safeguard-20b" | "meta-llama/llama-guard-4-12b" | "qwen/qwen3-32b" | "baidu/ernie-4.5-vl-424b-a47b" | "minimax/minimax-m2.1" | "qwen/qwen3-coder-next" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-next-80b-a3b-instruct" | "moonshotai/kimi-k2.5" | "moonshotai/kimi-k2-0905" | "giga-potato" | "corethink:free" | "giga-potato-thinking" | "morph-warp-grep-v2" | "eleutherai/llemma_7b" | "meituan/longcat-flash-chat" | "openai/gpt-5.2-codex" | "openai/o1-pro" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.4-pro" | "openai/gpt-3.5-turbo-16k" | "openai/gpt-4o:extended" | "openai/gpt-4-1106-preview" | "openai/gpt-5-pro" | "openai/gpt-3.5-turbo-0613" | "openai/gpt-5-image-mini" | "openai/gpt-4-0314" | "openai/gpt-audio" | "openai/gpt-4-turbo" | "openai/gpt-5.3-codex" | "openai/o3-mini-high" | "openai/gpt-4-turbo-preview" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo" | "openai/gpt-3.5-turbo-instruct" | "openai/gpt-5.1-chat" | "openai/gpt-5.4" | "openai/gpt-4o-audio-preview" | "openai/gpt-5-chat" | "openai/gpt-4o-search-preview" | "openai/o4-mini-high" | "openai/gpt-4o-2024-05-13" | "openai/gpt-4" | "openai/gpt-5.3-chat" | "openai/gpt-4o-2024-11-20" | "openai/gpt-5.2-chat" | "openai/gpt-5.2" | "openai/o4-mini-deep-research" | "openai/gpt-audio-mini" | "openai/gpt-5.1" | "openai/gpt-4o-mini-search-preview" | "openai/gpt-4o-mini-2024-07-18" | "openai/o3-pro" | "openai/gpt-4o-2024-08-06" | "openai/gpt-5-image" | "openai/gpt-5-codex" | "openai/o3-deep-research" | "openai/gpt-5.1-codex" | "openai/gpt-5.2-pro" | "prime-intellect/intellect-3" | "microsoft/wizardlm-2-8x22b" | "cohere/command-r-08-2024" | "cohere/command-r-plus-08-2024" | "cohere/command-r7b-12-2024" | "cohere/command-a" | "kwaipilot/kat-coder-pro" | "switchpoint/router" | "morph/morph-v3-large" | "morph/morph-v3-fast" | "x-ai/grok-4-fast" | "x-ai/grok-4.20-beta" | "x-ai/grok-4.1-fast" | "x-ai/grok-code-fast-1" | "x-ai/grok-4.20-multi-agent-beta" | "x-ai/grok-3-mini" | "x-ai/grok-3-beta" | "x-ai/grok-code-fast-1:optimized:free" | "x-ai/grok-3" | "x-ai/grok-3-mini-beta" | "anthropic/claude-opus-4.6" | "anthropic/claude-haiku-4.5" | "anthropic/claude-3.7-sonnet:thinking" | "anthropic/claude-3.7-sonnet" | "anthropic/claude-sonnet-4.6" | "anthropic/claude-3.5-haiku" | "anthropic/claude-opus-4.5" | "anthropic/claude-3.5-sonnet" | "anthropic/claude-3-haiku" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-opus-4" | "alpindale/goliath-120b" | "relace/relace-search" | "relace/relace-apply-3" | "sao10k/l3.1-70b-hanami-x1" | "sao10k/l3-lunaris-8b" | "sao10k/l3.3-euryale-70b" | "sao10k/l3.1-euryale-70b" | "sao10k/l3-euryale-70b" | "upstage/solar-pro-3" | "mancer/weaver" | "deepseek/deepseek-v3.1-terminus" | "deepseek/deepseek-r1-distill-llama-70b" | "deepseek/deepseek-chat" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-chat-v3.1" | "deepseek/deepseek-v3.2-speciale" | "deepseek/deepseek-r1-distill-qwen-32b" | "deepseek/deepseek-chat-v3-0324" | "deepseek/deepseek-v3.2" | "nvidia/llama-3.1-nemotron-70b-instruct" | "nvidia/llama-3.3-nemotron-super-49b-v1.5" | "nvidia/nemotron-nano-12b-v2-vl" | "nvidia/nemotron-nano-9b-v2" | "nvidia/nemotron-3-super-120b-a12b:free" | "nvidia/nemotron-3-nano-30b-a3b" | "gryphe/mythomax-l2-13b" | "z-ai/glm-4.7-flash" | "z-ai/glm-4.5" | "z-ai/glm-4.6" | "z-ai/glm-4.6v" | "z-ai/glm-4.5-air" | "z-ai/glm-4.5v" | "z-ai/glm-4.7" | "z-ai/glm-4-32b" | "nex-agi/deepseek-v3.1-nex-n1" | "allenai/olmo-3.1-32b-instruct" | "allenai/olmo-2-0325-32b-instruct" | "allenai/olmo-3-32b-think" | "allenai/olmo-3-7b-think" | "allenai/olmo-3.1-32b-think" | "allenai/olmo-3-7b-instruct" | "allenai/molmo-2-8b" | "google/gemini-2.5-flash-lite" | "google/gemini-2.5-flash-lite-preview-09-2025" | "google/gemma-2-9b-it" | "google/gemini-3.1-pro-preview" | "google/gemini-3-pro-preview" | "google/gemma-3-27b-it" | "google/gemma-3-4b-it" | "google/gemma-3n-e4b-it" | "google/gemini-2.5-pro-preview-05-06" | "google/gemini-2.0-flash-001" | "google/gemini-2.0-flash-lite-001" | "google/gemini-3-flash-preview" | "google/gemini-2.5-pro-preview" | "google/gemini-3.1-pro-preview-customtools" | "google/gemma-2-27b-it" | "google/gemini-3.1-flash-lite-preview" | "google/gemini-3.1-flash-image-preview" | "google/gemini-2.5-flash-image" | "google/gemma-3-12b-it" | "google/gemini-3-pro-image-preview" | "undi95/remm-slerp-l2-13b" | "amazon/nova-lite-v1" | "amazon/nova-2-lite-v1" | "amazon/nova-micro-v1" | "amazon/nova-pro-v1" | "amazon/nova-premier-v1" | "baidu/ernie-4.5-21b-a3b" | "baidu/ernie-4.5-300b-a47b" | "baidu/ernie-4.5-21b-a3b-thinking" | "baidu/ernie-4.5-vl-28b-a3b" | "ibm-granite/granite-4.0-h-micro" | "kilo/auto" | "kilo/auto-free" | "kilo/auto-small" | "meta-llama/llama-3.3-70b-instruct" | "meta-llama/llama-3.2-1b-instruct" | "meta-llama/llama-3.1-405b-instruct" | "meta-llama/llama-3.1-8b-instruct" | "meta-llama/llama-3.2-11b-vision-instruct" | "meta-llama/llama-4-scout" | "meta-llama/llama-3-8b-instruct" | "meta-llama/llama-3-70b-instruct" | "meta-llama/llama-4-maverick" | "meta-llama/llama-guard-3-8b" | "meta-llama/llama-3.2-3b-instruct" | "meta-llama/llama-3.1-70b-instruct" | "meta-llama/llama-3.1-405b" | "openrouter/hunter-alpha" | "openrouter/free" | "openrouter/bodybuilder" | "openrouter/auto" | "openrouter/healer-alpha" | "perplexity/sonar-deep-research" | "perplexity/sonar" | "perplexity/sonar-reasoning-pro" | "perplexity/sonar-pro" | "perplexity/sonar-pro-search" | "essentialai/rnj-1-instruct" | "arcee-ai/coder-large" | "arcee-ai/trinity-large-preview:free" | "arcee-ai/virtuoso-large" | "arcee-ai/trinity-mini" | "arcee-ai/spotlight" | "arcee-ai/maestro-reasoning" | "bytedance/ui-tars-1.5-7b" | "kilo-auto/free" | "kilo-auto/balanced" | "kilo-auto/small" | "kilo-auto/frontier" | "nousresearch/hermes-2-pro-llama-3-8b" | "nousresearch/hermes-4-70b" | "nousresearch/hermes-3-llama-3.1-70b" | "nousresearch/hermes-4-405b" | "nousresearch/hermes-3-llama-3.1-405b" | "minimax/minimax-m2-her" | "minimax/minimax-01" | "minimax/minimax-m2" | "minimax/minimax-m2.5" | "minimax/minimax-m1" | "minimax/minimax-m2.5:free" | "qwen/qwen3-coder-30b-a3b-instruct" | "qwen/qwen3-8b" | "qwen/qwen3.5-9b" | "qwen/qwen-turbo" | "qwen/qwen-vl-max" | "qwen/qwen3-vl-235b-a22b-instruct" | "qwen/qwq-32b" | "qwen/qwen-vl-plus" | "qwen/qwen-max" | "qwen/qwen3.5-flash-02-23" | "qwen/qwen3-vl-32b-instruct" | "qwen/qwen2.5-coder-7b-instruct" | "qwen/qwen3-14b" | "qwen/qwen3-30b-a3b-thinking-2507" | "qwen/qwen3.5-plus-02-15" | "qwen/qwen3.5-35b-a3b" | "qwen/qwen3.5-122b-a10b" | "qwen/qwen3-30b-a3b" | "qwen/qwen-plus" | "qwen/qwen3-vl-8b-thinking" | "qwen/qwen3.5-27b" | "qwen/qwen2.5-vl-32b-instruct" | "qwen/qwen-2.5-7b-instruct" | "qwen/qwen-plus-2025-07-28:thinking" | "qwen/qwen2.5-vl-72b-instruct" | "qwen/qwen3-235b-a22b-2507" | "qwen/qwen3-vl-8b-instruct" | "qwen/qwen3-vl-30b-a3b-instruct" | "qwen/qwen3-max-thinking" | "qwen/qwen3-235b-a22b" | "qwen/qwen3-vl-30b-a3b-thinking" | "qwen/qwen-plus-2025-07-28" | "qwen/qwen3.5-397b-a17b" | "qwen/qwen3-coder-plus" | "qwen/qwen3-max" | "qwen/qwen3-vl-235b-a22b-thinking" | "qwen/qwen-2.5-72b-instruct" | "qwen/qwen3-coder-flash" | "qwen/qwen-2.5-vl-7b-instruct" | "qwen/qwen-2.5-coder-32b-instruct" | "qwen/qwen3-30b-a3b-instruct-2507" | "xiaomi/mimo-v2-flash" | "stepfun/step-3.5-flash" | "stepfun/step-3.5-flash:free" | "alfredpros/codellama-7b-instruct-solidity" | "ai21/jamba-large-1.7" | "liquid/lfm-2.2-6b" | "liquid/lfm2-8b-a1b" | "liquid/lfm-2-24b-a2b" | "aion-labs/aion-2.0" | "aion-labs/aion-rp-llama-3.1-8b" | "aion-labs/aion-1.0-mini" | "aion-labs/aion-1.0" | "moonshotai/kimi-k2-thinking" | "tencent/hunyuan-a13b-instruct" | "alibaba/tongyi-deepresearch-30b-a3b" | "bytedance-seed/seed-2.0-mini" | "bytedance-seed/seed-1.6-flash" | "bytedance-seed/seed-1.6" | "bytedance-seed/seed-2.0-lite" | "inflection/inflection-3-pi" | "inflection/inflection-3-productivity" | "writer/palmyra-x5" | "inception/mercury-2" | "inception/mercury-coder" | "inception/mercury" | "anthracite-org/magnum-v4-72b" | "thedrummer/skyfall-36b-v2" | "thedrummer/rocinante-12b" | "thedrummer/cydonia-24b-v4.1" | "thedrummer/unslopnemo-12b" | "tngtech/deepseek-r1t2-chimera" | "deepcogito/cogito-v2.1-671b" | "mistralai/mistral-medium-3.1" | "mistralai/mixtral-8x22b-instruct" | "mistralai/devstral-medium" | "mistralai/mistral-7b-instruct-v0.1" | "mistralai/mistral-medium-3" | "mistralai/devstral-2512" | "mistralai/mixtral-8x7b-instruct" | "mistralai/mistral-small-24b-instruct-2501" | "mistralai/ministral-8b-2512" | "mistralai/ministral-14b-2512" | "mistralai/mistral-large-2411" | "mistralai/codestral-2508" | "mistralai/mistral-large-2407" | "mistralai/devstral-small" | "mistralai/mistral-small-creative" | "mistralai/voxtral-small-24b-2507" | "mistralai/mistral-nemo" | "mistralai/mistral-large" | "mistralai/ministral-3b-2512" | "mistralai/mistral-small-3.2-24b-instruct" | "mistralai/pixtral-large-2411" | "mistralai/mistral-saba" | "mistralai/mistral-small-3.1-24b-instruct" | "mistralai/mistral-large-2512">;
206
278
  /**
207
279
  * Kimi For Coding Provider
208
280
  * @see {@link https://www.kimi.com/coding/docs/en/third-party-agents.html}
@@ -210,7 +282,15 @@ declare const ioNet: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "meta-
210
282
  * - baseURL - `https://api.kimi.com/coding/v1`
211
283
  * - apiKey - `KIMI_API_KEY`
212
284
  */
213
- declare const kimiForCoding: ModelProvider & ChatProvider<"kimi-k2-thinking">;
285
+ declare const kimiForCoding: ModelProvider & ChatProvider<"kimi-k2-thinking" | "k2p5">;
286
+ /**
287
+ * KUAE Cloud Coding Plan Provider
288
+ * @see {@link https://docs.mthreads.com/kuaecloud/kuaecloud-doc-online/coding_plan/}
289
+ * @remarks
290
+ * - baseURL - `https://coding-plan-endpoint.kuaecloud.net/v1`
291
+ * - apiKey - `KUAE_API_KEY`
292
+ */
293
+ declare const kuaeCloudCodingPlan: ModelProvider & ChatProvider<"GLM-4.7">;
214
294
  /**
215
295
  * Llama Provider
216
296
  * @see {@link https://llama.developer.meta.com/docs/models}
@@ -218,7 +298,15 @@ declare const kimiForCoding: ModelProvider & ChatProvider<"kimi-k2-thinking">;
218
298
  * - baseURL - `https://api.llama.com/compat/v1/`
219
299
  * - apiKey - `LLAMA_API_KEY`
220
300
  */
221
- declare const llama: ModelProvider & ChatProvider<"llama-3.3-70b-instruct" | "llama-3.3-8b-instruct" | "llama-4-maverick-17b-128e-instruct-fp8" | "llama-4-scout-17b-16e-instruct-fp8" | "groq-llama-4-maverick-17b-128e-instruct" | "cerebras-llama-4-scout-17b-16e-instruct" | "cerebras-llama-4-maverick-17b-128e-instruct">;
301
+ declare const llama: ModelProvider & ChatProvider<"llama-3.3-70b-instruct" | "cerebras-llama-4-scout-17b-16e-instruct" | "cerebras-llama-4-maverick-17b-128e-instruct" | "groq-llama-4-maverick-17b-128e-instruct" | "llama-4-scout-17b-16e-instruct-fp8" | "llama-3.3-8b-instruct" | "llama-4-maverick-17b-128e-instruct-fp8">;
302
+ /**
303
+ * LLM Gateway Provider
304
+ * @see {@link https://llmgateway.io/docs}
305
+ * @remarks
306
+ * - baseURL - `https://api.llmgateway.io/v1`
307
+ * - apiKey - `LLMGATEWAY_API_KEY`
308
+ */
309
+ declare const llmgateway: ModelProvider & ChatProvider<"gpt-5.2-codex" | "claude-opus-4-5-20251101" | "o3-mini" | "gpt-5.2-chat-latest" | "gpt-5" | "claude-opus-4-20250514" | "gemini-3.1-pro-preview" | "claude-3-7-sonnet-20250219" | "kimi-k2.5" | "grok-4-fast-non-reasoning" | "grok-code-fast-1" | "gpt-5.3-codex" | "gpt-5-mini" | "claude-opus-4-6" | "claude-sonnet-4-5-20250929" | "gpt-4o-mini" | "claude-sonnet-4-6" | "gpt-4.1" | "gpt-5.4" | "o3" | "grok-4-1-fast-non-reasoning" | "gpt-5.3-chat-latest" | "claude-sonnet-4-20250514" | "grok-4-0709" | "gemini-3-flash-preview" | "gemini-2.5-pro" | "claude-opus-4-1-20250805" | "gemini-2.5-flash" | "gpt-5.2" | "gpt-5.1" | "gpt-4.1-mini" | "gpt-5-nano" | "gemini-3.1-flash-lite-preview" | "claude-haiku-4-5-20251001" | "qwen3-max" | "gpt-4.1-nano" | "o4-mini" | "gpt-5.1-codex" | "qwen3-coder-30b-a3b-instruct" | "qwen-turbo" | "qwen-vl-max" | "qwen-vl-plus" | "qwen-max" | "qwq-plus" | "qwen-plus" | "qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-next-80b-a3b-instruct" | "qwen3-vl-plus" | "qwen-omni-turbo" | "qwen3-coder-plus" | "qwen-flash" | "qwen2-5-vl-72b-instruct" | "qwen3-32b" | "qwen3-coder-flash" | "kimi-k2-thinking" | "glm-5" | "deepseek-r1-0528" | "qwen3-coder-next" | "glm-4.7" | "qwen3-max-2026-01-23" | "gpt-oss-120b" | "devstral-2512" | "glm-4.7-flash" | "minimax-m2" | "glm-4.5" | "minimax-m2.1" | "glm-4.5-air" | "claude-haiku-4-5" | "minimax-m2.5" | "gpt-4o" | "grok-4-1-fast-reasoning" | "gpt-oss-20b" | "claude-sonnet-4-5" | "gpt-5.1-codex-mini" | "gpt-5.4-mini" | "gemini-2.5-flash-lite" | "gemini-2.5-flash-lite-preview-09-2025" | "gemma-3n-e2b-it" | "gemma-3-4b-it" | "gemma-3n-e4b-it" | "gemini-2.0-flash-lite" | "gemini-2.0-flash" | "gemini-3.1-flash-image-preview" | "gemini-2.5-flash-image" | "gemma-3-12b-it" | "gemini-2.5-flash-image-preview" | "hermes-2-pro-llama-3-8b" | "gpt-5-pro" | "qwen3-vl-235b-a22b-instruct" | "llama-3.3-70b-instruct" | "grok-4" | "llama-3.1-8b-instruct" | "sonar" | "llama-4-scout" | "o1" | "sonar-reasoning-pro" | "claude-3-haiku-20240307" | "glm-4.6" | "sonar-pro" | "gpt-5-chat-latest" | "grok-4-fast-reasoning" | "grok-3" | "deepseek-v3.2" | "kimi-k2" | "qwen3-235b-a22b-thinking-2507" | "gpt-5.2-pro" | "grok-4-20-multi-agent-beta-0309" | "grok-imagine-image" | "qwen3-235b-a22b-instruct-2507" | "gpt-5.4-pro" | "qwen-image" | "pixtral-large-latest" | "grok-4-fast" | "llama-3.1-nemotron-ultra-253b" | "claude-3-opus" | "qwen-max-latest" | "glm-image" | "gemma-3-1b-it" | "seedream-4-0" | "llama-guard-4-12b" | "gpt-4-turbo" | "qwen-image-max" | "mistral-large-latest" | "grok-4-20-beta-0309-non-reasoning" | "qwen-plus-latest" | "ministral-8b-2512" | "kimi-k2-thinking-turbo" | "ministral-14b-2512" | "seed-1-6-250615" | "qwen3-30b-a3b-thinking-2507" | "qwen-image-edit-plus" | "mistral-small-2506" | "qwen35-397b-a17b" | "codestral-2508" | "qwen-image-plus" | "glm-4.5-airx" | "seed-1-6-flash-250715" | "gpt-3.5-turbo" | "qwen3-vl-flash" | "qwen3-30b-a3b-fp8" | "minimax-text-01" | "qwen2-5-vl-32b-instruct" | "llama-3-8b-instruct" | "llama-3-70b-instruct" | "glm-4.7-flashx" | "gemini-pro-latest" | "veo-3.1-fast-generate-preview" | "gpt-4o-search-preview" | "llama-4-scout-17b-instruct" | "qwen3-235b-a22b-fp8" | "glm-4.6v-flashx" | "gpt-5.4-nano" | "claude-3-5-sonnet" | "glm-4.6v" | "qwen3-vl-8b-instruct" | "gpt-4" | "minimax-m2.7" | "auto" | "qwen3-vl-30b-a3b-instruct" | "qwen-image-edit-max" | "glm-4-32b-0414-128k" | "gpt-4o-mini-search-preview" | "deepseek-v3.1" | "custom" | "glm-4.5-flash" | "llama-3.2-3b-instruct" | "qwen-image-max-2025-12-30" | "cogview-4" | "llama-3.2-11b-instruct" | "veo-3.1-generate-preview" | "qwen3-vl-30b-a3b-thinking" | "llama-3.1-70b-instruct" | "minimax-m2.1-lightning" | "seed-1-6-250915" | "qwen3-32b-fp8" | "claude-3-5-haiku" | "mixtral-8x7b-instruct-together" | "qwen-coder-plus" | "glm-4.5v" | "minimax-m2.7-highspeed" | "seedream-4-5" | "grok-imagine-image-pro" | "gemma-3-27b" | "grok-4-20-beta-0309-reasoning" | "devstral-small-2507" | "ministral-3b-2512" | "gemma-2-27b-it-together" | "qwen3-4b-fp8" | "qwen3-vl-235b-a22b-thinking" | "llama-4-maverick-17b-instruct" | "claude-3-haiku" | "glm-4.5-x" | "gemini-3-pro-image-preview" | "qwen25-coder-7b" | "seed-1-8-251228" | "claude-3-5-sonnet-20241022" | "mistral-large-2512" | "minimax-m2.5-highspeed" | "glm-4.6v-flash" | "qwen3-30b-a3b-instruct-2507" | "grok-4-1-fast" | "claude-3-7-sonnet">;
222
310
  /**
223
311
  * LMStudio Provider
224
312
  * @see {@link https://lmstudio.ai/models}
@@ -226,7 +314,7 @@ declare const llama: ModelProvider & ChatProvider<"llama-3.3-70b-instruct" | "ll
226
314
  * - baseURL - `http://127.0.0.1:1234/v1`
227
315
  * - apiKey - `LMSTUDIO_API_KEY`
228
316
  */
229
- declare const lmstudio: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "qwen/qwen3-30b-a3b-2507" | "qwen/qwen3-coder-30b">;
317
+ declare const lmstudio: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "qwen/qwen3-coder-30b" | "qwen/qwen3-30b-a3b-2507">;
230
318
  /**
231
319
  * LucidQuery AI Provider
232
320
  * @see {@link https://lucidquery.com/api/docs}
@@ -236,21 +324,45 @@ declare const lmstudio: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "qwe
236
324
  */
237
325
  declare const lucidquery: ModelProvider & ChatProvider<"lucidquery-nexus-coder" | "lucidnova-rf1-100b">;
238
326
  /**
239
- * MiniMax Provider
327
+ * Meganova Provider
328
+ * @see {@link https://docs.meganova.ai}
329
+ * @remarks
330
+ * - baseURL - `https://api.meganova.ai/v1`
331
+ * - apiKey - `MEGANOVA_API_KEY`
332
+ */
333
+ declare const meganova: ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-V3.2" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "zai-org/GLM-4.7" | "zai-org/GLM-5" | "zai-org/GLM-4.6" | "MiniMaxAI/MiniMax-M2.5" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2.5" | "moonshotai/Kimi-K2-Thinking" | "meta-llama/Llama-3.3-70B-Instruct" | "mistralai/Mistral-Small-3.2-24B-Instruct-2506" | "Qwen/Qwen2.5-VL-32B-Instruct" | "XiaomiMiMo/MiMo-V2-Flash" | "MiniMaxAI/MiniMax-M2.1" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Mistral-Nemo-Instruct-2407" | "deepseek-ai/DeepSeek-V3.2-Exp" | "Qwen/Qwen3.5-Plus">;
334
+ /**
335
+ * MiniMax (minimax.io) Provider
240
336
  * @see {@link https://platform.minimax.io/docs/guides/quickstart}
241
337
  * @remarks
242
338
  * - baseURL - `https://api.minimax.io/v1/`
243
339
  * - apiKey - `MINIMAX_API_KEY`
244
340
  */
245
- declare const minimax: ModelProvider & ChatProvider<"MiniMax-M2" | "MiniMax-M2.1">;
341
+ declare const minimax: ModelProvider & ChatProvider<"MiniMax-M2.5" | "MiniMax-M2.7" | "MiniMax-M2.1" | "MiniMax-M2" | "MiniMax-M2.5-highspeed" | "MiniMax-M2.7-highspeed">;
246
342
  /**
247
- * MiniMax (China) Provider
343
+ * MiniMax (minimaxi.com) Provider
248
344
  * @see {@link https://platform.minimaxi.com/docs/guides/quickstart}
249
345
  * @remarks
250
346
  * - baseURL - `https://api.minimaxi.com/v1/`
251
347
  * - apiKey - `MINIMAX_API_KEY`
252
348
  */
253
- declare const minimaxCn: ModelProvider & ChatProvider<"MiniMax-M2" | "MiniMax-M2.1">;
349
+ declare const minimaxCn: ModelProvider & ChatProvider<"MiniMax-M2.5" | "MiniMax-M2.7" | "MiniMax-M2.1" | "MiniMax-M2" | "MiniMax-M2.5-highspeed" | "MiniMax-M2.7-highspeed">;
350
+ /**
351
+ * MiniMax Coding Plan (minimaxi.com) Provider
352
+ * @see {@link https://platform.minimaxi.com/docs/coding-plan/intro}
353
+ * @remarks
354
+ * - baseURL - `https://api.minimaxi.com/anthropic/v1`
355
+ * - apiKey - `MINIMAX_API_KEY`
356
+ */
357
+ declare const minimaxCnCodingPlan: ModelProvider & ChatProvider<"MiniMax-M2.5" | "MiniMax-M2.7" | "MiniMax-M2.1" | "MiniMax-M2" | "MiniMax-M2.5-highspeed" | "MiniMax-M2.7-highspeed">;
358
+ /**
359
+ * MiniMax Coding Plan (minimax.io) Provider
360
+ * @see {@link https://platform.minimax.io/docs/coding-plan/intro}
361
+ * @remarks
362
+ * - baseURL - `https://api.minimax.io/anthropic/v1`
363
+ * - apiKey - `MINIMAX_API_KEY`
364
+ */
365
+ declare const minimaxCodingPlan: ModelProvider & ChatProvider<"MiniMax-M2.5" | "MiniMax-M2.7" | "MiniMax-M2.1" | "MiniMax-M2" | "MiniMax-M2.5-highspeed" | "MiniMax-M2.7-highspeed">;
254
366
  /**
255
367
  * Mistral Provider
256
368
  * @see {@link https://docs.mistral.ai/getting-started/models/}
@@ -258,7 +370,15 @@ declare const minimaxCn: ModelProvider & ChatProvider<"MiniMax-M2" | "MiniMax-M2
258
370
  * - baseURL - `https://api.mistral.ai/v1/`
259
371
  * - apiKey - `MISTRAL_API_KEY`
260
372
  */
261
- declare const mistral: ModelProvider & EmbedProvider<string> & ChatProvider<"devstral-2512" | "mistral-nemo" | "mistral-large-2411" | "devstral-medium-2507" | "mistral-large-2512" | "open-mixtral-8x22b" | "ministral-8b-latest" | "pixtral-large-latest" | "mistral-small-2506" | "ministral-3b-latest" | "pixtral-12b" | "mistral-medium-2505" | "labs-devstral-small-2512" | "devstral-medium-latest" | "devstral-small-2505" | "mistral-medium-2508" | "mistral-embed" | "mistral-small-latest" | "magistral-small" | "devstral-small-2507" | "codestral-latest" | "open-mixtral-8x7b" | "open-mistral-7b" | "mistral-large-latest" | "mistral-medium-latest" | "magistral-medium-latest">;
373
+ declare const mistral: ModelProvider & EmbedProvider<string> & ChatProvider<"devstral-2512" | "mistral-large-2411" | "mistral-nemo" | "pixtral-large-latest" | "mistral-large-latest" | "mistral-small-2506" | "devstral-small-2507" | "mistral-large-2512" | "devstral-small-2505" | "mistral-small-2603" | "ministral-3b-latest" | "magistral-small" | "codestral-latest" | "devstral-medium-latest" | "mistral-embed" | "devstral-medium-2507" | "magistral-medium-latest" | "labs-devstral-small-2512" | "mistral-medium-latest" | "mistral-medium-2505" | "open-mixtral-8x22b" | "ministral-8b-latest" | "open-mixtral-8x7b" | "pixtral-12b" | "mistral-small-latest" | "open-mistral-7b" | "mistral-medium-2508">;
374
+ /**
375
+ * Moark Provider
376
+ * @see {@link https://moark.com/docs/openapi/v1#tag/%E6%96%87%E6%9C%AC%E7%94%9F%E6%88%90}
377
+ * @remarks
378
+ * - baseURL - `https://moark.com/v1`
379
+ * - apiKey - `MOARK_API_KEY`
380
+ */
381
+ declare const moark: ModelProvider & ChatProvider<"GLM-4.7" | "MiniMax-M2.1">;
262
382
  /**
263
383
  * ModelScope Provider
264
384
  * @see {@link https://modelscope.cn/docs/model-service/API-Inference/intro}
@@ -266,7 +386,7 @@ declare const mistral: ModelProvider & EmbedProvider<string> & ChatProvider<"dev
266
386
  * - baseURL - `https://api-inference.modelscope.cn/v1`
267
387
  * - apiKey - `MODELSCOPE_API_KEY`
268
388
  */
269
- declare const modelscope: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct">;
389
+ declare const modelscope: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct">;
270
390
  /**
271
391
  * Moonshot AI Provider
272
392
  * @see {@link https://platform.moonshot.ai/docs/api/chat}
@@ -274,7 +394,7 @@ declare const modelscope: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Ins
274
394
  * - baseURL - `https://api.moonshot.ai/v1`
275
395
  * - apiKey - `MOONSHOT_API_KEY`
276
396
  */
277
- declare const moonshotai: ModelProvider & ChatProvider<"kimi-k2-turbo-preview" | "kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
397
+ declare const moonshotai: ModelProvider & ChatProvider<"kimi-k2.5" | "kimi-k2-turbo-preview" | "kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-0905-preview" | "kimi-k2-0711-preview">;
278
398
  /**
279
399
  * Moonshot AI (China) Provider
280
400
  * @see {@link https://platform.moonshot.cn/docs/api/chat}
@@ -282,7 +402,7 @@ declare const moonshotai: ModelProvider & ChatProvider<"kimi-k2-turbo-preview" |
282
402
  * - baseURL - `https://api.moonshot.cn/v1`
283
403
  * - apiKey - `MOONSHOT_API_KEY`
284
404
  */
285
- declare const moonshotaiCn: ModelProvider & ChatProvider<"kimi-k2-turbo-preview" | "kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
405
+ declare const moonshotaiCn: ModelProvider & ChatProvider<"kimi-k2.5" | "kimi-k2-turbo-preview" | "kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-0905-preview" | "kimi-k2-0711-preview">;
286
406
  /**
287
407
  * Morph Provider
288
408
  * @see {@link https://docs.morphllm.com/api-reference/introduction}
@@ -290,7 +410,7 @@ declare const moonshotaiCn: ModelProvider & ChatProvider<"kimi-k2-turbo-preview"
290
410
  * - baseURL - `https://api.morphllm.com/v1`
291
411
  * - apiKey - `MORPH_API_KEY`
292
412
  */
293
- declare const morph: ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
413
+ declare const morph: ModelProvider & ChatProvider<"auto" | "morph-v3-large" | "morph-v3-fast">;
294
414
  /**
295
415
  * NanoGPT Provider
296
416
  * @see {@link https://docs.nano-gpt.com}
@@ -298,7 +418,7 @@ declare const morph: ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "m
298
418
  * - baseURL - `https://nano-gpt.com/api/v1`
299
419
  * - apiKey - `NANO_GPT_API_KEY`
300
420
  */
301
- declare const nanoGpt: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "zai-org/glm-4.7" | "qwen/qwen3-coder" | "deepseek/deepseek-r1" | "moonshotai/kimi-k2-instruct" | "moonshotai/kimi-k2-thinking" | "nousresearch/hermes-4-405b:thinking" | "nvidia/llama-3_3-nemotron-super-49b-v1_5" | "deepseek/deepseek-v3.2:thinking" | "minimax/minimax-m2.1" | "z-ai/glm-4.6:thinking" | "z-ai/glm-4.6" | "qwen/qwen3-235b-a22b-thinking-2507" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "meta-llama/llama-4-maverick" | "meta-llama/llama-3.3-70b-instruct" | "zai-org/glm-4.5-air" | "zai-org/glm-4.7:thinking" | "zai-org/glm-4.5-air:thinking">;
421
+ declare const nanoGpt: ModelProvider & ChatProvider<"claude-opus-4-5-20251101" | "claude-opus-4-20250514" | "claude-3-7-sonnet-20250219" | "claude-sonnet-4-5-20250929" | "claude-sonnet-4-20250514" | "gemini-2.5-pro" | "claude-opus-4-1-20250805" | "gemini-2.5-flash" | "claude-haiku-4-5-20251001" | "openai/gpt-oss-120b" | "zai-org/glm-5" | "zai-org/glm-4.7" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "qwen3-coder-30b-a3b-instruct" | "qwen-turbo" | "qwen-max" | "qvq-max" | "qwen-plus" | "qwq-32b" | "deepseek-r1" | "qwen-long" | "qwen3-max-2026-01-23" | "deepseek-ai/DeepSeek-V3.1" | "moonshotai/Kimi-K2-Instruct-0905" | "unsloth/gemma-3-27b-it" | "unsloth/gemma-3-4b-it" | "unsloth/gemma-3-12b-it" | "openai/gpt-oss-20b" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "command-a-reasoning-08-2025" | "deepseek-v3-0324" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-chat" | "deepseek-reasoner" | "openai/gpt-5" | "openai/gpt-5-mini" | "openai/gpt-4.1" | "openai/gpt-5-nano" | "gemini-3-pro-preview" | "openai/o3-mini" | "openai/gpt-4o" | "openai/gpt-4o-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4.1-mini" | "openai/gpt-4.1-nano" | "openai/o1-preview" | "openai/o4-mini" | "gemini-2.5-flash-lite" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-flash-preview-05-20" | "gemini-2.5-pro-preview-06-05" | "gemini-2.5-pro-preview-05-06" | "gemini-2.0-flash-lite" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-lite-preview-06-17" | "openai/gpt-oss-safeguard-20b" | "moonshotai/kimi-k2-instruct" | "sonar-deep-research" | "sonar" | "sonar-reasoning-pro" | "sonar-pro" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "zai-org/glm-4.7-flash" | "minimax/minimax-m2.1" | "moonshotai/kimi-k2.5" | "openai/gpt-5.2-codex" | "openai/o1-pro" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5-pro" | "openai/gpt-4-turbo" | "openai/o3-mini-high" | "openai/gpt-4-turbo-preview" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo" | "openai/gpt-5.1-chat" | "openai/gpt-4o-search-preview" | "openai/o4-mini-high" | "openai/gpt-4o-2024-11-20" | "openai/gpt-5.2-chat" | "openai/gpt-5.2" | "openai/o4-mini-deep-research" | "openai/gpt-5.1" | "openai/gpt-4o-mini-search-preview" | "openai/gpt-4o-2024-08-06" | "openai/gpt-5-codex" | "openai/o3-deep-research" | "openai/gpt-5.1-codex" | "openai/gpt-5.2-pro" | "microsoft/wizardlm-2-8x22b" | "cohere/command-r-plus-08-2024" | "x-ai/grok-4-fast" | "x-ai/grok-4.1-fast" | "x-ai/grok-code-fast-1" | "anthropic/claude-opus-4.6" | "anthropic/claude-sonnet-4.6" | "deepseek/deepseek-v3.2-speciale" | "deepseek/deepseek-v3.2" | "nvidia/nemotron-3-nano-30b-a3b" | "z-ai/glm-4.6" | "z-ai/glm-4.5v" | "nex-agi/deepseek-v3.1-nex-n1" | "allenai/olmo-3.1-32b-instruct" | "allenai/olmo-3-32b-think" | "allenai/olmo-3.1-32b-think" | "allenai/molmo-2-8b" | "google/gemini-3-flash-preview" | "undi95/remm-slerp-l2-13b" | "amazon/nova-lite-v1" | "amazon/nova-2-lite-v1" | "amazon/nova-micro-v1" | "amazon/nova-pro-v1" | "baidu/ernie-4.5-300b-a47b" | "baidu/ernie-4.5-vl-28b-a3b" | "meta-llama/llama-3.3-70b-instruct" | "meta-llama/llama-3.1-8b-instruct" | "meta-llama/llama-4-scout" | "meta-llama/llama-4-maverick" | "meta-llama/llama-3.2-3b-instruct" | "essentialai/rnj-1-instruct" | "arcee-ai/trinity-mini" | "minimax/minimax-m2-her" | "minimax/minimax-01" | "minimax/minimax-m2.5" | "qwen/qwen3.5-397b-a17b" | "xiaomi/mimo-v2-flash" | "aion-labs/aion-rp-llama-3.1-8b" | "aion-labs/aion-1.0-mini" | "aion-labs/aion-1.0" | "moonshotai/kimi-k2-thinking" | "inflection/inflection-3-pi" | "inflection/inflection-3-productivity" | "anthracite-org/magnum-v4-72b" | "deepcogito/cogito-v2.1-671b" | "mistralai/mistral-medium-3.1" | "mistralai/mistral-medium-3" | "mistralai/ministral-8b-2512" | "mistralai/ministral-14b-2512" | "mistralai/codestral-2508" | "mistralai/mistral-small-creative" | "mistralai/mistral-large" | "mistralai/ministral-3b-2512" | "mistralai/mistral-saba" | "qwen-image" | "qwen3-vl-235b-a22b-thinking" | "gemini-3-pro-image-preview" | "claude-3-5-sonnet-20241022" | "qwen3-30b-a3b-instruct-2507" | "MiniMax-M2" | "claude-opus-4-thinking" | "brave-research" | "jamba-large-1.7" | "azure-o3-mini" | "claude-sonnet-4-thinking:8192" | "ernie-x1-32k" | "exa-answer" | "KAT-Coder-Pro-V1" | "ernie-4.5-turbo-128k" | "deepclaude" | "Llama-3.3-70B-Forgotten-Abomination-v5.0" | "gemini-2.5-pro-exp-03-25" | "mistral-small-31-24b-instruct" | "claude-sonnet-4-thinking:1024" | "claude-opus-4-1-thinking" | "jamba-large-1.6" | "universal-summarizer" | "Llama-3.3-70B-Bigger-Body" | "doubao-seed-2-0-pro-260215" | "Llama-3.3-70B-Progenitor-V3.3" | "claude-opus-4-1-thinking:32768" | "doubao-seed-1-6-thinking-250615" | "Llama-3.3-70B-Fallen-v1" | "glm-zero-preview" | "Llama-3.3-70B-MS-Nevoria" | "glm-z1-air" | "GLM-4.5-Air-Derestricted-Iceblink-ReExtract" | "Llama-3.3-70B-Mokume-Gane-R1" | "doubao-1-5-thinking-vision-pro-250428" | "Llama-3.3+(3v3.3)-70B-TenyxChat-DaybreakStorywriter" | "QwQ-32B-ArliAI-RpR-v1" | "GLM-4.6-Derestricted-v5" | "glm-4" | "claude-opus-4-thinking:32000" | "Llama-3.3-70B-GeneticLemonade-Opus" | "doubao-1.5-pro-32k" | "Llama-3.3-70B-Forgotten-Safeword-3.6" | "grok-3-mini-fast-beta" | "Llama-3.3-70B-ArliAI-RPMax-v3" | "gemini-2.0-flash-thinking-exp-1219" | "MiniMax-M1" | "chroma" | "azure-o1" | "claude-3-7-sonnet-thinking:128000" | "venice-uncensored:web" | "deepseek-r1-sambanova" | "qwen25-vl-72b-instruct" | "brave-pro" | "glm-4-airx" | "Llama-3.3-70B-Cu-Mai-R1" | "deepseek-chat-cheaper" | "ernie-5.0-thinking-latest" | "claude-opus-4-1-thinking:1024" | "ernie-x1-32k-preview" | "Llama-3.3-70B-The-Omega-Directive-Unslop-v2.1" | "claude-sonnet-4-thinking:64000" | "glm-z1-airx" | "qwen3-vl-235b-a22b-instruct-original" | "yi-lightning" | "Llama-3.3-70B-Ignition-v0.1" | "kimi-k2-instruct-fast" | "gemini-2.5-flash-preview-05-20:thinking" | "auto-model-premium" | "glm-4.1v-thinking-flash" | "claude-3-5-haiku-20241022" | "auto-model" | "gemini-2.0-flash-001" | "Llama-3.3-70B-Mhnnn-x1" | "step-2-16k-exp" | "Llama-3.3-70B-Legion-V2.1" | "Llama-3.3+(3.1v3.3)-70B-New-Dawn-v1.1" | "doubao-1-5-thinking-pro-vision-250415" | "claude-sonnet-4-5-20250929-thinking" | "yi-large" | "Gemma-3-27B-Nidum-Uncensored" | "claude-opus-4-thinking:32768" | "Llama-3.3-70B-Cirrus-x1" | "Gemma-3-27B-CardProjector-v4" | "Qwen2.5-32B-EVA-v0.2" | "v0-1.5-lg" | "gemini-2.5-flash-preview-09-2025-thinking" | "azure-gpt-4-turbo" | "gemini-2.5-flash-nothinking" | "Gemma-3-27B-ArliAI-RPMax-v3" | "v0-1.0-md" | "step-3" | "brave" | "ernie-x1-turbo-32k" | "glm-4-air" | "GLM-4.5-Air-Derestricted" | "grok-3-fast-beta" | "claude-3-7-sonnet-thinking:1024" | "doubao-seed-code-preview-latest" | "grok-3-beta" | "claude-3-7-sonnet-reasoner" | "gemini-2.0-pro-exp-02-05" | "glm-4-long" | "venice-uncensored" | "Gemma-3-27B-Big-Tiger-v3" | "doubao-seed-2-0-mini-260215" | "gemini-2.0-flash-thinking-exp-01-21" | "Llama-3.3-70B-Damascus-R1" | "doubao-1-5-thinking-pro-250415" | "asi1-mini" | "doubao-seed-2-0-code-preview-260215" | "exa-research-pro" | "GLM-4.5-Air-Derestricted-Iceblink-v2" | "Llama-3.3-70B-Nova" | "Gemma-3-27B-it" | "claude-opus-4-thinking:8192" | "claude-3-5-sonnet-20240620" | "sarvan-medium" | "hunyuan-t1-latest" | "Llama-3.3-70B-RAWMAW" | "kimi-thinking-preview" | "claude-sonnet-4-thinking:32768" | "gemini-2.5-pro-preview-03-25" | "Llama-3.3-70B-Electra-R1" | "Llama-3.3-70B-ArliAI-RPMax-v2" | "GLM-4.5-Air-Derestricted-Steam" | "doubao-seed-1-8-251215" | "Llama-3.3-70B-The-Omega-Directive-Unslop-v2.0" | "GLM-4.5-Air-Derestricted-Steam-ReExtract" | "exa-research" | "azure-gpt-4o" | "study_gpt-chatgpt-4o-latest" | "Llama-3.3-70B-Aurora-Borealis" | "Baichuan4-Turbo" | "Baichuan4-Air" | "KAT-Coder-Exp-72B-1010" | "gemini-2.0-flash-exp-image-generation" | "Llama-3.3-70B-Dark-Ages-v0.1" | "claude-opus-4-1-thinking:8192" | "Magistral-Small-2506" | "Llama-3.3-70B-MiraiFanfare" | "glm-4-flash" | "Llama-3.3-70B-Shakudo" | "gemini-2.0-pro-reasoner" | "Llama-3.3-70B-Fallen-R1-v1" | "doubao-seed-1-6-flash-250615" | "claude-opus-4-5-20251101:thinking" | "Llama-3.3-70B-Strawberrylemonade-v1.2" | "Llama-3.3-70B-Magnum-v4-SE" | "doubao-seed-1-6-250615" | "Llama-3.3-70B-StrawberryLemonade-v1.0" | "Meta-Llama-3-1-8B-Instruct-FP8" | "claude-opus-4-1-thinking:32000" | "GLM-4.5-Air-Derestricted-Iceblink-v2-ReExtract" | "gemini-3-pro-preview-thinking" | "gemini-2.5-flash-lite-preview-09-2025-thinking" | "Llama-3.3-70B-Vulpecula-R1" | "doubao-1.5-vision-pro-32k" | "deepseek-reasoner-cheaper" | "ernie-x1.1-preview" | "KAT-Coder-Air-V1" | "mercury-coder-small" | "doubao-1.5-pro-256k" | "glm-4-plus" | "Baichuan-M2" | "auto-model-standard" | "Gemma-3-27B-it-Abliterated" | "learnlm-1.5-pro-experimental" | "claude-opus-4-thinking:1024" | "gemini-2.5-flash-preview-04-17:thinking" | "v0-1.5-md" | "Llama-3.3-70B-Electranova-v1.0" | "Llama-3.3+(3.1v3.3)-70B-Hanami-x1" | "phi-4-multimodal-instruct" | "ernie-4.5-8k-preview" | "claude-3-7-sonnet-thinking:32768" | "claude-3-7-sonnet-thinking:8192" | "Llama-3.3-70B-Incandescent-Malevolence" | "step-2-mini" | "auto-model-basic" | "claude-sonnet-4-thinking" | "Llama-3.3-70B-GeneticLemonade-Unleashed-v3" | "step-r1-v-mini" | "glm-4-plus-0111" | "z-image-turbo" | "Llama-3.3-70B-Sapphira-0.2" | "deepseek-math-v2" | "azure-gpt-4o-mini" | "Mistral-Nemo-12B-Instruct-2407" | "jamba-mini-1.6" | "Llama-3.3-70B-ArliAI-RPMax-v1.4" | "Llama-3.3-70B-Anthrobomination" | "glm-4-air-0111" | "jamba-mini" | "ernie-5.0-thinking-preview" | "Gemma-3-27B-Glitter" | "hidream" | "glm-4.1v-thinking-flashx" | "phi-4-mini-instruct" | "Llama-3.3-70B-Sapphira-0.1" | "yi-medium-200k" | "jamba-mini-1.7" | "fastgpt" | "GLM-4.5-Air-Derestricted-Iceblink" | "Llama-3.3-70B-Predatorial-Extasy" | "claude-3-7-sonnet-thinking" | "Llama-3.3-70B-Magnum-v4-SE-Cirrus-x1-SLERP" | "gemini-exp-1206" | "doubao-seed-2-0-lite-260215" | "jamba-large" | "hunyuan-turbos-20250226" | "ernie-4.5-turbo-vl-32k" | "grok-3-mini-beta" | "Llama-3.3-70B-Argunaut-1-SFT" | "abacusai/Dracarys-72B-Instruct" | "THUDM/GLM-Z1-32B-0414" | "THUDM/GLM-Z1-9B-0414" | "THUDM/GLM-4-32B-0414" | "THUDM/GLM-Z1-Rumination-32B-0414" | "THUDM/GLM-4-9B-0414" | "unsloth/gemma-3-1b-it" | "shisa-ai/shisa-v2-llama3.3-70b" | "shisa-ai/shisa-v2.1-llama3.3-70b" | "openai/o3-mini-low" | "openai/chatgpt-4o-latest" | "openai/gpt-5.1-chat-latest" | "openai/o3-pro-2025-06-10" | "openai/gpt-5-chat-latest" | "openai/gpt-5.1-2025-11-13" | "CrucibleLab/L3.3-70B-Loki-V2.0" | "microsoft/MAI-DS-R1-FP8" | "cohere/command-r" | "x-ai/grok-4-fast:thinking" | "x-ai/grok-4.1-fast-reasoning" | "x-ai/grok-4-07-09" | "anthropic/claude-opus-4.6:thinking:max" | "anthropic/claude-sonnet-4.6:thinking" | "anthropic/claude-opus-4.6:thinking:medium" | "anthropic/claude-opus-4.6:thinking" | "anthropic/claude-opus-4.6:thinking:low" | "raifle/sorcererlm-8x22b" | "nothingiisreal/L3.1-70B-Celeste-V0.1-BF16" | "NousResearch 2/hermes-4-70b" | "NousResearch 2/hermes-4-405b:thinking" | "NousResearch 2/hermes-3-llama-3.1-70b" | "NousResearch 2/Hermes-4-70B:thinking" | "NousResearch 2/hermes-4-405b" | "NousResearch 2/DeepHermes-3-Mistral-24B-Preview" | "deepseek/deepseek-v3.2:thinking" | "deepseek/deepseek-prover-v2-671b" | "zai-org/glm-5.1:thinking" | "zai-org/glm-5:thinking" | "zai-org/glm-5.1" | "NeverSleep/Llama-3-Lumimaid-70B-v0.1" | "NeverSleep/Lumimaid-v0.2-70B" | "nvidia/Llama-3.1-Nemotron-Ultra-253B-v1" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/Llama-3_3-Nemotron-Super-49B-v1_5" | "nvidia/Llama-3.1-Nemotron-70B-Instruct-HF" | "nvidia/Llama-3.3-Nemotron-Super-49B-v1" | "z-ai/glm-4.6:thinking" | "z-ai/glm-4.5v:thinking" | "stepfun-ai/step-3.5-flash:thinking" | "stepfun-ai/step-3.5-flash" | "cognitivecomputations/dolphin-2.9.2-qwen2-72b" | "TheDrummer 2/Cydonia-24B-v4.1" | "TheDrummer 2/UnslopNemo-12B-v4.1" | "TheDrummer 2/Cydonia-24B-v4.3" | "TheDrummer 2/skyfall-36b-v2" | "TheDrummer 2/Anubis-70B-v1" | "TheDrummer 2/Cydonia-24B-v2" | "TheDrummer 2/Cydonia-24B-v4" | "TheDrummer 2/Anubis-70B-v1.1" | "TheDrummer 2/Magidonia-24B-v4.3" | "TheDrummer 2/Rocinante-12B-v1.1" | "google/gemini-3-flash-preview-thinking" | "google/gemini-flash-1.5" | "meta-llama/llama-3.2-90b-vision-instruct" | "ReadyArt/MS3.2-The-Omega-Directive-24B-Unslop-v2.0" | "ReadyArt/The-Omega-Abomination-L-70B-v1.0" | "miromind-ai/mirothinker-v1.5-235b" | "Doctor-Shotgun/MS3.2-24B-Magnum-Diamond" | "LLM360/K2-Think" | "MiniMaxAI/MiniMax-M1-80k" | "failspy/Meta-Llama-3-70B-Instruct-abliterated-v3.5" | "pamanseau/OpenReasoning-Nemotron-32B" | "arcee-ai/trinity-large" | "deepseek-ai/DeepSeek-V3.1-Terminus:thinking" | "deepseek-ai/DeepSeek-V3.1:thinking" | "deepseek-ai/deepseek-v3.2-exp-thinking" | "deepseek-ai/deepseek-v3.2-exp" | "inflatebot/MN-12B-Mag-Mell-R1" | "MarinaraSpaghetti/NemoMix-Unleashed-12B" | "Alibaba-NLP/Tongyi-DeepResearch-30B-A3B" | "soob3123/GrayLine-Qwen3-8B" | "soob3123/amoral-gemma3-27B-v2" | "soob3123/Veiled-Calla-12B" | "minimax/minimax-m2.7" | "dmind/dmind-1" | "dmind/dmind-1-mini" | "VongolaChouko/Starcannon-Unleashed-12B-v1.0" | "xiaomi/mimo-v2-flash-thinking" | "xiaomi/mimo-v2-flash-thinking-original" | "xiaomi/mimo-v2-flash-original" | "Salesforce/Llama-xLAM-2-70b-fc-r" | "Gryphe/MythoMax-L2-13b" | "baseten/Kimi-K2-Instruct-FP4" | "Steelskull/L3.3-Nevoria-R1-70b" | "Steelskull/L3.3-Electra-R1-70b" | "Steelskull/L3.3-Cu-Mai-R1-70b" | "Steelskull/L3.3-MS-Evalebis-70b" | "Steelskull/L3.3-MS-Nevoria-70b" | "Steelskull/L3.3-MS-Evayale-70B" | "meganova-ai/manta-mini-1.0" | "meganova-ai/manta-pro-1.0" | "meganova-ai/manta-flash-1.0" | "meituan-longcat/LongCat-Flash-Chat-FP8" | "TEE/kimi-k2-thinking" | "TEE/kimi-k2.5" | "TEE/glm-4.7-flash" | "TEE/gemma-3-27b-it" | "TEE/kimi-k2.5-thinking" | "TEE/gpt-oss-120b" | "TEE/qwen3-coder" | "TEE/minimax-m2.1" | "TEE/qwen2.5-vl-72b-instruct" | "TEE/glm-4.6" | "TEE/deepseek-v3.1" | "TEE/llama3-3-70b" | "TEE/glm-5" | "TEE/qwen3.5-397b-a17b" | "TEE/glm-4.7" | "TEE/deepseek-r1-0528" | "TEE/gpt-oss-20b" | "TEE/deepseek-v3.2" | "TEE/qwen3-30b-a3b-instruct-2507" | "Infermatic/MN-12B-Inferor-v0.0" | "mlabonne/NeuralDaredevil-8B-abliterated" | "moonshotai/kimi-k2.5:thinking" | "moonshotai/kimi-k2-thinking-turbo-original" | "moonshotai/kimi-k2-thinking-original" | "moonshotai/kimi-k2-instruct-0711" | "moonshotai/Kimi-Dev-72B" | "tencent/Hunyuan-MT-7B" | "Envoid/Llama-3.05-NT-Storybreaker-Ministral-70B" | "Envoid/Llama-3.05-Nemotron-Tenyxchat-Storybreaker-70B" | "Tongyi-Zhiwen/QwenLong-L1-32B" | "featherless-ai/Qwerky-72B" | "Sao10K/L3.3-70B-Euryale-v2.3" | "Sao10K/L3.1-70B-Hanami-x1" | "Sao10K/L3.1-70B-Euryale-v2.2" | "Sao10K/L3-8B-Stheno-v3.2" | "GalrionSoftworks/MN-LooseCannon-12B-v1" | "LatitudeGames/Wayfarer-Large-70B-Llama-3.3" | "anthracite-org/magnum-v2-72b" | "tngtech/tng-r1t-chimera" | "deepcogito/cogito-v1-preview-qwen-32B" | "huihui-ai/DeepSeek-R1-Distill-Llama-70B-abliterated" | "huihui-ai/Llama-3.3-70B-Instruct-abliterated" | "huihui-ai/Qwen2.5-32B-Instruct-abliterated" | "huihui-ai/DeepSeek-R1-Distill-Qwen-32B-abliterated" | "huihui-ai/Llama-3.1-Nemotron-70B-Instruct-HF-abliterated" | "mistralai/mistral-7b-instruct" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mixtral-8x22b-instruct-v0.1" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "mistralai/mistral-tiny" | "mistralai/mixtral-8x7b-instruct-v0.1" | "EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2" | "EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1" | "EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2" | "EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.0">;
302
422
  /**
303
423
  * Nebius Token Factory Provider
304
424
  * @see {@link https://docs.tokenfactory.nebius.com/}
@@ -306,7 +426,15 @@ declare const nanoGpt: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "zai
306
426
  * - baseURL - `https://api.tokenfactory.nebius.com/v1`
307
427
  * - apiKey - `NEBIUS_API_KEY`
308
428
  */
309
- declare const nebius: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "zai-org/glm-4.5" | "openai/gpt-oss-20b" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-235b-a22b-thinking-2507" | "zai-org/glm-4.5-air" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "deepseek-ai/deepseek-v3">;
429
+ declare const nebius: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "meta-llama/Meta-Llama-3.1-8B-Instruct" | "deepseek-ai/DeepSeek-V3.2" | "Qwen/Qwen3-32B" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "zai-org/GLM-5" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2.5" | "moonshotai/Kimi-K2-Thinking" | "meta-llama/Llama-3.3-70B-Instruct" | "openai/gpt-oss-20b" | "NousResearch/Hermes-4-70B" | "zai-org/GLM-4.5-Air" | "zai-org/GLM-4.7-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "zai-org/GLM-4.5" | "MiniMaxAI/MiniMax-M2.1" | "deepseek-ai/DeepSeek-R1-0528" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "google/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "NousResearch/Hermes-4-405B" | "nvidia/Nemotron-Nano-V2-12b" | "nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B" | "nvidia/nemotron-3-super-120b-a12b" | "nvidia/Llama-3_1-Nemotron-Ultra-253B-v1" | "google/gemma-3-27b-it-fast" | "google/gemma-2-2b-it" | "google/gemma-2-9b-it-fast" | "PrimeIntellect/INTELLECT-3" | "meta-llama/Llama-Guard-3-8B" | "meta-llama/Meta-Llama-3.1-8B-Instruct-fast" | "meta-llama/Llama-3.3-70B-Instruct-fast" | "deepseek-ai/DeepSeek-R1-0528-fast" | "deepseek-ai/DeepSeek-V3-0324-fast" | "intfloat/e5-mistral-7b-instruct" | "black-forest-labs/flux-dev" | "black-forest-labs/flux-schnell" | "Qwen/Qwen2.5-Coder-7B-fast" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen3-32B-fast" | "BAAI/bge-en-icl" | "BAAI/bge-multilingual-gemma2" | "moonshotai/Kimi-K2.5-fast">;
430
+ /**
431
+ * Nova Provider
432
+ * @see {@link https://nova.amazon.com/dev/documentation}
433
+ * @remarks
434
+ * - baseURL - `https://api.nova.amazon.com/v1`
435
+ * - apiKey - `NOVA_API_KEY`
436
+ */
437
+ declare const nova: ModelProvider & ChatProvider<"nova-2-lite-v1" | "nova-2-pro-v1">;
310
438
  /**
311
439
  * NovitaAI Provider
312
440
  * @see {@link https://novita.ai/docs/guides/introduction}
@@ -314,7 +442,7 @@ declare const nebius: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "zai-
314
442
  * - baseURL - `https://api.novita.ai/openai`
315
443
  * - apiKey - `NOVITA_API_KEY`
316
444
  */
317
- declare const novitaAi: ModelProvider & ChatProvider<"deepseek/deepseek-v3.1" | "openai/gpt-oss-120b" | "zai-org/glm-4.7" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "openai/gpt-oss-20b" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-v3-0324" | "moonshotai/kimi-k2-instruct" | "meta-llama/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-thinking" | "minimax/minimax-m2.1" | "qwen/qwen3-235b-a22b-thinking-2507" | "meta-llama/llama-3.3-70b-instruct" | "zai-org/glm-4.5-air" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "baichuan/baichuan-m2-32b" | "moonshotai/kimi-k2-0905" | "nousresearch/hermes-2-pro-llama-3-8b" | "paddlepaddle/paddleocr-vl" | "kwaipilot/kat-coder" | "kwaipilot/kat-coder-pro" | "xiaomimimo/mimo-v2-flash" | "deepseek/deepseek-prover-v2-671b" | "deepseek/deepseek-r1-0528-qwen3-8b" | "deepseek/deepseek-v3.1-terminus" | "deepseek/deepseek-r1-turbo" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-r1-distill-llama-70b" | "deepseek/deepseek-ocr" | "deepseek/deepseek-v3.2" | "deepseek/deepseek-v3-turbo" | "sao10k/l3-8b-lunaris" | "sao10k/L3-8B-Stheno-v3.2" | "sao10k/l31-70b-euryale-v2.2" | "sao10k/l3-70b-euryale-v2.1" | "skywork/r1v4-lite" | "minimaxai/minimax-m1-80k" | "minimax/minimax-m2" | "google/gemma-3-27b-it" | "microsoft/wizardlm-2-8x22b" | "gryphe/mythomax-l2-13b" | "baidu/ernie-4.5-vl-28b-a3b-thinking" | "baidu/ernie-4.5-300b-a47b-paddle" | "baidu/ernie-4.5-21B-a3b" | "baidu/ernie-4.5-21B-a3b-thinking" | "baidu/ernie-4.5-vl-424b-a47b" | "baidu/ernie-4.5-vl-28b-a3b" | "qwen/qwen3-vl-30b-a3b-thinking" | "qwen/qwen3-omni-30b-a3b-thinking" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-8b-fp8" | "qwen/qwen2.5-vl-72b-instruct" | "qwen/qwen3-4b-fp8" | "qwen/qwen3-coder-30b-a3b-instruct" | "qwen/qwen3-vl-8b-instruct" | "qwen/qwen2.5-7b-instruct" | "qwen/qwen3-30b-a3b-fp8" | "qwen/qwen3-32b-fp8" | "qwen/qwen3-omni-30b-a3b-instruct" | "qwen/qwen-2.5-72b-instruct" | "qwen/qwen3-vl-235b-a22b-thinking" | "qwen/qwen-mt-plus" | "qwen/qwen3-max" | "qwen/qwen3-vl-235b-a22b-instruct" | "qwen/qwen3-vl-30b-a3b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-235b-a22b-fp8" | "mistralai/mistral-nemo" | "meta-llama/llama-3-70b-instruct" | "meta-llama/llama-3-8b-instruct" | "meta-llama/llama-3.1-8b-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct-fp8" | "zai-org/glm-4.5v" | "zai-org/glm-4.6v" | "zai-org/autoglm-phone-9b-multilingual">;
445
+ declare const novitaAi: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "deepseek/deepseek-v3.1" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "zai-org/glm-5" | "zai-org/glm-4.7" | "openai/gpt-oss-20b" | "deepseek/deepseek-v3-0324" | "deepseek/deepseek-r1-0528" | "meta-llama/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-instruct" | "zai-org/glm-4.7-flash" | "zai-org/glm-4.5v" | "baidu/ernie-4.5-300b-a47b-paddle" | "baidu/ernie-4.5-vl-424b-a47b" | "minimaxai/minimax-m1-80k" | "minimax/minimax-m2.1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-next" | "qwen/qwen3-30b-a3b-fp8" | "qwen/qwen3-235b-a22b-fp8" | "qwen/qwen3-coder-480b-a35b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-32b-fp8" | "moonshotai/kimi-k2.5" | "moonshotai/kimi-k2-0905" | "xiaomimimo/mimo-v2-flash" | "microsoft/wizardlm-2-8x22b" | "kwaipilot/kat-coder-pro" | "deepseek/deepseek-v3.1-terminus" | "deepseek/deepseek-r1-distill-llama-70b" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-v3.2" | "gryphe/mythomax-l2-13b" | "google/gemma-3-27b-it" | "baidu/ernie-4.5-vl-28b-a3b" | "meta-llama/llama-3.3-70b-instruct" | "meta-llama/llama-3.1-8b-instruct" | "meta-llama/llama-3-8b-instruct" | "meta-llama/llama-3-70b-instruct" | "nousresearch/hermes-2-pro-llama-3-8b" | "minimax/minimax-m2" | "minimax/minimax-m2.5" | "qwen/qwen3-coder-30b-a3b-instruct" | "qwen/qwen3-vl-235b-a22b-instruct" | "qwen/qwen2.5-vl-72b-instruct" | "qwen/qwen3-vl-8b-instruct" | "qwen/qwen3-vl-30b-a3b-instruct" | "qwen/qwen3-vl-30b-a3b-thinking" | "qwen/qwen3.5-397b-a17b" | "qwen/qwen3-max" | "qwen/qwen3-vl-235b-a22b-thinking" | "qwen/qwen-2.5-72b-instruct" | "moonshotai/kimi-k2-thinking" | "mistralai/mistral-nemo" | "deepseek/deepseek-prover-v2-671b" | "paddlepaddle/paddleocr-vl" | "kwaipilot/kat-coder" | "sao10k/l3-8b-lunaris" | "sao10k/l3-70b-euryale-v2.1" | "sao10k/l31-70b-euryale-v2.2" | "sao10k/L3-8B-Stheno-v3.2" | "deepseek/deepseek-r1-turbo" | "deepseek/deepseek-r1-0528-qwen3-8b" | "deepseek/deepseek-ocr-2" | "deepseek/deepseek-v3-turbo" | "deepseek/deepseek-ocr" | "zai-org/autoglm-phone-9b-multilingual" | "zai-org/glm-4.6v" | "zai-org/glm-4.5-air" | "skywork/r1v4-lite" | "baidu/ernie-4.5-21B-a3b-thinking" | "baidu/ernie-4.5-vl-28b-a3b-thinking" | "baidu/ernie-4.5-21B-a3b" | "meta-llama/llama-4-maverick-17b-128e-instruct-fp8" | "qwen/qwen-mt-plus" | "qwen/qwen3-8b-fp8" | "qwen/qwen3-omni-30b-a3b-instruct" | "qwen/qwen2.5-7b-instruct" | "qwen/qwen3-omni-30b-a3b-thinking" | "qwen/qwen3-4b-fp8" | "baichuan/baichuan-m2-32b">;
318
446
  /**
319
447
  * Nvidia Provider
320
448
  * @see {@link https://docs.api.nvidia.com/nim/}
@@ -322,7 +450,7 @@ declare const novitaAi: ModelProvider & ChatProvider<"deepseek/deepseek-v3.1" |
322
450
  * - baseURL - `https://integrate.api.nvidia.com/v1`
323
451
  * - apiKey - `NVIDIA_API_KEY`
324
452
  */
325
- declare const nvidia: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-mini-instruct" | "microsoft/phi-3-medium-4k-instruct" | "meta/llama-3.2-11b-vision-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "meta/llama-3.2-1b-instruct" | "moonshotai/kimi-k2-thinking" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "qwen/qwen3-coder-480b-a35b-instruct" | "google/gemma-3-27b-it" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/nemotron-3-nano-30b-a3b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.3-nemotron-super-49b-v1" | "nvidia/llama-3.1-nemotron-51b-instruct" | "nvidia/llama3-chatqa-1.5-70b" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "nvidia/llama-3.1-nemotron-70b-instruct" | "nvidia/nemotron-4-340b-instruct" | "nvidia/llama-3.3-nemotron-super-49b-v1.5" | "minimaxai/minimax-m2" | "google/gemma-3n-e2b-it" | "google/codegemma-1.1-7b" | "google/gemma-3n-e4b-it" | "google/gemma-2-2b-it" | "google/gemma-3-12b-it" | "google/codegemma-7b" | "google/gemma-3-1b-it" | "google/gemma-2-27b-it" | "microsoft/phi-3-vision-128k-instruct" | "openai/whisper-large-v3" | "qwen/qwen2.5-coder-32b-instruct" | "qwen/qwen2.5-coder-7b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwq-32b" | "mistralai/mamba-codestral-7b-v0.1" | "mistralai/mistral-large-2-instruct" | "mistralai/codestral-22b-instruct-v0.1" | "mistralai/mistral-small-3.1-24b-instruct-2503" | "meta/llama3-70b-instruct" | "meta/llama-4-maverick-17b-128e-instruct" | "meta/codellama-70b" | "meta/llama-3.1-405b-instruct" | "meta/llama3-8b-instruct" | "meta/llama-3.1-70b-instruct" | "deepseek-ai/deepseek-r1-0528" | "deepseek-ai/deepseek-r1" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "deepseek-ai/deepseek-coder-6.7b-instruct" | "black-forest-labs/flux.1-dev">;
453
+ declare const nvidia: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "openai/whisper-large-v3" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-4-mini-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/llama-3.2-11b-vision-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "meta/llama-3.2-1b-instruct" | "qwen/qwen3-coder-480b-a35b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-next-80b-a3b-instruct" | "moonshotai/kimi-k2.5" | "nvidia/llama-3.1-nemotron-70b-instruct" | "nvidia/llama-3.3-nemotron-super-49b-v1.5" | "nvidia/nemotron-3-nano-30b-a3b" | "google/gemma-3-27b-it" | "google/gemma-3n-e4b-it" | "google/gemma-2-27b-it" | "google/gemma-3-12b-it" | "qwen/qwq-32b" | "qwen/qwen2.5-coder-7b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwen3.5-397b-a17b" | "moonshotai/kimi-k2-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "stepfun-ai/step-3.5-flash" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "nvidia/nemotron-3-super-120b-a12b" | "google/gemma-2-2b-it" | "microsoft/phi-3-vision-128k-instruct" | "nvidia/llama3-chatqa-1.5-70b" | "nvidia/nemotron-4-340b-instruct" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-3.1-nemotron-51b-instruct" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-embed-nemotron-8b" | "nvidia/llama-3.3-nemotron-super-49b-v1" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "z-ai/glm5" | "z-ai/glm4.7" | "google/gemma-3-1b-it" | "google/gemma-3n-e2b-it" | "google/codegemma-1.1-7b" | "google/codegemma-7b" | "minimaxai/minimax-m2.1" | "minimaxai/minimax-m2.5" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-r1" | "deepseek-ai/deepseek-coder-6.7b-instruct" | "deepseek-ai/deepseek-v3.1" | "deepseek-ai/deepseek-r1-0528" | "deepseek-ai/deepseek-v3.2" | "qwen/qwen2.5-coder-32b-instruct" | "black-forest-labs/flux.1-dev" | "meta/codellama-70b" | "meta/llama-3.1-405b-instruct" | "meta/llama3-8b-instruct" | "meta/llama3-70b-instruct" | "meta/llama-4-maverick-17b-128e-instruct" | "meta/llama-3.1-70b-instruct" | "mistralai/mistral-large-2-instruct" | "mistralai/mamba-codestral-7b-v0.1" | "mistralai/codestral-22b-instruct-v0.1" | "mistralai/mistral-small-3.1-24b-instruct-2503">;
326
454
  /**
327
455
  * Ollama Cloud Provider
328
456
  * @see {@link https://docs.ollama.com/cloud}
@@ -330,7 +458,7 @@ declare const nvidia: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "micr
330
458
  * - baseURL - `https://ollama.com/v1`
331
459
  * - apiKey - `OLLAMA_API_KEY`
332
460
  */
333
- declare const ollamaCloud: ModelProvider & ChatProvider<"kimi-k2-thinking:cloud" | "qwen3-vl-235b-cloud" | "qwen3-coder:480b-cloud" | "gpt-oss:120b-cloud" | "deepseek-v3.1:671b-cloud" | "glm-4.6:cloud" | "cogito-2.1:671b-cloud" | "gpt-oss:20b-cloud" | "qwen3-vl-235b-instruct-cloud" | "kimi-k2:1t-cloud" | "minimax-m2:cloud" | "gemini-3-pro-preview:latest">;
461
+ declare const ollamaCloud: ModelProvider & ChatProvider<"kimi-k2.5" | "gemini-3-flash-preview" | "kimi-k2-thinking" | "glm-5" | "qwen3-coder-next" | "glm-4.7" | "minimax-m2" | "minimax-m2.1" | "minimax-m2.5" | "glm-4.6" | "deepseek-v3.2" | "minimax-m2.7" | "mistral-large-3:675b" | "qwen3-vl:235b-instruct" | "gpt-oss:120b" | "cogito-2.1:671b" | "nemotron-3-nano:30b" | "ministral-3:14b" | "devstral-small-2:24b" | "qwen3-next:80b" | "qwen3-vl:235b" | "gemma3:12b" | "gemma3:27b" | "qwen3-coder:480b" | "gpt-oss:20b" | "ministral-3:8b" | "gemma3:4b" | "qwen3.5:397b" | "nemotron-3-super" | "devstral-2:123b" | "rnj-1:8b" | "deepseek-v3.1:671b" | "ministral-3:3b" | "kimi-k2:1t">;
334
462
  /**
335
463
  * OpenAI Provider
336
464
  * @see {@link https://platform.openai.com/docs/models}
@@ -338,7 +466,7 @@ declare const ollamaCloud: ModelProvider & ChatProvider<"kimi-k2-thinking:cloud"
338
466
  * - baseURL - `https://api.openai.com/v1/`
339
467
  * - apiKey - `OPENAI_API_KEY`
340
468
  */
341
- declare const openai: ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1-nano" | "gpt-4.1-mini" | "o3-mini" | "gpt-5.2-chat-latest" | "gpt-5.1" | "gpt-5-nano" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "o3-pro" | "gpt-4o-2024-11-20" | "o3" | "gpt-4o-mini" | "gpt-5" | "gpt-5.2" | "gpt-5.1-chat-latest" | "gpt-4o" | "gpt-5.2-codex" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "gpt-5.1-codex-max" | "codex-mini-latest" | "o1" | "o1-mini" | "gpt-5-chat-latest" | "gpt-5-pro" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "o3-deep-research" | "gpt-3.5-turbo" | "gpt-5.2-pro" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "text-embedding-ada-002" | "o4-mini-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
469
+ declare const openai: ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-5.2-codex" | "o3-mini" | "gpt-5.2-chat-latest" | "gpt-5" | "gpt-5.3-codex" | "gpt-5-mini" | "gpt-4o-mini" | "gpt-5.1-codex-max" | "gpt-4.1" | "gpt-5.1-chat-latest" | "gpt-5.4" | "o3" | "gpt-5.3-chat-latest" | "gpt-4o-2024-11-20" | "gpt-5.2" | "gpt-5.1" | "gpt-4.1-mini" | "gpt-5-nano" | "o3-pro" | "gpt-4.1-nano" | "gpt-5-codex" | "o4-mini" | "gpt-5.1-codex" | "gpt-4o" | "gpt-5.1-codex-mini" | "gpt-5.4-mini" | "gpt-5-pro" | "o1" | "codex-mini-latest" | "gpt-5-chat-latest" | "o1-mini" | "gpt-5.2-pro" | "gpt-5.4-pro" | "gpt-4-turbo" | "gpt-3.5-turbo" | "gpt-5.4-nano" | "gpt-4" | "o1-pro" | "text-embedding-3-large" | "gpt-5.3-codex-spark" | "gpt-4o-2024-05-13" | "text-embedding-ada-002" | "o4-mini-deep-research" | "gpt-4o-2024-08-06" | "o1-preview" | "text-embedding-3-small" | "o3-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
342
470
  /**
343
471
  * OpenCode Zen Provider
344
472
  * @see {@link https://opencode.ai/docs/zen}
@@ -346,7 +474,15 @@ declare const openai: ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-
346
474
  * - baseURL - `https://opencode.ai/zen/v1`
347
475
  * - apiKey - `OPENCODE_API_KEY`
348
476
  */
349
- declare const opencode: ModelProvider & ChatProvider<"gpt-5.1" | "gpt-5-nano" | "gpt-5" | "gpt-5.2" | "kimi-k2-thinking" | "claude-sonnet-4" | "claude-opus-4-5" | "gpt-5.2-codex" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "gpt-5.1-codex-max" | "qwen3-coder" | "claude-opus-4-1" | "glm-4.6" | "kimi-k2" | "claude-haiku-4-5" | "gemini-3-pro" | "alpha-glm-4.7" | "claude-sonnet-4-5" | "alpha-gd4" | "big-pickle" | "claude-3-5-haiku" | "glm-4.7-free" | "grok-code" | "gemini-3-flash" | "minimax-m2.1-free">;
477
+ declare const opencode: ModelProvider & ChatProvider<"gpt-5.2-codex" | "gpt-5" | "kimi-k2.5" | "gpt-5.3-codex" | "claude-opus-4-6" | "gpt-5.1-codex-max" | "claude-sonnet-4-6" | "gpt-5.4" | "gpt-5.2" | "gpt-5.1" | "gpt-5-nano" | "gpt-5-codex" | "gpt-5.1-codex" | "kimi-k2-thinking" | "glm-5" | "glm-4.7" | "minimax-m2.1" | "claude-sonnet-4" | "claude-haiku-4-5" | "minimax-m2.5" | "claude-opus-4-5" | "claude-sonnet-4-5" | "gpt-5.1-codex-mini" | "gpt-5.4-mini" | "qwen3-coder" | "claude-opus-4-1" | "glm-4.6" | "kimi-k2" | "gpt-5.4-pro" | "gpt-5.4-nano" | "claude-3-5-haiku" | "gpt-5.3-codex-spark" | "mimo-v2-flash-free" | "big-pickle" | "minimax-m2.1-free" | "mimo-v2-omni-free" | "qwen3.6-plus-free" | "grok-code" | "kimi-k2.5-free" | "glm-5-free" | "gemini-3-pro" | "nemotron-3-super-free" | "minimax-m2.5-free" | "trinity-large-preview-free" | "gemini-3.1-pro" | "glm-4.7-free" | "gemini-3-flash" | "mimo-v2-pro-free">;
478
+ /**
479
+ * OpenCode Go Provider
480
+ * @see {@link https://opencode.ai/docs/zen}
481
+ * @remarks
482
+ * - baseURL - `https://opencode.ai/zen/go/v1`
483
+ * - apiKey - `OPENCODE_API_KEY`
484
+ */
485
+ declare const opencodeGo: ModelProvider & ChatProvider<"kimi-k2.5" | "glm-5" | "minimax-m2.5" | "minimax-m2.7">;
350
486
  /**
351
487
  * OVHcloud AI Endpoints Provider
352
488
  * @see {@link https://www.ovhcloud.com/en/public-cloud/ai-endpoints/catalog//}
@@ -354,7 +490,7 @@ declare const opencode: ModelProvider & ChatProvider<"gpt-5.1" | "gpt-5-nano" |
354
490
  * - baseURL - `https://oai.endpoints.kepler.ai.cloud.ovh.net/v1`
355
491
  * - apiKey - `OVHCLOUD_API_KEY`
356
492
  */
357
- declare const ovhcloud: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "gpt-oss-20b" | "mixtral-8x7b-instruct-v0.1" | "mistral-7b-instruct-v0.3" | "qwen2.5-vl-72b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen2.5-coder-32b-instruct" | "llava-next-mistral-7b" | "meta-llama-3_1-70b-instruct" | "meta-llama-3_3-70b-instruct">;
493
+ declare const ovhcloud: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "gpt-oss-20b" | "llama-3.1-8b-instruct" | "mistral-nemo-instruct-2407" | "qwen2.5-coder-32b-instruct" | "qwen2.5-vl-72b-instruct" | "mistral-7b-instruct-v0.3" | "mistral-small-3.2-24b-instruct-2506" | "mixtral-8x7b-instruct-v0.1" | "meta-llama-3_3-70b-instruct">;
358
494
  /**
359
495
  * Perplexity Provider
360
496
  * @see {@link https://docs.perplexity.ai}
@@ -362,7 +498,15 @@ declare const ovhcloud: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instru
362
498
  * - baseURL - `https://api.perplexity.ai/`
363
499
  * - apiKey - `PERPLEXITY_API_KEY`
364
500
  */
365
- declare const perplexity: ModelProvider & ChatProvider<"sonar" | "sonar-pro" | "sonar-reasoning-pro">;
501
+ declare const perplexity: ModelProvider & ChatProvider<"sonar-deep-research" | "sonar" | "sonar-reasoning-pro" | "sonar-pro">;
502
+ /**
503
+ * Perplexity Agent Provider
504
+ * @see {@link https://docs.perplexity.ai/docs/agent-api/models}
505
+ * @remarks
506
+ * - baseURL - `https://api.perplexity.ai/v1`
507
+ * - apiKey - `PERPLEXITY_API_KEY`
508
+ */
509
+ declare const perplexityAgent: ModelProvider & ChatProvider<"openai/gpt-5-mini" | "google/gemini-2.5-pro" | "google/gemini-2.5-flash" | "openai/gpt-5.4" | "openai/gpt-5.2" | "openai/gpt-5.1" | "google/gemini-3.1-pro-preview" | "google/gemini-3-flash-preview" | "perplexity/sonar" | "nvidia/nemotron-3-super-120b-a12b" | "anthropic/claude-opus-4-5" | "anthropic/claude-opus-4-6" | "anthropic/claude-sonnet-4-6" | "anthropic/claude-sonnet-4-5" | "anthropic/claude-haiku-4-5" | "xai/grok-4-1-fast-non-reasoning">;
366
510
  /**
367
511
  * Poe Provider
368
512
  * @see {@link https://creator.poe.com/docs/external-applications/openai-compatible-api}
@@ -370,7 +514,7 @@ declare const perplexity: ModelProvider & ChatProvider<"sonar" | "sonar-pro" | "
370
514
  * - baseURL - `https://api.poe.com/v1`
371
515
  * - apiKey - `POE_API_KEY`
372
516
  */
373
- declare const poe: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-4.1-fast-non-reasoning" | "ideogramai/ideogram" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram-v2" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "poetools/claude-code" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "google/gemini-deep-research" | "google/nano-banana" | "google/imagen-4" | "google/imagen-3" | "google/imagen-4-ultra" | "google/gemini-2.0-flash-lite" | "google/gemini-3-pro" | "google/veo-3.1" | "google/imagen-3-fast" | "google/lyria" | "google/gemini-2.0-flash" | "google/gemini-2.5-flash-lite" | "google/veo-3" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-2" | "google/gemini-3-flash" | "google/nano-banana-pro" | "google/veo-3.1-fast" | "openai/gpt-5.2-instant" | "openai/sora-2" | "openai/o1-pro" | "openai/gpt-5.1-codex" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-4-classic" | "openai/gpt-5-chat" | "openai/o3-deep-research" | "openai/gpt-4o-search" | "openai/gpt-image-1.5" | "openai/gpt-image-1-mini" | "openai/gpt-3.5-turbo" | "openai/gpt-5.2-pro" | "openai/o3-mini-high" | "openai/chatgpt-4o-latest" | "openai/gpt-4-turbo" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1-instant" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-4o-aug" | "openai/o3-pro" | "openai/gpt-image-1" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo-instruct" | "openai/o4-mini-deep-research" | "openai/gpt-4-classic-0314" | "openai/dall-e-3" | "openai/sora-2-pro" | "openai/gpt-5-pro" | "openai/gpt-5.2" | "openai/gpt-4o-mini-search" | "stabilityai/stablediffusionxl" | "topazlabs-co/topazlabs" | "lumalabs/ray2" | "lumalabs/dream-machine" | "anthropic/claude-opus-3" | "anthropic/claude-opus-4" | "anthropic/claude-sonnet-3.7-reasoning" | "anthropic/claude-opus-4-search" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-haiku-3.5-search" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4-reasoning" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7-search" | "anthropic/claude-opus-4-reasoning" | "anthropic/claude-sonnet-3.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-haiku-3.5" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-sonnet-4-search" | "trytako/tako" | "novita/glm-4.7" | "novita/kimi-k2-thinking" | "novita/kat-coder-pro" | "novita/glm-4.6" | "novita/minimax-m2.1" | "novita/glm-4.6v" | "cerebras/gpt-oss-120b-cs" | "cerebras/zai-glm-4.6-cs">;
517
+ declare const poe: ModelProvider & ChatProvider<"openai/gpt-5" | "openai/gpt-5-mini" | "openai/gpt-4.1" | "openai/gpt-5-nano" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "google/gemini-2.5-pro" | "google/gemini-2.5-flash" | "openai/o3-mini" | "openai/gpt-4o" | "openai/gpt-4o-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4.1-mini" | "openai/gpt-4.1-nano" | "openai/o4-mini" | "xai/grok-3-mini" | "xai/grok-3" | "openai/gpt-5.2-codex" | "openai/o1-pro" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.4-pro" | "openai/gpt-5-pro" | "openai/gpt-4-turbo" | "openai/gpt-5.3-codex" | "openai/o3-mini-high" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo" | "openai/gpt-3.5-turbo-instruct" | "openai/gpt-5.4" | "openai/gpt-5-chat" | "openai/gpt-5.2" | "openai/o4-mini-deep-research" | "openai/gpt-5.1" | "openai/o3-pro" | "openai/gpt-5-codex" | "openai/o3-deep-research" | "openai/gpt-5.1-codex" | "openai/gpt-5.2-pro" | "anthropic/claude-opus-4.6" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.6" | "anthropic/claude-opus-4.5" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-opus-4" | "google/gemini-2.5-flash-lite" | "openai/chatgpt-4o-latest" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "openai/sora-2" | "openai/gpt-5.4-mini" | "openai/gpt-image-1.5" | "openai/gpt-image-1-mini" | "openai/dall-e-3" | "openai/gpt-4-classic" | "openai/gpt-4o-aug" | "openai/gpt-4o-mini-search" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-5.4-nano" | "openai/gpt-5.1-instant" | "openai/gpt-5.2-instant" | "openai/gpt-4-classic-0314" | "openai/gpt-4o-search" | "openai/gpt-image-1" | "openai/gpt-5.3-instant" | "openai/sora-2-pro" | "anthropic/claude-haiku-3.5" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-3.5" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-4.1-fast-non-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4.20-multi-agent" | "xai/grok-4-fast-reasoning" | "stabilityai/stablediffusionxl" | "trytako/tako" | "google/lyria" | "google/imagen-4-ultra" | "google/nano-banana-pro" | "google/imagen-3-fast" | "google/imagen-3" | "google/gemini-2.0-flash-lite" | "google/veo-3.1-fast" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-3.1" | "google/imagen-4" | "google/veo-3" | "google/gemini-2.0-flash" | "google/gemini-3-pro" | "google/gemini-deep-research" | "google/veo-2" | "google/gemini-3.1-pro" | "google/nano-banana" | "google/gemini-3-flash" | "google/gemini-3.1-flash-lite" | "poetools/claude-code" | "novita/glm-4.7-n" | "novita/kimi-k2-thinking" | "novita/kimi-k2.5" | "novita/glm-4.7-flash" | "novita/minimax-m2.1" | "novita/glm-4.6" | "novita/glm-4.6v" | "novita/glm-4.7" | "novita/deepseek-v3.2" | "topazlabs-co/topazlabs" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "lumalabs/ray2" | "cerebras/gpt-oss-120b-cs" | "cerebras/llama-3.3-70b-cs" | "cerebras/qwen3-235b-2507-cs" | "cerebras/llama-3.1-8b-cs" | "cerebras/qwen3-32b-cs" | "ideogramai/ideogram-v2" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram">;
374
518
  /**
375
519
  * Privatemode AI Provider
376
520
  * @see {@link https://docs.privatemode.ai/api/overview}
@@ -378,7 +522,23 @@ declare const poe: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "goo
378
522
  * - baseURL - `http://localhost:8080/v1`
379
523
  * - apiKey - `PRIVATEMODE_API_KEY`
380
524
  */
381
- declare const privatemodeAi: ModelProvider & ChatProvider<"gpt-oss-120b" | "whisper-large-v3" | "qwen3-embedding-4b" | "gemma-3-27b" | "qwen3-coder-30b-a3b">;
525
+ declare const privatemodeAi: ModelProvider & ChatProvider<"gpt-oss-120b" | "whisper-large-v3" | "gemma-3-27b" | "qwen3-embedding-4b" | "qwen3-coder-30b-a3b">;
526
+ /**
527
+ * QiHang Provider
528
+ * @see {@link https://www.qhaigc.net/docs}
529
+ * @remarks
530
+ * - baseURL - `https://api.qhaigc.net/v1`
531
+ * - apiKey - `QIHANG_API_KEY`
532
+ */
533
+ declare const qihangAi: ModelProvider & ChatProvider<"gpt-5.2-codex" | "claude-opus-4-5-20251101" | "gpt-5-mini" | "claude-sonnet-4-5-20250929" | "gemini-3-flash-preview" | "gemini-2.5-flash" | "gpt-5.2" | "claude-haiku-4-5-20251001" | "gemini-3-pro-preview">;
534
+ /**
535
+ * Qiniu Provider
536
+ * @see {@link https://developer.qiniu.com/aitokenapi}
537
+ * @remarks
538
+ * - baseURL - `https://api.qnaigc.com/v1`
539
+ * - apiKey - `QINIU_API_KEY`
540
+ */
541
+ declare const qiniuAi: ModelProvider & ChatProvider<"gemini-2.5-pro" | "gemini-2.5-flash" | "qwen3-max" | "qwen-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-next-80b-a3b-instruct" | "qwen3-235b-a22b" | "qwen3.5-397b-a17b" | "qwen3-32b" | "deepseek-r1" | "deepseek-v3" | "deepseek-r1-0528" | "gpt-oss-120b" | "deepseek-v3-0324" | "glm-4.5" | "glm-4.5-air" | "openai/gpt-5" | "z-ai/glm-5" | "gpt-oss-20b" | "gemini-2.5-flash-lite" | "gemini-2.0-flash-lite" | "gemini-2.0-flash" | "gemini-2.5-flash-image" | "claude-4.5-opus" | "qwen3-30b-a3b" | "claude-3.7-sonnet" | "claude-3.5-haiku" | "claude-4.5-sonnet" | "claude-4.5-haiku" | "qwen3-max-preview" | "kimi-k2" | "qwen3-235b-a22b-thinking-2507" | "minimax/minimax-m2.1" | "moonshotai/kimi-k2.5" | "moonshotai/kimi-k2-0905" | "meituan/longcat-flash-chat" | "openai/gpt-5.2" | "x-ai/grok-4-fast" | "x-ai/grok-4.1-fast" | "x-ai/grok-code-fast-1" | "deepseek/deepseek-v3.1-terminus" | "deepseek/deepseek-v3.2-exp" | "z-ai/glm-4.6" | "z-ai/glm-4.7" | "minimax/minimax-m2" | "minimax/minimax-m2.5" | "xiaomi/mimo-v2-flash" | "stepfun/step-3.5-flash" | "moonshotai/kimi-k2-thinking" | "qwen3-235b-a22b-instruct-2507" | "qwen3-30b-a3b-thinking-2507" | "deepseek-v3.1" | "qwen3-vl-30b-a3b-thinking" | "qwen3-30b-a3b-instruct-2507" | "doubao-1.5-pro-32k" | "MiniMax-M1" | "x-ai/grok-4.1-fast-reasoning" | "qwen2.5-vl-72b-instruct" | "doubao-1.5-vision-pro" | "claude-4.0-sonnet" | "gemini-3.0-flash-preview" | "gemini-3.0-pro-image-preview" | "doubao-seed-2.0-lite" | "doubao-seed-2.0-mini" | "doubao-seed-2.0-pro" | "doubao-seed-2.0-code" | "doubao-seed-1.6" | "doubao-seed-1.6-flash" | "gemini-3.0-pro-preview" | "qwen-max-2025-01-25" | "claude-4.0-opus" | "qwen-vl-max-2025-01-25" | "kling-v2-6" | "doubao-seed-1.6-thinking" | "doubao-1.5-thinking-pro" | "claude-4.1-opus" | "qwen2.5-vl-7b-instruct" | "claude-3.5-sonnet" | "mimo-v2-flash" | "meituan/longcat-flash-lite" | "x-ai/grok-4-fast-non-reasoning" | "x-ai/grok-4.1-fast-non-reasoning" | "x-ai/grok-4-fast-reasoning" | "deepseek/deepseek-v3.1-terminus-thinking" | "deepseek/deepseek-v3.2-exp-thinking" | "deepseek/deepseek-math-v2" | "deepseek/deepseek-v3.2-251201" | "z-ai/autoglm-phone-9b" | "stepfun-ai/gelab-zero-4b-preview" | "minimax/minimax-m2.5-highspeed">;
382
542
  /**
383
543
  * Requesty Provider
384
544
  * @see {@link https://requesty.ai/solution/llm-routing/models}
@@ -386,7 +546,7 @@ declare const privatemodeAi: ModelProvider & ChatProvider<"gpt-oss-120b" | "whis
386
546
  * - baseURL - `https://router.requesty.ai/v1`
387
547
  * - apiKey - `REQUESTY_API_KEY`
388
548
  */
389
- declare const requesty: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-sonnet-4" | "openai/gpt-4.1-mini" | "openai/o4-mini" | "openai/gpt-4o-mini" | "xai/grok-4" | "anthropic/claude-opus-4" | "xai/grok-4-fast" | "google/gemini-3-flash-preview" | "google/gemini-3-pro-preview" | "anthropic/claude-opus-4-1" | "anthropic/claude-haiku-4-5" | "anthropic/claude-opus-4-5" | "anthropic/claude-sonnet-4-5" | "anthropic/claude-3-7-sonnet">;
549
+ declare const requesty: ModelProvider & ChatProvider<"openai/gpt-5" | "openai/gpt-5-mini" | "openai/gpt-4.1" | "openai/gpt-5-nano" | "anthropic/claude-sonnet-4" | "google/gemini-2.5-pro" | "google/gemini-2.5-flash" | "openai/gpt-4o-mini" | "openai/gpt-4.1-mini" | "openai/o4-mini" | "openai/gpt-5.2-codex" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.4-pro" | "openai/gpt-5-pro" | "openai/gpt-5.3-codex" | "openai/gpt-5.1-codex-max" | "openai/gpt-5.1-chat" | "openai/gpt-5.4" | "openai/gpt-5-chat" | "openai/gpt-5.2-chat" | "openai/gpt-5.2" | "openai/gpt-5.1" | "openai/gpt-5-image" | "openai/gpt-5-codex" | "openai/gpt-5.1-codex" | "openai/gpt-5.2-pro" | "anthropic/claude-opus-4" | "google/gemini-3-pro-preview" | "google/gemini-3-flash-preview" | "anthropic/claude-opus-4-5" | "anthropic/claude-opus-4-6" | "anthropic/claude-sonnet-4-6" | "anthropic/claude-sonnet-4-5" | "anthropic/claude-haiku-4-5" | "xai/grok-4" | "anthropic/claude-opus-4-1" | "anthropic/claude-3-7-sonnet" | "xai/grok-4-fast">;
390
550
  /**
391
551
  * Scaleway Provider
392
552
  * @see {@link https://www.scaleway.com/en/docs/generative-apis/}
@@ -394,7 +554,7 @@ declare const requesty: ModelProvider & ChatProvider<"google/gemini-2.5-flash" |
394
554
  * - baseURL - `https://api.scaleway.ai/v1`
395
555
  * - apiKey - `SCALEWAY_API_KEY`
396
556
  */
397
- declare const scaleway: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "whisper-large-v3" | "qwen3-235b-a22b-instruct-2507" | "devstral-2-123b-instruct-2512" | "pixtral-12b-2409" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
557
+ declare const scaleway: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "qwen3.5-397b-a17b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "gemma-3-27b-it" | "whisper-large-v3" | "llama-3.3-70b-instruct" | "llama-3.1-8b-instruct" | "qwen3-235b-a22b-instruct-2507" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen3-embedding-8b" | "devstral-2-123b-instruct-2512" | "bge-multilingual-gemma2" | "pixtral-12b-2409" | "voxtral-small-24b-2507">;
398
558
  /**
399
559
  * SiliconFlow Provider
400
560
  * @see {@link https://cloud.siliconflow.com/models}
@@ -402,7 +562,7 @@ declare const scaleway: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instru
402
562
  * - baseURL - `https://api.siliconflow.com/v1`
403
563
  * - apiKey - `SILICONFLOW_API_KEY`
404
564
  */
405
- declare const siliconflow: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"openai/gpt-oss-120b" | "meta-llama/Meta-Llama-3.1-8B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/QwQ-32B" | "Qwen/Qwen3-32B" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-V3" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2" | "zai-org/GLM-4.5" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "inclusionAI/Ling-mini-2.0" | "inclusionAI/Ling-flash-2.0" | "inclusionAI/Ring-flash-2.0" | "moonshotai/Kimi-Dev-72B" | "tencent/Hunyuan-MT-7B" | "tencent/Hunyuan-A13B-Instruct" | "MiniMaxAI/MiniMax-M1-80k" | "THUDM/GLM-4-32B-0414" | "THUDM/GLM-4.1V-9B-Thinking" | "THUDM/GLM-Z1-9B-0414" | "THUDM/GLM-4-9B-0414" | "THUDM/GLM-Z1-32B-0414" | "stepfun-ai/step3" | "nex-agi/DeepSeek-V3.1-Nex-N1" | "baidu/ERNIE-4.5-300B-A47B" | "ByteDance-Seed/Seed-OSS-36B-Instruct" | "Qwen/Qwen3-VL-30B-A3B-Instruct" | "Qwen/Qwen3-Omni-30B-A3B-Captioner" | "Qwen/Qwen3-8B" | "Qwen/Qwen3-Omni-30B-A3B-Instruct" | "Qwen/Qwen3-VL-8B-Thinking" | "Qwen/Qwen2.5-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct-128K" | "Qwen/Qwen2.5-7B-Instruct" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen2.5-VL-7B-Instruct" | "Qwen/Qwen3-VL-8B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-VL-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Thinking" | "Qwen/Qwen3-Omni-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Instruct" | "Qwen/Qwen2.5-14B-Instruct" | "zai-org/GLM-4.5V" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" | "deepseek-ai/deepseek-vl2" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B" | "deepseek-ai/DeepSeek-V3.2-Exp" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B" | "deepseek-ai/DeepSeek-V3.1">;
565
+ declare const siliconflow: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"openai/gpt-oss-120b" | "meta-llama/Meta-Llama-3.1-8B-Instruct" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2" | "Qwen/Qwen3-32B" | "Qwen/QwQ-32B" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "zai-org/GLM-4.7" | "zai-org/GLM-5" | "zai-org/GLM-4.6" | "MiniMaxAI/MiniMax-M2.5" | "deepseek-ai/DeepSeek-V3.1" | "moonshotai/Kimi-K2.5" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "openai/gpt-oss-20b" | "zai-org/GLM-4.5-Air" | "zai-org/GLM-4.6V" | "deepseek-ai/DeepSeek-V3" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen3-14B" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "zai-org/GLM-4.5" | "MiniMaxAI/MiniMax-M2.1" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-VL-30B-A3B-Instruct" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/DeepSeek-V3.2-Exp" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "THUDM/GLM-Z1-32B-0414" | "THUDM/GLM-Z1-9B-0414" | "THUDM/GLM-4-32B-0414" | "THUDM/GLM-4-9B-0414" | "tencent/Hunyuan-MT-7B" | "Qwen/Qwen2.5-VL-72B-Instruct" | "zai-org/GLM-4.5V" | "nex-agi/DeepSeek-V3.1-Nex-N1" | "stepfun-ai/Step-3.5-Flash" | "baidu/ERNIE-4.5-300B-A47B" | "deepseek-ai/deepseek-vl2" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B" | "ByteDance-Seed/Seed-OSS-36B-Instruct" | "Qwen/Qwen3-Omni-30B-A3B-Thinking" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen2.5-14B-Instruct" | "Qwen/Qwen2.5-72B-Instruct-128K" | "Qwen/Qwen3-VL-8B-Thinking" | "Qwen/Qwen3-Omni-30B-A3B-Captioner" | "Qwen/Qwen3-VL-30B-A3B-Thinking" | "Qwen/Qwen2.5-VL-7B-Instruct" | "Qwen/Qwen3-VL-8B-Instruct" | "Qwen/Qwen3-8B" | "Qwen/Qwen3-VL-32B-Instruct" | "Qwen/Qwen3-Omni-30B-A3B-Instruct" | "Qwen/Qwen2.5-7B-Instruct" | "Qwen/Qwen3-VL-32B-Thinking" | "Qwen/Qwen2.5-32B-Instruct" | "inclusionAI/Ling-flash-2.0" | "inclusionAI/Ring-flash-2.0" | "inclusionAI/Ling-mini-2.0" | "tencent/Hunyuan-A13B-Instruct">;
406
566
  /**
407
567
  * SiliconFlow (China) Provider
408
568
  * @see {@link https://cloud.siliconflow.com/models}
@@ -410,7 +570,23 @@ declare const siliconflow: ModelProvider & EmbedProvider<string> & SpeechProvide
410
570
  * - baseURL - `https://api.siliconflow.cn/v1`
411
571
  * - apiKey - `SILICONFLOW_CN_API_KEY`
412
572
  */
413
- declare const siliconflowCn: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/QwQ-32B" | "Qwen/Qwen3-32B" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-V3" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2" | "zai-org/GLM-4.5" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "inclusionAI/Ling-mini-2.0" | "inclusionAI/Ling-flash-2.0" | "inclusionAI/Ring-flash-2.0" | "moonshotai/Kimi-Dev-72B" | "tencent/Hunyuan-MT-7B" | "tencent/Hunyuan-A13B-Instruct" | "MiniMaxAI/MiniMax-M1-80k" | "THUDM/GLM-4-32B-0414" | "THUDM/GLM-4.1V-9B-Thinking" | "THUDM/GLM-Z1-9B-0414" | "THUDM/GLM-4-9B-0414" | "THUDM/GLM-Z1-32B-0414" | "stepfun-ai/step3" | "nex-agi/DeepSeek-V3.1-Nex-N1" | "baidu/ERNIE-4.5-300B-A47B" | "ByteDance-Seed/Seed-OSS-36B-Instruct" | "Qwen/Qwen3-VL-30B-A3B-Instruct" | "Qwen/Qwen3-Omni-30B-A3B-Captioner" | "Qwen/Qwen3-8B" | "Qwen/Qwen3-Omni-30B-A3B-Instruct" | "Qwen/Qwen3-VL-8B-Thinking" | "Qwen/Qwen2.5-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct-128K" | "Qwen/Qwen2.5-7B-Instruct" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen2.5-VL-7B-Instruct" | "Qwen/Qwen3-VL-8B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-VL-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Thinking" | "Qwen/Qwen3-Omni-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Instruct" | "Qwen/Qwen2.5-14B-Instruct" | "zai-org/GLM-4.5V" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" | "deepseek-ai/deepseek-vl2" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B" | "Kwaipilot/KAT-Dev" | "ascend-tribe/pangu-pro-moe">;
573
+ declare const siliconflowCn: ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2" | "Qwen/Qwen3-32B" | "Qwen/QwQ-32B" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "zai-org/GLM-4.6" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.5-Air" | "zai-org/GLM-4.6V" | "deepseek-ai/DeepSeek-V3" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen3-14B" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-VL-30B-A3B-Instruct" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "Qwen/Qwen3.5-397B-A17B" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "THUDM/GLM-Z1-32B-0414" | "THUDM/GLM-Z1-9B-0414" | "THUDM/GLM-4-32B-0414" | "THUDM/GLM-4-9B-0414" | "tencent/Hunyuan-MT-7B" | "Qwen/Qwen2.5-VL-72B-Instruct" | "zai-org/GLM-4.5V" | "stepfun-ai/Step-3.5-Flash" | "baidu/ERNIE-4.5-300B-A47B" | "deepseek-ai/deepseek-vl2" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B" | "ByteDance-Seed/Seed-OSS-36B-Instruct" | "Qwen/Qwen3-Omni-30B-A3B-Thinking" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen2.5-14B-Instruct" | "Qwen/Qwen2.5-72B-Instruct-128K" | "Qwen/Qwen3-VL-8B-Thinking" | "Qwen/Qwen3-Omni-30B-A3B-Captioner" | "Qwen/Qwen3-VL-30B-A3B-Thinking" | "Qwen/Qwen3-VL-8B-Instruct" | "Qwen/Qwen3-8B" | "Qwen/Qwen3-VL-32B-Instruct" | "Qwen/Qwen3-Omni-30B-A3B-Instruct" | "Qwen/Qwen2.5-7B-Instruct" | "Qwen/Qwen3-VL-32B-Thinking" | "Qwen/Qwen2.5-32B-Instruct" | "inclusionAI/Ling-flash-2.0" | "inclusionAI/Ring-flash-2.0" | "inclusionAI/Ling-mini-2.0" | "tencent/Hunyuan-A13B-Instruct" | "Pro/zai-org/GLM-4.7" | "Pro/zai-org/GLM-5" | "Pro/MiniMaxAI/MiniMax-M2.1" | "Pro/MiniMaxAI/MiniMax-M2.5" | "Pro/deepseek-ai/DeepSeek-R1" | "Pro/deepseek-ai/DeepSeek-V3.1-Terminus" | "Pro/deepseek-ai/DeepSeek-V3.2" | "Pro/deepseek-ai/DeepSeek-V3" | "Pro/moonshotai/Kimi-K2.5" | "Pro/moonshotai/Kimi-K2-Instruct-0905" | "Pro/moonshotai/Kimi-K2-Thinking" | "PaddlePaddle/PaddleOCR-VL-1.5" | "PaddlePaddle/PaddleOCR-VL" | "deepseek-ai/DeepSeek-OCR" | "Qwen/Qwen3.5-35B-A3B" | "Qwen/Qwen3.5-122B-A10B" | "Qwen/Qwen3.5-9B" | "Qwen/Qwen3.5-27B" | "Qwen/Qwen3.5-4B" | "ascend-tribe/pangu-pro-moe" | "Kwaipilot/KAT-Dev">;
574
+ /**
575
+ * STACKIT Provider
576
+ * @see {@link https://docs.stackit.cloud/products/data-and-ai/ai-model-serving/basics/available-shared-models}
577
+ * @remarks
578
+ * - baseURL - `https://api.openai-compat.model-serving.eu01.onstackit.cloud/v1`
579
+ * - apiKey - `STACKIT_API_KEY`
580
+ */
581
+ declare const stackit: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "google/gemma-3-27b-it" | "intfloat/e5-mistral-7b-instruct" | "cortecs/Llama-3.3-70B-Instruct-FP8-Dynamic" | "neuralmagic/Mistral-Nemo-Instruct-2407-FP8" | "neuralmagic/Meta-Llama-3.1-8B-Instruct-FP8" | "Qwen/Qwen3-VL-235B-A22B-Instruct-FP8" | "Qwen/Qwen3-VL-Embedding-8B">;
582
+ /**
583
+ * StepFun Provider
584
+ * @see {@link https://platform.stepfun.com/docs/zh/overview/concept}
585
+ * @remarks
586
+ * - baseURL - `https://api.stepfun.com/v1`
587
+ * - apiKey - `STEPFUN_API_KEY`
588
+ */
589
+ declare const stepfun: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"step-1-32k" | "step-3.5-flash" | "step-2-16k">;
414
590
  /**
415
591
  * submodel Provider
416
592
  * @see {@link https://submodel.gitbook.io}
@@ -418,15 +594,23 @@ declare const siliconflowCn: ModelProvider & ChatProvider<"openai/gpt-oss-120b"
418
594
  * - baseURL - `https://llm.submodel.ai/v1`
419
595
  * - apiKey - `SUBMODEL_INSTAGEN_ACCESS_KEY`
420
596
  */
421
- declare const submodel: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "zai-org/GLM-4.5-FP8" | "deepseek-ai/DeepSeek-V3-0324">;
597
+ declare const submodel: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324" | "zai-org/GLM-4.5-Air" | "zai-org/GLM-4.5-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "deepseek-ai/DeepSeek-R1-0528" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8">;
422
598
  /**
423
599
  * Synthetic Provider
424
600
  * @see {@link https://synthetic.new/pricing}
425
601
  * @remarks
426
- * - baseURL - `https://api.synthetic.new/v1`
602
+ * - baseURL - `https://api.synthetic.new/openai/v1`
427
603
  * - apiKey - `SYNTHETIC_API_KEY`
428
604
  */
429
- declare const synthetic: ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:MiniMaxAI/MiniMax-M2.1" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.7" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
605
+ declare const synthetic: ModelProvider & ChatProvider<"hf:moonshotai/Kimi-K2.5" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:zai-org/GLM-4.7" | "hf:zai-org/GLM-4.7-Flash" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:nvidia/Kimi-K2.5-NVFP4" | "hf:MiniMaxAI/MiniMax-M2.1" | "hf:MiniMaxAI/MiniMax-M2.5" | "hf:MiniMaxAI/MiniMax-M2" | "hf:openai/gpt-oss-120b" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3.5-397B-A17B" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Instruct-2507">;
606
+ /**
607
+ * Tencent Coding Plan (China) Provider
608
+ * @see {@link https://cloud.tencent.com/document/product/1772/128947}
609
+ * @remarks
610
+ * - baseURL - `https://api.lkeap.cloud.tencent.com/coding/v3`
611
+ * - apiKey - `TENCENT_CODING_PLAN_API_KEY`
612
+ */
613
+ declare const tencentCodingPlan: ModelProvider & ChatProvider<"kimi-k2.5" | "glm-5" | "minimax-m2.5" | "hunyuan-2.0-instruct" | "hunyuan-t1" | "hunyuan-2.0-thinking" | "tc-code-latest" | "hunyuan-turbos">;
430
614
  /**
431
615
  * Upstage Provider
432
616
  * @see {@link https://developers.upstage.ai/docs/apis/chat}
@@ -434,7 +618,7 @@ declare const synthetic: ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-I
434
618
  * - baseURL - `https://api.upstage.ai/v1/solar`
435
619
  * - apiKey - `UPSTAGE_API_KEY`
436
620
  */
437
- declare const upstage: ModelProvider & ChatProvider<"solar-mini" | "solar-pro3" | "solar-pro2">;
621
+ declare const upstage: ModelProvider & ChatProvider<"solar-pro3" | "solar-pro2" | "solar-mini">;
438
622
  /**
439
623
  * Venice AI Provider
440
624
  * @see {@link https://docs.venice.ai}
@@ -442,7 +626,7 @@ declare const upstage: ModelProvider & ChatProvider<"solar-mini" | "solar-pro3"
442
626
  * - baseURL - `https://api.venice.ai/api/v1`
443
627
  * - apiKey - `VENICE_API_KEY`
444
628
  */
445
- declare const venice: ModelProvider & ChatProvider<"gemini-3-flash-preview" | "grok-code-fast-1" | "gemini-3-pro-preview" | "qwen3-coder-480b-a35b-instruct" | "kimi-k2-thinking" | "deepseek-v3.2" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "grok-41-fast" | "claude-opus-45" | "mistral-31-24b" | "zai-org-glm-4.7" | "venice-uncensored" | "openai-gpt-52" | "qwen3-4b" | "llama-3.3-70b" | "claude-sonnet-45" | "openai-gpt-oss-120b" | "llama-3.2-3b" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "zai-org-glm-4.6v" | "minimax-m21" | "qwen3-next-80b" | "openai-gpt-52-codex">;
629
+ declare const venice: ModelProvider & ChatProvider<"grok-code-fast-1" | "claude-opus-4-6" | "claude-sonnet-4-6" | "gemini-3-flash-preview" | "qwen3-coder-480b-a35b-instruct" | "qwen3-vl-235b-a22b" | "kimi-k2-thinking" | "gemini-3-1-pro-preview" | "deepseek-v3.2" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "venice-uncensored" | "qwen3-coder-480b-a35b-instruct-turbo" | "aion-labs.aion-2-0" | "qwen3-next-80b" | "zai-org-glm-4.7" | "openai-gpt-53-codex" | "qwen3-5-35b-a3b" | "olafangensan-glm-4.7-flash-heretic" | "openai-gpt-4o-2024-11-20" | "grok-4-20-beta" | "grok-4-20-multi-agent-beta" | "openai-gpt-54-mini" | "openai-gpt-54-pro" | "kimi-k2-5" | "claude-opus-45" | "llama-3.2-3b" | "qwen3-5-9b" | "minimax-m27" | "openai-gpt-oss-120b" | "openai-gpt-52" | "zai-org-glm-4.6" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "claude-sonnet-45" | "llama-3.3-70b" | "zai-org-glm-5" | "minimax-m25" | "openai-gpt-54" | "nvidia-nemotron-3-nano-30b-a3b" | "venice-uncensored-role-play" | "grok-41-fast" | "openai-gpt-4o-mini-2024-07-18" | "zai-org-glm-4.7-flash" | "mistral-small-3-2-24b-instruct" | "openai-gpt-52-codex" | "minimax-m21">;
446
630
  /**
447
631
  * Vivgrid Provider
448
632
  * @see {@link https://docs.vivgrid.com/models}
@@ -450,7 +634,7 @@ declare const venice: ModelProvider & ChatProvider<"gemini-3-flash-preview" | "g
450
634
  * - baseURL - `https://api.vivgrid.com/v1`
451
635
  * - apiKey - `VIVGRID_API_KEY`
452
636
  */
453
- declare const vivgrid: ModelProvider & ChatProvider<"gpt-5.1-codex">;
637
+ declare const vivgrid: ModelProvider & ChatProvider<"gpt-5.2-codex" | "gemini-3.1-pro-preview" | "gpt-5-mini" | "gpt-5.1-codex-max" | "gpt-5.4" | "gemini-3.1-flash-lite-preview" | "gpt-5.1-codex" | "glm-5" | "deepseek-v3.2">;
454
638
  /**
455
639
  * Vultr Provider
456
640
  * @see {@link https://api.vultrinference.com/}
@@ -458,15 +642,15 @@ declare const vivgrid: ModelProvider & ChatProvider<"gpt-5.1-codex">;
458
642
  * - baseURL - `https://api.vultrinference.com/v1`
459
643
  * - apiKey - `VULTR_API_KEY`
460
644
  */
461
- declare const vultr: ModelProvider & ChatProvider<"deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-instruct" | "qwen2.5-coder-32b-instruct">;
645
+ declare const vultr: ModelProvider & ChatProvider<"MiniMax-M2.5" | "Kimi-K2.5" | "GLM-5-FP8" | "DeepSeek-V3.2">;
462
646
  /**
463
647
  * Weights & Biases Provider
464
- * @see {@link https://weave-docs.wandb.ai/guides/integrations/inference/}
648
+ * @see {@link https://docs.wandb.ai/guides/integrations/inference/}
465
649
  * @remarks
466
650
  * - baseURL - `https://api.inference.wandb.ai/v1`
467
651
  * - apiKey - `WANDB_API_KEY`
468
652
  */
469
- declare const wandb: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "moonshotai/Kimi-K2-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "meta-llama/Llama-3.3-70B-Instruct" | "deepseek-ai/DeepSeek-V3-0324" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct">;
653
+ declare const wandb: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "MiniMaxAI/MiniMax-M2.5" | "deepseek-ai/DeepSeek-V3.1" | "moonshotai/Kimi-K2.5" | "meta-llama/Llama-3.3-70B-Instruct" | "openai/gpt-oss-20b" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "meta-llama/Llama-4-Scout-17B-16E-Instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-3.1-70B-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "microsoft/Phi-4-mini-instruct" | "zai-org/GLM-5-FP8" | "nvidia/NVIDIA-Nemotron-3-Super-120B-A12B-FP8" | "OpenPipe/Qwen3-14B-Instruct">;
470
654
  /**
471
655
  * xAI Provider
472
656
  * @see {@link https://docs.x.ai/docs/models}
@@ -474,7 +658,7 @@ declare const wandb: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct
474
658
  * - baseURL - `https://api.x.ai/v1/`
475
659
  * - apiKey - `XAI_API_KEY`
476
660
  */
477
- declare const xai: ModelProvider & ChatProvider<"grok-4-fast-non-reasoning" | "grok-code-fast-1" | "grok-4-1-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
661
+ declare const xai: ModelProvider & ChatProvider<"grok-4-fast-non-reasoning" | "grok-code-fast-1" | "grok-4-1-fast-non-reasoning" | "grok-4" | "grok-3-mini" | "grok-3" | "grok-4-fast" | "grok-4-1-fast" | "grok-2-vision-1212" | "grok-2-latest" | "grok-3-mini-fast" | "grok-beta" | "grok-2-vision-latest" | "grok-2" | "grok-3-fast-latest" | "grok-2-1212" | "grok-2-vision" | "grok-3-latest" | "grok-4.20-multi-agent-0309" | "grok-3-fast" | "grok-3-mini-fast-latest" | "grok-3-mini-latest" | "grok-4.20-0309-reasoning" | "grok-4.20-0309-non-reasoning" | "grok-vision-beta">;
478
662
  /**
479
663
  * Xiaomi Provider
480
664
  * @see {@link https://platform.xiaomimimo.com/#/docs}
@@ -482,7 +666,7 @@ declare const xai: ModelProvider & ChatProvider<"grok-4-fast-non-reasoning" | "g
482
666
  * - baseURL - `https://api.xiaomimimo.com/v1`
483
667
  * - apiKey - `XIAOMI_API_KEY`
484
668
  */
485
- declare const xiaomi: ModelProvider & ChatProvider<"mimo-v2-flash">;
669
+ declare const xiaomi: ModelProvider & ChatProvider<"mimo-v2-flash" | "mimo-v2-pro" | "mimo-v2-omni">;
486
670
  /**
487
671
  * Z.AI Provider
488
672
  * @see {@link https://docs.z.ai/guides/overview/pricing}
@@ -490,7 +674,7 @@ declare const xiaomi: ModelProvider & ChatProvider<"mimo-v2-flash">;
490
674
  * - baseURL - `https://api.z.ai/api/paas/v4`
491
675
  * - apiKey - `ZHIPU_API_KEY`
492
676
  */
493
- declare const zai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
677
+ declare const zai: ModelProvider & ChatProvider<"glm-5" | "glm-4.7" | "glm-4.7-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.6" | "glm-4.7-flashx" | "glm-4.6v" | "glm-4.5-flash" | "glm-4.5v" | "glm-5-turbo">;
494
678
  /**
495
679
  * Z.AI Coding Plan Provider
496
680
  * @see {@link https://docs.z.ai/devpack/overview}
@@ -498,7 +682,7 @@ declare const zai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5
498
682
  * - baseURL - `https://api.z.ai/api/coding/paas/v4`
499
683
  * - apiKey - `ZHIPU_API_KEY`
500
684
  */
501
- declare const zaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
685
+ declare const zaiCodingPlan: ModelProvider & ChatProvider<"glm-5" | "glm-4.7" | "glm-4.7-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.6" | "glm-4.7-flashx" | "glm-4.6v" | "glm-4.5-flash" | "glm-4.5v" | "glm-5-turbo" | "glm-5.1">;
502
686
  /**
503
687
  * ZenMux Provider
504
688
  * @see {@link https://docs.zenmux.ai}
@@ -506,7 +690,7 @@ declare const zaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7"
506
690
  * - baseURL - `https://zenmux.ai/api/v1`
507
691
  * - apiKey - `ZENMUX_API_KEY`
508
692
  */
509
- declare const zenmux: ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "moonshotai/kimi-k2-thinking" | "minimax/minimax-m2.1" | "z-ai/glm-4.6" | "moonshotai/kimi-k2-0905" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-v3.2" | "minimax/minimax-m2" | "google/gemini-2.5-flash-lite" | "openai/gpt-5.1-codex" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-5.2" | "anthropic/claude-opus-4" | "anthropic/claude-haiku-4.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-sonnet-4.5" | "google/gemini-3-flash-preview" | "google/gemini-3-pro-preview" | "stepfun/step-3" | "moonshotai/kimi-k2-thinking-turbo" | "xiaomi/mimo-v2-flash-free" | "xiaomi/mimo-v2-flash" | "x-ai/grok-code-fast-1" | "x-ai/grok-4.1-fast-non-reasoning" | "x-ai/grok-4-fast" | "x-ai/grok-4.1-fast" | "deepseek/deepseek-chat" | "deepseek/deepseek-reasoner" | "google/gemini-3-flash-preview-free" | "volcengine/doubao-seed-code" | "volcengine/doubao-seed-1.8" | "openai/gpt-5.1-chat" | "baidu/ernie-5.0-thinking-preview" | "inclusionai/ring-1t" | "inclusionai/ling-1t" | "z-ai/glm-4.7" | "z-ai/glm-4.6v-flash-free" | "z-ai/glm-4.6v-flash" | "z-ai/glm-4.5" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6v" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1-free" | "kuaishou/kat-coder-pro-v1">;
693
+ declare const zenmux: ModelProvider & ChatProvider<"openai/gpt-5" | "x-ai/grok-4" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "z-ai/glm-5" | "google/gemini-2.5-pro" | "google/gemini-2.5-flash" | "minimax/minimax-m2.1" | "moonshotai/kimi-k2.5" | "moonshotai/kimi-k2-0905" | "openai/gpt-5.2-codex" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.4-pro" | "openai/gpt-5.3-codex" | "openai/gpt-5.1-chat" | "openai/gpt-5.4" | "openai/gpt-5.3-chat" | "openai/gpt-5.2" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-5.1-codex" | "openai/gpt-5.2-pro" | "x-ai/grok-4-fast" | "x-ai/grok-4.1-fast" | "x-ai/grok-code-fast-1" | "anthropic/claude-opus-4.6" | "anthropic/claude-haiku-4.5" | "anthropic/claude-3.7-sonnet" | "anthropic/claude-sonnet-4.6" | "anthropic/claude-3.5-haiku" | "anthropic/claude-opus-4.5" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-opus-4" | "deepseek/deepseek-chat" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-v3.2" | "z-ai/glm-4.5" | "z-ai/glm-4.6" | "z-ai/glm-4.6v" | "z-ai/glm-4.5-air" | "z-ai/glm-4.7" | "google/gemini-2.5-flash-lite" | "google/gemini-3.1-pro-preview" | "google/gemini-3-pro-preview" | "google/gemini-3-flash-preview" | "google/gemini-3.1-flash-lite-preview" | "minimax/minimax-m2" | "minimax/minimax-m2.5" | "qwen/qwen3-coder-plus" | "qwen/qwen3-max" | "xiaomi/mimo-v2-flash" | "stepfun/step-3.5-flash" | "moonshotai/kimi-k2-thinking" | "minimax/minimax-m2.7" | "openai/gpt-5.4-mini" | "openai/gpt-5.4-nano" | "x-ai/grok-4.1-fast-non-reasoning" | "x-ai/grok-4.2-fast" | "x-ai/grok-4.2-fast-non-reasoning" | "volcengine/doubao-seed-2.0-lite" | "volcengine/doubao-seed-2.0-mini" | "volcengine/doubao-seed-2.0-pro" | "volcengine/doubao-seed-2.0-code" | "volcengine/doubao-seed-1.8" | "volcengine/doubao-seed-code" | "z-ai/glm-4.7-flash-free" | "z-ai/glm-5-turbo" | "z-ai/glm-4.7-flashx" | "z-ai/glm-4.6v-flash-free" | "z-ai/glm-4.6v-flash" | "inclusionai/ring-1t" | "inclusionai/ling-1t" | "baidu/ernie-5.0-thinking-preview" | "minimax/minimax-m2.5-lightning" | "minimax/minimax-m2.7-highspeed" | "qwen/qwen3.5-plus" | "qwen/qwen3.5-flash" | "xiaomi/mimo-v2-flash-free" | "xiaomi/mimo-v2-pro" | "xiaomi/mimo-v2-omni" | "stepfun/step-3.5-flash-free" | "stepfun/step-3" | "kuaishou/kat-coder-pro-v1-free" | "kuaishou/kat-coder-pro-v1" | "moonshotai/kimi-k2-thinking-turbo">;
510
694
  /**
511
695
  * Zhipu AI Provider
512
696
  * @see {@link https://docs.z.ai/guides/overview/pricing}
@@ -514,7 +698,7 @@ declare const zenmux: ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemin
514
698
  * - baseURL - `https://open.bigmodel.cn/api/paas/v4`
515
699
  * - apiKey - `ZHIPU_API_KEY`
516
700
  */
517
- declare const zhipuai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
701
+ declare const zhipuai: ModelProvider & ChatProvider<"glm-5" | "glm-4.7" | "glm-4.7-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.6" | "glm-4.7-flashx" | "glm-4.6v" | "glm-4.5-flash" | "glm-4.5v">;
518
702
  /**
519
703
  * Zhipu AI Coding Plan Provider
520
704
  * @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
@@ -522,15 +706,7 @@ declare const zhipuai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm
522
706
  * - baseURL - `https://open.bigmodel.cn/api/coding/paas/v4`
523
707
  * - apiKey - `ZHIPU_API_KEY`
524
708
  */
525
- declare const zhipuaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
526
- /**
527
- * StepFun Provider
528
- * @see {@link https://www.stepfun.com}
529
- * @remarks
530
- * - baseURL - `https://api.stepfun.com/v1/`
531
- * - apiKey - `STEPFUN_API_KEY`
532
- */
533
- declare const stepfun: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
709
+ declare const zhipuaiCodingPlan: ModelProvider & ChatProvider<"glm-5" | "glm-4.7" | "glm-4.7-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.6" | "glm-4.7-flashx" | "glm-4.6v" | "glm-4.5-flash" | "glm-4.5v" | "glm-4.6v-flash" | "glm-5-turbo" | "glm-5.1">;
534
710
  /**
535
711
  * Tencent Hunyuan Provider
536
712
  * @see {@link https://cloud.tencent.com/document/product/1729}
@@ -589,6 +765,6 @@ declare const openrouter: ModelProvider & ChatProviderWithExtraOptions<Openroute
589
765
  */
590
766
  declare const togetherai: ModelProvider & EmbedProvider<string> & ChatProvider<TogetheraiModels>;
591
767
  /** @deprecated use `novitaAi` instead. */
592
- declare const novita: ModelProvider & ChatProvider<"deepseek/deepseek-v3.1" | "openai/gpt-oss-120b" | "zai-org/glm-4.7" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "openai/gpt-oss-20b" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-v3-0324" | "moonshotai/kimi-k2-instruct" | "meta-llama/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-thinking" | "minimax/minimax-m2.1" | "qwen/qwen3-235b-a22b-thinking-2507" | "meta-llama/llama-3.3-70b-instruct" | "zai-org/glm-4.5-air" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "baichuan/baichuan-m2-32b" | "moonshotai/kimi-k2-0905" | "nousresearch/hermes-2-pro-llama-3-8b" | "paddlepaddle/paddleocr-vl" | "kwaipilot/kat-coder" | "kwaipilot/kat-coder-pro" | "xiaomimimo/mimo-v2-flash" | "deepseek/deepseek-prover-v2-671b" | "deepseek/deepseek-r1-0528-qwen3-8b" | "deepseek/deepseek-v3.1-terminus" | "deepseek/deepseek-r1-turbo" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-r1-distill-llama-70b" | "deepseek/deepseek-ocr" | "deepseek/deepseek-v3.2" | "deepseek/deepseek-v3-turbo" | "sao10k/l3-8b-lunaris" | "sao10k/L3-8B-Stheno-v3.2" | "sao10k/l31-70b-euryale-v2.2" | "sao10k/l3-70b-euryale-v2.1" | "skywork/r1v4-lite" | "minimaxai/minimax-m1-80k" | "minimax/minimax-m2" | "google/gemma-3-27b-it" | "microsoft/wizardlm-2-8x22b" | "gryphe/mythomax-l2-13b" | "baidu/ernie-4.5-vl-28b-a3b-thinking" | "baidu/ernie-4.5-300b-a47b-paddle" | "baidu/ernie-4.5-21B-a3b" | "baidu/ernie-4.5-21B-a3b-thinking" | "baidu/ernie-4.5-vl-424b-a47b" | "baidu/ernie-4.5-vl-28b-a3b" | "qwen/qwen3-vl-30b-a3b-thinking" | "qwen/qwen3-omni-30b-a3b-thinking" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-8b-fp8" | "qwen/qwen2.5-vl-72b-instruct" | "qwen/qwen3-4b-fp8" | "qwen/qwen3-coder-30b-a3b-instruct" | "qwen/qwen3-vl-8b-instruct" | "qwen/qwen2.5-7b-instruct" | "qwen/qwen3-30b-a3b-fp8" | "qwen/qwen3-32b-fp8" | "qwen/qwen3-omni-30b-a3b-instruct" | "qwen/qwen-2.5-72b-instruct" | "qwen/qwen3-vl-235b-a22b-thinking" | "qwen/qwen-mt-plus" | "qwen/qwen3-max" | "qwen/qwen3-vl-235b-a22b-instruct" | "qwen/qwen3-vl-30b-a3b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-235b-a22b-fp8" | "mistralai/mistral-nemo" | "meta-llama/llama-3-70b-instruct" | "meta-llama/llama-3-8b-instruct" | "meta-llama/llama-3.1-8b-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct-fp8" | "zai-org/glm-4.5v" | "zai-org/glm-4.6v" | "zai-org/autoglm-phone-9b-multilingual">;
768
+ declare const novita: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "deepseek/deepseek-v3.1" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "zai-org/glm-5" | "zai-org/glm-4.7" | "openai/gpt-oss-20b" | "deepseek/deepseek-v3-0324" | "deepseek/deepseek-r1-0528" | "meta-llama/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-instruct" | "zai-org/glm-4.7-flash" | "zai-org/glm-4.5v" | "baidu/ernie-4.5-300b-a47b-paddle" | "baidu/ernie-4.5-vl-424b-a47b" | "minimaxai/minimax-m1-80k" | "minimax/minimax-m2.1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-next" | "qwen/qwen3-30b-a3b-fp8" | "qwen/qwen3-235b-a22b-fp8" | "qwen/qwen3-coder-480b-a35b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-32b-fp8" | "moonshotai/kimi-k2.5" | "moonshotai/kimi-k2-0905" | "xiaomimimo/mimo-v2-flash" | "microsoft/wizardlm-2-8x22b" | "kwaipilot/kat-coder-pro" | "deepseek/deepseek-v3.1-terminus" | "deepseek/deepseek-r1-distill-llama-70b" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-v3.2" | "gryphe/mythomax-l2-13b" | "google/gemma-3-27b-it" | "baidu/ernie-4.5-vl-28b-a3b" | "meta-llama/llama-3.3-70b-instruct" | "meta-llama/llama-3.1-8b-instruct" | "meta-llama/llama-3-8b-instruct" | "meta-llama/llama-3-70b-instruct" | "nousresearch/hermes-2-pro-llama-3-8b" | "minimax/minimax-m2" | "minimax/minimax-m2.5" | "qwen/qwen3-coder-30b-a3b-instruct" | "qwen/qwen3-vl-235b-a22b-instruct" | "qwen/qwen2.5-vl-72b-instruct" | "qwen/qwen3-vl-8b-instruct" | "qwen/qwen3-vl-30b-a3b-instruct" | "qwen/qwen3-vl-30b-a3b-thinking" | "qwen/qwen3.5-397b-a17b" | "qwen/qwen3-max" | "qwen/qwen3-vl-235b-a22b-thinking" | "qwen/qwen-2.5-72b-instruct" | "moonshotai/kimi-k2-thinking" | "mistralai/mistral-nemo" | "deepseek/deepseek-prover-v2-671b" | "paddlepaddle/paddleocr-vl" | "kwaipilot/kat-coder" | "sao10k/l3-8b-lunaris" | "sao10k/l3-70b-euryale-v2.1" | "sao10k/l31-70b-euryale-v2.2" | "sao10k/L3-8B-Stheno-v3.2" | "deepseek/deepseek-r1-turbo" | "deepseek/deepseek-r1-0528-qwen3-8b" | "deepseek/deepseek-ocr-2" | "deepseek/deepseek-v3-turbo" | "deepseek/deepseek-ocr" | "zai-org/autoglm-phone-9b-multilingual" | "zai-org/glm-4.6v" | "zai-org/glm-4.5-air" | "skywork/r1v4-lite" | "baidu/ernie-4.5-21B-a3b-thinking" | "baidu/ernie-4.5-vl-28b-a3b-thinking" | "baidu/ernie-4.5-21B-a3b" | "meta-llama/llama-4-maverick-17b-128e-instruct-fp8" | "qwen/qwen-mt-plus" | "qwen/qwen3-8b-fp8" | "qwen/qwen3-omni-30b-a3b-instruct" | "qwen/qwen2.5-7b-instruct" | "qwen/qwen3-omni-30b-a3b-thinking" | "qwen/qwen3-4b-fp8" | "baichuan/baichuan-m2-32b">;
593
769
 
594
- export { abacus, alibaba, alibabaCn, anthropic, bailing, baseten, cerebras, chutes, cohere, cortecs, deepinfra, deepseek, fastrouter, featherless, fireworks, firmware, friendli, githubCopilot, githubModels, google, groq, helicone, huggingface, iflowcn, inception, inference, ioNet, kimiForCoding, litellm, llama, lmstudio, lucidquery, minimax, minimaxCn, mistral, modelscope, moonshotai, moonshotaiCn, morph, nanoGpt, nebius, novita, novitaAi, nvidia, ollama, ollamaCloud, openai, opencode, openrouter, ovhcloud, perplexity, poe, privatemodeAi, requesty, scaleway, siliconflow, siliconflowCn, stepfun, submodel, synthetic, tencentHunyuan, togetherai, upstage, venice, vivgrid, vultr, wandb, xai, xiaomi, zai, zaiCodingPlan, zenmux, zhipuai, zhipuaiCodingPlan };
770
+ export { abacus, alibaba, alibabaCn, alibabaCodingPlan, alibabaCodingPlanCn, anthropic, bailing, baseten, berget, cerebras, chutes, cloudferroSherlock, cohere, cortecs, deepinfra, deepseek, dinference, drun, evroc, fastrouter, featherless, fireworks, firmware, friendli, githubCopilot, githubModels, google, groq, helicone, huggingface, iflowcn, inception, inference, ioNet, jiekou, kilo, kimiForCoding, kuaeCloudCodingPlan, litellm, llama, llmgateway, lmstudio, lucidquery, meganova, minimax, minimaxCn, minimaxCnCodingPlan, minimaxCodingPlan, mistral, moark, modelscope, moonshotai, moonshotaiCn, morph, nanoGpt, nebius, nova, novita, novitaAi, nvidia, ollama, ollamaCloud, openai, opencode, opencodeGo, openrouter, ovhcloud, perplexity, perplexityAgent, poe, privatemodeAi, qihangAi, qiniuAi, requesty, scaleway, siliconflow, siliconflowCn, stackit, stepfun, submodel, synthetic, tencentCodingPlan, tencentHunyuan, togetherai, upstage, venice, vivgrid, vultr, wandb, xai, xiaomi, zai, zaiCodingPlan, zenmux, zhipuai, zhipuaiCodingPlan };