@xsai-ext/providers 0.4.1 → 0.4.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.ts CHANGED
@@ -1,8 +1,16 @@
1
1
  import { C as ChatProvider, M as ModelProvider, E as EmbedProvider, I as ImageProvider, S as SpeechProvider, T as TranscriptionProvider, a as ChatProviderWithExtraOptions } from './types-DW1hvH0W.js';
2
- import { A as AnthropicOptions, O as OpenRouterOptions } from './anthropic-Bel15KQV.js';
3
- import { A as AnthropicModels, O as OpenrouterModels, T as TogetheraiModels } from './types-Czo87gCh.js';
2
+ import { A as AnthropicOptions, O as OpenRouterOptions } from './anthropic-C6HzSBof.js';
3
+ import { A as AnthropicModels, O as OpenrouterModels, T as TogetheraiModels } from './types-B2HZyaCH.js';
4
4
  import '@xsai/shared';
5
5
 
6
+ /**
7
+ * Abacus Provider
8
+ * @see {@link https://abacus.ai/help/api}
9
+ * @remarks
10
+ * - baseURL - `https://routellm.abacus.ai/v1`
11
+ * - apiKey - `ABACUS_API_KEY`
12
+ */
13
+ declare const abacus: ChatProvider<"gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "gemini-2.0-flash-001" | "gemini-3-flash-preview" | "route-llm" | "grok-code-fast-1" | "kimi-k2-turbo-preview" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "gpt-4.1-mini" | "claude-opus-4-5-20251101" | "qwen-2.5-coder-32b" | "claude-sonnet-4-5-20250929" | "grok-4-0709" | "o3-mini" | "gpt-5.2-chat-latest" | "gemini-2.0-pro-exp-02-05" | "gpt-5.1" | "gpt-5-nano" | "claude-sonnet-4-20250514" | "gpt-4.1" | "o4-mini" | "claude-opus-4-20250514" | "gpt-5-mini" | "o3-pro" | "claude-3-7-sonnet-20250219" | "gemini-2.5-pro" | "gpt-4o-2024-11-20" | "o3" | "gpt-4o-mini" | "qwen3-max" | "gpt-5" | "grok-4-1-fast-non-reasoning" | "llama-3.3-70b-versatile" | "claude-opus-4-1-20250805" | "gpt-5.2" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "deepseek/deepseek-v3.1" | "openai/gpt-oss-120b" | "meta-llama/Meta-Llama-3.1-8B-Instruct" | "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "meta-llama/Meta-Llama-3.1-70B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/QwQ-32B" | "Qwen/Qwen3-32B" | "Qwen/qwen3-coder-480b-a35b-instruct" | "zai-org/glm-4.7" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2"> & ModelProvider;
6
14
  /**
7
15
  * Alibaba Provider
8
16
  * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
@@ -10,7 +18,7 @@ import '@xsai/shared';
10
18
  * - baseURL - `https://dashscope-intl.aliyuncs.com/compatible-mode/v1`
11
19
  * - apiKey - `DASHSCOPE_API_KEY`
12
20
  */
13
- declare const alibaba: ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus"> & ModelProvider;
21
+ declare const alibaba: ModelProvider & ChatProvider<"qwen3-max" | "qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus">;
14
22
  /**
15
23
  * Alibaba (China) Provider
16
24
  * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
@@ -18,7 +26,7 @@ declare const alibaba: ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen
18
26
  * - baseURL - `https://dashscope.aliyuncs.com/compatible-mode/v1`
19
27
  * - apiKey - `DASHSCOPE_API_KEY`
20
28
  */
21
- declare const alibabaCn: ModelProvider & ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
29
+ declare const alibabaCn: ModelProvider & ChatProvider<"qwen3-max" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
22
30
  /**
23
31
  * Bailing Provider
24
32
  * @see {@link https://alipaytbox.yuque.com/sxs0ba/ling/intro}
@@ -34,7 +42,7 @@ declare const bailing: ModelProvider & ChatProvider<"Ling-1T" | "Ring-1T">;
34
42
  * - baseURL - `https://inference.baseten.co/v1`
35
43
  * - apiKey - `BASETEN_API_KEY`
36
44
  */
37
- declare const baseten: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2">;
45
+ declare const baseten: ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "zai-org/GLM-4.6">;
38
46
  /**
39
47
  * Cerebras Provider
40
48
  * @see {@link https://inference-docs.cerebras.ai/models/overview}
@@ -42,7 +50,7 @@ declare const baseten: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct
42
50
  * - baseURL - `https://api.cerebras.ai/v1/`
43
51
  * - apiKey - `CEREBRAS_API_KEY`
44
52
  */
45
- declare const cerebras: ModelProvider & ChatProvider<"qwen-3-235b-a22b-instruct-2507" | "zai-glm-4.6" | "gpt-oss-120b"> & EmbedProvider<string>;
53
+ declare const cerebras: ModelProvider & ChatProvider<"zai-glm-4.7" | "qwen-3-235b-a22b-instruct-2507" | "gpt-oss-120b"> & EmbedProvider<string>;
46
54
  /**
47
55
  * Chutes Provider
48
56
  * @see {@link https://llm.chutes.ai/v1/models}
@@ -50,7 +58,15 @@ declare const cerebras: ModelProvider & ChatProvider<"qwen-3-235b-a22b-instruct-
50
58
  * - baseURL - `https://llm.chutes.ai/v1`
51
59
  * - apiKey - `CHUTES_API_KEY`
52
60
  */
53
- declare const chutes: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2" | "NousResearch/Hermes-4.3-36B" | "NousResearch/Hermes-4-70B" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4-405B-FP8" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "rednote-hilab/dots.ocr" | "MiniMaxAI/MiniMax-M2" | "ArliAI/QwQ-32B-ArliAI-RpR-v1" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "tngtech/TNG-R1T-Chimera-TEE" | "OpenGVLab/InternVL3-78B" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "Alibaba-NLP/Tongyi-DeepResearch-30B-A3B" | "mistralai/Devstral-2-123B-Instruct-2512" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "unsloth/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen3-32B" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-V3" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
61
+ declare const chutes: ModelProvider & ChatProvider<"Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-32B" | "moonshotai/Kimi-K2-Instruct-0905" | "NousResearch/Hermes-4.3-36B" | "NousResearch/Hermes-4-70B" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4-405B-FP8-TEE" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "rednote-hilab/dots.ocr" | "moonshotai/Kimi-K2-Thinking-TEE" | "MiniMaxAI/MiniMax-M2.1-TEE" | "nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "tngtech/TNG-R1T-Chimera-TEE" | "XiaomiMiMo/MiMo-V2-Flash" | "OpenGVLab/InternVL3-78B-TEE" | "openai/gpt-oss-120b-TEE" | "openai/gpt-oss-20b" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "mistralai/Devstral-2-123B-Instruct-2512" | "mistralai/Devstral-2-123B-Instruct-2512-TEE" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "unsloth/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3Guard-Gen-0.6B" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-VL-72B-Instruct-TEE" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-235B-A22B-Instruct-2507-TEE" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8-TEE" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "miromind-ai/MiroThinker-v1.5-235B" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.5-TEE" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.7-TEE" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-V3-0324-TEE" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-V3.1-Terminus-TEE" | "deepseek-ai/DeepSeek-V3" | "deepseek-ai/DeepSeek-R1-TEE" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-R1-0528-TEE" | "deepseek-ai/DeepSeek-V3.2-TEE" | "deepseek-ai/DeepSeek-V3.1-TEE">;
62
+ /**
63
+ * Cohere Provider
64
+ * @see {@link https://docs.cohere.com/docs/models}
65
+ * @remarks
66
+ * - baseURL - `https://api.cohere.ai/compatibility/v1/`
67
+ * - apiKey - `COHERE_API_KEY`
68
+ */
69
+ declare const cohere: ModelProvider & EmbedProvider<string> & ChatProvider<"command-a-translate-08-2025" | "command-a-03-2025" | "command-r-08-2024" | "command-r-plus-08-2024" | "command-r7b-12-2024" | "command-a-reasoning-08-2025" | "command-a-vision-07-2025">;
54
70
  /**
55
71
  * Cortecs Provider
56
72
  * @see {@link https://api.cortecs.ai/v1/models}
@@ -58,7 +74,7 @@ declare const chutes: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-
58
74
  * - baseURL - `https://api.cortecs.ai/v1`
59
75
  * - apiKey - `CORTECS_API_KEY`
60
76
  */
61
- declare const cortecs: ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "gpt-4.1" | "gemini-2.5-pro" | "devstral-small-2512" | "claude-sonnet-4" | "llama-3.1-405b-instruct">;
77
+ declare const cortecs: ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro" | "qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "devstral-small-2512" | "claude-sonnet-4" | "llama-3.1-405b-instruct">;
62
78
  /**
63
79
  * Deep Infra Provider
64
80
  * @see {@link https://deepinfra.com/models}
@@ -66,7 +82,7 @@ declare const cortecs: ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instr
66
82
  * - baseURL - `https://api.deepinfra.com/v1/openai/`
67
83
  * - apiKey - `DEEPINFRA_API_KEY`
68
84
  */
69
- declare const deepinfra: ModelProvider & EmbedProvider<string> & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "zai-org/GLM-4.5" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo">;
85
+ declare const deepinfra: ModelProvider & EmbedProvider<string> & ChatProvider<"openai/gpt-oss-120b" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "openai/gpt-oss-20b" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2" | "MiniMaxAI/MiniMax-M2.1" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo" | "zai-org/GLM-4.5">;
70
86
  /**
71
87
  * DeepSeek Provider
72
88
  * @see {@link https://platform.deepseek.com/api-docs/pricing}
@@ -82,7 +98,7 @@ declare const deepseek: ModelProvider & ChatProvider<"deepseek-chat" | "deepseek
82
98
  * - baseURL - `https://go.fastrouter.ai/api/v1`
83
99
  * - apiKey - `FASTROUTER_API_KEY`
84
100
  */
85
- declare const fastrouter: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
101
+ declare const fastrouter: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "openai/gpt-oss-20b" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
86
102
  /**
87
103
  * Fireworks AI Provider
88
104
  * @see {@link https://fireworks.ai/docs/}
@@ -90,7 +106,23 @@ declare const fastrouter: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "o
90
106
  * - baseURL - `https://api.fireworks.ai/inference/v1/`
91
107
  * - apiKey - `FIREWORKS_API_KEY`
92
108
  */
93
- declare const fireworks: ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
109
+ declare const fireworks: ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/deepseek-v3p2" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/minimax-m2p1" | "accounts/fireworks/models/glm-4p7" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/glm-4p6" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
110
+ /**
111
+ * Firmware Provider
112
+ * @see {@link https://docs.firmware.ai}
113
+ * @remarks
114
+ * - baseURL - `https://app.firmware.ai/api/v1`
115
+ * - apiKey - `FIRMWARE_API_KEY`
116
+ */
117
+ declare const firmware: ModelProvider & ChatProvider<"grok-4-fast-non-reasoning" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "claude-sonnet-4-5-20250929" | "gpt-5-nano" | "gpt-5-mini" | "gemini-2.5-pro" | "gpt-5" | "gpt-5.2" | "claude-haiku-4-5-20251001" | "deepseek-chat" | "deepseek-reasoner" | "grok-4-fast-reasoning" | "deepseek-coder" | "claude-opus-4-5" | "gpt-4o">;
118
+ /**
119
+ * Friendli Provider
120
+ * @see {@link https://friendli.ai/docs/guides/serverless_endpoints/introduction}
121
+ * @remarks
122
+ * - baseURL - `https://api.friendli.ai/serverless/v1`
123
+ * - apiKey - `FRIENDLI_TOKEN`
124
+ */
125
+ declare const friendli: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-32B" | "zai-org/GLM-4.6" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "meta-llama-3.3-70b-instruct" | "meta-llama-3.1-8b-instruct" | "LGAI-EXAONE/K-EXAONE-236B-A23B" | "LGAI-EXAONE/EXAONE-4.0.1-32B" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
94
126
  /**
95
127
  * GitHub Copilot Provider
96
128
  * @see {@link https://docs.github.com/en/copilot}
@@ -98,7 +130,7 @@ declare const fireworks: ModelProvider & ChatProvider<"accounts/fireworks/models
98
130
  * - baseURL - `https://api.githubcopilot.com`
99
131
  * - apiKey - `GITHUB_TOKEN`
100
132
  */
101
- declare const githubCopilot: ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "gemini-2.0-flash-001" | "claude-opus-4" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gpt-5.1-codex" | "claude-haiku-4.5" | "gemini-3-pro-preview" | "oswe-vscode-prime" | "claude-3.5-sonnet" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "claude-opus-41" | "gpt-5-mini" | "claude-3.7-sonnet" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "claude-3.7-sonnet-thought" | "claude-opus-4.5" | "gpt-5.2" | "claude-sonnet-4.5">;
133
+ declare const githubCopilot: ModelProvider & ChatProvider<"gemini-2.0-flash-001" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gemini-3-pro-preview" | "o3-mini" | "gpt-5.1" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "gemini-2.5-pro" | "o3" | "gpt-5" | "gpt-5.2" | "claude-sonnet-4" | "gpt-4o" | "claude-opus-4" | "gpt-5.2-codex" | "gpt-5.1-codex" | "claude-haiku-4.5" | "claude-3.5-sonnet" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "claude-opus-41" | "claude-3.7-sonnet" | "gpt-5.1-codex-max" | "claude-3.7-sonnet-thought" | "claude-opus-4.5" | "claude-sonnet-4.5">;
102
134
  /**
103
135
  * GitHub Models Provider
104
136
  * @see {@link https://docs.github.com/en/github-models}
@@ -114,7 +146,7 @@ declare const githubModels: ModelProvider & ChatProvider<"openai/gpt-4.1" | "cor
114
146
  * - baseURL - `https://generativelanguage.googleapis.com/v1beta/openai/`
115
147
  * - apiKey - `GOOGLE_GENERATIVE_AI_API_KEY or GEMINI_API_KEY`
116
148
  */
117
- declare const google: ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-2.5-pro" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-2.5-flash" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
149
+ declare const google: ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-3-flash-preview" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "gemini-2.5-pro" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
118
150
  /**
119
151
  * Groq Provider
120
152
  * @see {@link https://console.groq.com/docs/models}
@@ -122,7 +154,7 @@ declare const google: ModelProvider & EmbedProvider<string> & ChatProvider<"gemi
122
154
  * - baseURL - `https://api.groq.com/openai/v1/`
123
155
  * - apiKey - `GROQ_API_KEY`
124
156
  */
125
- declare const groq: ModelProvider & EmbedProvider<string> & ChatProvider<"deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b">;
157
+ declare const groq: ModelProvider & EmbedProvider<string> & ChatProvider<"llama-3.3-70b-versatile" | "openai/gpt-oss-120b" | "deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b">;
126
158
  /**
127
159
  * Helicone Provider
128
160
  * @see {@link https://helicone.ai/models}
@@ -130,7 +162,7 @@ declare const groq: ModelProvider & EmbedProvider<string> & ChatProvider<"deepse
130
162
  * - baseURL - `https://ai-gateway.helicone.ai/v1`
131
163
  * - apiKey - `HELICONE_API_KEY`
132
164
  */
133
- declare const helicone: ModelProvider & ChatProvider<"qwen3-next-80b-a3b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-thinking" | "gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "deepseek-reasoner" | "claude-opus-4" | "grok-code-fast-1" | "gpt-5.1-codex" | "gemini-3-pro-preview" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "claude-3.7-sonnet" | "o3" | "gpt-5" | "gemini-2.5-flash" | "gemini-2.5-flash-lite" | "llama-3.1-8b-instant" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "qwen3-coder" | "grok-4-fast-reasoning" | "claude-opus-4-1" | "grok-4" | "llama-4-maverick" | "llama-prompt-guard-2-86m" | "grok-4-1-fast-reasoning" | "claude-4.5-haiku" | "llama-3.1-8b-instruct-turbo" | "gpt-4.1-mini-2025-04-14" | "llama-guard-4" | "llama-3.1-8b-instruct" | "gpt-4.1-mini" | "deepseek-v3.1-terminus" | "llama-prompt-guard-2-22m" | "claude-3.5-sonnet-v2" | "sonar-deep-research" | "claude-sonnet-4-5-20250929" | "grok-3" | "mistral-small" | "kimi-k2-0711" | "chatgpt-4o-latest" | "kimi-k2-0905" | "sonar-reasoning" | "llama-3.3-70b-instruct" | "claude-4.5-sonnet" | "codex-mini-latest" | "gpt-5-nano" | "deepseek-tng-r1t2-chimera" | "claude-4.5-opus" | "sonar" | "glm-4.6" | "qwen3-235b-a22b-thinking" | "hermes-2-pro-llama-3-8b" | "o1" | "grok-3-mini" | "sonar-pro" | "o1-mini" | "claude-3-haiku-20240307" | "o3-pro" | "qwen2.5-coder-7b-fast" | "gemma-3-12b-it" | "mistral-nemo" | "gpt-oss-20b" | "claude-3.5-haiku" | "gpt-5-chat-latest" | "gpt-4o-mini" | "sonar-reasoning-pro" | "qwen3-vl-235b-a22b-instruct" | "qwen3-30b-a3b" | "deepseek-v3.2" | "grok-4-1-fast-non-reasoning" | "gpt-5-pro" | "mistral-large-2411" | "claude-opus-4-1-20250805" | "ernie-4.5-21b-a3b-thinking" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "llama-4-scout">;
165
+ declare const helicone: ModelProvider & ChatProvider<"gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "grok-code-fast-1" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "gpt-4.1-mini" | "claude-sonnet-4-5-20250929" | "o3-mini" | "gpt-5.1" | "gpt-5-nano" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "o3-pro" | "gemini-2.5-pro" | "o3" | "gpt-4o-mini" | "gpt-5" | "grok-4-1-fast-non-reasoning" | "llama-3.3-70b-versatile" | "claude-opus-4-1-20250805" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "qwen3-next-80b-a3b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-thinking" | "claude-sonnet-4" | "deepseek-reasoner" | "grok-4-fast-reasoning" | "gpt-4o" | "claude-opus-4" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "claude-3.7-sonnet" | "gemini-2.5-flash-lite" | "llama-3.1-8b-instant" | "gemma2-9b-it" | "qwen3-coder" | "claude-opus-4-1" | "grok-4" | "llama-4-maverick" | "llama-prompt-guard-2-86m" | "grok-4-1-fast-reasoning" | "claude-4.5-haiku" | "llama-3.1-8b-instruct-turbo" | "gpt-4.1-mini-2025-04-14" | "llama-guard-4" | "llama-3.1-8b-instruct" | "deepseek-v3.1-terminus" | "llama-prompt-guard-2-22m" | "claude-3.5-sonnet-v2" | "sonar-deep-research" | "grok-3" | "mistral-small" | "kimi-k2-0711" | "chatgpt-4o-latest" | "kimi-k2-0905" | "sonar-reasoning" | "llama-3.3-70b-instruct" | "claude-4.5-sonnet" | "codex-mini-latest" | "deepseek-tng-r1t2-chimera" | "claude-4.5-opus" | "sonar" | "glm-4.6" | "qwen3-235b-a22b-thinking" | "hermes-2-pro-llama-3-8b" | "o1" | "grok-3-mini" | "sonar-pro" | "o1-mini" | "claude-3-haiku-20240307" | "qwen2.5-coder-7b-fast" | "gemma-3-12b-it" | "mistral-nemo" | "gpt-oss-20b" | "claude-3.5-haiku" | "gpt-5-chat-latest" | "sonar-reasoning-pro" | "qwen3-vl-235b-a22b-instruct" | "qwen3-30b-a3b" | "deepseek-v3.2" | "gpt-5-pro" | "mistral-large-2411" | "ernie-4.5-21b-a3b-thinking" | "llama-4-scout">;
134
166
  /**
135
167
  * Hugging Face Provider
136
168
  * @see {@link https://huggingface.co/docs/inference-providers}
@@ -138,7 +170,7 @@ declare const helicone: ModelProvider & ChatProvider<"qwen3-next-80b-a3b-instruc
138
170
  * - baseURL - `https://router.huggingface.co/v1`
139
171
  * - apiKey - `HF_TOKEN`
140
172
  */
141
- declare const huggingface: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "MiniMaxAI/MiniMax-M2" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/Deepseek-V3-0324">;
173
+ declare const huggingface: ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "XiaomiMiMo/MiMo-V2-Flash" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2.1" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/DeepSeek-R1-0528">;
142
174
  /**
143
175
  * iFlow Provider
144
176
  * @see {@link https://platform.iflow.cn/en/docs}
@@ -146,7 +178,7 @@ declare const huggingface: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Inst
146
178
  * - baseURL - `https://apis.iflow.cn/v1`
147
179
  * - apiKey - `IFLOW_API_KEY`
148
180
  */
149
- declare const iflowcn: ModelProvider & ChatProvider<"qwen3-vl-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "kimi-k2-thinking" | "qwen3-coder" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "kimi-k2" | "deepseek-v3.1" | "minimax-m2" | "qwen3-235b" | "deepseek-v3.2-chat" | "qwen3-235b-a22b-thinking-2507" | "tstars2.0" | "qwen3-235b-a22b-instruct" | "qwen3-max-preview">;
181
+ declare const iflowcn: ModelProvider & ChatProvider<"qwen3-max" | "qwen3-vl-plus" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "kimi-k2" | "qwen3-235b" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct" | "qwen3-max-preview">;
150
182
  /**
151
183
  * Inception Provider
152
184
  * @see {@link https://platform.inceptionlabs.ai/docs}
@@ -170,7 +202,7 @@ declare const inference: ModelProvider & ChatProvider<"meta/llama-3.2-11b-vision
170
202
  * - baseURL - `https://api.intelligence.io.solutions/api/v1`
171
203
  * - apiKey - `IOINTELLIGENCE_API_KEY`
172
204
  */
173
- declare const ioNet: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506" | "mistralai/Mistral-Large-Instruct-2411" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar">;
205
+ declare const ioNet: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506" | "mistralai/Mistral-Large-Instruct-2411" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar">;
174
206
  /**
175
207
  * Kimi For Coding Provider
176
208
  * @see {@link https://www.kimi.com/coding/docs/en/third-party-agents.html}
@@ -210,7 +242,7 @@ declare const lucidquery: ModelProvider & ChatProvider<"lucidquery-nexus-coder"
210
242
  * - baseURL - `https://api.minimax.io/v1/`
211
243
  * - apiKey - `MINIMAX_API_KEY`
212
244
  */
213
- declare const minimax: ModelProvider & ChatProvider<"MiniMax-M2">;
245
+ declare const minimax: ModelProvider & ChatProvider<"MiniMax-M2" | "MiniMax-M2.1">;
214
246
  /**
215
247
  * MiniMax (China) Provider
216
248
  * @see {@link https://platform.minimaxi.com/docs/guides/quickstart}
@@ -218,7 +250,7 @@ declare const minimax: ModelProvider & ChatProvider<"MiniMax-M2">;
218
250
  * - baseURL - `https://api.minimaxi.com/v1/`
219
251
  * - apiKey - `MINIMAX_API_KEY`
220
252
  */
221
- declare const minimaxCn: ModelProvider & ChatProvider<"MiniMax-M2">;
253
+ declare const minimaxCn: ModelProvider & ChatProvider<"MiniMax-M2" | "MiniMax-M2.1">;
222
254
  /**
223
255
  * Mistral Provider
224
256
  * @see {@link https://docs.mistral.ai/getting-started/models/}
@@ -234,7 +266,7 @@ declare const mistral: ModelProvider & EmbedProvider<string> & ChatProvider<"dev
234
266
  * - baseURL - `https://api-inference.modelscope.cn/v1`
235
267
  * - apiKey - `MODELSCOPE_API_KEY`
236
268
  */
237
- declare const modelscope: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507">;
269
+ declare const modelscope: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct">;
238
270
  /**
239
271
  * Moonshot AI Provider
240
272
  * @see {@link https://platform.moonshot.ai/docs/api/chat}
@@ -242,7 +274,7 @@ declare const modelscope: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Ins
242
274
  * - baseURL - `https://api.moonshot.ai/v1`
243
275
  * - apiKey - `MOONSHOT_API_KEY`
244
276
  */
245
- declare const moonshotai: ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
277
+ declare const moonshotai: ModelProvider & ChatProvider<"kimi-k2-turbo-preview" | "kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
246
278
  /**
247
279
  * Moonshot AI (China) Provider
248
280
  * @see {@link https://platform.moonshot.cn/docs/api/chat}
@@ -250,7 +282,7 @@ declare const moonshotai: ModelProvider & ChatProvider<"kimi-k2-thinking" | "kim
250
282
  * - baseURL - `https://api.moonshot.cn/v1`
251
283
  * - apiKey - `MOONSHOT_API_KEY`
252
284
  */
253
- declare const moonshotaiCn: ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
285
+ declare const moonshotaiCn: ModelProvider & ChatProvider<"kimi-k2-turbo-preview" | "kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
254
286
  /**
255
287
  * Morph Provider
256
288
  * @see {@link https://docs.morphllm.com/api-reference/introduction}
@@ -259,6 +291,14 @@ declare const moonshotaiCn: ModelProvider & ChatProvider<"kimi-k2-thinking" | "k
259
291
  * - apiKey - `MORPH_API_KEY`
260
292
  */
261
293
  declare const morph: ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
294
+ /**
295
+ * NanoGPT Provider
296
+ * @see {@link https://docs.nano-gpt.com}
297
+ * @remarks
298
+ * - baseURL - `https://nano-gpt.com/api/v1`
299
+ * - apiKey - `NANO_GPT_API_KEY`
300
+ */
301
+ declare const nanoGpt: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "zai-org/glm-4.7" | "qwen/qwen3-coder" | "deepseek/deepseek-r1" | "moonshotai/kimi-k2-instruct" | "moonshotai/kimi-k2-thinking" | "nousresearch/hermes-4-405b:thinking" | "nvidia/llama-3_3-nemotron-super-49b-v1_5" | "deepseek/deepseek-v3.2:thinking" | "minimax/minimax-m2.1" | "z-ai/glm-4.6:thinking" | "z-ai/glm-4.6" | "qwen/qwen3-235b-a22b-thinking-2507" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "meta-llama/llama-4-maverick" | "meta-llama/llama-3.3-70b-instruct" | "zai-org/glm-4.5-air" | "zai-org/glm-4.7:thinking" | "zai-org/glm-4.5-air:thinking">;
262
302
  /**
263
303
  * Nebius Token Factory Provider
264
304
  * @see {@link https://docs.tokenfactory.nebius.com/}
@@ -266,7 +306,15 @@ declare const morph: ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "m
266
306
  * - baseURL - `https://api.tokenfactory.nebius.com/v1`
267
307
  * - apiKey - `NEBIUS_API_KEY`
268
308
  */
269
- declare const nebius: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2-instruct" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "zai-org/glm-4.5" | "zai-org/glm-4.5-air" | "deepseek-ai/deepseek-v3">;
309
+ declare const nebius: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "zai-org/glm-4.5" | "openai/gpt-oss-20b" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-235b-a22b-thinking-2507" | "zai-org/glm-4.5-air" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "deepseek-ai/deepseek-v3">;
310
+ /**
311
+ * NovitaAI Provider
312
+ * @see {@link https://novita.ai/docs/guides/introduction}
313
+ * @remarks
314
+ * - baseURL - `https://api.novita.ai/openai`
315
+ * - apiKey - `NOVITA_API_KEY`
316
+ */
317
+ declare const novitaAi: ModelProvider & ChatProvider<"deepseek/deepseek-v3.1" | "openai/gpt-oss-120b" | "zai-org/glm-4.7" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "openai/gpt-oss-20b" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-v3-0324" | "moonshotai/kimi-k2-instruct" | "meta-llama/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-thinking" | "minimax/minimax-m2.1" | "qwen/qwen3-235b-a22b-thinking-2507" | "meta-llama/llama-3.3-70b-instruct" | "zai-org/glm-4.5-air" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "baichuan/baichuan-m2-32b" | "moonshotai/kimi-k2-0905" | "nousresearch/hermes-2-pro-llama-3-8b" | "paddlepaddle/paddleocr-vl" | "kwaipilot/kat-coder" | "kwaipilot/kat-coder-pro" | "xiaomimimo/mimo-v2-flash" | "deepseek/deepseek-prover-v2-671b" | "deepseek/deepseek-r1-0528-qwen3-8b" | "deepseek/deepseek-v3.1-terminus" | "deepseek/deepseek-r1-turbo" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-r1-distill-llama-70b" | "deepseek/deepseek-ocr" | "deepseek/deepseek-v3.2" | "deepseek/deepseek-v3-turbo" | "sao10k/l3-8b-lunaris" | "sao10k/L3-8B-Stheno-v3.2" | "sao10k/l31-70b-euryale-v2.2" | "sao10k/l3-70b-euryale-v2.1" | "skywork/r1v4-lite" | "minimaxai/minimax-m1-80k" | "minimax/minimax-m2" | "google/gemma-3-27b-it" | "microsoft/wizardlm-2-8x22b" | "gryphe/mythomax-l2-13b" | "baidu/ernie-4.5-vl-28b-a3b-thinking" | "baidu/ernie-4.5-300b-a47b-paddle" | "baidu/ernie-4.5-21B-a3b" | "baidu/ernie-4.5-21B-a3b-thinking" | "baidu/ernie-4.5-vl-424b-a47b" | "baidu/ernie-4.5-vl-28b-a3b" | "qwen/qwen3-vl-30b-a3b-thinking" | "qwen/qwen3-omni-30b-a3b-thinking" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-8b-fp8" | "qwen/qwen2.5-vl-72b-instruct" | "qwen/qwen3-4b-fp8" | "qwen/qwen3-coder-30b-a3b-instruct" | "qwen/qwen3-vl-8b-instruct" | "qwen/qwen2.5-7b-instruct" | "qwen/qwen3-30b-a3b-fp8" | "qwen/qwen3-32b-fp8" | "qwen/qwen3-omni-30b-a3b-instruct" | "qwen/qwen-2.5-72b-instruct" | "qwen/qwen3-vl-235b-a22b-thinking" | "qwen/qwen-mt-plus" | "qwen/qwen3-max" | "qwen/qwen3-vl-235b-a22b-instruct" | "qwen/qwen3-vl-30b-a3b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-235b-a22b-fp8" | "mistralai/mistral-nemo" | "meta-llama/llama-3-70b-instruct" | "meta-llama/llama-3-8b-instruct" | "meta-llama/llama-3.1-8b-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct-fp8" | "zai-org/glm-4.5v" | "zai-org/glm-4.6v" | "zai-org/autoglm-phone-9b-multilingual">;
270
318
  /**
271
319
  * Nvidia Provider
272
320
  * @see {@link https://docs.api.nvidia.com/nim/}
@@ -274,7 +322,7 @@ declare const nebius: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "opena
274
322
  * - baseURL - `https://integrate.api.nvidia.com/v1`
275
323
  * - apiKey - `NVIDIA_API_KEY`
276
324
  */
277
- declare const nvidia: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-4-mini-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-coder-480b-a35b-instruct" | "moonshotai/kimi-k2-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "minimaxai/minimax-m2" | "google/gemma-3-27b-it" | "openai/whisper-large-v3" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwen3-next-80b-a3b-thinking" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "black-forest-labs/flux.1-dev">;
325
+ declare const nvidia: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-mini-instruct" | "microsoft/phi-3-medium-4k-instruct" | "meta/llama-3.2-11b-vision-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "meta/llama-3.2-1b-instruct" | "moonshotai/kimi-k2-thinking" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "qwen/qwen3-coder-480b-a35b-instruct" | "google/gemma-3-27b-it" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/nemotron-3-nano-30b-a3b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.3-nemotron-super-49b-v1" | "nvidia/llama-3.1-nemotron-51b-instruct" | "nvidia/llama3-chatqa-1.5-70b" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "nvidia/llama-3.1-nemotron-70b-instruct" | "nvidia/nemotron-4-340b-instruct" | "nvidia/llama-3.3-nemotron-super-49b-v1.5" | "minimaxai/minimax-m2" | "google/gemma-3n-e2b-it" | "google/codegemma-1.1-7b" | "google/gemma-3n-e4b-it" | "google/gemma-2-2b-it" | "google/gemma-3-12b-it" | "google/codegemma-7b" | "google/gemma-3-1b-it" | "google/gemma-2-27b-it" | "microsoft/phi-3-vision-128k-instruct" | "openai/whisper-large-v3" | "qwen/qwen2.5-coder-32b-instruct" | "qwen/qwen2.5-coder-7b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwq-32b" | "mistralai/mamba-codestral-7b-v0.1" | "mistralai/mistral-large-2-instruct" | "mistralai/codestral-22b-instruct-v0.1" | "mistralai/mistral-small-3.1-24b-instruct-2503" | "meta/llama3-70b-instruct" | "meta/llama-4-maverick-17b-128e-instruct" | "meta/codellama-70b" | "meta/llama-3.1-405b-instruct" | "meta/llama3-8b-instruct" | "meta/llama-3.1-70b-instruct" | "deepseek-ai/deepseek-r1-0528" | "deepseek-ai/deepseek-r1" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "deepseek-ai/deepseek-coder-6.7b-instruct" | "black-forest-labs/flux.1-dev">;
278
326
  /**
279
327
  * Ollama Cloud Provider
280
328
  * @see {@link https://docs.ollama.com/cloud}
@@ -290,7 +338,7 @@ declare const ollamaCloud: ModelProvider & ChatProvider<"kimi-k2-thinking:cloud"
290
338
  * - baseURL - `https://api.openai.com/v1/`
291
339
  * - apiKey - `OPENAI_API_KEY`
292
340
  */
293
- declare const openai: ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "gpt-5.2" | "gpt-4.1-nano" | "gpt-4.1-mini" | "codex-mini-latest" | "gpt-5-nano" | "o1" | "o1-mini" | "o3-pro" | "gpt-5-chat-latest" | "gpt-4o-mini" | "gpt-5-pro" | "gpt-5.1-chat-latest" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "o3-deep-research" | "gpt-3.5-turbo" | "gpt-5.2-pro" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "gpt-5.2-chat-latest" | "text-embedding-ada-002" | "gpt-4o-2024-11-20" | "o4-mini-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
341
+ declare const openai: ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1-nano" | "gpt-4.1-mini" | "o3-mini" | "gpt-5.2-chat-latest" | "gpt-5.1" | "gpt-5-nano" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "o3-pro" | "gpt-4o-2024-11-20" | "o3" | "gpt-4o-mini" | "gpt-5" | "gpt-5.2" | "gpt-5.1-chat-latest" | "gpt-4o" | "gpt-5.2-codex" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "gpt-5.1-codex-max" | "codex-mini-latest" | "o1" | "o1-mini" | "gpt-5-chat-latest" | "gpt-5-pro" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "o3-deep-research" | "gpt-3.5-turbo" | "gpt-5.2-pro" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "text-embedding-ada-002" | "o4-mini-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
294
342
  /**
295
343
  * OpenCode Zen Provider
296
344
  * @see {@link https://opencode.ai/docs/zen}
@@ -298,7 +346,7 @@ declare const openai: ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-
298
346
  * - baseURL - `https://opencode.ai/zen/v1`
299
347
  * - apiKey - `OPENCODE_API_KEY`
300
348
  */
301
- declare const opencode: ModelProvider & ChatProvider<"kimi-k2-thinking" | "claude-sonnet-4" | "gpt-5.1-codex" | "gpt-5.1" | "gpt-5-codex" | "gpt-5.1-codex-max" | "gpt-5" | "gpt-5.2" | "qwen3-coder" | "claude-opus-4-1" | "gpt-5-nano" | "glm-4.6" | "kimi-k2" | "claude-haiku-4-5" | "claude-opus-4-5" | "gemini-3-pro" | "claude-sonnet-4-5" | "alpha-gd4" | "big-pickle" | "claude-3-5-haiku" | "grok-code" | "gemini-3-flash" | "alpha-doubao-seed-code" | "minimax-m2.1">;
349
+ declare const opencode: ModelProvider & ChatProvider<"gpt-5.1" | "gpt-5-nano" | "gpt-5" | "gpt-5.2" | "kimi-k2-thinking" | "claude-sonnet-4" | "claude-opus-4-5" | "gpt-5.2-codex" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "gpt-5.1-codex-max" | "qwen3-coder" | "claude-opus-4-1" | "glm-4.6" | "kimi-k2" | "claude-haiku-4-5" | "gemini-3-pro" | "alpha-glm-4.7" | "claude-sonnet-4-5" | "alpha-gd4" | "big-pickle" | "claude-3-5-haiku" | "glm-4.7-free" | "grok-code" | "gemini-3-flash" | "minimax-m2.1-free">;
302
350
  /**
303
351
  * OVHcloud AI Endpoints Provider
304
352
  * @see {@link https://www.ovhcloud.com/en/public-cloud/ai-endpoints/catalog//}
@@ -314,7 +362,7 @@ declare const ovhcloud: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instru
314
362
  * - baseURL - `https://api.perplexity.ai/`
315
363
  * - apiKey - `PERPLEXITY_API_KEY`
316
364
  */
317
- declare const perplexity: ModelProvider & ChatProvider<"sonar-reasoning" | "sonar" | "sonar-pro" | "sonar-reasoning-pro">;
365
+ declare const perplexity: ModelProvider & ChatProvider<"sonar" | "sonar-pro" | "sonar-reasoning-pro">;
318
366
  /**
319
367
  * Poe Provider
320
368
  * @see {@link https://creator.poe.com/docs/external-applications/openai-compatible-api}
@@ -322,7 +370,15 @@ declare const perplexity: ModelProvider & ChatProvider<"sonar-reasoning" | "sona
322
370
  * - baseURL - `https://api.poe.com/v1`
323
371
  * - apiKey - `POE_API_KEY`
324
372
  */
325
- declare const poe: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-2" | "xai/grok-4.1-fast-non-reasoning" | "ideogramai/ideogram" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram-v2" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "poetools/claude-code" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "google/gemini-deep-research" | "google/nano-banana" | "google/imagen-4" | "google/imagen-3" | "google/imagen-4-ultra" | "google/gemini-2.0-flash-lite" | "google/gemini-3-pro" | "google/veo-3.1" | "google/imagen-3-fast" | "google/lyria" | "google/gemini-2.0-flash" | "google/gemini-2.5-flash-lite" | "google/veo-3" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-2" | "google/nano-banana-pro" | "google/veo-3.1-fast" | "openai/gpt-5.2-instant" | "openai/sora-2" | "openai/o1-pro" | "openai/gpt-5.1-codex" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-4-classic" | "openai/gpt-5-chat" | "openai/o3-deep-research" | "openai/gpt-4o-search" | "openai/gpt-image-1-mini" | "openai/gpt-3.5-turbo" | "openai/gpt-5.2-pro" | "openai/o3-mini-high" | "openai/chatgpt-4o-latest" | "openai/gpt-4-turbo" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1-instant" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-4o-aug" | "openai/o3-pro" | "openai/gpt-image-1" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo-instruct" | "openai/o4-mini-deep-research" | "openai/gpt-4-classic-0314" | "openai/dall-e-3" | "openai/sora-2-pro" | "openai/gpt-5-pro" | "openai/gpt-5.2" | "openai/gpt-4o-mini-search" | "stabilityai/stablediffusionxl" | "topazlabs-co/topazlabs" | "lumalabs/ray2" | "lumalabs/dream-machine" | "anthropic/claude-opus-3" | "anthropic/claude-opus-4" | "anthropic/claude-sonnet-3.7-reasoning" | "anthropic/claude-opus-4-search" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-haiku-3.5-search" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4-reasoning" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7-search" | "anthropic/claude-opus-4-reasoning" | "anthropic/claude-sonnet-3.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-haiku-3.5" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-sonnet-4-search" | "trytako/tako" | "novita/kimi-k2-thinking" | "novita/glm-4.6">;
373
+ declare const poe: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-4.1-fast-non-reasoning" | "ideogramai/ideogram" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram-v2" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "poetools/claude-code" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "google/gemini-deep-research" | "google/nano-banana" | "google/imagen-4" | "google/imagen-3" | "google/imagen-4-ultra" | "google/gemini-2.0-flash-lite" | "google/gemini-3-pro" | "google/veo-3.1" | "google/imagen-3-fast" | "google/lyria" | "google/gemini-2.0-flash" | "google/gemini-2.5-flash-lite" | "google/veo-3" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-2" | "google/gemini-3-flash" | "google/nano-banana-pro" | "google/veo-3.1-fast" | "openai/gpt-5.2-instant" | "openai/sora-2" | "openai/o1-pro" | "openai/gpt-5.1-codex" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-4-classic" | "openai/gpt-5-chat" | "openai/o3-deep-research" | "openai/gpt-4o-search" | "openai/gpt-image-1.5" | "openai/gpt-image-1-mini" | "openai/gpt-3.5-turbo" | "openai/gpt-5.2-pro" | "openai/o3-mini-high" | "openai/chatgpt-4o-latest" | "openai/gpt-4-turbo" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1-instant" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-4o-aug" | "openai/o3-pro" | "openai/gpt-image-1" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo-instruct" | "openai/o4-mini-deep-research" | "openai/gpt-4-classic-0314" | "openai/dall-e-3" | "openai/sora-2-pro" | "openai/gpt-5-pro" | "openai/gpt-5.2" | "openai/gpt-4o-mini-search" | "stabilityai/stablediffusionxl" | "topazlabs-co/topazlabs" | "lumalabs/ray2" | "lumalabs/dream-machine" | "anthropic/claude-opus-3" | "anthropic/claude-opus-4" | "anthropic/claude-sonnet-3.7-reasoning" | "anthropic/claude-opus-4-search" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-haiku-3.5-search" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4-reasoning" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7-search" | "anthropic/claude-opus-4-reasoning" | "anthropic/claude-sonnet-3.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-haiku-3.5" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-sonnet-4-search" | "trytako/tako" | "novita/glm-4.7" | "novita/kimi-k2-thinking" | "novita/kat-coder-pro" | "novita/glm-4.6" | "novita/minimax-m2.1" | "novita/glm-4.6v" | "cerebras/gpt-oss-120b-cs" | "cerebras/zai-glm-4.6-cs">;
374
+ /**
375
+ * Privatemode AI Provider
376
+ * @see {@link https://docs.privatemode.ai/api/overview}
377
+ * @remarks
378
+ * - baseURL - `http://localhost:8080/v1`
379
+ * - apiKey - `PRIVATEMODE_API_KEY`
380
+ */
381
+ declare const privatemodeAi: ModelProvider & ChatProvider<"gpt-oss-120b" | "whisper-large-v3" | "qwen3-embedding-4b" | "gemma-3-27b" | "qwen3-coder-30b-a3b">;
326
382
  /**
327
383
  * Requesty Provider
328
384
  * @see {@link https://requesty.ai/solution/llm-routing/models}
@@ -338,7 +394,7 @@ declare const requesty: ModelProvider & ChatProvider<"google/gemini-2.5-flash" |
338
394
  * - baseURL - `https://api.scaleway.ai/v1`
339
395
  * - apiKey - `SCALEWAY_API_KEY`
340
396
  */
341
- declare const scaleway: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen3-235b-a22b-instruct-2507" | "pixtral-12b-2409" | "whisper-large-v3" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
397
+ declare const scaleway: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "whisper-large-v3" | "qwen3-235b-a22b-instruct-2507" | "devstral-2-123b-instruct-2512" | "pixtral-12b-2409" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
342
398
  /**
343
399
  * SiliconFlow Provider
344
400
  * @see {@link https://cloud.siliconflow.com/models}
@@ -346,15 +402,15 @@ declare const scaleway: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instru
346
402
  * - baseURL - `https://api.siliconflow.com/v1`
347
403
  * - apiKey - `SILICONFLOW_API_KEY`
348
404
  */
349
- declare const siliconflow: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
405
+ declare const siliconflow: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"openai/gpt-oss-120b" | "meta-llama/Meta-Llama-3.1-8B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/QwQ-32B" | "Qwen/Qwen3-32B" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-V3" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2" | "zai-org/GLM-4.5" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "inclusionAI/Ling-mini-2.0" | "inclusionAI/Ling-flash-2.0" | "inclusionAI/Ring-flash-2.0" | "moonshotai/Kimi-Dev-72B" | "tencent/Hunyuan-MT-7B" | "tencent/Hunyuan-A13B-Instruct" | "MiniMaxAI/MiniMax-M1-80k" | "THUDM/GLM-4-32B-0414" | "THUDM/GLM-4.1V-9B-Thinking" | "THUDM/GLM-Z1-9B-0414" | "THUDM/GLM-4-9B-0414" | "THUDM/GLM-Z1-32B-0414" | "stepfun-ai/step3" | "nex-agi/DeepSeek-V3.1-Nex-N1" | "baidu/ERNIE-4.5-300B-A47B" | "ByteDance-Seed/Seed-OSS-36B-Instruct" | "Qwen/Qwen3-VL-30B-A3B-Instruct" | "Qwen/Qwen3-Omni-30B-A3B-Captioner" | "Qwen/Qwen3-8B" | "Qwen/Qwen3-Omni-30B-A3B-Instruct" | "Qwen/Qwen3-VL-8B-Thinking" | "Qwen/Qwen2.5-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct-128K" | "Qwen/Qwen2.5-7B-Instruct" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen2.5-VL-7B-Instruct" | "Qwen/Qwen3-VL-8B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-VL-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Thinking" | "Qwen/Qwen3-Omni-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Instruct" | "Qwen/Qwen2.5-14B-Instruct" | "zai-org/GLM-4.5V" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" | "deepseek-ai/deepseek-vl2" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B" | "deepseek-ai/DeepSeek-V3.2-Exp" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B" | "deepseek-ai/DeepSeek-V3.1">;
350
406
  /**
351
407
  * SiliconFlow (China) Provider
352
408
  * @see {@link https://cloud.siliconflow.com/models}
353
409
  * @remarks
354
410
  * - baseURL - `https://api.siliconflow.cn/v1`
355
- * - apiKey - `SILICONFLOW_API_KEY`
411
+ * - apiKey - `SILICONFLOW_CN_API_KEY`
356
412
  */
357
- declare const siliconflowCn: ModelProvider & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
413
+ declare const siliconflowCn: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/QwQ-32B" | "Qwen/Qwen3-32B" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-V3" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2" | "zai-org/GLM-4.5" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "inclusionAI/Ling-mini-2.0" | "inclusionAI/Ling-flash-2.0" | "inclusionAI/Ring-flash-2.0" | "moonshotai/Kimi-Dev-72B" | "tencent/Hunyuan-MT-7B" | "tencent/Hunyuan-A13B-Instruct" | "MiniMaxAI/MiniMax-M1-80k" | "THUDM/GLM-4-32B-0414" | "THUDM/GLM-4.1V-9B-Thinking" | "THUDM/GLM-Z1-9B-0414" | "THUDM/GLM-4-9B-0414" | "THUDM/GLM-Z1-32B-0414" | "stepfun-ai/step3" | "nex-agi/DeepSeek-V3.1-Nex-N1" | "baidu/ERNIE-4.5-300B-A47B" | "ByteDance-Seed/Seed-OSS-36B-Instruct" | "Qwen/Qwen3-VL-30B-A3B-Instruct" | "Qwen/Qwen3-Omni-30B-A3B-Captioner" | "Qwen/Qwen3-8B" | "Qwen/Qwen3-Omni-30B-A3B-Instruct" | "Qwen/Qwen3-VL-8B-Thinking" | "Qwen/Qwen2.5-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct-128K" | "Qwen/Qwen2.5-7B-Instruct" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen2.5-VL-7B-Instruct" | "Qwen/Qwen3-VL-8B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-VL-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Thinking" | "Qwen/Qwen3-Omni-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Instruct" | "Qwen/Qwen2.5-14B-Instruct" | "zai-org/GLM-4.5V" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" | "deepseek-ai/deepseek-vl2" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B" | "Kwaipilot/KAT-Dev" | "ascend-tribe/pangu-pro-moe">;
358
414
  /**
359
415
  * submodel Provider
360
416
  * @see {@link https://submodel.gitbook.io}
@@ -362,7 +418,7 @@ declare const siliconflowCn: ModelProvider & ChatProvider<"qwen-qwq-32b" | "deep
362
418
  * - baseURL - `https://llm.submodel.ai/v1`
363
419
  * - apiKey - `SUBMODEL_INSTAGEN_ACCESS_KEY`
364
420
  */
365
- declare const submodel: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324" | "zai-org/GLM-4.5-FP8">;
421
+ declare const submodel: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "zai-org/GLM-4.5-FP8" | "deepseek-ai/DeepSeek-V3-0324">;
366
422
  /**
367
423
  * Synthetic Provider
368
424
  * @see {@link https://synthetic.new/pricing}
@@ -370,15 +426,15 @@ declare const submodel: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qw
370
426
  * - baseURL - `https://api.synthetic.new/v1`
371
427
  * - apiKey - `SYNTHETIC_API_KEY`
372
428
  */
373
- declare const synthetic: ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
429
+ declare const synthetic: ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:MiniMaxAI/MiniMax-M2.1" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.7" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
374
430
  /**
375
431
  * Upstage Provider
376
432
  * @see {@link https://developers.upstage.ai/docs/apis/chat}
377
433
  * @remarks
378
- * - baseURL - `https://api.upstage.ai`
434
+ * - baseURL - `https://api.upstage.ai/v1/solar`
379
435
  * - apiKey - `UPSTAGE_API_KEY`
380
436
  */
381
- declare const upstage: ModelProvider & ChatProvider<"solar-mini" | "solar-pro2">;
437
+ declare const upstage: ModelProvider & ChatProvider<"solar-mini" | "solar-pro3" | "solar-pro2">;
382
438
  /**
383
439
  * Venice AI Provider
384
440
  * @see {@link https://docs.venice.ai}
@@ -386,7 +442,15 @@ declare const upstage: ModelProvider & ChatProvider<"solar-mini" | "solar-pro2">
386
442
  * - baseURL - `https://api.venice.ai/api/v1`
387
443
  * - apiKey - `VENICE_API_KEY`
388
444
  */
389
- declare const venice: ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "kimi-k2-thinking" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "deepseek-v3.2" | "qwen3-235b" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "openai-gpt-oss-120b" | "zai-org-glm-4.6" | "grok-41-fast" | "claude-opus-45" | "mistral-31-24b" | "venice-uncensored" | "openai-gpt-52" | "qwen3-4b" | "llama-3.3-70b" | "devstral-2-2512" | "llama-3.2-3b" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "zai-org-glm-4.6v" | "qwen3-next-80b">;
445
+ declare const venice: ModelProvider & ChatProvider<"gemini-3-flash-preview" | "grok-code-fast-1" | "gemini-3-pro-preview" | "qwen3-coder-480b-a35b-instruct" | "kimi-k2-thinking" | "deepseek-v3.2" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "grok-41-fast" | "claude-opus-45" | "mistral-31-24b" | "zai-org-glm-4.7" | "venice-uncensored" | "openai-gpt-52" | "qwen3-4b" | "llama-3.3-70b" | "claude-sonnet-45" | "openai-gpt-oss-120b" | "llama-3.2-3b" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "zai-org-glm-4.6v" | "minimax-m21" | "qwen3-next-80b" | "openai-gpt-52-codex">;
446
+ /**
447
+ * Vivgrid Provider
448
+ * @see {@link https://docs.vivgrid.com/models}
449
+ * @remarks
450
+ * - baseURL - `https://api.vivgrid.com/v1`
451
+ * - apiKey - `VIVGRID_API_KEY`
452
+ */
453
+ declare const vivgrid: ModelProvider & ChatProvider<"gpt-5.1-codex">;
390
454
  /**
391
455
  * Vultr Provider
392
456
  * @see {@link https://api.vultrinference.com/}
@@ -402,7 +466,7 @@ declare const vultr: ModelProvider & ChatProvider<"deepseek-r1-distill-qwen-32b"
402
466
  * - baseURL - `https://api.inference.wandb.ai/v1`
403
467
  * - apiKey - `WANDB_API_KEY`
404
468
  */
405
- declare const wandb: ModelProvider & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2-Instruct" | "meta-llama/Llama-3.3-70B-Instruct" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
469
+ declare const wandb: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "moonshotai/Kimi-K2-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "meta-llama/Llama-3.3-70B-Instruct" | "deepseek-ai/DeepSeek-V3-0324" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct">;
406
470
  /**
407
471
  * xAI Provider
408
472
  * @see {@link https://docs.x.ai/docs/models}
@@ -410,7 +474,7 @@ declare const wandb: ModelProvider & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-In
410
474
  * - baseURL - `https://api.x.ai/v1/`
411
475
  * - apiKey - `XAI_API_KEY`
412
476
  */
413
- declare const xai: ModelProvider & ChatProvider<"grok-code-fast-1" | "grok-4-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-4-1-fast-non-reasoning" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
477
+ declare const xai: ModelProvider & ChatProvider<"grok-4-fast-non-reasoning" | "grok-code-fast-1" | "grok-4-1-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
414
478
  /**
415
479
  * Xiaomi Provider
416
480
  * @see {@link https://platform.xiaomimimo.com/#/docs}
@@ -426,7 +490,7 @@ declare const xiaomi: ModelProvider & ChatProvider<"mimo-v2-flash">;
426
490
  * - baseURL - `https://api.z.ai/api/paas/v4`
427
491
  * - apiKey - `ZHIPU_API_KEY`
428
492
  */
429
- declare const zai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
493
+ declare const zai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
430
494
  /**
431
495
  * Z.AI Coding Plan Provider
432
496
  * @see {@link https://docs.z.ai/devpack/overview}
@@ -434,7 +498,7 @@ declare const zai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "g
434
498
  * - baseURL - `https://api.z.ai/api/coding/paas/v4`
435
499
  * - apiKey - `ZHIPU_API_KEY`
436
500
  */
437
- declare const zaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
501
+ declare const zaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
438
502
  /**
439
503
  * ZenMux Provider
440
504
  * @see {@link https://docs.zenmux.ai}
@@ -442,7 +506,7 @@ declare const zaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-f
442
506
  * - baseURL - `https://zenmux.ai/api/v1`
443
507
  * - apiKey - `ZENMUX_API_KEY`
444
508
  */
445
- declare const zenmux: ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "moonshotai/kimi-k2-thinking" | "openai/gpt-5-codex" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.5" | "moonshotai/kimi-k2-thinking-turbo" | "moonshotai/kimi-k2-0905" | "xiaomi/mimo-v2-flash" | "x-ai/grok-4-fast-non-reasoning" | "x-ai/grok-code-fast-1" | "x-ai/grok-4-fast" | "deepseek/deepseek-chat" | "minimax/minimax-m2" | "inclusionai/ring-1t" | "inclusionai/lint-1t" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1">;
509
+ declare const zenmux: ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "moonshotai/kimi-k2-thinking" | "minimax/minimax-m2.1" | "z-ai/glm-4.6" | "moonshotai/kimi-k2-0905" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-v3.2" | "minimax/minimax-m2" | "google/gemini-2.5-flash-lite" | "openai/gpt-5.1-codex" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-5.2" | "anthropic/claude-opus-4" | "anthropic/claude-haiku-4.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-sonnet-4.5" | "google/gemini-3-flash-preview" | "google/gemini-3-pro-preview" | "stepfun/step-3" | "moonshotai/kimi-k2-thinking-turbo" | "xiaomi/mimo-v2-flash-free" | "xiaomi/mimo-v2-flash" | "x-ai/grok-code-fast-1" | "x-ai/grok-4.1-fast-non-reasoning" | "x-ai/grok-4-fast" | "x-ai/grok-4.1-fast" | "deepseek/deepseek-chat" | "deepseek/deepseek-reasoner" | "google/gemini-3-flash-preview-free" | "volcengine/doubao-seed-code" | "volcengine/doubao-seed-1.8" | "openai/gpt-5.1-chat" | "baidu/ernie-5.0-thinking-preview" | "inclusionai/ring-1t" | "inclusionai/ling-1t" | "z-ai/glm-4.7" | "z-ai/glm-4.6v-flash-free" | "z-ai/glm-4.6v-flash" | "z-ai/glm-4.5" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6v" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1-free" | "kuaishou/kat-coder-pro-v1">;
446
510
  /**
447
511
  * Zhipu AI Provider
448
512
  * @see {@link https://docs.z.ai/guides/overview/pricing}
@@ -450,7 +514,7 @@ declare const zenmux: ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemin
450
514
  * - baseURL - `https://open.bigmodel.cn/api/paas/v4`
451
515
  * - apiKey - `ZHIPU_API_KEY`
452
516
  */
453
- declare const zhipuai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
517
+ declare const zhipuai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
454
518
  /**
455
519
  * Zhipu AI Coding Plan Provider
456
520
  * @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
@@ -458,15 +522,7 @@ declare const zhipuai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash"
458
522
  * - baseURL - `https://open.bigmodel.cn/api/coding/paas/v4`
459
523
  * - apiKey - `ZHIPU_API_KEY`
460
524
  */
461
- declare const zhipuaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
462
- /**
463
- * Novita AI Provider
464
- * @see {@link https://novita.ai/docs/guides/llm-api#api-integration}
465
- * @remarks
466
- * - baseURL - `https://api.novita.ai/v3/openai/`
467
- * - apiKey - `NOVITA_API_KEY`
468
- */
469
- declare const novita: ModelProvider & ChatProvider<string>;
525
+ declare const zhipuaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
470
526
  /**
471
527
  * StepFun Provider
472
528
  * @see {@link https://www.stepfun.com}
@@ -532,5 +588,7 @@ declare const openrouter: ModelProvider & ChatProviderWithExtraOptions<Openroute
532
588
  * - apiKey - `TOGETHER_API_KEY`
533
589
  */
534
590
  declare const togetherai: ModelProvider & EmbedProvider<string> & ChatProvider<TogetheraiModels>;
591
+ /** @deprecated use `novitaAi` instead. */
592
+ declare const novita: ModelProvider & ChatProvider<"deepseek/deepseek-v3.1" | "openai/gpt-oss-120b" | "zai-org/glm-4.7" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "openai/gpt-oss-20b" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-v3-0324" | "moonshotai/kimi-k2-instruct" | "meta-llama/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-thinking" | "minimax/minimax-m2.1" | "qwen/qwen3-235b-a22b-thinking-2507" | "meta-llama/llama-3.3-70b-instruct" | "zai-org/glm-4.5-air" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "baichuan/baichuan-m2-32b" | "moonshotai/kimi-k2-0905" | "nousresearch/hermes-2-pro-llama-3-8b" | "paddlepaddle/paddleocr-vl" | "kwaipilot/kat-coder" | "kwaipilot/kat-coder-pro" | "xiaomimimo/mimo-v2-flash" | "deepseek/deepseek-prover-v2-671b" | "deepseek/deepseek-r1-0528-qwen3-8b" | "deepseek/deepseek-v3.1-terminus" | "deepseek/deepseek-r1-turbo" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-r1-distill-llama-70b" | "deepseek/deepseek-ocr" | "deepseek/deepseek-v3.2" | "deepseek/deepseek-v3-turbo" | "sao10k/l3-8b-lunaris" | "sao10k/L3-8B-Stheno-v3.2" | "sao10k/l31-70b-euryale-v2.2" | "sao10k/l3-70b-euryale-v2.1" | "skywork/r1v4-lite" | "minimaxai/minimax-m1-80k" | "minimax/minimax-m2" | "google/gemma-3-27b-it" | "microsoft/wizardlm-2-8x22b" | "gryphe/mythomax-l2-13b" | "baidu/ernie-4.5-vl-28b-a3b-thinking" | "baidu/ernie-4.5-300b-a47b-paddle" | "baidu/ernie-4.5-21B-a3b" | "baidu/ernie-4.5-21B-a3b-thinking" | "baidu/ernie-4.5-vl-424b-a47b" | "baidu/ernie-4.5-vl-28b-a3b" | "qwen/qwen3-vl-30b-a3b-thinking" | "qwen/qwen3-omni-30b-a3b-thinking" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-8b-fp8" | "qwen/qwen2.5-vl-72b-instruct" | "qwen/qwen3-4b-fp8" | "qwen/qwen3-coder-30b-a3b-instruct" | "qwen/qwen3-vl-8b-instruct" | "qwen/qwen2.5-7b-instruct" | "qwen/qwen3-30b-a3b-fp8" | "qwen/qwen3-32b-fp8" | "qwen/qwen3-omni-30b-a3b-instruct" | "qwen/qwen-2.5-72b-instruct" | "qwen/qwen3-vl-235b-a22b-thinking" | "qwen/qwen-mt-plus" | "qwen/qwen3-max" | "qwen/qwen3-vl-235b-a22b-instruct" | "qwen/qwen3-vl-30b-a3b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-235b-a22b-fp8" | "mistralai/mistral-nemo" | "meta-llama/llama-3-70b-instruct" | "meta-llama/llama-3-8b-instruct" | "meta-llama/llama-3.1-8b-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct-fp8" | "zai-org/glm-4.5v" | "zai-org/glm-4.6v" | "zai-org/autoglm-phone-9b-multilingual">;
535
593
 
536
- export { alibaba, alibabaCn, anthropic, bailing, baseten, cerebras, chutes, cortecs, deepinfra, deepseek, fastrouter, featherless, fireworks, githubCopilot, githubModels, google, groq, helicone, huggingface, iflowcn, inception, inference, ioNet, kimiForCoding, litellm, llama, lmstudio, lucidquery, minimax, minimaxCn, mistral, modelscope, moonshotai, moonshotaiCn, morph, nebius, novita, nvidia, ollama, ollamaCloud, openai, opencode, openrouter, ovhcloud, perplexity, poe, requesty, scaleway, siliconflow, siliconflowCn, stepfun, submodel, synthetic, tencentHunyuan, togetherai, upstage, venice, vultr, wandb, xai, xiaomi, zai, zaiCodingPlan, zenmux, zhipuai, zhipuaiCodingPlan };
594
+ export { abacus, alibaba, alibabaCn, anthropic, bailing, baseten, cerebras, chutes, cohere, cortecs, deepinfra, deepseek, fastrouter, featherless, fireworks, firmware, friendli, githubCopilot, githubModels, google, groq, helicone, huggingface, iflowcn, inception, inference, ioNet, kimiForCoding, litellm, llama, lmstudio, lucidquery, minimax, minimaxCn, mistral, modelscope, moonshotai, moonshotaiCn, morph, nanoGpt, nebius, novita, novitaAi, nvidia, ollama, ollamaCloud, openai, opencode, openrouter, ovhcloud, perplexity, poe, privatemodeAi, requesty, scaleway, siliconflow, siliconflowCn, stepfun, submodel, synthetic, tencentHunyuan, togetherai, upstage, venice, vivgrid, vultr, wandb, xai, xiaomi, zai, zaiCodingPlan, zenmux, zhipuai, zhipuaiCodingPlan };