@xsai-ext/providers 0.4.0-beta.9 → 0.4.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.ts CHANGED
@@ -1,86 +1,96 @@
1
- import * as _xsai_ext_shared_providers from '@xsai-ext/shared-providers';
2
- import { A as AnthropicModels, O as OpenrouterModels, a as OpenRouterOptions, T as TogetheraiModels } from './openrouter-DCc0QhRt.js';
1
+ import { C as ChatProvider, M as ModelProvider, E as EmbedProvider, I as ImageProvider, S as SpeechProvider, T as TranscriptionProvider, a as ChatProviderWithExtraOptions } from './types-DW1hvH0W.js';
2
+ import { A as AnthropicOptions, O as OpenRouterOptions } from './anthropic-Bel15KQV.js';
3
+ import { A as AnthropicModels, O as OpenrouterModels, T as TogetheraiModels } from './types-Czo87gCh.js';
4
+ import '@xsai/shared';
3
5
 
4
6
  /**
5
- * Moonshot AI (China) Provider
6
- * @see {@link https://platform.moonshot.cn/docs/api/chat}
7
+ * Alibaba Provider
8
+ * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
7
9
  * @remarks
8
- * - baseURL - `https://api.moonshot.cn/v1`
9
- * - apiKey - `MOONSHOT_API_KEY`
10
+ * - baseURL - `https://dashscope-intl.aliyuncs.com/compatible-mode/v1`
11
+ * - apiKey - `DASHSCOPE_API_KEY`
10
12
  */
11
- declare const moonshotaiCn: _xsai_ext_shared_providers.ChatProvider<"kimi-k2-0905-preview" | "kimi-k2-0711-preview" | "kimi-k2-turbo-preview"> & _xsai_ext_shared_providers.ModelProvider;
13
+ declare const alibaba: ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus"> & ModelProvider;
12
14
  /**
13
- * LucidQuery AI Provider
14
- * @see {@link https://lucidquery.com/api/docs}
15
+ * Alibaba (China) Provider
16
+ * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
15
17
  * @remarks
16
- * - baseURL - `https://lucidquery.com/api/v1`
17
- * - apiKey - `LUCIDQUERY_API_KEY`
18
+ * - baseURL - `https://dashscope.aliyuncs.com/compatible-mode/v1`
19
+ * - apiKey - `DASHSCOPE_API_KEY`
18
20
  */
19
- declare const lucidquery: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"lucidquery-nexus-coder" | "lucidnova-rf1-100b">;
21
+ declare const alibabaCn: ModelProvider & ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
20
22
  /**
21
- * Moonshot AI Provider
22
- * @see {@link https://platform.moonshot.ai/docs/api/chat}
23
+ * Bailing Provider
24
+ * @see {@link https://alipaytbox.yuque.com/sxs0ba/ling/intro}
23
25
  * @remarks
24
- * - baseURL - `https://api.moonshot.ai/v1`
25
- * - apiKey - `MOONSHOT_API_KEY`
26
+ * - baseURL - `https://api.tbox.cn/api/llm/v1/chat/completions`
27
+ * - apiKey - `BAILING_API_TOKEN`
26
28
  */
27
- declare const moonshotai: _xsai_ext_shared_providers.ChatProvider<"kimi-k2-0905-preview" | "kimi-k2-0711-preview" | "kimi-k2-turbo-preview"> & _xsai_ext_shared_providers.ModelProvider;
29
+ declare const bailing: ModelProvider & ChatProvider<"Ling-1T" | "Ring-1T">;
28
30
  /**
29
- * Z.AI Coding Plan Provider
30
- * @see {@link https://docs.z.ai/devpack/overview}
31
+ * Baseten Provider
32
+ * @see {@link https://docs.baseten.co/development/model-apis/overview}
31
33
  * @remarks
32
- * - baseURL - `https://api.z.ai/api/coding/paas/v4`
33
- * - apiKey - `ZHIPU_API_KEY`
34
+ * - baseURL - `https://inference.baseten.co/v1`
35
+ * - apiKey - `BASETEN_API_KEY`
34
36
  */
35
- declare const zaiCodingPlan: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6">;
37
+ declare const baseten: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2">;
36
38
  /**
37
- * Alibaba Provider
38
- * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
39
+ * Cerebras Provider
40
+ * @see {@link https://inference-docs.cerebras.ai/models/overview}
39
41
  * @remarks
40
- * - baseURL - `https://dashscope-intl.aliyuncs.com/compatible-mode/v1`
41
- * - apiKey - `DASHSCOPE_API_KEY`
42
+ * - baseURL - `https://api.cerebras.ai/v1/`
43
+ * - apiKey - `CEREBRAS_API_KEY`
42
44
  */
43
- declare const alibaba: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus">;
45
+ declare const cerebras: ModelProvider & ChatProvider<"qwen-3-235b-a22b-instruct-2507" | "zai-glm-4.6" | "gpt-oss-120b"> & EmbedProvider<string>;
44
46
  /**
45
- * xAI Provider
46
- * @see {@link https://docs.x.ai/docs/models}
47
+ * Chutes Provider
48
+ * @see {@link https://llm.chutes.ai/v1/models}
47
49
  * @remarks
48
- * - baseURL - `https://api.x.ai/v1/`
49
- * - apiKey - `XAI_API_KEY`
50
+ * - baseURL - `https://llm.chutes.ai/v1`
51
+ * - apiKey - `CHUTES_API_KEY`
50
52
  */
51
- declare const xai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"grok-4-fast-non-reasoning" | "grok-3-fast" | "grok-4" | "grok-2-vision" | "grok-code-fast-1" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-3" | "grok-4-fast" | "grok-2-latest" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-3-mini" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
53
+ declare const chutes: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2" | "NousResearch/Hermes-4.3-36B" | "NousResearch/Hermes-4-70B" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4-405B-FP8" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "rednote-hilab/dots.ocr" | "MiniMaxAI/MiniMax-M2" | "ArliAI/QwQ-32B-ArliAI-RpR-v1" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "tngtech/TNG-R1T-Chimera-TEE" | "OpenGVLab/InternVL3-78B" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "Alibaba-NLP/Tongyi-DeepResearch-30B-A3B" | "mistralai/Devstral-2-123B-Instruct-2512" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "unsloth/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen3-32B" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-V3" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
52
54
  /**
53
- * Vultr Provider
54
- * @see {@link https://api.vultrinference.com/}
55
+ * Cortecs Provider
56
+ * @see {@link https://api.cortecs.ai/v1/models}
55
57
  * @remarks
56
- * - baseURL - `https://api.vultrinference.com/v1`
57
- * - apiKey - `VULTR_API_KEY`
58
+ * - baseURL - `https://api.cortecs.ai/v1`
59
+ * - apiKey - `CORTECS_API_KEY`
58
60
  */
59
- declare const vultr: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"deepseek-r1-distill-qwen-32b" | "qwen2.5-coder-32b-instruct" | "kimi-k2-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b">;
61
+ declare const cortecs: ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "gpt-4.1" | "gemini-2.5-pro" | "devstral-small-2512" | "claude-sonnet-4" | "llama-3.1-405b-instruct">;
60
62
  /**
61
- * Nvidia Provider
62
- * @see {@link https://docs.api.nvidia.com/nim/}
63
+ * Deep Infra Provider
64
+ * @see {@link https://deepinfra.com/models}
63
65
  * @remarks
64
- * - baseURL - `https://integrate.api.nvidia.com/v1`
65
- * - apiKey - `NVIDIA_API_KEY`
66
+ * - baseURL - `https://api.deepinfra.com/v1/openai/`
67
+ * - apiKey - `DEEPINFRA_API_KEY`
66
68
  */
67
- declare const nvidia: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "minimaxai/minimax-m2" | "google/gemma-3-27b-it" | "microsoft/phi-4-mini-instruct" | "openai/whisper-large-v3" | "openai/gpt-oss-120b" | "qwen/qwen3-235b-a22b" | "qwen/qwen3-coder-480b-a35b-instruct" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "black-forest-labs/flux.1-dev">;
69
+ declare const deepinfra: ModelProvider & EmbedProvider<string> & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "zai-org/GLM-4.5" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo">;
68
70
  /**
69
- * Upstage Provider
70
- * @see {@link https://developers.upstage.ai/docs/apis/chat}
71
+ * DeepSeek Provider
72
+ * @see {@link https://platform.deepseek.com/api-docs/pricing}
71
73
  * @remarks
72
- * - baseURL - `https://api.upstage.ai`
73
- * - apiKey - `UPSTAGE_API_KEY`
74
+ * - baseURL - `https://api.deepseek.com`
75
+ * - apiKey - `DEEPSEEK_API_KEY`
74
76
  */
75
- declare const upstage: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"solar-mini" | "solar-pro2">;
77
+ declare const deepseek: ModelProvider & ChatProvider<"deepseek-chat" | "deepseek-reasoner">;
76
78
  /**
77
- * Groq Provider
78
- * @see {@link https://console.groq.com/docs/models}
79
+ * FastRouter Provider
80
+ * @see {@link https://fastrouter.ai/models}
79
81
  * @remarks
80
- * - baseURL - `https://api.groq.com/openai/v1/`
81
- * - apiKey - `GROQ_API_KEY`
82
+ * - baseURL - `https://go.fastrouter.ai/api/v1`
83
+ * - apiKey - `FASTROUTER_API_KEY`
82
84
  */
83
- declare const groq: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"deepseek-r1-distill-llama-70b" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "openai/gpt-oss-120b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "openai/gpt-oss-20b" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b"> & _xsai_ext_shared_providers.EmbedProvider<string>;
85
+ declare const fastrouter: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
86
+ /**
87
+ * Fireworks AI Provider
88
+ * @see {@link https://fireworks.ai/docs/}
89
+ * @remarks
90
+ * - baseURL - `https://api.fireworks.ai/inference/v1/`
91
+ * - apiKey - `FIREWORKS_API_KEY`
92
+ */
93
+ declare const fireworks: ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
84
94
  /**
85
95
  * GitHub Copilot Provider
86
96
  * @see {@link https://docs.github.com/en/copilot}
@@ -88,127 +98,191 @@ declare const groq: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_
88
98
  * - baseURL - `https://api.githubcopilot.com`
89
99
  * - apiKey - `GITHUB_TOKEN`
90
100
  */
91
- declare const githubCopilot: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"grok-code-fast-1" | "gemini-2.0-flash-001" | "claude-opus-4" | "claude-haiku-4.5" | "claude-3.5-sonnet" | "o3-mini" | "gpt-5-codex" | "gpt-4o" | "gpt-4.1" | "o4-mini" | "claude-opus-41" | "gpt-5-mini" | "claude-3.7-sonnet" | "gemini-2.5-pro" | "o3" | "claude-sonnet-4" | "gpt-5" | "claude-3.7-sonnet-thought" | "claude-sonnet-4.5">;
101
+ declare const githubCopilot: ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "gemini-2.0-flash-001" | "claude-opus-4" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gpt-5.1-codex" | "claude-haiku-4.5" | "gemini-3-pro-preview" | "oswe-vscode-prime" | "claude-3.5-sonnet" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "claude-opus-41" | "gpt-5-mini" | "claude-3.7-sonnet" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "claude-3.7-sonnet-thought" | "claude-opus-4.5" | "gpt-5.2" | "claude-sonnet-4.5">;
92
102
  /**
93
- * Mistral Provider
94
- * @see {@link https://docs.mistral.ai/getting-started/models/}
103
+ * GitHub Models Provider
104
+ * @see {@link https://docs.github.com/en/github-models}
95
105
  * @remarks
96
- * - baseURL - `https://api.mistral.ai/v1/`
97
- * - apiKey - `MISTRAL_API_KEY`
106
+ * - baseURL - `https://models.github.ai/inference`
107
+ * - apiKey - `GITHUB_TOKEN`
98
108
  */
99
- declare const mistral: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<"devstral-medium-2507" | "open-mixtral-8x22b" | "ministral-8b-latest" | "pixtral-large-latest" | "ministral-3b-latest" | "pixtral-12b" | "mistral-medium-2505" | "devstral-small-2505" | "mistral-medium-2508" | "mistral-small-latest" | "magistral-small" | "devstral-small-2507" | "codestral-latest" | "open-mixtral-8x7b" | "mistral-nemo" | "open-mistral-7b" | "mistral-large-latest" | "mistral-medium-latest" | "magistral-medium-latest">;
109
+ declare const githubModels: ModelProvider & ChatProvider<"openai/gpt-4.1" | "core42/jais-30b-chat" | "xai/grok-3" | "xai/grok-3-mini" | "cohere/cohere-command-r-08-2024" | "cohere/cohere-command-a" | "cohere/cohere-command-r-plus-08-2024" | "cohere/cohere-command-r" | "cohere/cohere-command-r-plus" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-r1" | "deepseek/deepseek-v3-0324" | "mistral-ai/mistral-medium-2505" | "mistral-ai/ministral-3b" | "mistral-ai/mistral-nemo" | "mistral-ai/mistral-large-2411" | "mistral-ai/codestral-2501" | "mistral-ai/mistral-small-2503" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-mini-4k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-4" | "microsoft/phi-4-mini-reasoning" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-mini-instruct" | "microsoft/phi-4-multimodal-instruct" | "microsoft/phi-3-mini-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-mini-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-4-reasoning" | "microsoft/mai-ds-r1" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o1-preview" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o1-mini" | "openai/o3" | "openai/gpt-4o-mini" | "meta/llama-3.2-11b-vision-instruct" | "meta/meta-llama-3.1-405b-instruct" | "meta/llama-4-maverick-17b-128e-instruct-fp8" | "meta/meta-llama-3-70b-instruct" | "meta/meta-llama-3.1-70b-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-3.2-90b-vision-instruct" | "meta/meta-llama-3-8b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/meta-llama-3.1-8b-instruct" | "ai21-labs/ai21-jamba-1.5-large" | "ai21-labs/ai21-jamba-1.5-mini">;
100
110
  /**
101
- * Nebius AI Studio Provider
102
- * @see {@link https://docs.studio.nebius.com/quickstart}
111
+ * Google Provider
112
+ * @see {@link https://ai.google.dev/gemini-api/docs/pricing}
103
113
  * @remarks
104
- * - baseURL - `https://api.studio.nebius.com/v1/`
105
- * - apiKey - `NEBIUS_API_KEY`
114
+ * - baseURL - `https://generativelanguage.googleapis.com/v1beta/openai/`
115
+ * - apiKey - `GOOGLE_GENERATIVE_AI_API_KEY or GEMINI_API_KEY`
106
116
  */
107
- declare const nebius: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"moonshotai/kimi-k2-instruct" | "openai/gpt-oss-120b" | "qwen/qwen3-coder-480b-a35b-instruct" | "openai/gpt-oss-20b" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-235b-a22b-thinking-2507" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "zai-org/glm-4.5" | "zai-org/glm-4.5-air" | "deepseek-ai/deepseek-v3">;
117
+ declare const google: ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-2.5-pro" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-2.5-flash" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
108
118
  /**
109
- * DeepSeek Provider
110
- * @see {@link https://platform.deepseek.com/api-docs/pricing}
119
+ * Groq Provider
120
+ * @see {@link https://console.groq.com/docs/models}
111
121
  * @remarks
112
- * - baseURL - `https://api.deepseek.com`
113
- * - apiKey - `DEEPSEEK_API_KEY`
122
+ * - baseURL - `https://api.groq.com/openai/v1/`
123
+ * - apiKey - `GROQ_API_KEY`
114
124
  */
115
- declare const deepseek: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"deepseek-chat" | "deepseek-reasoner">;
125
+ declare const groq: ModelProvider & EmbedProvider<string> & ChatProvider<"deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b">;
116
126
  /**
117
- * Alibaba (China) Provider
118
- * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
127
+ * Helicone Provider
128
+ * @see {@link https://helicone.ai/models}
119
129
  * @remarks
120
- * - baseURL - `https://dashscope.aliyuncs.com/compatible-mode/v1`
121
- * - apiKey - `DASHSCOPE_API_KEY`
130
+ * - baseURL - `https://ai-gateway.helicone.ai/v1`
131
+ * - apiKey - `HELICONE_API_KEY`
122
132
  */
123
- declare const alibabaCn: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
133
+ declare const helicone: ModelProvider & ChatProvider<"qwen3-next-80b-a3b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-thinking" | "gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "deepseek-reasoner" | "claude-opus-4" | "grok-code-fast-1" | "gpt-5.1-codex" | "gemini-3-pro-preview" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "claude-3.7-sonnet" | "o3" | "gpt-5" | "gemini-2.5-flash" | "gemini-2.5-flash-lite" | "llama-3.1-8b-instant" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "qwen3-coder" | "grok-4-fast-reasoning" | "claude-opus-4-1" | "grok-4" | "llama-4-maverick" | "llama-prompt-guard-2-86m" | "grok-4-1-fast-reasoning" | "claude-4.5-haiku" | "llama-3.1-8b-instruct-turbo" | "gpt-4.1-mini-2025-04-14" | "llama-guard-4" | "llama-3.1-8b-instruct" | "gpt-4.1-mini" | "deepseek-v3.1-terminus" | "llama-prompt-guard-2-22m" | "claude-3.5-sonnet-v2" | "sonar-deep-research" | "claude-sonnet-4-5-20250929" | "grok-3" | "mistral-small" | "kimi-k2-0711" | "chatgpt-4o-latest" | "kimi-k2-0905" | "sonar-reasoning" | "llama-3.3-70b-instruct" | "claude-4.5-sonnet" | "codex-mini-latest" | "gpt-5-nano" | "deepseek-tng-r1t2-chimera" | "claude-4.5-opus" | "sonar" | "glm-4.6" | "qwen3-235b-a22b-thinking" | "hermes-2-pro-llama-3-8b" | "o1" | "grok-3-mini" | "sonar-pro" | "o1-mini" | "claude-3-haiku-20240307" | "o3-pro" | "qwen2.5-coder-7b-fast" | "gemma-3-12b-it" | "mistral-nemo" | "gpt-oss-20b" | "claude-3.5-haiku" | "gpt-5-chat-latest" | "gpt-4o-mini" | "sonar-reasoning-pro" | "qwen3-vl-235b-a22b-instruct" | "qwen3-30b-a3b" | "deepseek-v3.2" | "grok-4-1-fast-non-reasoning" | "gpt-5-pro" | "mistral-large-2411" | "claude-opus-4-1-20250805" | "ernie-4.5-21b-a3b-thinking" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "llama-4-scout">;
124
134
  /**
125
- * Venice AI Provider
126
- * @see {@link https://docs.venice.ai}
135
+ * Hugging Face Provider
136
+ * @see {@link https://huggingface.co/docs/inference-providers}
127
137
  * @remarks
128
- * - baseURL - `https://api.venice.ai/api/v1`
129
- * - apiKey - `VENICE_API_KEY`
138
+ * - baseURL - `https://router.huggingface.co/v1`
139
+ * - apiKey - `HF_TOKEN`
130
140
  */
131
- declare const venice: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"dolphin-2.9.2-qwen2-72b" | "mistral-31-24b" | "venice-uncensored" | "qwen-2.5-vl" | "qwen3-235b" | "qwen-2.5-qwq-32b" | "deepseek-coder-v2-lite" | "qwen3-4b" | "llama-3.3-70b" | "qwen-2.5-coder-32b" | "deepseek-r1-671b" | "llama-3.2-3b" | "llama-3.1-405b">;
141
+ declare const huggingface: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "MiniMaxAI/MiniMax-M2" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/Deepseek-V3-0324">;
132
142
  /**
133
- * Chutes Provider
134
- * @see {@link https://llm.chutes.ai/v1/models}
143
+ * iFlow Provider
144
+ * @see {@link https://platform.iflow.cn/en/docs}
135
145
  * @remarks
136
- * - baseURL - `https://llm.chutes.ai/v1`
137
- * - apiKey - `CHUTES_API_KEY`
146
+ * - baseURL - `https://apis.iflow.cn/v1`
147
+ * - apiKey - `IFLOW_API_KEY`
138
148
  */
139
- declare const chutes: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"openai/gpt-oss-120b" | "moonshotai/Kimi-K2-Instruct-75k" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-VL-A3B-Thinking" | "MiniMaxAI/MiniMax-M2" | "meituan-longcat/LongCat-Flash-Chat-FP8" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "zai-org/GLM-4.5" | "zai-org/GLM-4.6-FP8" | "zai-org/GLM-4.6-turbo" | "zai-org/GLM-4.6" | "zai-org/GLM-4.5-FP8" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.2-Exp" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.1:THINKING" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
149
+ declare const iflowcn: ModelProvider & ChatProvider<"qwen3-vl-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "kimi-k2-thinking" | "qwen3-coder" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "kimi-k2" | "deepseek-v3.1" | "minimax-m2" | "qwen3-235b" | "deepseek-v3.2-chat" | "qwen3-235b-a22b-thinking-2507" | "tstars2.0" | "qwen3-235b-a22b-instruct" | "qwen3-max-preview">;
140
150
  /**
141
- * Cortecs Provider
142
- * @see {@link https://api.cortecs.ai/v1/models}
151
+ * Inception Provider
152
+ * @see {@link https://platform.inceptionlabs.ai/docs}
143
153
  * @remarks
144
- * - baseURL - `https://api.cortecs.ai/v1`
145
- * - apiKey - `CORTECS_API_KEY`
154
+ * - baseURL - `https://api.inceptionlabs.ai/v1/`
155
+ * - apiKey - `INCEPTION_API_KEY`
146
156
  */
147
- declare const cortecs: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"qwen3-coder-480b-a35b-instruct" | "qwen3-32b" | "kimi-k2-instruct" | "gpt-oss-120b" | "gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "nova-pro-v1" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "llama-3.1-405b-instruct">;
157
+ declare const inception: ModelProvider & ChatProvider<"mercury-coder" | "mercury">;
148
158
  /**
149
- * GitHub Models Provider
150
- * @see {@link https://docs.github.com/en/github-models}
159
+ * Inference Provider
160
+ * @see {@link https://inference.net/models}
151
161
  * @remarks
152
- * - baseURL - `https://models.github.ai/inference`
153
- * - apiKey - `GITHUB_TOKEN`
162
+ * - baseURL - `https://inference.net/v1`
163
+ * - apiKey - `INFERENCE_API_KEY`
154
164
  */
155
- declare const githubModels: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"microsoft/phi-4-mini-instruct" | "core42/jais-30b-chat" | "xai/grok-3" | "xai/grok-3-mini" | "cohere/cohere-command-r-08-2024" | "cohere/cohere-command-a" | "cohere/cohere-command-r-plus-08-2024" | "cohere/cohere-command-r" | "cohere/cohere-command-r-plus" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-r1" | "deepseek/deepseek-v3-0324" | "mistral-ai/mistral-medium-2505" | "mistral-ai/ministral-3b" | "mistral-ai/mistral-nemo" | "mistral-ai/mistral-large-2411" | "mistral-ai/codestral-2501" | "mistral-ai/mistral-small-2503" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-mini-4k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-4" | "microsoft/phi-4-mini-reasoning" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-mini-instruct" | "microsoft/phi-4-multimodal-instruct" | "microsoft/phi-3-mini-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-4-reasoning" | "microsoft/mai-ds-r1" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o1-preview" | "openai/o3-mini" | "openai/gpt-4o" | "openai/gpt-4.1" | "openai/o4-mini" | "openai/o1" | "openai/o1-mini" | "openai/o3" | "openai/gpt-4o-mini" | "meta/llama-3.2-11b-vision-instruct" | "meta/meta-llama-3.1-405b-instruct" | "meta/llama-4-maverick-17b-128e-instruct-fp8" | "meta/meta-llama-3-70b-instruct" | "meta/meta-llama-3.1-70b-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-3.2-90b-vision-instruct" | "meta/meta-llama-3-8b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/meta-llama-3.1-8b-instruct" | "ai21-labs/ai21-jamba-1.5-large" | "ai21-labs/ai21-jamba-1.5-mini">;
165
+ declare const inference: ModelProvider & ChatProvider<"meta/llama-3.2-11b-vision-instruct" | "mistral/mistral-nemo-12b-instruct" | "google/gemma-3" | "osmosis/osmosis-structure-0.6b" | "qwen/qwen3-embedding-4b" | "qwen/qwen-2.5-7b-vision-instruct" | "meta/llama-3.1-8b-instruct" | "meta/llama-3.2-3b-instruct" | "meta/llama-3.2-1b-instruct">;
156
166
  /**
157
- * Baseten Provider
158
- * @see {@link https://docs.baseten.co/development/model-apis/overview}
167
+ * IO.NET Provider
168
+ * @see {@link https://io.net/docs/guides/intelligence/io-intelligence}
159
169
  * @remarks
160
- * - baseURL - `https://inference.baseten.co/v1`
161
- * - apiKey - `BASETEN_API_KEY`
170
+ * - baseURL - `https://api.intelligence.io.solutions/api/v1`
171
+ * - apiKey - `IOINTELLIGENCE_API_KEY`
162
172
  */
163
- declare const baseten: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "zai-org/GLM-4.6" | "Qwen3/Qwen3-Coder-480B-A35B-Instruct">;
173
+ declare const ioNet: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506" | "mistralai/Mistral-Large-Instruct-2411" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar">;
164
174
  /**
165
- * Hugging Face Provider
166
- * @see {@link https://huggingface.co/docs/inference-providers}
175
+ * Kimi For Coding Provider
176
+ * @see {@link https://www.kimi.com/coding/docs/en/third-party-agents.html}
167
177
  * @remarks
168
- * - baseURL - `https://router.huggingface.co/v1`
169
- * - apiKey - `HF_TOKEN`
178
+ * - baseURL - `https://api.kimi.com/coding/v1`
179
+ * - apiKey - `KIMI_API_KEY`
170
180
  */
171
- declare const huggingface: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "MiniMaxAI/MiniMax-M2" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "zai-org/GLM-4.5" | "zai-org/GLM-4.6" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "deepseek-ai/Deepseek-V3-0324">;
181
+ declare const kimiForCoding: ModelProvider & ChatProvider<"kimi-k2-thinking">;
172
182
  /**
173
- * OpenCode Zen Provider
174
- * @see {@link https://opencode.ai/docs/zen}
183
+ * Llama Provider
184
+ * @see {@link https://llama.developer.meta.com/docs/models}
175
185
  * @remarks
176
- * - baseURL - `https://opencode.ai/zen/v1`
177
- * - apiKey - `OPENCODE_API_KEY`
186
+ * - baseURL - `https://api.llama.com/compat/v1/`
187
+ * - apiKey - `LLAMA_API_KEY`
178
188
  */
179
- declare const opencode: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.6" | "gpt-5-codex" | "claude-sonnet-4" | "gpt-5" | "qwen3-coder" | "claude-opus-4-1" | "kimi-k2" | "claude-haiku-4-5" | "minimax-m2" | "claude-sonnet-4-5" | "an-gbt" | "big-pickle" | "claude-3-5-haiku" | "grok-code">;
189
+ declare const llama: ModelProvider & ChatProvider<"llama-3.3-70b-instruct" | "llama-3.3-8b-instruct" | "llama-4-maverick-17b-128e-instruct-fp8" | "llama-4-scout-17b-16e-instruct-fp8" | "groq-llama-4-maverick-17b-128e-instruct" | "cerebras-llama-4-scout-17b-16e-instruct" | "cerebras-llama-4-maverick-17b-128e-instruct">;
180
190
  /**
181
- * FastRouter Provider
182
- * @see {@link https://fastrouter.ai/models}
191
+ * LMStudio Provider
192
+ * @see {@link https://lmstudio.ai/models}
183
193
  * @remarks
184
- * - baseURL - `https://go.fastrouter.ai/api/v1`
185
- * - apiKey - `FASTROUTER_API_KEY`
194
+ * - baseURL - `http://127.0.0.1:1234/v1`
195
+ * - apiKey - `LMSTUDIO_API_KEY`
186
196
  */
187
- declare const fastrouter: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"openai/gpt-oss-120b" | "openai/gpt-oss-20b" | "openai/gpt-4.1" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
197
+ declare const lmstudio: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "qwen/qwen3-30b-a3b-2507" | "qwen/qwen3-coder-30b">;
188
198
  /**
189
- * Google Provider
190
- * @see {@link https://ai.google.dev/gemini-api/docs/pricing}
199
+ * LucidQuery AI Provider
200
+ * @see {@link https://lucidquery.com/api/docs}
191
201
  * @remarks
192
- * - baseURL - `https://generativelanguage.googleapis.com/v1beta/openai/`
193
- * - apiKey - `GOOGLE_GENERATIVE_AI_API_KEY or GEMINI_API_KEY`
202
+ * - baseURL - `https://lucidquery.com/api/v1`
203
+ * - apiKey - `LUCIDQUERY_API_KEY`
194
204
  */
195
- declare const google: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<"gemini-2.5-pro" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-2.5-flash" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
205
+ declare const lucidquery: ModelProvider & ChatProvider<"lucidquery-nexus-coder" | "lucidnova-rf1-100b">;
196
206
  /**
197
- * Inception Provider
198
- * @see {@link https://platform.inceptionlabs.ai/docs}
207
+ * MiniMax Provider
208
+ * @see {@link https://platform.minimax.io/docs/guides/quickstart}
199
209
  * @remarks
200
- * - baseURL - `https://api.inceptionlabs.ai/v1/`
201
- * - apiKey - `INCEPTION_API_KEY`
210
+ * - baseURL - `https://api.minimax.io/v1/`
211
+ * - apiKey - `MINIMAX_API_KEY`
202
212
  */
203
- declare const inception: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"mercury-coder" | "mercury">;
213
+ declare const minimax: ModelProvider & ChatProvider<"MiniMax-M2">;
204
214
  /**
205
- * Weights & Biases Provider
206
- * @see {@link https://weave-docs.wandb.ai/guides/integrations/inference/}
215
+ * MiniMax (China) Provider
216
+ * @see {@link https://platform.minimaxi.com/docs/guides/quickstart}
207
217
  * @remarks
208
- * - baseURL - `https://api.inference.wandb.ai/v1`
209
- * - apiKey - `WANDB_API_KEY`
218
+ * - baseURL - `https://api.minimaxi.com/v1/`
219
+ * - apiKey - `MINIMAX_API_KEY`
210
220
  */
211
- declare const wandb: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
221
+ declare const minimaxCn: ModelProvider & ChatProvider<"MiniMax-M2">;
222
+ /**
223
+ * Mistral Provider
224
+ * @see {@link https://docs.mistral.ai/getting-started/models/}
225
+ * @remarks
226
+ * - baseURL - `https://api.mistral.ai/v1/`
227
+ * - apiKey - `MISTRAL_API_KEY`
228
+ */
229
+ declare const mistral: ModelProvider & EmbedProvider<string> & ChatProvider<"devstral-2512" | "mistral-nemo" | "mistral-large-2411" | "devstral-medium-2507" | "mistral-large-2512" | "open-mixtral-8x22b" | "ministral-8b-latest" | "pixtral-large-latest" | "mistral-small-2506" | "ministral-3b-latest" | "pixtral-12b" | "mistral-medium-2505" | "labs-devstral-small-2512" | "devstral-medium-latest" | "devstral-small-2505" | "mistral-medium-2508" | "mistral-embed" | "mistral-small-latest" | "magistral-small" | "devstral-small-2507" | "codestral-latest" | "open-mixtral-8x7b" | "open-mistral-7b" | "mistral-large-latest" | "mistral-medium-latest" | "magistral-medium-latest">;
230
+ /**
231
+ * ModelScope Provider
232
+ * @see {@link https://modelscope.cn/docs/model-service/API-Inference/intro}
233
+ * @remarks
234
+ * - baseURL - `https://api-inference.modelscope.cn/v1`
235
+ * - apiKey - `MODELSCOPE_API_KEY`
236
+ */
237
+ declare const modelscope: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507">;
238
+ /**
239
+ * Moonshot AI Provider
240
+ * @see {@link https://platform.moonshot.ai/docs/api/chat}
241
+ * @remarks
242
+ * - baseURL - `https://api.moonshot.ai/v1`
243
+ * - apiKey - `MOONSHOT_API_KEY`
244
+ */
245
+ declare const moonshotai: ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
246
+ /**
247
+ * Moonshot AI (China) Provider
248
+ * @see {@link https://platform.moonshot.cn/docs/api/chat}
249
+ * @remarks
250
+ * - baseURL - `https://api.moonshot.cn/v1`
251
+ * - apiKey - `MOONSHOT_API_KEY`
252
+ */
253
+ declare const moonshotaiCn: ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
254
+ /**
255
+ * Morph Provider
256
+ * @see {@link https://docs.morphllm.com/api-reference/introduction}
257
+ * @remarks
258
+ * - baseURL - `https://api.morphllm.com/v1`
259
+ * - apiKey - `MORPH_API_KEY`
260
+ */
261
+ declare const morph: ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
262
+ /**
263
+ * Nebius Token Factory Provider
264
+ * @see {@link https://docs.tokenfactory.nebius.com/}
265
+ * @remarks
266
+ * - baseURL - `https://api.tokenfactory.nebius.com/v1`
267
+ * - apiKey - `NEBIUS_API_KEY`
268
+ */
269
+ declare const nebius: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2-instruct" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "zai-org/glm-4.5" | "zai-org/glm-4.5-air" | "deepseek-ai/deepseek-v3">;
270
+ /**
271
+ * Nvidia Provider
272
+ * @see {@link https://docs.api.nvidia.com/nim/}
273
+ * @remarks
274
+ * - baseURL - `https://integrate.api.nvidia.com/v1`
275
+ * - apiKey - `NVIDIA_API_KEY`
276
+ */
277
+ declare const nvidia: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-4-mini-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-coder-480b-a35b-instruct" | "moonshotai/kimi-k2-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "minimaxai/minimax-m2" | "google/gemma-3-27b-it" | "openai/whisper-large-v3" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwen3-next-80b-a3b-thinking" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "black-forest-labs/flux.1-dev">;
278
+ /**
279
+ * Ollama Cloud Provider
280
+ * @see {@link https://docs.ollama.com/cloud}
281
+ * @remarks
282
+ * - baseURL - `https://ollama.com/v1`
283
+ * - apiKey - `OLLAMA_API_KEY`
284
+ */
285
+ declare const ollamaCloud: ModelProvider & ChatProvider<"kimi-k2-thinking:cloud" | "qwen3-vl-235b-cloud" | "qwen3-coder:480b-cloud" | "gpt-oss:120b-cloud" | "deepseek-v3.1:671b-cloud" | "glm-4.6:cloud" | "cogito-2.1:671b-cloud" | "gpt-oss:20b-cloud" | "qwen3-vl-235b-instruct-cloud" | "kimi-k2:1t-cloud" | "minimax-m2:cloud" | "gemini-3-pro-preview:latest">;
212
286
  /**
213
287
  * OpenAI Provider
214
288
  * @see {@link https://platform.openai.com/docs/models}
@@ -216,15 +290,23 @@ declare const wandb: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared
216
290
  * - baseURL - `https://api.openai.com/v1/`
217
291
  * - apiKey - `OPENAI_API_KEY`
218
292
  */
219
- declare const openai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<"o3-mini" | "gpt-5-codex" | "gpt-4o" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "o3" | "gpt-5" | "gpt-4.1-nano" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "gpt-4.1-mini" | "o3-deep-research" | "gpt-3.5-turbo" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "codex-mini-latest" | "gpt-5-nano" | "o1" | "o1-mini" | "text-embedding-ada-002" | "o3-pro" | "gpt-4o-2024-11-20" | "o4-mini-deep-research" | "gpt-5-chat-latest" | "gpt-4o-mini" | "gpt-5-pro"> & _xsai_ext_shared_providers.ImageProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string>;
293
+ declare const openai: ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "gpt-5.2" | "gpt-4.1-nano" | "gpt-4.1-mini" | "codex-mini-latest" | "gpt-5-nano" | "o1" | "o1-mini" | "o3-pro" | "gpt-5-chat-latest" | "gpt-4o-mini" | "gpt-5-pro" | "gpt-5.1-chat-latest" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "o3-deep-research" | "gpt-3.5-turbo" | "gpt-5.2-pro" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "gpt-5.2-chat-latest" | "text-embedding-ada-002" | "gpt-4o-2024-11-20" | "o4-mini-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
220
294
  /**
221
- * Zhipu AI Coding Plan Provider
222
- * @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
295
+ * OpenCode Zen Provider
296
+ * @see {@link https://opencode.ai/docs/zen}
223
297
  * @remarks
224
- * - baseURL - `https://open.bigmodel.cn/api/coding/paas/v4`
225
- * - apiKey - `ZHIPU_API_KEY`
298
+ * - baseURL - `https://opencode.ai/zen/v1`
299
+ * - apiKey - `OPENCODE_API_KEY`
300
+ */
301
+ declare const opencode: ModelProvider & ChatProvider<"kimi-k2-thinking" | "claude-sonnet-4" | "gpt-5.1-codex" | "gpt-5.1" | "gpt-5-codex" | "gpt-5.1-codex-max" | "gpt-5" | "gpt-5.2" | "qwen3-coder" | "claude-opus-4-1" | "gpt-5-nano" | "glm-4.6" | "kimi-k2" | "claude-haiku-4-5" | "claude-opus-4-5" | "gemini-3-pro" | "claude-sonnet-4-5" | "alpha-gd4" | "big-pickle" | "claude-3-5-haiku" | "grok-code" | "gemini-3-flash" | "alpha-doubao-seed-code" | "minimax-m2.1">;
302
+ /**
303
+ * OVHcloud AI Endpoints Provider
304
+ * @see {@link https://www.ovhcloud.com/en/public-cloud/ai-endpoints/catalog//}
305
+ * @remarks
306
+ * - baseURL - `https://oai.endpoints.kepler.ai.cloud.ovh.net/v1`
307
+ * - apiKey - `OVHCLOUD_API_KEY`
226
308
  */
227
- declare const zhipuaiCodingPlan: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6">;
309
+ declare const ovhcloud: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "gpt-oss-20b" | "mixtral-8x7b-instruct-v0.1" | "mistral-7b-instruct-v0.3" | "qwen2.5-vl-72b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen2.5-coder-32b-instruct" | "llava-next-mistral-7b" | "meta-llama-3_1-70b-instruct" | "meta-llama-3_3-70b-instruct">;
228
310
  /**
229
311
  * Perplexity Provider
230
312
  * @see {@link https://docs.perplexity.ai}
@@ -232,47 +314,47 @@ declare const zhipuaiCodingPlan: _xsai_ext_shared_providers.ModelProvider & _xsa
232
314
  * - baseURL - `https://api.perplexity.ai/`
233
315
  * - apiKey - `PERPLEXITY_API_KEY`
234
316
  */
235
- declare const perplexity: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"sonar-reasoning" | "sonar" | "sonar-pro" | "sonar-reasoning-pro">;
317
+ declare const perplexity: ModelProvider & ChatProvider<"sonar-reasoning" | "sonar" | "sonar-pro" | "sonar-reasoning-pro">;
236
318
  /**
237
- * ZenMux Provider
238
- * @see {@link https://docs.zenmux.ai}
319
+ * Poe Provider
320
+ * @see {@link https://creator.poe.com/docs/external-applications/openai-compatible-api}
239
321
  * @remarks
240
- * - baseURL - `https://zenmux.ai/api/v1`
241
- * - apiKey - `ZENMUX_API_KEY`
322
+ * - baseURL - `https://api.poe.com/v1`
323
+ * - apiKey - `POE_API_KEY`
242
324
  */
243
- declare const zenmux: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "moonshotai/kimi-k2-0905" | "x-ai/grok-4-fast-non-reasoning" | "x-ai/grok-code-fast-1" | "x-ai/grok-4-fast" | "deepseek/deepseek-chat" | "openai/gpt-5-codex" | "inclusionai/ring-1t" | "inclusionai/lint-1t" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.5">;
325
+ declare const poe: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-2" | "xai/grok-4.1-fast-non-reasoning" | "ideogramai/ideogram" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram-v2" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "poetools/claude-code" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "google/gemini-deep-research" | "google/nano-banana" | "google/imagen-4" | "google/imagen-3" | "google/imagen-4-ultra" | "google/gemini-2.0-flash-lite" | "google/gemini-3-pro" | "google/veo-3.1" | "google/imagen-3-fast" | "google/lyria" | "google/gemini-2.0-flash" | "google/gemini-2.5-flash-lite" | "google/veo-3" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-2" | "google/nano-banana-pro" | "google/veo-3.1-fast" | "openai/gpt-5.2-instant" | "openai/sora-2" | "openai/o1-pro" | "openai/gpt-5.1-codex" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-4-classic" | "openai/gpt-5-chat" | "openai/o3-deep-research" | "openai/gpt-4o-search" | "openai/gpt-image-1-mini" | "openai/gpt-3.5-turbo" | "openai/gpt-5.2-pro" | "openai/o3-mini-high" | "openai/chatgpt-4o-latest" | "openai/gpt-4-turbo" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1-instant" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-4o-aug" | "openai/o3-pro" | "openai/gpt-image-1" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo-instruct" | "openai/o4-mini-deep-research" | "openai/gpt-4-classic-0314" | "openai/dall-e-3" | "openai/sora-2-pro" | "openai/gpt-5-pro" | "openai/gpt-5.2" | "openai/gpt-4o-mini-search" | "stabilityai/stablediffusionxl" | "topazlabs-co/topazlabs" | "lumalabs/ray2" | "lumalabs/dream-machine" | "anthropic/claude-opus-3" | "anthropic/claude-opus-4" | "anthropic/claude-sonnet-3.7-reasoning" | "anthropic/claude-opus-4-search" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-haiku-3.5-search" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4-reasoning" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7-search" | "anthropic/claude-opus-4-reasoning" | "anthropic/claude-sonnet-3.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-haiku-3.5" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-sonnet-4-search" | "trytako/tako" | "novita/kimi-k2-thinking" | "novita/glm-4.6">;
244
326
  /**
245
- * iFlow Provider
246
- * @see {@link https://platform.iflow.cn/en/docs}
327
+ * Requesty Provider
328
+ * @see {@link https://requesty.ai/solution/llm-routing/models}
247
329
  * @remarks
248
- * - baseURL - `https://apis.iflow.cn/v1`
249
- * - apiKey - `IFLOW_API_KEY`
330
+ * - baseURL - `https://router.requesty.ai/v1`
331
+ * - apiKey - `REQUESTY_API_KEY`
250
332
  */
251
- declare const iflowcn: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.6" | "qwen3-vl-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "qwen3-235b" | "qwen3-coder" | "kimi-k2" | "deepseek-v3.1" | "kimi-k2-0905" | "qwen3-235b-a22b-thinking-2507" | "tstars2.0" | "qwen3-235b-a22b-instruct" | "deepseek-v3.2" | "qwen3-max-preview">;
333
+ declare const requesty: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-sonnet-4" | "openai/gpt-4.1-mini" | "openai/o4-mini" | "openai/gpt-4o-mini" | "xai/grok-4" | "anthropic/claude-opus-4" | "xai/grok-4-fast" | "google/gemini-3-flash-preview" | "google/gemini-3-pro-preview" | "anthropic/claude-opus-4-1" | "anthropic/claude-haiku-4-5" | "anthropic/claude-opus-4-5" | "anthropic/claude-sonnet-4-5" | "anthropic/claude-3-7-sonnet">;
252
334
  /**
253
- * Synthetic Provider
254
- * @see {@link https://synthetic.new/pricing}
335
+ * Scaleway Provider
336
+ * @see {@link https://www.scaleway.com/en/docs/generative-apis/}
255
337
  * @remarks
256
- * - baseURL - `https://api.synthetic.new/v1`
257
- * - apiKey - `SYNTHETIC_API_KEY`
338
+ * - baseURL - `https://api.scaleway.ai/v1`
339
+ * - apiKey - `SCALEWAY_API_KEY`
258
340
  */
259
- declare const synthetic: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
341
+ declare const scaleway: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen3-235b-a22b-instruct-2507" | "pixtral-12b-2409" | "whisper-large-v3" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
260
342
  /**
261
- * Deep Infra Provider
262
- * @see {@link https://deepinfra.com/models}
343
+ * SiliconFlow Provider
344
+ * @see {@link https://cloud.siliconflow.com/models}
263
345
  * @remarks
264
- * - baseURL - `https://api.deepinfra.com/v1/openai/`
265
- * - apiKey - `DEEPINFRA_API_KEY`
346
+ * - baseURL - `https://api.siliconflow.com/v1`
347
+ * - apiKey - `SILICONFLOW_API_KEY`
266
348
  */
267
- declare const deepinfra: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<"zai-org/GLM-4.5" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo">;
349
+ declare const siliconflow: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
268
350
  /**
269
- * Zhipu AI Provider
270
- * @see {@link https://docs.z.ai/guides/overview/pricing}
351
+ * SiliconFlow (China) Provider
352
+ * @see {@link https://cloud.siliconflow.com/models}
271
353
  * @remarks
272
- * - baseURL - `https://open.bigmodel.cn/api/paas/v4`
273
- * - apiKey - `ZHIPU_API_KEY`
354
+ * - baseURL - `https://api.siliconflow.cn/v1`
355
+ * - apiKey - `SILICONFLOW_API_KEY`
274
356
  */
275
- declare const zhipuai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6">;
357
+ declare const siliconflowCn: ModelProvider & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
276
358
  /**
277
359
  * submodel Provider
278
360
  * @see {@link https://submodel.gitbook.io}
@@ -280,103 +362,103 @@ declare const zhipuai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
280
362
  * - baseURL - `https://llm.submodel.ai/v1`
281
363
  * - apiKey - `SUBMODEL_INSTAGEN_ACCESS_KEY`
282
364
  */
283
- declare const submodel: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-FP8" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
365
+ declare const submodel: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324" | "zai-org/GLM-4.5-FP8">;
284
366
  /**
285
- * Z.AI Provider
286
- * @see {@link https://docs.z.ai/guides/overview/pricing}
367
+ * Synthetic Provider
368
+ * @see {@link https://synthetic.new/pricing}
287
369
  * @remarks
288
- * - baseURL - `https://api.z.ai/api/paas/v4`
289
- * - apiKey - `ZHIPU_API_KEY`
370
+ * - baseURL - `https://api.synthetic.new/v1`
371
+ * - apiKey - `SYNTHETIC_API_KEY`
290
372
  */
291
- declare const zai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6">;
373
+ declare const synthetic: ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
292
374
  /**
293
- * Inference Provider
294
- * @see {@link https://inference.net/models}
375
+ * Upstage Provider
376
+ * @see {@link https://developers.upstage.ai/docs/apis/chat}
295
377
  * @remarks
296
- * - baseURL - `https://inference.net/v1`
297
- * - apiKey - `INFERENCE_API_KEY`
378
+ * - baseURL - `https://api.upstage.ai`
379
+ * - apiKey - `UPSTAGE_API_KEY`
298
380
  */
299
- declare const inference: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"meta/llama-3.2-11b-vision-instruct" | "mistral/mistral-nemo-12b-instruct" | "google/gemma-3" | "osmosis/osmosis-structure-0.6b" | "qwen/qwen3-embedding-4b" | "qwen/qwen-2.5-7b-vision-instruct" | "meta/llama-3.1-8b-instruct" | "meta/llama-3.2-3b-instruct" | "meta/llama-3.2-1b-instruct">;
381
+ declare const upstage: ModelProvider & ChatProvider<"solar-mini" | "solar-pro2">;
300
382
  /**
301
- * Requesty Provider
302
- * @see {@link https://requesty.ai/solution/llm-routing/models}
383
+ * Venice AI Provider
384
+ * @see {@link https://docs.venice.ai}
303
385
  * @remarks
304
- * - baseURL - `https://router.requesty.ai/v1`
305
- * - apiKey - `REQUESTY_API_KEY`
386
+ * - baseURL - `https://api.venice.ai/api/v1`
387
+ * - apiKey - `VENICE_API_KEY`
306
388
  */
307
- declare const requesty: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"openai/gpt-4.1-mini" | "openai/gpt-4.1" | "openai/o4-mini" | "openai/gpt-4o-mini" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4" | "anthropic/claude-3-7-sonnet" | "anthropic/claude-4-sonnet-20250522" | "anthropic/claude-opus-4-1-20250805">;
389
+ declare const venice: ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "kimi-k2-thinking" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "deepseek-v3.2" | "qwen3-235b" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "openai-gpt-oss-120b" | "zai-org-glm-4.6" | "grok-41-fast" | "claude-opus-45" | "mistral-31-24b" | "venice-uncensored" | "openai-gpt-52" | "qwen3-4b" | "llama-3.3-70b" | "devstral-2-2512" | "llama-3.2-3b" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "zai-org-glm-4.6v" | "qwen3-next-80b">;
308
390
  /**
309
- * Morph Provider
310
- * @see {@link https://docs.morphllm.com/api-reference/introduction}
391
+ * Vultr Provider
392
+ * @see {@link https://api.vultrinference.com/}
311
393
  * @remarks
312
- * - baseURL - `https://api.morphllm.com/v1`
313
- * - apiKey - `MORPH_API_KEY`
394
+ * - baseURL - `https://api.vultrinference.com/v1`
395
+ * - apiKey - `VULTR_API_KEY`
314
396
  */
315
- declare const morph: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
397
+ declare const vultr: ModelProvider & ChatProvider<"deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-instruct" | "qwen2.5-coder-32b-instruct">;
316
398
  /**
317
- * LMStudio Provider
318
- * @see {@link https://lmstudio.ai/models}
399
+ * Weights & Biases Provider
400
+ * @see {@link https://weave-docs.wandb.ai/guides/integrations/inference/}
319
401
  * @remarks
320
- * - baseURL - `http://127.0.0.1:1234/v1`
321
- * - apiKey - `LMSTUDIO_API_KEY`
402
+ * - baseURL - `https://api.inference.wandb.ai/v1`
403
+ * - apiKey - `WANDB_API_KEY`
322
404
  */
323
- declare const lmstudio: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"openai/gpt-oss-20b" | "qwen/qwen3-30b-a3b-2507" | "qwen/qwen3-coder-30b">;
405
+ declare const wandb: ModelProvider & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2-Instruct" | "meta-llama/Llama-3.3-70B-Instruct" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
324
406
  /**
325
- * Fireworks AI Provider
326
- * @see {@link https://fireworks.ai/docs/}
407
+ * xAI Provider
408
+ * @see {@link https://docs.x.ai/docs/models}
327
409
  * @remarks
328
- * - baseURL - `https://api.fireworks.ai/inference/v1/`
329
- * - apiKey - `FIREWORKS_API_KEY`
410
+ * - baseURL - `https://api.x.ai/v1/`
411
+ * - apiKey - `XAI_API_KEY`
330
412
  */
331
- declare const fireworks: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
413
+ declare const xai: ModelProvider & ChatProvider<"grok-code-fast-1" | "grok-4-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-4-1-fast-non-reasoning" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
332
414
  /**
333
- * ModelScope Provider
334
- * @see {@link https://modelscope.cn/docs/model-service/API-Inference/intro}
415
+ * Xiaomi Provider
416
+ * @see {@link https://platform.xiaomimimo.com/#/docs}
335
417
  * @remarks
336
- * - baseURL - `https://api-inference.modelscope.cn/v1`
337
- * - apiKey - `MODELSCOPE_API_KEY`
418
+ * - baseURL - `https://api.xiaomimimo.com/v1`
419
+ * - apiKey - `XIAOMI_API_KEY`
338
420
  */
339
- declare const modelscope: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507">;
421
+ declare const xiaomi: ModelProvider & ChatProvider<"mimo-v2-flash">;
340
422
  /**
341
- * Llama Provider
342
- * @see {@link https://llama.developer.meta.com/docs/models}
423
+ * Z.AI Provider
424
+ * @see {@link https://docs.z.ai/guides/overview/pricing}
343
425
  * @remarks
344
- * - baseURL - `https://api.llama.com/compat/v1/`
345
- * - apiKey - `LLAMA_API_KEY`
426
+ * - baseURL - `https://api.z.ai/api/paas/v4`
427
+ * - apiKey - `ZHIPU_API_KEY`
346
428
  */
347
- declare const llama: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"llama-3.3-8b-instruct" | "llama-4-maverick-17b-128e-instruct-fp8" | "llama-3.3-70b-instruct" | "llama-4-scout-17b-16e-instruct-fp8" | "groq-llama-4-maverick-17b-128e-instruct" | "cerebras-llama-4-scout-17b-16e-instruct" | "cerebras-llama-4-maverick-17b-128e-instruct">;
429
+ declare const zai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
348
430
  /**
349
- * Scaleway Provider
350
- * @see {@link https://www.scaleway.com/en/docs/generative-apis/}
431
+ * Z.AI Coding Plan Provider
432
+ * @see {@link https://docs.z.ai/devpack/overview}
351
433
  * @remarks
352
- * - baseURL - `https://api.scaleway.ai/v1`
353
- * - apiKey - `SCALEWAY_API_KEY`
434
+ * - baseURL - `https://api.z.ai/api/coding/paas/v4`
435
+ * - apiKey - `ZHIPU_API_KEY`
354
436
  */
355
- declare const scaleway: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.3-70b-instruct" | "qwen3-235b-a22b-instruct-2507" | "pixtral-12b-2409" | "llama-3.1-8b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "whisper-large-v3" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
437
+ declare const zaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
356
438
  /**
357
- * Cerebras Provider
358
- * @see {@link https://inference-docs.cerebras.ai/models/overview}
439
+ * ZenMux Provider
440
+ * @see {@link https://docs.zenmux.ai}
359
441
  * @remarks
360
- * - baseURL - `https://api.cerebras.ai/v1/`
361
- * - apiKey - `CEREBRAS_API_KEY`
442
+ * - baseURL - `https://zenmux.ai/api/v1`
443
+ * - apiKey - `ZENMUX_API_KEY`
362
444
  */
363
- declare const cerebras: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<"gpt-oss-120b" | "qwen-3-235b-a22b-instruct-2507" | "zai-glm-4.6" | "qwen-3-coder-480b">;
445
+ declare const zenmux: ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "moonshotai/kimi-k2-thinking" | "openai/gpt-5-codex" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.5" | "moonshotai/kimi-k2-thinking-turbo" | "moonshotai/kimi-k2-0905" | "xiaomi/mimo-v2-flash" | "x-ai/grok-4-fast-non-reasoning" | "x-ai/grok-code-fast-1" | "x-ai/grok-4-fast" | "deepseek/deepseek-chat" | "minimax/minimax-m2" | "inclusionai/ring-1t" | "inclusionai/lint-1t" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1">;
364
446
  /**
365
- * Minimax Provider
366
- * @see {@link https://platform.minimax.io/docs/api-reference/text-openai-api}
447
+ * Zhipu AI Provider
448
+ * @see {@link https://docs.z.ai/guides/overview/pricing}
367
449
  * @remarks
368
- * - baseURL - `https://api.minimax.io/v1/`
369
- * - apiKey - `MINIMAX_API_KEY`
450
+ * - baseURL - `https://open.bigmodel.cn/api/paas/v4`
451
+ * - apiKey - `ZHIPU_API_KEY`
370
452
  */
371
- declare const minimax: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<string>;
453
+ declare const zhipuai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
372
454
  /**
373
- * Minimaxi Provider
374
- * @see {@link https://platform.minimaxi.com/docs/api-reference/text-openai-api}
455
+ * Zhipu AI Coding Plan Provider
456
+ * @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
375
457
  * @remarks
376
- * - baseURL - `https://api.minimaxi.com/v1/`
377
- * - apiKey - `MINIMAX_API_KEY`
458
+ * - baseURL - `https://open.bigmodel.cn/api/coding/paas/v4`
459
+ * - apiKey - `ZHIPU_API_KEY`
378
460
  */
379
- declare const minimaxi: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<string>;
461
+ declare const zhipuaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
380
462
  /**
381
463
  * Novita AI Provider
382
464
  * @see {@link https://novita.ai/docs/guides/llm-api#api-integration}
@@ -384,15 +466,7 @@ declare const minimaxi: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
384
466
  * - baseURL - `https://api.novita.ai/v3/openai/`
385
467
  * - apiKey - `NOVITA_API_KEY`
386
468
  */
387
- declare const novita: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<string>;
388
- /**
389
- * SiliconFlow Provider
390
- * @see {@link https://docs.siliconflow.com/en/userguide/quickstart#4-3-call-via-openai-interface}
391
- * @remarks
392
- * - baseURL - `https://api.siliconflow.cn/v1/`
393
- * - apiKey - `SILICON_FLOW_API_KEY`
394
- */
395
- declare const siliconFlow: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string> & _xsai_ext_shared_providers.ChatProvider<string>;
469
+ declare const novita: ModelProvider & ChatProvider<string>;
396
470
  /**
397
471
  * StepFun Provider
398
472
  * @see {@link https://www.stepfun.com}
@@ -400,7 +474,7 @@ declare const siliconFlow: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_
400
474
  * - baseURL - `https://api.stepfun.com/v1/`
401
475
  * - apiKey - `STEPFUN_API_KEY`
402
476
  */
403
- declare const stepfun: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string> & _xsai_ext_shared_providers.ChatProvider<string>;
477
+ declare const stepfun: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
404
478
  /**
405
479
  * Tencent Hunyuan Provider
406
480
  * @see {@link https://cloud.tencent.com/document/product/1729}
@@ -408,7 +482,23 @@ declare const stepfun: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
408
482
  * - baseURL - `https://api.hunyuan.cloud.tencent.com/v1/`
409
483
  * - apiKey - `TENCENT_HUNYUAN_API_KEY`
410
484
  */
411
- declare const tencentHunyuan: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<string>;
485
+ declare const tencentHunyuan: ModelProvider & EmbedProvider<string> & ChatProvider<string>;
486
+ /**
487
+ * Ollama Provider
488
+ * @see {@link https://docs.ollama.com}
489
+ * @remarks
490
+ * - baseURL - `http://localhost:11434/v1/`
491
+ * - apiKey - `OLLAMA_API_KEY`
492
+ */
493
+ declare const ollama: ModelProvider & EmbedProvider<string> & ChatProvider<string>;
494
+ /**
495
+ * LiteLLM Provider
496
+ * @see {@link https://docs.litellm.ai}
497
+ * @remarks
498
+ * - baseURL - `http://localhost:4000/v1/`
499
+ * - apiKey - `LITELLM_API_KEY`
500
+ */
501
+ declare const litellm: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
412
502
 
413
503
  /**
414
504
  * Anthropic Provider
@@ -417,7 +507,7 @@ declare const tencentHunyuan: _xsai_ext_shared_providers.ModelProvider & _xsai_e
417
507
  * - baseURL - `https://api.anthropic.com/v1/`
418
508
  * - apiKey - `ANTHROPIC_API_KEY`
419
509
  */
420
- declare const anthropic: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<AnthropicModels>;
510
+ declare const anthropic: ModelProvider & ChatProviderWithExtraOptions<AnthropicModels, AnthropicOptions>;
421
511
  /**
422
512
  * Featherless AI Provider
423
513
  * @see {@link https://featherless.ai/models}
@@ -425,7 +515,7 @@ declare const anthropic: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sh
425
515
  * - baseURL - `https://api.featherless.ai/v1/`
426
516
  * - apiKey - `FEATHERLESS_API_KEY`
427
517
  */
428
- declare const featherless: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<string>;
518
+ declare const featherless: ModelProvider & ChatProvider<string>;
429
519
  /**
430
520
  * OpenRouter Provider
431
521
  * @see {@link https://openrouter.ai/models}
@@ -433,7 +523,7 @@ declare const featherless: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_
433
523
  * - baseURL - `https://openrouter.ai/api/v1/`
434
524
  * - apiKey - `OPENROUTER_API_KEY`
435
525
  */
436
- declare const openrouter: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProviderWithExtraOptions<OpenrouterModels, OpenRouterOptions>;
526
+ declare const openrouter: ModelProvider & ChatProviderWithExtraOptions<OpenrouterModels, OpenRouterOptions>;
437
527
  /**
438
528
  * Together AI Provider
439
529
  * @see {@link https://docs.together.ai/docs/serverless-models}
@@ -441,6 +531,6 @@ declare const openrouter: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_s
441
531
  * - baseURL - `https://api.together.xyz/v1/`
442
532
  * - apiKey - `TOGETHER_API_KEY`
443
533
  */
444
- declare const togetherai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<TogetheraiModels>;
534
+ declare const togetherai: ModelProvider & EmbedProvider<string> & ChatProvider<TogetheraiModels>;
445
535
 
446
- export { alibaba, alibabaCn, anthropic, baseten, cerebras, chutes, cortecs, deepinfra, deepseek, fastrouter, featherless, fireworks, githubCopilot, githubModels, google, groq, huggingface, iflowcn, inception, inference, llama, lmstudio, lucidquery, minimax, minimaxi, mistral, modelscope, moonshotai, moonshotaiCn, morph, nebius, novita, nvidia, openai, opencode, openrouter, perplexity, requesty, scaleway, siliconFlow, stepfun, submodel, synthetic, tencentHunyuan, togetherai, upstage, venice, vultr, wandb, xai, zai, zaiCodingPlan, zenmux, zhipuai, zhipuaiCodingPlan };
536
+ export { alibaba, alibabaCn, anthropic, bailing, baseten, cerebras, chutes, cortecs, deepinfra, deepseek, fastrouter, featherless, fireworks, githubCopilot, githubModels, google, groq, helicone, huggingface, iflowcn, inception, inference, ioNet, kimiForCoding, litellm, llama, lmstudio, lucidquery, minimax, minimaxCn, mistral, modelscope, moonshotai, moonshotaiCn, morph, nebius, novita, nvidia, ollama, ollamaCloud, openai, opencode, openrouter, ovhcloud, perplexity, poe, requesty, scaleway, siliconflow, siliconflowCn, stepfun, submodel, synthetic, tencentHunyuan, togetherai, upstage, venice, vultr, wandb, xai, xiaomi, zai, zaiCodingPlan, zenmux, zhipuai, zhipuaiCodingPlan };