@xsai-ext/providers 0.4.0-beta.9 → 0.4.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/create.d.ts CHANGED
@@ -1,336 +1,329 @@
1
- import * as _xsai_ext_shared_providers from '@xsai-ext/shared-providers';
2
- import { A as AnthropicModels, b as AzureModels, T as TogetheraiModels, C as CloudflareWorkersAiModels } from './openrouter-DCc0QhRt.js';
3
- export { c as createOpenRouter } from './openrouter-DCc0QhRt.js';
1
+ import { C as ChatProvider, M as ModelProvider, E as EmbedProvider, I as ImageProvider, S as SpeechProvider, T as TranscriptionProvider } from './types-DW1hvH0W.js';
2
+ export { c as createAnthropic, a as createOpenRouter } from './anthropic-Bel15KQV.js';
3
+ import { T as TogetheraiModels } from './types-Czo87gCh.js';
4
+ import '@xsai/shared';
4
5
 
5
6
  /**
6
- * Create a Moonshot AI (China) Provider
7
- * @see {@link https://platform.moonshot.cn/docs/api/chat}
7
+ * Create a Alibaba Provider
8
+ * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
8
9
  */
9
- declare const createMoonshotaiCn: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ChatProvider<"kimi-k2-0905-preview" | "kimi-k2-0711-preview" | "kimi-k2-turbo-preview"> & _xsai_ext_shared_providers.ModelProvider;
10
+ declare const createAlibaba: (apiKey: string, baseURL?: string) => ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus"> & ModelProvider;
10
11
  /**
11
- * Create a LucidQuery AI Provider
12
- * @see {@link https://lucidquery.com/api/docs}
12
+ * Create a Alibaba (China) Provider
13
+ * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
13
14
  */
14
- declare const createLucidquery: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"lucidquery-nexus-coder" | "lucidnova-rf1-100b">;
15
+ declare const createAlibabaCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
15
16
  /**
16
- * Create a Moonshot AI Provider
17
- * @see {@link https://platform.moonshot.ai/docs/api/chat}
17
+ * Create a Bailing Provider
18
+ * @see {@link https://alipaytbox.yuque.com/sxs0ba/ling/intro}
18
19
  */
19
- declare const createMoonshotai: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ChatProvider<"kimi-k2-0905-preview" | "kimi-k2-0711-preview" | "kimi-k2-turbo-preview"> & _xsai_ext_shared_providers.ModelProvider;
20
+ declare const createBailing: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Ling-1T" | "Ring-1T">;
20
21
  /**
21
- * Create a Z.AI Coding Plan Provider
22
- * @see {@link https://docs.z.ai/devpack/overview}
22
+ * Create a Baseten Provider
23
+ * @see {@link https://docs.baseten.co/development/model-apis/overview}
23
24
  */
24
- declare const createZaiCodingPlan: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6">;
25
+ declare const createBaseten: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2">;
25
26
  /**
26
- * Create a Alibaba Provider
27
- * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
27
+ * Create a Cerebras Provider
28
+ * @see {@link https://inference-docs.cerebras.ai/models/overview}
28
29
  */
29
- declare const createAlibaba: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus">;
30
+ declare const createCerebras: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen-3-235b-a22b-instruct-2507" | "zai-glm-4.6" | "gpt-oss-120b"> & EmbedProvider<string>;
30
31
  /**
31
- * Create a xAI Provider
32
- * @see {@link https://docs.x.ai/docs/models}
32
+ * Create a Chutes Provider
33
+ * @see {@link https://llm.chutes.ai/v1/models}
33
34
  */
34
- declare const createXai: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"grok-4-fast-non-reasoning" | "grok-3-fast" | "grok-4" | "grok-2-vision" | "grok-code-fast-1" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-3" | "grok-4-fast" | "grok-2-latest" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-3-mini" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
35
+ declare const createChutes: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2" | "NousResearch/Hermes-4.3-36B" | "NousResearch/Hermes-4-70B" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4-405B-FP8" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "rednote-hilab/dots.ocr" | "MiniMaxAI/MiniMax-M2" | "ArliAI/QwQ-32B-ArliAI-RpR-v1" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "tngtech/TNG-R1T-Chimera-TEE" | "OpenGVLab/InternVL3-78B" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "Alibaba-NLP/Tongyi-DeepResearch-30B-A3B" | "mistralai/Devstral-2-123B-Instruct-2512" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "unsloth/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen3-32B" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-V3" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
35
36
  /**
36
- * Create a Vultr Provider
37
- * @see {@link https://api.vultrinference.com/}
37
+ * Create a Cortecs Provider
38
+ * @see {@link https://api.cortecs.ai/v1/models}
38
39
  */
39
- declare const createVultr: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"deepseek-r1-distill-qwen-32b" | "qwen2.5-coder-32b-instruct" | "kimi-k2-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b">;
40
+ declare const createCortecs: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "gpt-4.1" | "gemini-2.5-pro" | "devstral-small-2512" | "claude-sonnet-4" | "llama-3.1-405b-instruct">;
40
41
  /**
41
- * Create a Nvidia Provider
42
- * @see {@link https://docs.api.nvidia.com/nim/}
42
+ * Create a Deep Infra Provider
43
+ * @see {@link https://deepinfra.com/models}
43
44
  */
44
- declare const createNvidia: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "minimaxai/minimax-m2" | "google/gemma-3-27b-it" | "microsoft/phi-4-mini-instruct" | "openai/whisper-large-v3" | "openai/gpt-oss-120b" | "qwen/qwen3-235b-a22b" | "qwen/qwen3-coder-480b-a35b-instruct" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "black-forest-labs/flux.1-dev">;
45
+ declare const createDeepinfra: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "zai-org/GLM-4.5" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo">;
45
46
  /**
46
- * Create a Upstage Provider
47
- * @see {@link https://developers.upstage.ai/docs/apis/chat}
47
+ * Create a DeepSeek Provider
48
+ * @see {@link https://platform.deepseek.com/api-docs/pricing}
48
49
  */
49
- declare const createUpstage: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"solar-mini" | "solar-pro2">;
50
+ declare const createDeepSeek: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"deepseek-chat" | "deepseek-reasoner">;
50
51
  /**
51
- * Create a Groq Provider
52
- * @see {@link https://console.groq.com/docs/models}
52
+ * Create a FastRouter Provider
53
+ * @see {@link https://fastrouter.ai/models}
54
+ */
55
+ declare const createFastrouter: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
56
+ /**
57
+ * Create a Fireworks AI Provider
58
+ * @see {@link https://fireworks.ai/docs/}
53
59
  */
54
- declare const createGroq: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"deepseek-r1-distill-llama-70b" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "openai/gpt-oss-120b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "openai/gpt-oss-20b" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b"> & _xsai_ext_shared_providers.EmbedProvider<string>;
60
+ declare const createFireworks: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
55
61
  /**
56
62
  * Create a GitHub Copilot Provider
57
63
  * @see {@link https://docs.github.com/en/copilot}
58
64
  */
59
- declare const createGithubCopilot: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"grok-code-fast-1" | "gemini-2.0-flash-001" | "claude-opus-4" | "claude-haiku-4.5" | "claude-3.5-sonnet" | "o3-mini" | "gpt-5-codex" | "gpt-4o" | "gpt-4.1" | "o4-mini" | "claude-opus-41" | "gpt-5-mini" | "claude-3.7-sonnet" | "gemini-2.5-pro" | "o3" | "claude-sonnet-4" | "gpt-5" | "claude-3.7-sonnet-thought" | "claude-sonnet-4.5">;
65
+ declare const createGithubCopilot: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "gemini-2.0-flash-001" | "claude-opus-4" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gpt-5.1-codex" | "claude-haiku-4.5" | "gemini-3-pro-preview" | "oswe-vscode-prime" | "claude-3.5-sonnet" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "claude-opus-41" | "gpt-5-mini" | "claude-3.7-sonnet" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "claude-3.7-sonnet-thought" | "claude-opus-4.5" | "gpt-5.2" | "claude-sonnet-4.5">;
60
66
  /**
61
- * Create a Mistral Provider
62
- * @see {@link https://docs.mistral.ai/getting-started/models/}
67
+ * Create a GitHub Models Provider
68
+ * @see {@link https://docs.github.com/en/github-models}
69
+ */
70
+ declare const createGithubModels: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-4.1" | "core42/jais-30b-chat" | "xai/grok-3" | "xai/grok-3-mini" | "cohere/cohere-command-r-08-2024" | "cohere/cohere-command-a" | "cohere/cohere-command-r-plus-08-2024" | "cohere/cohere-command-r" | "cohere/cohere-command-r-plus" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-r1" | "deepseek/deepseek-v3-0324" | "mistral-ai/mistral-medium-2505" | "mistral-ai/ministral-3b" | "mistral-ai/mistral-nemo" | "mistral-ai/mistral-large-2411" | "mistral-ai/codestral-2501" | "mistral-ai/mistral-small-2503" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-mini-4k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-4" | "microsoft/phi-4-mini-reasoning" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-mini-instruct" | "microsoft/phi-4-multimodal-instruct" | "microsoft/phi-3-mini-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-mini-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-4-reasoning" | "microsoft/mai-ds-r1" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o1-preview" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o1-mini" | "openai/o3" | "openai/gpt-4o-mini" | "meta/llama-3.2-11b-vision-instruct" | "meta/meta-llama-3.1-405b-instruct" | "meta/llama-4-maverick-17b-128e-instruct-fp8" | "meta/meta-llama-3-70b-instruct" | "meta/meta-llama-3.1-70b-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-3.2-90b-vision-instruct" | "meta/meta-llama-3-8b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/meta-llama-3.1-8b-instruct" | "ai21-labs/ai21-jamba-1.5-large" | "ai21-labs/ai21-jamba-1.5-mini">;
71
+ /**
72
+ * Create a Google Provider
73
+ * @see {@link https://ai.google.dev/gemini-api/docs/pricing}
63
74
  */
64
- declare const createMistral: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<"devstral-medium-2507" | "open-mixtral-8x22b" | "ministral-8b-latest" | "pixtral-large-latest" | "ministral-3b-latest" | "pixtral-12b" | "mistral-medium-2505" | "devstral-small-2505" | "mistral-medium-2508" | "mistral-small-latest" | "magistral-small" | "devstral-small-2507" | "codestral-latest" | "open-mixtral-8x7b" | "mistral-nemo" | "open-mistral-7b" | "mistral-large-latest" | "mistral-medium-latest" | "magistral-medium-latest">;
75
+ declare const createGoogleGenerativeAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-2.5-pro" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-2.5-flash" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
65
76
  /**
66
- * Create a Nebius AI Studio Provider
67
- * @see {@link https://docs.studio.nebius.com/quickstart}
77
+ * Create a Groq Provider
78
+ * @see {@link https://console.groq.com/docs/models}
68
79
  */
69
- declare const createNebius: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"moonshotai/kimi-k2-instruct" | "openai/gpt-oss-120b" | "qwen/qwen3-coder-480b-a35b-instruct" | "openai/gpt-oss-20b" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-235b-a22b-thinking-2507" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "zai-org/glm-4.5" | "zai-org/glm-4.5-air" | "deepseek-ai/deepseek-v3">;
80
+ declare const createGroq: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b">;
70
81
  /**
71
- * Create a DeepSeek Provider
72
- * @see {@link https://platform.deepseek.com/api-docs/pricing}
82
+ * Create a Helicone Provider
83
+ * @see {@link https://helicone.ai/models}
73
84
  */
74
- declare const createDeepSeek: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"deepseek-chat" | "deepseek-reasoner">;
85
+ declare const createHelicone: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-next-80b-a3b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-thinking" | "gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "deepseek-reasoner" | "claude-opus-4" | "grok-code-fast-1" | "gpt-5.1-codex" | "gemini-3-pro-preview" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "claude-3.7-sonnet" | "o3" | "gpt-5" | "gemini-2.5-flash" | "gemini-2.5-flash-lite" | "llama-3.1-8b-instant" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "qwen3-coder" | "grok-4-fast-reasoning" | "claude-opus-4-1" | "grok-4" | "llama-4-maverick" | "llama-prompt-guard-2-86m" | "grok-4-1-fast-reasoning" | "claude-4.5-haiku" | "llama-3.1-8b-instruct-turbo" | "gpt-4.1-mini-2025-04-14" | "llama-guard-4" | "llama-3.1-8b-instruct" | "gpt-4.1-mini" | "deepseek-v3.1-terminus" | "llama-prompt-guard-2-22m" | "claude-3.5-sonnet-v2" | "sonar-deep-research" | "claude-sonnet-4-5-20250929" | "grok-3" | "mistral-small" | "kimi-k2-0711" | "chatgpt-4o-latest" | "kimi-k2-0905" | "sonar-reasoning" | "llama-3.3-70b-instruct" | "claude-4.5-sonnet" | "codex-mini-latest" | "gpt-5-nano" | "deepseek-tng-r1t2-chimera" | "claude-4.5-opus" | "sonar" | "glm-4.6" | "qwen3-235b-a22b-thinking" | "hermes-2-pro-llama-3-8b" | "o1" | "grok-3-mini" | "sonar-pro" | "o1-mini" | "claude-3-haiku-20240307" | "o3-pro" | "qwen2.5-coder-7b-fast" | "gemma-3-12b-it" | "mistral-nemo" | "gpt-oss-20b" | "claude-3.5-haiku" | "gpt-5-chat-latest" | "gpt-4o-mini" | "sonar-reasoning-pro" | "qwen3-vl-235b-a22b-instruct" | "qwen3-30b-a3b" | "deepseek-v3.2" | "grok-4-1-fast-non-reasoning" | "gpt-5-pro" | "mistral-large-2411" | "claude-opus-4-1-20250805" | "ernie-4.5-21b-a3b-thinking" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "llama-4-scout">;
75
86
  /**
76
- * Create a Alibaba (China) Provider
77
- * @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
87
+ * Create a Hugging Face Provider
88
+ * @see {@link https://huggingface.co/docs/inference-providers}
78
89
  */
79
- declare const createAlibabaCn: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
90
+ declare const createHuggingface: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "MiniMaxAI/MiniMax-M2" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/Deepseek-V3-0324">;
80
91
  /**
81
- * Create a Venice AI Provider
82
- * @see {@link https://docs.venice.ai}
92
+ * Create a iFlow Provider
93
+ * @see {@link https://platform.iflow.cn/en/docs}
83
94
  */
84
- declare const createVenice: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"dolphin-2.9.2-qwen2-72b" | "mistral-31-24b" | "venice-uncensored" | "qwen-2.5-vl" | "qwen3-235b" | "qwen-2.5-qwq-32b" | "deepseek-coder-v2-lite" | "qwen3-4b" | "llama-3.3-70b" | "qwen-2.5-coder-32b" | "deepseek-r1-671b" | "llama-3.2-3b" | "llama-3.1-405b">;
95
+ declare const createIflowcn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-vl-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "kimi-k2-thinking" | "qwen3-coder" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "kimi-k2" | "deepseek-v3.1" | "minimax-m2" | "qwen3-235b" | "deepseek-v3.2-chat" | "qwen3-235b-a22b-thinking-2507" | "tstars2.0" | "qwen3-235b-a22b-instruct" | "qwen3-max-preview">;
85
96
  /**
86
- * Create a Chutes Provider
87
- * @see {@link https://llm.chutes.ai/v1/models}
97
+ * Create a Inception Provider
98
+ * @see {@link https://platform.inceptionlabs.ai/docs}
88
99
  */
89
- declare const createChutes: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"openai/gpt-oss-120b" | "moonshotai/Kimi-K2-Instruct-75k" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-VL-A3B-Thinking" | "MiniMaxAI/MiniMax-M2" | "meituan-longcat/LongCat-Flash-Chat-FP8" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "zai-org/GLM-4.5" | "zai-org/GLM-4.6-FP8" | "zai-org/GLM-4.6-turbo" | "zai-org/GLM-4.6" | "zai-org/GLM-4.5-FP8" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.2-Exp" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.1:THINKING" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
100
+ declare const createInception: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"mercury-coder" | "mercury">;
90
101
  /**
91
- * Create a Cortecs Provider
92
- * @see {@link https://api.cortecs.ai/v1/models}
102
+ * Create a Inference Provider
103
+ * @see {@link https://inference.net/models}
93
104
  */
94
- declare const createCortecs: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"qwen3-coder-480b-a35b-instruct" | "qwen3-32b" | "kimi-k2-instruct" | "gpt-oss-120b" | "gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "nova-pro-v1" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "llama-3.1-405b-instruct">;
105
+ declare const createInference: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"meta/llama-3.2-11b-vision-instruct" | "mistral/mistral-nemo-12b-instruct" | "google/gemma-3" | "osmosis/osmosis-structure-0.6b" | "qwen/qwen3-embedding-4b" | "qwen/qwen-2.5-7b-vision-instruct" | "meta/llama-3.1-8b-instruct" | "meta/llama-3.2-3b-instruct" | "meta/llama-3.2-1b-instruct">;
95
106
  /**
96
- * Create a GitHub Models Provider
97
- * @see {@link https://docs.github.com/en/github-models}
107
+ * Create a IO.NET Provider
108
+ * @see {@link https://io.net/docs/guides/intelligence/io-intelligence}
98
109
  */
99
- declare const createGithubModels: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"microsoft/phi-4-mini-instruct" | "core42/jais-30b-chat" | "xai/grok-3" | "xai/grok-3-mini" | "cohere/cohere-command-r-08-2024" | "cohere/cohere-command-a" | "cohere/cohere-command-r-plus-08-2024" | "cohere/cohere-command-r" | "cohere/cohere-command-r-plus" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-r1" | "deepseek/deepseek-v3-0324" | "mistral-ai/mistral-medium-2505" | "mistral-ai/ministral-3b" | "mistral-ai/mistral-nemo" | "mistral-ai/mistral-large-2411" | "mistral-ai/codestral-2501" | "mistral-ai/mistral-small-2503" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-mini-4k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-4" | "microsoft/phi-4-mini-reasoning" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-mini-instruct" | "microsoft/phi-4-multimodal-instruct" | "microsoft/phi-3-mini-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-4-reasoning" | "microsoft/mai-ds-r1" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o1-preview" | "openai/o3-mini" | "openai/gpt-4o" | "openai/gpt-4.1" | "openai/o4-mini" | "openai/o1" | "openai/o1-mini" | "openai/o3" | "openai/gpt-4o-mini" | "meta/llama-3.2-11b-vision-instruct" | "meta/meta-llama-3.1-405b-instruct" | "meta/llama-4-maverick-17b-128e-instruct-fp8" | "meta/meta-llama-3-70b-instruct" | "meta/meta-llama-3.1-70b-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-3.2-90b-vision-instruct" | "meta/meta-llama-3-8b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/meta-llama-3.1-8b-instruct" | "ai21-labs/ai21-jamba-1.5-large" | "ai21-labs/ai21-jamba-1.5-mini">;
110
+ declare const createIoNet: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506" | "mistralai/Mistral-Large-Instruct-2411" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar">;
100
111
  /**
101
- * Create a Baseten Provider
102
- * @see {@link https://docs.baseten.co/development/model-apis/overview}
112
+ * Create a Kimi For Coding Provider
113
+ * @see {@link https://www.kimi.com/coding/docs/en/third-party-agents.html}
103
114
  */
104
- declare const createBaseten: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "zai-org/GLM-4.6" | "Qwen3/Qwen3-Coder-480B-A35B-Instruct">;
115
+ declare const createKimiForCoding: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking">;
105
116
  /**
106
- * Create a Hugging Face Provider
107
- * @see {@link https://huggingface.co/docs/inference-providers}
117
+ * Create a Llama Provider
118
+ * @see {@link https://llama.developer.meta.com/docs/models}
108
119
  */
109
- declare const createHuggingface: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "MiniMaxAI/MiniMax-M2" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "zai-org/GLM-4.5" | "zai-org/GLM-4.6" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "deepseek-ai/Deepseek-V3-0324">;
120
+ declare const createLlama: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"llama-3.3-70b-instruct" | "llama-3.3-8b-instruct" | "llama-4-maverick-17b-128e-instruct-fp8" | "llama-4-scout-17b-16e-instruct-fp8" | "groq-llama-4-maverick-17b-128e-instruct" | "cerebras-llama-4-scout-17b-16e-instruct" | "cerebras-llama-4-maverick-17b-128e-instruct">;
110
121
  /**
111
- * Create a OpenCode Zen Provider
112
- * @see {@link https://opencode.ai/docs/zen}
122
+ * Create a LMStudio Provider
123
+ * @see {@link https://lmstudio.ai/models}
113
124
  */
114
- declare const createOpencode: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.6" | "gpt-5-codex" | "claude-sonnet-4" | "gpt-5" | "qwen3-coder" | "claude-opus-4-1" | "kimi-k2" | "claude-haiku-4-5" | "minimax-m2" | "claude-sonnet-4-5" | "an-gbt" | "big-pickle" | "claude-3-5-haiku" | "grok-code">;
125
+ declare const createLmstudio: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "qwen/qwen3-30b-a3b-2507" | "qwen/qwen3-coder-30b">;
115
126
  /**
116
- * Create a FastRouter Provider
117
- * @see {@link https://fastrouter.ai/models}
127
+ * Create a LucidQuery AI Provider
128
+ * @see {@link https://lucidquery.com/api/docs}
118
129
  */
119
- declare const createFastrouter: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"openai/gpt-oss-120b" | "openai/gpt-oss-20b" | "openai/gpt-4.1" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
130
+ declare const createLucidquery: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"lucidquery-nexus-coder" | "lucidnova-rf1-100b">;
120
131
  /**
121
- * Create a Google Provider
122
- * @see {@link https://ai.google.dev/gemini-api/docs/pricing}
132
+ * Create a MiniMax Provider
133
+ * @see {@link https://platform.minimax.io/docs/guides/quickstart}
123
134
  */
124
- declare const createGoogleGenerativeAI: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<"gemini-2.5-pro" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-2.5-flash" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
135
+ declare const createMinimax: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2">;
125
136
  /**
126
- * Create a Inception Provider
127
- * @see {@link https://platform.inceptionlabs.ai/docs}
137
+ * Create a MiniMax (China) Provider
138
+ * @see {@link https://platform.minimaxi.com/docs/guides/quickstart}
128
139
  */
129
- declare const createInception: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"mercury-coder" | "mercury">;
140
+ declare const createMinimaxCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2">;
130
141
  /**
131
- * Create a Weights & Biases Provider
132
- * @see {@link https://weave-docs.wandb.ai/guides/integrations/inference/}
142
+ * Create a Mistral Provider
143
+ * @see {@link https://docs.mistral.ai/getting-started/models/}
144
+ */
145
+ declare const createMistral: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"devstral-2512" | "mistral-nemo" | "mistral-large-2411" | "devstral-medium-2507" | "mistral-large-2512" | "open-mixtral-8x22b" | "ministral-8b-latest" | "pixtral-large-latest" | "mistral-small-2506" | "ministral-3b-latest" | "pixtral-12b" | "mistral-medium-2505" | "labs-devstral-small-2512" | "devstral-medium-latest" | "devstral-small-2505" | "mistral-medium-2508" | "mistral-embed" | "mistral-small-latest" | "magistral-small" | "devstral-small-2507" | "codestral-latest" | "open-mixtral-8x7b" | "open-mistral-7b" | "mistral-large-latest" | "mistral-medium-latest" | "magistral-medium-latest">;
146
+ /**
147
+ * Create a ModelScope Provider
148
+ * @see {@link https://modelscope.cn/docs/model-service/API-Inference/intro}
149
+ */
150
+ declare const createModelscope: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507">;
151
+ /**
152
+ * Create a Moonshot AI Provider
153
+ * @see {@link https://platform.moonshot.ai/docs/api/chat}
154
+ */
155
+ declare const createMoonshotai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
156
+ /**
157
+ * Create a Moonshot AI (China) Provider
158
+ * @see {@link https://platform.moonshot.cn/docs/api/chat}
159
+ */
160
+ declare const createMoonshotaiCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
161
+ /**
162
+ * Create a Morph Provider
163
+ * @see {@link https://docs.morphllm.com/api-reference/introduction}
164
+ */
165
+ declare const createMorph: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
166
+ /**
167
+ * Create a Nebius Token Factory Provider
168
+ * @see {@link https://docs.tokenfactory.nebius.com/}
169
+ */
170
+ declare const createNebius: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2-instruct" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "zai-org/glm-4.5" | "zai-org/glm-4.5-air" | "deepseek-ai/deepseek-v3">;
171
+ /**
172
+ * Create a Nvidia Provider
173
+ * @see {@link https://docs.api.nvidia.com/nim/}
174
+ */
175
+ declare const createNvidia: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-4-mini-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-coder-480b-a35b-instruct" | "moonshotai/kimi-k2-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "minimaxai/minimax-m2" | "google/gemma-3-27b-it" | "openai/whisper-large-v3" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwen3-next-80b-a3b-thinking" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "black-forest-labs/flux.1-dev">;
176
+ /**
177
+ * Create a Ollama Cloud Provider
178
+ * @see {@link https://docs.ollama.com/cloud}
133
179
  */
134
- declare const createWandb: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
180
+ declare const createOllamaCloud: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking:cloud" | "qwen3-vl-235b-cloud" | "qwen3-coder:480b-cloud" | "gpt-oss:120b-cloud" | "deepseek-v3.1:671b-cloud" | "glm-4.6:cloud" | "cogito-2.1:671b-cloud" | "gpt-oss:20b-cloud" | "qwen3-vl-235b-instruct-cloud" | "kimi-k2:1t-cloud" | "minimax-m2:cloud" | "gemini-3-pro-preview:latest">;
135
181
  /**
136
182
  * Create a OpenAI Provider
137
183
  * @see {@link https://platform.openai.com/docs/models}
138
184
  */
139
- declare const createOpenAI: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<"o3-mini" | "gpt-5-codex" | "gpt-4o" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "o3" | "gpt-5" | "gpt-4.1-nano" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "gpt-4.1-mini" | "o3-deep-research" | "gpt-3.5-turbo" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "codex-mini-latest" | "gpt-5-nano" | "o1" | "o1-mini" | "text-embedding-ada-002" | "o3-pro" | "gpt-4o-2024-11-20" | "o4-mini-deep-research" | "gpt-5-chat-latest" | "gpt-4o-mini" | "gpt-5-pro"> & _xsai_ext_shared_providers.ImageProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string>;
185
+ declare const createOpenAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "gpt-5.2" | "gpt-4.1-nano" | "gpt-4.1-mini" | "codex-mini-latest" | "gpt-5-nano" | "o1" | "o1-mini" | "o3-pro" | "gpt-5-chat-latest" | "gpt-4o-mini" | "gpt-5-pro" | "gpt-5.1-chat-latest" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "o3-deep-research" | "gpt-3.5-turbo" | "gpt-5.2-pro" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "gpt-5.2-chat-latest" | "text-embedding-ada-002" | "gpt-4o-2024-11-20" | "o4-mini-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
140
186
  /**
141
- * Create a Zhipu AI Coding Plan Provider
142
- * @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
187
+ * Create a OpenCode Zen Provider
188
+ * @see {@link https://opencode.ai/docs/zen}
189
+ */
190
+ declare const createOpencode: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking" | "claude-sonnet-4" | "gpt-5.1-codex" | "gpt-5.1" | "gpt-5-codex" | "gpt-5.1-codex-max" | "gpt-5" | "gpt-5.2" | "qwen3-coder" | "claude-opus-4-1" | "gpt-5-nano" | "glm-4.6" | "kimi-k2" | "claude-haiku-4-5" | "claude-opus-4-5" | "gemini-3-pro" | "claude-sonnet-4-5" | "alpha-gd4" | "big-pickle" | "claude-3-5-haiku" | "grok-code" | "gemini-3-flash" | "alpha-doubao-seed-code" | "minimax-m2.1">;
191
+ /**
192
+ * Create a OVHcloud AI Endpoints Provider
193
+ * @see {@link https://www.ovhcloud.com/en/public-cloud/ai-endpoints/catalog//}
143
194
  */
144
- declare const createZhipuaiCodingPlan: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6">;
195
+ declare const createOvhcloud: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "gpt-oss-20b" | "mixtral-8x7b-instruct-v0.1" | "mistral-7b-instruct-v0.3" | "qwen2.5-vl-72b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen2.5-coder-32b-instruct" | "llava-next-mistral-7b" | "meta-llama-3_1-70b-instruct" | "meta-llama-3_3-70b-instruct">;
145
196
  /**
146
197
  * Create a Perplexity Provider
147
198
  * @see {@link https://docs.perplexity.ai}
148
199
  */
149
- declare const createPerplexity: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"sonar-reasoning" | "sonar" | "sonar-pro" | "sonar-reasoning-pro">;
200
+ declare const createPerplexity: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"sonar-reasoning" | "sonar" | "sonar-pro" | "sonar-reasoning-pro">;
150
201
  /**
151
- * Create a ZenMux Provider
152
- * @see {@link https://docs.zenmux.ai}
202
+ * Create a Poe Provider
203
+ * @see {@link https://creator.poe.com/docs/external-applications/openai-compatible-api}
153
204
  */
154
- declare const createZenmux: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "moonshotai/kimi-k2-0905" | "x-ai/grok-4-fast-non-reasoning" | "x-ai/grok-code-fast-1" | "x-ai/grok-4-fast" | "deepseek/deepseek-chat" | "openai/gpt-5-codex" | "inclusionai/ring-1t" | "inclusionai/lint-1t" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.5">;
205
+ declare const createPoe: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-2" | "xai/grok-4.1-fast-non-reasoning" | "ideogramai/ideogram" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram-v2" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "poetools/claude-code" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "google/gemini-deep-research" | "google/nano-banana" | "google/imagen-4" | "google/imagen-3" | "google/imagen-4-ultra" | "google/gemini-2.0-flash-lite" | "google/gemini-3-pro" | "google/veo-3.1" | "google/imagen-3-fast" | "google/lyria" | "google/gemini-2.0-flash" | "google/gemini-2.5-flash-lite" | "google/veo-3" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-2" | "google/nano-banana-pro" | "google/veo-3.1-fast" | "openai/gpt-5.2-instant" | "openai/sora-2" | "openai/o1-pro" | "openai/gpt-5.1-codex" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-4-classic" | "openai/gpt-5-chat" | "openai/o3-deep-research" | "openai/gpt-4o-search" | "openai/gpt-image-1-mini" | "openai/gpt-3.5-turbo" | "openai/gpt-5.2-pro" | "openai/o3-mini-high" | "openai/chatgpt-4o-latest" | "openai/gpt-4-turbo" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1-instant" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-4o-aug" | "openai/o3-pro" | "openai/gpt-image-1" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo-instruct" | "openai/o4-mini-deep-research" | "openai/gpt-4-classic-0314" | "openai/dall-e-3" | "openai/sora-2-pro" | "openai/gpt-5-pro" | "openai/gpt-5.2" | "openai/gpt-4o-mini-search" | "stabilityai/stablediffusionxl" | "topazlabs-co/topazlabs" | "lumalabs/ray2" | "lumalabs/dream-machine" | "anthropic/claude-opus-3" | "anthropic/claude-opus-4" | "anthropic/claude-sonnet-3.7-reasoning" | "anthropic/claude-opus-4-search" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-haiku-3.5-search" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4-reasoning" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7-search" | "anthropic/claude-opus-4-reasoning" | "anthropic/claude-sonnet-3.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-haiku-3.5" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-sonnet-4-search" | "trytako/tako" | "novita/kimi-k2-thinking" | "novita/glm-4.6">;
155
206
  /**
156
- * Create a iFlow Provider
157
- * @see {@link https://platform.iflow.cn/en/docs}
207
+ * Create a Requesty Provider
208
+ * @see {@link https://requesty.ai/solution/llm-routing/models}
158
209
  */
159
- declare const createIflowcn: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.6" | "qwen3-vl-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "qwen3-235b" | "qwen3-coder" | "kimi-k2" | "deepseek-v3.1" | "kimi-k2-0905" | "qwen3-235b-a22b-thinking-2507" | "tstars2.0" | "qwen3-235b-a22b-instruct" | "deepseek-v3.2" | "qwen3-max-preview">;
210
+ declare const createRequesty: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-sonnet-4" | "openai/gpt-4.1-mini" | "openai/o4-mini" | "openai/gpt-4o-mini" | "xai/grok-4" | "anthropic/claude-opus-4" | "xai/grok-4-fast" | "google/gemini-3-flash-preview" | "google/gemini-3-pro-preview" | "anthropic/claude-opus-4-1" | "anthropic/claude-haiku-4-5" | "anthropic/claude-opus-4-5" | "anthropic/claude-sonnet-4-5" | "anthropic/claude-3-7-sonnet">;
160
211
  /**
161
- * Create a Synthetic Provider
162
- * @see {@link https://synthetic.new/pricing}
212
+ * Create a Scaleway Provider
213
+ * @see {@link https://www.scaleway.com/en/docs/generative-apis/}
163
214
  */
164
- declare const createSynthetic: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
215
+ declare const createScaleway: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen3-235b-a22b-instruct-2507" | "pixtral-12b-2409" | "whisper-large-v3" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
165
216
  /**
166
- * Create a Deep Infra Provider
167
- * @see {@link https://deepinfra.com/models}
217
+ * Create a SiliconFlow Provider
218
+ * @see {@link https://cloud.siliconflow.com/models}
168
219
  */
169
- declare const createDeepinfra: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<"zai-org/GLM-4.5" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo">;
220
+ declare const createSiliconFlow: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
170
221
  /**
171
- * Create a Zhipu AI Provider
172
- * @see {@link https://docs.z.ai/guides/overview/pricing}
222
+ * Create a SiliconFlow (China) Provider
223
+ * @see {@link https://cloud.siliconflow.com/models}
173
224
  */
174
- declare const createZhipuai: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6">;
225
+ declare const createSiliconflowCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
175
226
  /**
176
227
  * Create a submodel Provider
177
228
  * @see {@link https://submodel.gitbook.io}
178
229
  */
179
- declare const createSubmodel: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-FP8" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
230
+ declare const createSubmodel: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324" | "zai-org/GLM-4.5-FP8">;
180
231
  /**
181
- * Create a Z.AI Provider
182
- * @see {@link https://docs.z.ai/guides/overview/pricing}
232
+ * Create a Synthetic Provider
233
+ * @see {@link https://synthetic.new/pricing}
183
234
  */
184
- declare const createZai: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6">;
235
+ declare const createSynthetic: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
185
236
  /**
186
- * Create a Inference Provider
187
- * @see {@link https://inference.net/models}
237
+ * Create a Upstage Provider
238
+ * @see {@link https://developers.upstage.ai/docs/apis/chat}
188
239
  */
189
- declare const createInference: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"meta/llama-3.2-11b-vision-instruct" | "mistral/mistral-nemo-12b-instruct" | "google/gemma-3" | "osmosis/osmosis-structure-0.6b" | "qwen/qwen3-embedding-4b" | "qwen/qwen-2.5-7b-vision-instruct" | "meta/llama-3.1-8b-instruct" | "meta/llama-3.2-3b-instruct" | "meta/llama-3.2-1b-instruct">;
240
+ declare const createUpstage: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"solar-mini" | "solar-pro2">;
190
241
  /**
191
- * Create a Requesty Provider
192
- * @see {@link https://requesty.ai/solution/llm-routing/models}
242
+ * Create a Venice AI Provider
243
+ * @see {@link https://docs.venice.ai}
193
244
  */
194
- declare const createRequesty: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"openai/gpt-4.1-mini" | "openai/gpt-4.1" | "openai/o4-mini" | "openai/gpt-4o-mini" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4" | "anthropic/claude-3-7-sonnet" | "anthropic/claude-4-sonnet-20250522" | "anthropic/claude-opus-4-1-20250805">;
245
+ declare const createVenice: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "kimi-k2-thinking" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "deepseek-v3.2" | "qwen3-235b" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "openai-gpt-oss-120b" | "zai-org-glm-4.6" | "grok-41-fast" | "claude-opus-45" | "mistral-31-24b" | "venice-uncensored" | "openai-gpt-52" | "qwen3-4b" | "llama-3.3-70b" | "devstral-2-2512" | "llama-3.2-3b" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "zai-org-glm-4.6v" | "qwen3-next-80b">;
195
246
  /**
196
- * Create a Morph Provider
197
- * @see {@link https://docs.morphllm.com/api-reference/introduction}
247
+ * Create a Vultr Provider
248
+ * @see {@link https://api.vultrinference.com/}
198
249
  */
199
- declare const createMorph: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
250
+ declare const createVultr: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-instruct" | "qwen2.5-coder-32b-instruct">;
200
251
  /**
201
- * Create a LMStudio Provider
202
- * @see {@link https://lmstudio.ai/models}
252
+ * Create a Weights & Biases Provider
253
+ * @see {@link https://weave-docs.wandb.ai/guides/integrations/inference/}
203
254
  */
204
- declare const createLmstudio: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"openai/gpt-oss-20b" | "qwen/qwen3-30b-a3b-2507" | "qwen/qwen3-coder-30b">;
255
+ declare const createWandb: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2-Instruct" | "meta-llama/Llama-3.3-70B-Instruct" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
205
256
  /**
206
- * Create a Fireworks AI Provider
207
- * @see {@link https://fireworks.ai/docs/}
257
+ * Create a xAI Provider
258
+ * @see {@link https://docs.x.ai/docs/models}
208
259
  */
209
- declare const createFireworks: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
260
+ declare const createXai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"grok-code-fast-1" | "grok-4-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-4-1-fast-non-reasoning" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
210
261
  /**
211
- * Create a ModelScope Provider
212
- * @see {@link https://modelscope.cn/docs/model-service/API-Inference/intro}
262
+ * Create a Xiaomi Provider
263
+ * @see {@link https://platform.xiaomimimo.com/#/docs}
213
264
  */
214
- declare const createModelscope: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507">;
265
+ declare const createXiaomi: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"mimo-v2-flash">;
215
266
  /**
216
- * Create a Llama Provider
217
- * @see {@link https://llama.developer.meta.com/docs/models}
267
+ * Create a Z.AI Provider
268
+ * @see {@link https://docs.z.ai/guides/overview/pricing}
218
269
  */
219
- declare const createLlama: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"llama-3.3-8b-instruct" | "llama-4-maverick-17b-128e-instruct-fp8" | "llama-3.3-70b-instruct" | "llama-4-scout-17b-16e-instruct-fp8" | "groq-llama-4-maverick-17b-128e-instruct" | "cerebras-llama-4-scout-17b-16e-instruct" | "cerebras-llama-4-maverick-17b-128e-instruct">;
270
+ declare const createZai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
220
271
  /**
221
- * Create a Scaleway Provider
222
- * @see {@link https://www.scaleway.com/en/docs/generative-apis/}
272
+ * Create a Z.AI Coding Plan Provider
273
+ * @see {@link https://docs.z.ai/devpack/overview}
223
274
  */
224
- declare const createScaleway: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.3-70b-instruct" | "qwen3-235b-a22b-instruct-2507" | "pixtral-12b-2409" | "llama-3.1-8b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "whisper-large-v3" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
275
+ declare const createZaiCodingPlan: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
225
276
  /**
226
- * Create a Cerebras Provider
227
- * @see {@link https://inference-docs.cerebras.ai/models/overview}
277
+ * Create a ZenMux Provider
278
+ * @see {@link https://docs.zenmux.ai}
228
279
  */
229
- declare const createCerebras: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<"gpt-oss-120b" | "qwen-3-235b-a22b-instruct-2507" | "zai-glm-4.6" | "qwen-3-coder-480b">;
280
+ declare const createZenmux: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "moonshotai/kimi-k2-thinking" | "openai/gpt-5-codex" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.5" | "moonshotai/kimi-k2-thinking-turbo" | "moonshotai/kimi-k2-0905" | "xiaomi/mimo-v2-flash" | "x-ai/grok-4-fast-non-reasoning" | "x-ai/grok-code-fast-1" | "x-ai/grok-4-fast" | "deepseek/deepseek-chat" | "minimax/minimax-m2" | "inclusionai/ring-1t" | "inclusionai/lint-1t" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1">;
230
281
  /**
231
- * Create a Minimax Provider
232
- * @see {@link https://platform.minimax.io/docs/api-reference/text-openai-api}
282
+ * Create a Zhipu AI Provider
283
+ * @see {@link https://docs.z.ai/guides/overview/pricing}
233
284
  */
234
- declare const createMinimax: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<string>;
285
+ declare const createZhipuai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
235
286
  /**
236
- * Create a Minimaxi Provider
237
- * @see {@link https://platform.minimaxi.com/docs/api-reference/text-openai-api}
287
+ * Create a Zhipu AI Coding Plan Provider
288
+ * @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
238
289
  */
239
- declare const createMinimaxi: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<string>;
290
+ declare const createZhipuaiCodingPlan: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
240
291
  /**
241
292
  * Create a Novita AI Provider
242
293
  * @see {@link https://novita.ai/docs/guides/llm-api#api-integration}
243
294
  */
244
- declare const createNovita: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<string>;
245
- /**
246
- * Create a SiliconFlow Provider
247
- * @see {@link https://docs.siliconflow.com/en/userguide/quickstart#4-3-call-via-openai-interface}
248
- */
249
- declare const createSiliconFlow: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string> & _xsai_ext_shared_providers.ChatProvider<string>;
295
+ declare const createNovita: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<string>;
250
296
  /**
251
297
  * Create a StepFun Provider
252
298
  * @see {@link https://www.stepfun.com}
253
299
  */
254
- declare const createStepfun: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string> & _xsai_ext_shared_providers.ChatProvider<string>;
300
+ declare const createStepfun: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
255
301
  /**
256
302
  * Create a Tencent Hunyuan Provider
257
303
  * @see {@link https://cloud.tencent.com/document/product/1729}
258
304
  */
259
- declare const createTencentHunyuan: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<string>;
260
-
305
+ declare const createTencentHunyuan: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<string>;
261
306
  /**
262
- * Create a Anthropic Provider
263
- * @see {@link https://docs.claude.com/en/api/openai-sdk}
307
+ * Create a Ollama Provider
308
+ * @see {@link https://docs.ollama.com}
264
309
  */
265
- declare const createAnthropic: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<AnthropicModels>;
266
-
267
- interface CreateAzureOptions {
268
- /**
269
- * The static API key or AD access token fetcher for authorization.
270
- *
271
- * If passed in as a function, it is treated as an accessTokenFetcher.
272
- *
273
- * @see {@link https://learn.microsoft.com/en-us/azure/api-management/api-management-authenticate-authorize-azure-openai}
274
- */
275
- apiKey: (() => Promise<string> | string) | string;
276
- /**
277
- * The Azure API version to use (`api-version` param).
278
- *
279
- * Notice: Different deployment over different time may have different API versions, please
280
- * follow the exact prompt from either [Azure AI Foundry](https://ai.azure.com/) or Azure OpenAI service
281
- * to get the correct API version from the Azure OpenAI Service endpoint.
282
- *
283
- * On Azure AI Foundry portal, you can go to https://ai.azure.com/build/overview > Choose the project >
284
- * Overview > Endpoints and keys > Included capabilities > Azure OpenAI Service to get the correct endpoint.
285
- *
286
- * @see {@link https://learn.microsoft.com/en-us/azure/ai-services/openai/reference#rest-api-versioning}
287
- */
288
- apiVersion?: string;
289
- /**
290
- * Azure resource name.
291
- *
292
- * On Azure AI Foundry portal, you can go to https://ai.azure.com/build/overview > Choose the project >
293
- * Overview > Endpoints and keys > Included capabilities > Azure OpenAI Service to get the correct endpoint.
294
- *
295
- * @see {@link https://learn.microsoft.com/en-us/azure/ai-services/openai/reference#uri-parameters}
296
- */
297
- resourceName: string;
298
- }
299
- /**
300
- * Create a Azure Provider
301
- * @see {@link https://ai.azure.com/explore/models}
302
- * @see {@link https://learn.microsoft.com/en-us/azure/ai-foundry/model-inference/concepts/endpoints?tabs=rest#routing}
303
- * @remarks
304
- * For Azure AI services, you can have multiple deployments of the same model with different names.
305
- *
306
- * Please pass your deployment name as the `model` parameter. By default, Azure will use the model name
307
- * as the deployment name when deploying a model.
308
- */
309
- declare const createAzure: (options: CreateAzureOptions) => Promise<_xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string> & _xsai_ext_shared_providers.ChatProvider<AzureModels>>;
310
+ declare const createOllama: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<string>;
311
+ /**
312
+ * Create a LiteLLM Provider
313
+ * @see {@link https://docs.litellm.ai}
314
+ */
315
+ declare const createLitellm: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
310
316
 
311
317
  /**
312
318
  * Create a Featherless AI Provider
313
319
  * @see {@link https://featherless.ai/models}
314
320
  */
315
- declare const createFeatherless: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<string>;
316
-
317
- /** @deprecated use `createFeatherless` instead. */
318
- declare const createFatherless: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<string>;
319
- /** @deprecated use `createAlibabaCn` instead. */
320
- declare const createQwen: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
321
- /** @deprecated use `createXai` instead. */
322
- declare const createXAI: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"grok-4-fast-non-reasoning" | "grok-3-fast" | "grok-4" | "grok-2-vision" | "grok-code-fast-1" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-3" | "grok-4-fast" | "grok-2-latest" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-3-mini" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
321
+ declare const createFeatherless: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<string>;
323
322
 
324
323
  /**
325
324
  * Create a Together AI Provider
326
325
  * @see {@link https://docs.together.ai/docs/serverless-models}
327
326
  */
328
- declare const createTogetherAI: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<TogetheraiModels>;
329
-
330
- /**
331
- * Create a Workers AI Provider
332
- * @see {@link https://developers.cloudflare.com/workers-ai}
333
- */
334
- declare const createWorkersAI: (apiKey: string, accountId: string) => _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<CloudflareWorkersAiModels>;
327
+ declare const createTogetherAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<TogetheraiModels>;
335
328
 
336
- export { createAlibaba, createAlibabaCn, createAnthropic, createAzure, createBaseten, createCerebras, createChutes, createCortecs, createDeepSeek, createDeepinfra, createFastrouter, createFatherless, createFeatherless, createFireworks, createGithubCopilot, createGithubModels, createGoogleGenerativeAI, createGroq, createHuggingface, createIflowcn, createInception, createInference, createLlama, createLmstudio, createLucidquery, createMinimax, createMinimaxi, createMistral, createModelscope, createMoonshotai, createMoonshotaiCn, createMorph, createNebius, createNovita, createNvidia, createOpenAI, createOpencode, createPerplexity, createQwen, createRequesty, createScaleway, createSiliconFlow, createStepfun, createSubmodel, createSynthetic, createTencentHunyuan, createTogetherAI, createUpstage, createVenice, createVultr, createWandb, createWorkersAI, createXAI, createXai, createZai, createZaiCodingPlan, createZenmux, createZhipuai, createZhipuaiCodingPlan };
329
+ export { createAlibaba, createAlibabaCn, createBailing, createBaseten, createCerebras, createChutes, createCortecs, createDeepSeek, createDeepinfra, createFastrouter, createFeatherless, createFireworks, createGithubCopilot, createGithubModels, createGoogleGenerativeAI, createGroq, createHelicone, createHuggingface, createIflowcn, createInception, createInference, createIoNet, createKimiForCoding, createLitellm, createLlama, createLmstudio, createLucidquery, createMinimax, createMinimaxCn, createMistral, createModelscope, createMoonshotai, createMoonshotaiCn, createMorph, createNebius, createNovita, createNvidia, createOllama, createOllamaCloud, createOpenAI, createOpencode, createOvhcloud, createPerplexity, createPoe, createRequesty, createScaleway, createSiliconFlow, createSiliconflowCn, createStepfun, createSubmodel, createSynthetic, createTencentHunyuan, createTogetherAI, createUpstage, createVenice, createVultr, createWandb, createXai, createXiaomi, createZai, createZaiCodingPlan, createZenmux, createZhipuai, createZhipuaiCodingPlan };