@xsai-ext/providers 0.4.0-beta.11 → 0.4.0-beta.13
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/anthropic-Bel15KQV.d.ts +111 -0
- package/dist/create.d.ts +73 -122
- package/dist/create.js +4 -37
- package/dist/index.d.ts +72 -70
- package/dist/index.js +2 -2
- package/dist/player2-D788vRq_.js +47 -0
- package/dist/special/create.d.ts +66 -0
- package/dist/special/create.js +37 -0
- package/dist/special/index.d.ts +12 -0
- package/dist/special/index.js +6 -0
- package/dist/{together-ai-BNlmR9ei.js → together-ai-D5anbRie.js} +2 -2
- package/dist/{anthropic-B6NN_uBS.d.ts → types-Czo87gCh.d.ts} +1 -111
- package/dist/types-DW1hvH0W.d.ts +38 -0
- package/dist/utils/index.d.ts +52 -0
- package/dist/utils/index.js +37 -0
- package/package.json +16 -4
package/dist/index.d.ts
CHANGED
|
@@ -1,5 +1,7 @@
|
|
|
1
|
-
import
|
|
2
|
-
import { A as
|
|
1
|
+
import { C as ChatProvider, M as ModelProvider, E as EmbedProvider, I as ImageProvider, S as SpeechProvider, T as TranscriptionProvider, a as ChatProviderWithExtraOptions } from './types-DW1hvH0W.js';
|
|
2
|
+
import { A as AnthropicOptions, O as OpenRouterOptions } from './anthropic-Bel15KQV.js';
|
|
3
|
+
import { A as AnthropicModels, O as OpenrouterModels, T as TogetheraiModels } from './types-Czo87gCh.js';
|
|
4
|
+
import '@xsai/shared';
|
|
3
5
|
|
|
4
6
|
/**
|
|
5
7
|
* Alibaba Provider
|
|
@@ -8,7 +10,7 @@ import { A as AnthropicModels, a as AnthropicOptions, O as OpenrouterModels, b a
|
|
|
8
10
|
* - baseURL - `https://dashscope-intl.aliyuncs.com/compatible-mode/v1`
|
|
9
11
|
* - apiKey - `DASHSCOPE_API_KEY`
|
|
10
12
|
*/
|
|
11
|
-
declare const alibaba:
|
|
13
|
+
declare const alibaba: ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus"> & ModelProvider;
|
|
12
14
|
/**
|
|
13
15
|
* Alibaba (China) Provider
|
|
14
16
|
* @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
|
|
@@ -16,7 +18,7 @@ declare const alibaba: _xsai_ext_shared_providers.ChatProvider<"qwen3-livetransl
|
|
|
16
18
|
* - baseURL - `https://dashscope.aliyuncs.com/compatible-mode/v1`
|
|
17
19
|
* - apiKey - `DASHSCOPE_API_KEY`
|
|
18
20
|
*/
|
|
19
|
-
declare const alibabaCn:
|
|
21
|
+
declare const alibabaCn: ModelProvider & ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
|
|
20
22
|
/**
|
|
21
23
|
* Bailing Provider
|
|
22
24
|
* @see {@link https://alipaytbox.yuque.com/sxs0ba/ling/intro}
|
|
@@ -24,7 +26,7 @@ declare const alibabaCn: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sh
|
|
|
24
26
|
* - baseURL - `https://api.tbox.cn/api/llm/v1/chat/completions`
|
|
25
27
|
* - apiKey - `BAILING_API_TOKEN`
|
|
26
28
|
*/
|
|
27
|
-
declare const bailing:
|
|
29
|
+
declare const bailing: ModelProvider & ChatProvider<"Ling-1T" | "Ring-1T">;
|
|
28
30
|
/**
|
|
29
31
|
* Baseten Provider
|
|
30
32
|
* @see {@link https://docs.baseten.co/development/model-apis/overview}
|
|
@@ -32,7 +34,7 @@ declare const bailing: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
32
34
|
* - baseURL - `https://inference.baseten.co/v1`
|
|
33
35
|
* - apiKey - `BASETEN_API_KEY`
|
|
34
36
|
*/
|
|
35
|
-
declare const baseten:
|
|
37
|
+
declare const baseten: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2">;
|
|
36
38
|
/**
|
|
37
39
|
* Cerebras Provider
|
|
38
40
|
* @see {@link https://inference-docs.cerebras.ai/models/overview}
|
|
@@ -40,7 +42,7 @@ declare const baseten: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
40
42
|
* - baseURL - `https://api.cerebras.ai/v1/`
|
|
41
43
|
* - apiKey - `CEREBRAS_API_KEY`
|
|
42
44
|
*/
|
|
43
|
-
declare const cerebras:
|
|
45
|
+
declare const cerebras: ModelProvider & ChatProvider<"qwen-3-235b-a22b-instruct-2507" | "zai-glm-4.6" | "gpt-oss-120b"> & EmbedProvider<string>;
|
|
44
46
|
/**
|
|
45
47
|
* Chutes Provider
|
|
46
48
|
* @see {@link https://llm.chutes.ai/v1/models}
|
|
@@ -48,7 +50,7 @@ declare const cerebras: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
|
|
|
48
50
|
* - baseURL - `https://llm.chutes.ai/v1`
|
|
49
51
|
* - apiKey - `CHUTES_API_KEY`
|
|
50
52
|
*/
|
|
51
|
-
declare const chutes:
|
|
53
|
+
declare const chutes: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2" | "NousResearch/Hermes-4.3-36B" | "NousResearch/Hermes-4-70B" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4-405B-FP8" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "rednote-hilab/dots.ocr" | "MiniMaxAI/MiniMax-M2" | "ArliAI/QwQ-32B-ArliAI-RpR-v1" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "tngtech/TNG-R1T-Chimera-TEE" | "OpenGVLab/InternVL3-78B" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "Alibaba-NLP/Tongyi-DeepResearch-30B-A3B" | "mistralai/Devstral-2-123B-Instruct-2512" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "unsloth/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen3-32B" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-V3" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
|
|
52
54
|
/**
|
|
53
55
|
* Cortecs Provider
|
|
54
56
|
* @see {@link https://api.cortecs.ai/v1/models}
|
|
@@ -56,7 +58,7 @@ declare const chutes: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_share
|
|
|
56
58
|
* - baseURL - `https://api.cortecs.ai/v1`
|
|
57
59
|
* - apiKey - `CORTECS_API_KEY`
|
|
58
60
|
*/
|
|
59
|
-
declare const cortecs:
|
|
61
|
+
declare const cortecs: ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "gpt-4.1" | "gemini-2.5-pro" | "devstral-small-2512" | "claude-sonnet-4" | "llama-3.1-405b-instruct">;
|
|
60
62
|
/**
|
|
61
63
|
* Deep Infra Provider
|
|
62
64
|
* @see {@link https://deepinfra.com/models}
|
|
@@ -64,7 +66,7 @@ declare const cortecs: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
64
66
|
* - baseURL - `https://api.deepinfra.com/v1/openai/`
|
|
65
67
|
* - apiKey - `DEEPINFRA_API_KEY`
|
|
66
68
|
*/
|
|
67
|
-
declare const deepinfra:
|
|
69
|
+
declare const deepinfra: ModelProvider & EmbedProvider<string> & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "zai-org/GLM-4.5" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo">;
|
|
68
70
|
/**
|
|
69
71
|
* DeepSeek Provider
|
|
70
72
|
* @see {@link https://platform.deepseek.com/api-docs/pricing}
|
|
@@ -72,7 +74,7 @@ declare const deepinfra: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sh
|
|
|
72
74
|
* - baseURL - `https://api.deepseek.com`
|
|
73
75
|
* - apiKey - `DEEPSEEK_API_KEY`
|
|
74
76
|
*/
|
|
75
|
-
declare const deepseek:
|
|
77
|
+
declare const deepseek: ModelProvider & ChatProvider<"deepseek-chat" | "deepseek-reasoner">;
|
|
76
78
|
/**
|
|
77
79
|
* FastRouter Provider
|
|
78
80
|
* @see {@link https://fastrouter.ai/models}
|
|
@@ -80,7 +82,7 @@ declare const deepseek: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
|
|
|
80
82
|
* - baseURL - `https://go.fastrouter.ai/api/v1`
|
|
81
83
|
* - apiKey - `FASTROUTER_API_KEY`
|
|
82
84
|
*/
|
|
83
|
-
declare const fastrouter:
|
|
85
|
+
declare const fastrouter: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
|
|
84
86
|
/**
|
|
85
87
|
* Fireworks AI Provider
|
|
86
88
|
* @see {@link https://fireworks.ai/docs/}
|
|
@@ -88,7 +90,7 @@ declare const fastrouter: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_s
|
|
|
88
90
|
* - baseURL - `https://api.fireworks.ai/inference/v1/`
|
|
89
91
|
* - apiKey - `FIREWORKS_API_KEY`
|
|
90
92
|
*/
|
|
91
|
-
declare const fireworks:
|
|
93
|
+
declare const fireworks: ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
|
|
92
94
|
/**
|
|
93
95
|
* GitHub Copilot Provider
|
|
94
96
|
* @see {@link https://docs.github.com/en/copilot}
|
|
@@ -96,7 +98,7 @@ declare const fireworks: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sh
|
|
|
96
98
|
* - baseURL - `https://api.githubcopilot.com`
|
|
97
99
|
* - apiKey - `GITHUB_TOKEN`
|
|
98
100
|
*/
|
|
99
|
-
declare const githubCopilot:
|
|
101
|
+
declare const githubCopilot: ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "gemini-2.0-flash-001" | "claude-opus-4" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gpt-5.1-codex" | "claude-haiku-4.5" | "gemini-3-pro-preview" | "oswe-vscode-prime" | "claude-3.5-sonnet" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "claude-opus-41" | "gpt-5-mini" | "claude-3.7-sonnet" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "claude-3.7-sonnet-thought" | "claude-opus-4.5" | "gpt-5.2" | "claude-sonnet-4.5">;
|
|
100
102
|
/**
|
|
101
103
|
* GitHub Models Provider
|
|
102
104
|
* @see {@link https://docs.github.com/en/github-models}
|
|
@@ -104,7 +106,7 @@ declare const githubCopilot: _xsai_ext_shared_providers.ModelProvider & _xsai_ex
|
|
|
104
106
|
* - baseURL - `https://models.github.ai/inference`
|
|
105
107
|
* - apiKey - `GITHUB_TOKEN`
|
|
106
108
|
*/
|
|
107
|
-
declare const githubModels:
|
|
109
|
+
declare const githubModels: ModelProvider & ChatProvider<"openai/gpt-4.1" | "core42/jais-30b-chat" | "xai/grok-3" | "xai/grok-3-mini" | "cohere/cohere-command-r-08-2024" | "cohere/cohere-command-a" | "cohere/cohere-command-r-plus-08-2024" | "cohere/cohere-command-r" | "cohere/cohere-command-r-plus" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-r1" | "deepseek/deepseek-v3-0324" | "mistral-ai/mistral-medium-2505" | "mistral-ai/ministral-3b" | "mistral-ai/mistral-nemo" | "mistral-ai/mistral-large-2411" | "mistral-ai/codestral-2501" | "mistral-ai/mistral-small-2503" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-mini-4k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-4" | "microsoft/phi-4-mini-reasoning" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-mini-instruct" | "microsoft/phi-4-multimodal-instruct" | "microsoft/phi-3-mini-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-mini-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-4-reasoning" | "microsoft/mai-ds-r1" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o1-preview" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o1-mini" | "openai/o3" | "openai/gpt-4o-mini" | "meta/llama-3.2-11b-vision-instruct" | "meta/meta-llama-3.1-405b-instruct" | "meta/llama-4-maverick-17b-128e-instruct-fp8" | "meta/meta-llama-3-70b-instruct" | "meta/meta-llama-3.1-70b-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-3.2-90b-vision-instruct" | "meta/meta-llama-3-8b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/meta-llama-3.1-8b-instruct" | "ai21-labs/ai21-jamba-1.5-large" | "ai21-labs/ai21-jamba-1.5-mini">;
|
|
108
110
|
/**
|
|
109
111
|
* Google Provider
|
|
110
112
|
* @see {@link https://ai.google.dev/gemini-api/docs/pricing}
|
|
@@ -112,7 +114,7 @@ declare const githubModels: _xsai_ext_shared_providers.ModelProvider & _xsai_ext
|
|
|
112
114
|
* - baseURL - `https://generativelanguage.googleapis.com/v1beta/openai/`
|
|
113
115
|
* - apiKey - `GOOGLE_GENERATIVE_AI_API_KEY or GEMINI_API_KEY`
|
|
114
116
|
*/
|
|
115
|
-
declare const google:
|
|
117
|
+
declare const google: ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-2.5-pro" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-2.5-flash" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
|
|
116
118
|
/**
|
|
117
119
|
* Groq Provider
|
|
118
120
|
* @see {@link https://console.groq.com/docs/models}
|
|
@@ -120,7 +122,7 @@ declare const google: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_share
|
|
|
120
122
|
* - baseURL - `https://api.groq.com/openai/v1/`
|
|
121
123
|
* - apiKey - `GROQ_API_KEY`
|
|
122
124
|
*/
|
|
123
|
-
declare const groq:
|
|
125
|
+
declare const groq: ModelProvider & EmbedProvider<string> & ChatProvider<"deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b">;
|
|
124
126
|
/**
|
|
125
127
|
* Helicone Provider
|
|
126
128
|
* @see {@link https://helicone.ai/models}
|
|
@@ -128,7 +130,7 @@ declare const groq: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_
|
|
|
128
130
|
* - baseURL - `https://ai-gateway.helicone.ai/v1`
|
|
129
131
|
* - apiKey - `HELICONE_API_KEY`
|
|
130
132
|
*/
|
|
131
|
-
declare const helicone:
|
|
133
|
+
declare const helicone: ModelProvider & ChatProvider<"qwen3-next-80b-a3b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-thinking" | "gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "deepseek-reasoner" | "claude-opus-4" | "grok-code-fast-1" | "gpt-5.1-codex" | "gemini-3-pro-preview" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "claude-3.7-sonnet" | "o3" | "gpt-5" | "gemini-2.5-flash" | "gemini-2.5-flash-lite" | "llama-3.1-8b-instant" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "qwen3-coder" | "grok-4-fast-reasoning" | "claude-opus-4-1" | "grok-4" | "llama-4-maverick" | "llama-prompt-guard-2-86m" | "grok-4-1-fast-reasoning" | "claude-4.5-haiku" | "llama-3.1-8b-instruct-turbo" | "gpt-4.1-mini-2025-04-14" | "llama-guard-4" | "llama-3.1-8b-instruct" | "gpt-4.1-mini" | "deepseek-v3.1-terminus" | "llama-prompt-guard-2-22m" | "claude-3.5-sonnet-v2" | "sonar-deep-research" | "claude-sonnet-4-5-20250929" | "grok-3" | "mistral-small" | "kimi-k2-0711" | "chatgpt-4o-latest" | "kimi-k2-0905" | "sonar-reasoning" | "llama-3.3-70b-instruct" | "claude-4.5-sonnet" | "codex-mini-latest" | "gpt-5-nano" | "deepseek-tng-r1t2-chimera" | "claude-4.5-opus" | "sonar" | "glm-4.6" | "qwen3-235b-a22b-thinking" | "hermes-2-pro-llama-3-8b" | "o1" | "grok-3-mini" | "sonar-pro" | "o1-mini" | "claude-3-haiku-20240307" | "o3-pro" | "qwen2.5-coder-7b-fast" | "gemma-3-12b-it" | "mistral-nemo" | "gpt-oss-20b" | "claude-3.5-haiku" | "gpt-5-chat-latest" | "gpt-4o-mini" | "sonar-reasoning-pro" | "qwen3-vl-235b-a22b-instruct" | "qwen3-30b-a3b" | "deepseek-v3.2" | "grok-4-1-fast-non-reasoning" | "gpt-5-pro" | "mistral-large-2411" | "claude-opus-4-1-20250805" | "ernie-4.5-21b-a3b-thinking" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "llama-4-scout">;
|
|
132
134
|
/**
|
|
133
135
|
* Hugging Face Provider
|
|
134
136
|
* @see {@link https://huggingface.co/docs/inference-providers}
|
|
@@ -136,7 +138,7 @@ declare const helicone: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
|
|
|
136
138
|
* - baseURL - `https://router.huggingface.co/v1`
|
|
137
139
|
* - apiKey - `HF_TOKEN`
|
|
138
140
|
*/
|
|
139
|
-
declare const huggingface:
|
|
141
|
+
declare const huggingface: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "MiniMaxAI/MiniMax-M2" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/Deepseek-V3-0324">;
|
|
140
142
|
/**
|
|
141
143
|
* iFlow Provider
|
|
142
144
|
* @see {@link https://platform.iflow.cn/en/docs}
|
|
@@ -144,7 +146,7 @@ declare const huggingface: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_
|
|
|
144
146
|
* - baseURL - `https://apis.iflow.cn/v1`
|
|
145
147
|
* - apiKey - `IFLOW_API_KEY`
|
|
146
148
|
*/
|
|
147
|
-
declare const iflowcn:
|
|
149
|
+
declare const iflowcn: ModelProvider & ChatProvider<"qwen3-vl-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "kimi-k2-thinking" | "qwen3-coder" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "kimi-k2" | "deepseek-v3.1" | "minimax-m2" | "qwen3-235b" | "deepseek-v3.2-chat" | "qwen3-235b-a22b-thinking-2507" | "tstars2.0" | "qwen3-235b-a22b-instruct" | "qwen3-max-preview">;
|
|
148
150
|
/**
|
|
149
151
|
* Inception Provider
|
|
150
152
|
* @see {@link https://platform.inceptionlabs.ai/docs}
|
|
@@ -152,7 +154,7 @@ declare const iflowcn: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
152
154
|
* - baseURL - `https://api.inceptionlabs.ai/v1/`
|
|
153
155
|
* - apiKey - `INCEPTION_API_KEY`
|
|
154
156
|
*/
|
|
155
|
-
declare const inception:
|
|
157
|
+
declare const inception: ModelProvider & ChatProvider<"mercury-coder" | "mercury">;
|
|
156
158
|
/**
|
|
157
159
|
* Inference Provider
|
|
158
160
|
* @see {@link https://inference.net/models}
|
|
@@ -160,7 +162,7 @@ declare const inception: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sh
|
|
|
160
162
|
* - baseURL - `https://inference.net/v1`
|
|
161
163
|
* - apiKey - `INFERENCE_API_KEY`
|
|
162
164
|
*/
|
|
163
|
-
declare const inference:
|
|
165
|
+
declare const inference: ModelProvider & ChatProvider<"meta/llama-3.2-11b-vision-instruct" | "mistral/mistral-nemo-12b-instruct" | "google/gemma-3" | "osmosis/osmosis-structure-0.6b" | "qwen/qwen3-embedding-4b" | "qwen/qwen-2.5-7b-vision-instruct" | "meta/llama-3.1-8b-instruct" | "meta/llama-3.2-3b-instruct" | "meta/llama-3.2-1b-instruct">;
|
|
164
166
|
/**
|
|
165
167
|
* IO.NET Provider
|
|
166
168
|
* @see {@link https://io.net/docs/guides/intelligence/io-intelligence}
|
|
@@ -168,7 +170,7 @@ declare const inference: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sh
|
|
|
168
170
|
* - baseURL - `https://api.intelligence.io.solutions/api/v1`
|
|
169
171
|
* - apiKey - `IOINTELLIGENCE_API_KEY`
|
|
170
172
|
*/
|
|
171
|
-
declare const ioNet:
|
|
173
|
+
declare const ioNet: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506" | "mistralai/Mistral-Large-Instruct-2411" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar">;
|
|
172
174
|
/**
|
|
173
175
|
* Kimi For Coding Provider
|
|
174
176
|
* @see {@link https://www.kimi.com/coding/docs/en/third-party-agents.html}
|
|
@@ -176,7 +178,7 @@ declare const ioNet: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared
|
|
|
176
178
|
* - baseURL - `https://api.kimi.com/coding/v1`
|
|
177
179
|
* - apiKey - `KIMI_API_KEY`
|
|
178
180
|
*/
|
|
179
|
-
declare const kimiForCoding:
|
|
181
|
+
declare const kimiForCoding: ModelProvider & ChatProvider<"kimi-k2-thinking">;
|
|
180
182
|
/**
|
|
181
183
|
* Llama Provider
|
|
182
184
|
* @see {@link https://llama.developer.meta.com/docs/models}
|
|
@@ -184,7 +186,7 @@ declare const kimiForCoding: _xsai_ext_shared_providers.ModelProvider & _xsai_ex
|
|
|
184
186
|
* - baseURL - `https://api.llama.com/compat/v1/`
|
|
185
187
|
* - apiKey - `LLAMA_API_KEY`
|
|
186
188
|
*/
|
|
187
|
-
declare const llama:
|
|
189
|
+
declare const llama: ModelProvider & ChatProvider<"llama-3.3-70b-instruct" | "llama-3.3-8b-instruct" | "llama-4-maverick-17b-128e-instruct-fp8" | "llama-4-scout-17b-16e-instruct-fp8" | "groq-llama-4-maverick-17b-128e-instruct" | "cerebras-llama-4-scout-17b-16e-instruct" | "cerebras-llama-4-maverick-17b-128e-instruct">;
|
|
188
190
|
/**
|
|
189
191
|
* LMStudio Provider
|
|
190
192
|
* @see {@link https://lmstudio.ai/models}
|
|
@@ -192,7 +194,7 @@ declare const llama: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared
|
|
|
192
194
|
* - baseURL - `http://127.0.0.1:1234/v1`
|
|
193
195
|
* - apiKey - `LMSTUDIO_API_KEY`
|
|
194
196
|
*/
|
|
195
|
-
declare const lmstudio:
|
|
197
|
+
declare const lmstudio: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "qwen/qwen3-30b-a3b-2507" | "qwen/qwen3-coder-30b">;
|
|
196
198
|
/**
|
|
197
199
|
* LucidQuery AI Provider
|
|
198
200
|
* @see {@link https://lucidquery.com/api/docs}
|
|
@@ -200,7 +202,7 @@ declare const lmstudio: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
|
|
|
200
202
|
* - baseURL - `https://lucidquery.com/api/v1`
|
|
201
203
|
* - apiKey - `LUCIDQUERY_API_KEY`
|
|
202
204
|
*/
|
|
203
|
-
declare const lucidquery:
|
|
205
|
+
declare const lucidquery: ModelProvider & ChatProvider<"lucidquery-nexus-coder" | "lucidnova-rf1-100b">;
|
|
204
206
|
/**
|
|
205
207
|
* MiniMax Provider
|
|
206
208
|
* @see {@link https://platform.minimax.io/docs/guides/quickstart}
|
|
@@ -208,7 +210,7 @@ declare const lucidquery: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_s
|
|
|
208
210
|
* - baseURL - `https://api.minimax.io/v1/`
|
|
209
211
|
* - apiKey - `MINIMAX_API_KEY`
|
|
210
212
|
*/
|
|
211
|
-
declare const minimax:
|
|
213
|
+
declare const minimax: ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
212
214
|
/**
|
|
213
215
|
* MiniMax (China) Provider
|
|
214
216
|
* @see {@link https://platform.minimaxi.com/docs/guides/quickstart}
|
|
@@ -216,7 +218,7 @@ declare const minimax: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
216
218
|
* - baseURL - `https://api.minimaxi.com/v1/`
|
|
217
219
|
* - apiKey - `MINIMAX_API_KEY`
|
|
218
220
|
*/
|
|
219
|
-
declare const minimaxCn:
|
|
221
|
+
declare const minimaxCn: ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
220
222
|
/**
|
|
221
223
|
* Mistral Provider
|
|
222
224
|
* @see {@link https://docs.mistral.ai/getting-started/models/}
|
|
@@ -224,7 +226,7 @@ declare const minimaxCn: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sh
|
|
|
224
226
|
* - baseURL - `https://api.mistral.ai/v1/`
|
|
225
227
|
* - apiKey - `MISTRAL_API_KEY`
|
|
226
228
|
*/
|
|
227
|
-
declare const mistral:
|
|
229
|
+
declare const mistral: ModelProvider & EmbedProvider<string> & ChatProvider<"devstral-2512" | "mistral-nemo" | "mistral-large-2411" | "devstral-medium-2507" | "mistral-large-2512" | "open-mixtral-8x22b" | "ministral-8b-latest" | "pixtral-large-latest" | "mistral-small-2506" | "ministral-3b-latest" | "pixtral-12b" | "mistral-medium-2505" | "labs-devstral-small-2512" | "devstral-medium-latest" | "devstral-small-2505" | "mistral-medium-2508" | "mistral-embed" | "mistral-small-latest" | "magistral-small" | "devstral-small-2507" | "codestral-latest" | "open-mixtral-8x7b" | "open-mistral-7b" | "mistral-large-latest" | "mistral-medium-latest" | "magistral-medium-latest">;
|
|
228
230
|
/**
|
|
229
231
|
* ModelScope Provider
|
|
230
232
|
* @see {@link https://modelscope.cn/docs/model-service/API-Inference/intro}
|
|
@@ -232,7 +234,7 @@ declare const mistral: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
232
234
|
* - baseURL - `https://api-inference.modelscope.cn/v1`
|
|
233
235
|
* - apiKey - `MODELSCOPE_API_KEY`
|
|
234
236
|
*/
|
|
235
|
-
declare const modelscope:
|
|
237
|
+
declare const modelscope: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507">;
|
|
236
238
|
/**
|
|
237
239
|
* Moonshot AI Provider
|
|
238
240
|
* @see {@link https://platform.moonshot.ai/docs/api/chat}
|
|
@@ -240,7 +242,7 @@ declare const modelscope: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_s
|
|
|
240
242
|
* - baseURL - `https://api.moonshot.ai/v1`
|
|
241
243
|
* - apiKey - `MOONSHOT_API_KEY`
|
|
242
244
|
*/
|
|
243
|
-
declare const moonshotai:
|
|
245
|
+
declare const moonshotai: ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
|
|
244
246
|
/**
|
|
245
247
|
* Moonshot AI (China) Provider
|
|
246
248
|
* @see {@link https://platform.moonshot.cn/docs/api/chat}
|
|
@@ -248,7 +250,7 @@ declare const moonshotai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_s
|
|
|
248
250
|
* - baseURL - `https://api.moonshot.cn/v1`
|
|
249
251
|
* - apiKey - `MOONSHOT_API_KEY`
|
|
250
252
|
*/
|
|
251
|
-
declare const moonshotaiCn:
|
|
253
|
+
declare const moonshotaiCn: ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
|
|
252
254
|
/**
|
|
253
255
|
* Morph Provider
|
|
254
256
|
* @see {@link https://docs.morphllm.com/api-reference/introduction}
|
|
@@ -256,7 +258,7 @@ declare const moonshotaiCn: _xsai_ext_shared_providers.ModelProvider & _xsai_ext
|
|
|
256
258
|
* - baseURL - `https://api.morphllm.com/v1`
|
|
257
259
|
* - apiKey - `MORPH_API_KEY`
|
|
258
260
|
*/
|
|
259
|
-
declare const morph:
|
|
261
|
+
declare const morph: ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
|
|
260
262
|
/**
|
|
261
263
|
* Nebius Token Factory Provider
|
|
262
264
|
* @see {@link https://docs.tokenfactory.nebius.com/}
|
|
@@ -264,7 +266,7 @@ declare const morph: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared
|
|
|
264
266
|
* - baseURL - `https://api.tokenfactory.nebius.com/v1`
|
|
265
267
|
* - apiKey - `NEBIUS_API_KEY`
|
|
266
268
|
*/
|
|
267
|
-
declare const nebius:
|
|
269
|
+
declare const nebius: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2-instruct" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "zai-org/glm-4.5" | "zai-org/glm-4.5-air" | "deepseek-ai/deepseek-v3">;
|
|
268
270
|
/**
|
|
269
271
|
* Nvidia Provider
|
|
270
272
|
* @see {@link https://docs.api.nvidia.com/nim/}
|
|
@@ -272,7 +274,7 @@ declare const nebius: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_share
|
|
|
272
274
|
* - baseURL - `https://integrate.api.nvidia.com/v1`
|
|
273
275
|
* - apiKey - `NVIDIA_API_KEY`
|
|
274
276
|
*/
|
|
275
|
-
declare const nvidia:
|
|
277
|
+
declare const nvidia: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-4-mini-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-coder-480b-a35b-instruct" | "moonshotai/kimi-k2-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "minimaxai/minimax-m2" | "google/gemma-3-27b-it" | "openai/whisper-large-v3" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwen3-next-80b-a3b-thinking" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "black-forest-labs/flux.1-dev">;
|
|
276
278
|
/**
|
|
277
279
|
* Ollama Cloud Provider
|
|
278
280
|
* @see {@link https://docs.ollama.com/cloud}
|
|
@@ -280,7 +282,7 @@ declare const nvidia: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_share
|
|
|
280
282
|
* - baseURL - `https://ollama.com/v1`
|
|
281
283
|
* - apiKey - `OLLAMA_API_KEY`
|
|
282
284
|
*/
|
|
283
|
-
declare const ollamaCloud:
|
|
285
|
+
declare const ollamaCloud: ModelProvider & ChatProvider<"kimi-k2-thinking:cloud" | "qwen3-vl-235b-cloud" | "qwen3-coder:480b-cloud" | "gpt-oss:120b-cloud" | "deepseek-v3.1:671b-cloud" | "glm-4.6:cloud" | "cogito-2.1:671b-cloud" | "gpt-oss:20b-cloud" | "qwen3-vl-235b-instruct-cloud" | "kimi-k2:1t-cloud" | "minimax-m2:cloud" | "gemini-3-pro-preview:latest">;
|
|
284
286
|
/**
|
|
285
287
|
* OpenAI Provider
|
|
286
288
|
* @see {@link https://platform.openai.com/docs/models}
|
|
@@ -288,7 +290,7 @@ declare const ollamaCloud: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_
|
|
|
288
290
|
* - baseURL - `https://api.openai.com/v1/`
|
|
289
291
|
* - apiKey - `OPENAI_API_KEY`
|
|
290
292
|
*/
|
|
291
|
-
declare const openai:
|
|
293
|
+
declare const openai: ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "gpt-5.2" | "gpt-4.1-nano" | "gpt-4.1-mini" | "codex-mini-latest" | "gpt-5-nano" | "o1" | "o1-mini" | "o3-pro" | "gpt-5-chat-latest" | "gpt-4o-mini" | "gpt-5-pro" | "gpt-5.1-chat-latest" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "o3-deep-research" | "gpt-3.5-turbo" | "gpt-5.2-pro" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "gpt-5.2-chat-latest" | "text-embedding-ada-002" | "gpt-4o-2024-11-20" | "o4-mini-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
|
|
292
294
|
/**
|
|
293
295
|
* OpenCode Zen Provider
|
|
294
296
|
* @see {@link https://opencode.ai/docs/zen}
|
|
@@ -296,7 +298,7 @@ declare const openai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_share
|
|
|
296
298
|
* - baseURL - `https://opencode.ai/zen/v1`
|
|
297
299
|
* - apiKey - `OPENCODE_API_KEY`
|
|
298
300
|
*/
|
|
299
|
-
declare const opencode:
|
|
301
|
+
declare const opencode: ModelProvider & ChatProvider<"kimi-k2-thinking" | "claude-sonnet-4" | "gpt-5.1-codex" | "gpt-5.1" | "gpt-5-codex" | "gpt-5.1-codex-max" | "gpt-5" | "gpt-5.2" | "qwen3-coder" | "claude-opus-4-1" | "gpt-5-nano" | "glm-4.6" | "kimi-k2" | "claude-haiku-4-5" | "claude-opus-4-5" | "gemini-3-pro" | "claude-sonnet-4-5" | "alpha-gd4" | "big-pickle" | "claude-3-5-haiku" | "grok-code" | "gemini-3-flash" | "alpha-doubao-seed-code" | "minimax-m2.1">;
|
|
300
302
|
/**
|
|
301
303
|
* OVHcloud AI Endpoints Provider
|
|
302
304
|
* @see {@link https://www.ovhcloud.com/en/public-cloud/ai-endpoints/catalog//}
|
|
@@ -304,7 +306,7 @@ declare const opencode: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
|
|
|
304
306
|
* - baseURL - `https://oai.endpoints.kepler.ai.cloud.ovh.net/v1`
|
|
305
307
|
* - apiKey - `OVHCLOUD_API_KEY`
|
|
306
308
|
*/
|
|
307
|
-
declare const ovhcloud:
|
|
309
|
+
declare const ovhcloud: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "gpt-oss-20b" | "mixtral-8x7b-instruct-v0.1" | "mistral-7b-instruct-v0.3" | "qwen2.5-vl-72b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen2.5-coder-32b-instruct" | "llava-next-mistral-7b" | "meta-llama-3_1-70b-instruct" | "meta-llama-3_3-70b-instruct">;
|
|
308
310
|
/**
|
|
309
311
|
* Perplexity Provider
|
|
310
312
|
* @see {@link https://docs.perplexity.ai}
|
|
@@ -312,7 +314,7 @@ declare const ovhcloud: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
|
|
|
312
314
|
* - baseURL - `https://api.perplexity.ai/`
|
|
313
315
|
* - apiKey - `PERPLEXITY_API_KEY`
|
|
314
316
|
*/
|
|
315
|
-
declare const perplexity:
|
|
317
|
+
declare const perplexity: ModelProvider & ChatProvider<"sonar-reasoning" | "sonar" | "sonar-pro" | "sonar-reasoning-pro">;
|
|
316
318
|
/**
|
|
317
319
|
* Poe Provider
|
|
318
320
|
* @see {@link https://creator.poe.com/docs/external-applications/openai-compatible-api}
|
|
@@ -320,7 +322,7 @@ declare const perplexity: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_s
|
|
|
320
322
|
* - baseURL - `https://api.poe.com/v1`
|
|
321
323
|
* - apiKey - `POE_API_KEY`
|
|
322
324
|
*/
|
|
323
|
-
declare const poe:
|
|
325
|
+
declare const poe: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-2" | "xai/grok-4.1-fast-non-reasoning" | "ideogramai/ideogram" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram-v2" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "poetools/claude-code" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "google/gemini-deep-research" | "google/nano-banana" | "google/imagen-4" | "google/imagen-3" | "google/imagen-4-ultra" | "google/gemini-2.0-flash-lite" | "google/gemini-3-pro" | "google/veo-3.1" | "google/imagen-3-fast" | "google/lyria" | "google/gemini-2.0-flash" | "google/gemini-2.5-flash-lite" | "google/veo-3" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-2" | "google/nano-banana-pro" | "google/veo-3.1-fast" | "openai/gpt-5.2-instant" | "openai/sora-2" | "openai/o1-pro" | "openai/gpt-5.1-codex" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-4-classic" | "openai/gpt-5-chat" | "openai/o3-deep-research" | "openai/gpt-4o-search" | "openai/gpt-image-1-mini" | "openai/gpt-3.5-turbo" | "openai/gpt-5.2-pro" | "openai/o3-mini-high" | "openai/chatgpt-4o-latest" | "openai/gpt-4-turbo" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1-instant" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-4o-aug" | "openai/o3-pro" | "openai/gpt-image-1" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo-instruct" | "openai/o4-mini-deep-research" | "openai/gpt-4-classic-0314" | "openai/dall-e-3" | "openai/sora-2-pro" | "openai/gpt-5-pro" | "openai/gpt-5.2" | "openai/gpt-4o-mini-search" | "stabilityai/stablediffusionxl" | "topazlabs-co/topazlabs" | "lumalabs/ray2" | "lumalabs/dream-machine" | "anthropic/claude-opus-3" | "anthropic/claude-opus-4" | "anthropic/claude-sonnet-3.7-reasoning" | "anthropic/claude-opus-4-search" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-haiku-3.5-search" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4-reasoning" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7-search" | "anthropic/claude-opus-4-reasoning" | "anthropic/claude-sonnet-3.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-haiku-3.5" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-sonnet-4-search" | "trytako/tako" | "novita/kimi-k2-thinking" | "novita/glm-4.6">;
|
|
324
326
|
/**
|
|
325
327
|
* Requesty Provider
|
|
326
328
|
* @see {@link https://requesty.ai/solution/llm-routing/models}
|
|
@@ -328,7 +330,7 @@ declare const poe: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_p
|
|
|
328
330
|
* - baseURL - `https://router.requesty.ai/v1`
|
|
329
331
|
* - apiKey - `REQUESTY_API_KEY`
|
|
330
332
|
*/
|
|
331
|
-
declare const requesty:
|
|
333
|
+
declare const requesty: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-sonnet-4" | "openai/gpt-4.1-mini" | "openai/o4-mini" | "openai/gpt-4o-mini" | "xai/grok-4" | "anthropic/claude-opus-4" | "xai/grok-4-fast" | "google/gemini-3-flash-preview" | "google/gemini-3-pro-preview" | "anthropic/claude-opus-4-1" | "anthropic/claude-haiku-4-5" | "anthropic/claude-opus-4-5" | "anthropic/claude-sonnet-4-5" | "anthropic/claude-3-7-sonnet">;
|
|
332
334
|
/**
|
|
333
335
|
* Scaleway Provider
|
|
334
336
|
* @see {@link https://www.scaleway.com/en/docs/generative-apis/}
|
|
@@ -336,7 +338,7 @@ declare const requesty: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
|
|
|
336
338
|
* - baseURL - `https://api.scaleway.ai/v1`
|
|
337
339
|
* - apiKey - `SCALEWAY_API_KEY`
|
|
338
340
|
*/
|
|
339
|
-
declare const scaleway:
|
|
341
|
+
declare const scaleway: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen3-235b-a22b-instruct-2507" | "pixtral-12b-2409" | "whisper-large-v3" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
|
|
340
342
|
/**
|
|
341
343
|
* SiliconFlow Provider
|
|
342
344
|
* @see {@link https://cloud.siliconflow.com/models}
|
|
@@ -344,7 +346,7 @@ declare const scaleway: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
|
|
|
344
346
|
* - baseURL - `https://api.siliconflow.com/v1`
|
|
345
347
|
* - apiKey - `SILICONFLOW_API_KEY`
|
|
346
348
|
*/
|
|
347
|
-
declare const siliconflow:
|
|
349
|
+
declare const siliconflow: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
|
|
348
350
|
/**
|
|
349
351
|
* SiliconFlow (China) Provider
|
|
350
352
|
* @see {@link https://cloud.siliconflow.com/models}
|
|
@@ -352,7 +354,7 @@ declare const siliconflow: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_
|
|
|
352
354
|
* - baseURL - `https://api.siliconflow.cn/v1`
|
|
353
355
|
* - apiKey - `SILICONFLOW_API_KEY`
|
|
354
356
|
*/
|
|
355
|
-
declare const siliconflowCn:
|
|
357
|
+
declare const siliconflowCn: ModelProvider & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
|
|
356
358
|
/**
|
|
357
359
|
* submodel Provider
|
|
358
360
|
* @see {@link https://submodel.gitbook.io}
|
|
@@ -360,7 +362,7 @@ declare const siliconflowCn: _xsai_ext_shared_providers.ModelProvider & _xsai_ex
|
|
|
360
362
|
* - baseURL - `https://llm.submodel.ai/v1`
|
|
361
363
|
* - apiKey - `SUBMODEL_INSTAGEN_ACCESS_KEY`
|
|
362
364
|
*/
|
|
363
|
-
declare const submodel:
|
|
365
|
+
declare const submodel: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324" | "zai-org/GLM-4.5-FP8">;
|
|
364
366
|
/**
|
|
365
367
|
* Synthetic Provider
|
|
366
368
|
* @see {@link https://synthetic.new/pricing}
|
|
@@ -368,7 +370,7 @@ declare const submodel: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
|
|
|
368
370
|
* - baseURL - `https://api.synthetic.new/v1`
|
|
369
371
|
* - apiKey - `SYNTHETIC_API_KEY`
|
|
370
372
|
*/
|
|
371
|
-
declare const synthetic:
|
|
373
|
+
declare const synthetic: ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
|
|
372
374
|
/**
|
|
373
375
|
* Upstage Provider
|
|
374
376
|
* @see {@link https://developers.upstage.ai/docs/apis/chat}
|
|
@@ -376,7 +378,7 @@ declare const synthetic: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sh
|
|
|
376
378
|
* - baseURL - `https://api.upstage.ai`
|
|
377
379
|
* - apiKey - `UPSTAGE_API_KEY`
|
|
378
380
|
*/
|
|
379
|
-
declare const upstage:
|
|
381
|
+
declare const upstage: ModelProvider & ChatProvider<"solar-mini" | "solar-pro2">;
|
|
380
382
|
/**
|
|
381
383
|
* Venice AI Provider
|
|
382
384
|
* @see {@link https://docs.venice.ai}
|
|
@@ -384,7 +386,7 @@ declare const upstage: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
384
386
|
* - baseURL - `https://api.venice.ai/api/v1`
|
|
385
387
|
* - apiKey - `VENICE_API_KEY`
|
|
386
388
|
*/
|
|
387
|
-
declare const venice:
|
|
389
|
+
declare const venice: ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "kimi-k2-thinking" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "deepseek-v3.2" | "qwen3-235b" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "openai-gpt-oss-120b" | "zai-org-glm-4.6" | "grok-41-fast" | "claude-opus-45" | "mistral-31-24b" | "venice-uncensored" | "openai-gpt-52" | "qwen3-4b" | "llama-3.3-70b" | "devstral-2-2512" | "llama-3.2-3b" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "zai-org-glm-4.6v" | "qwen3-next-80b">;
|
|
388
390
|
/**
|
|
389
391
|
* Vultr Provider
|
|
390
392
|
* @see {@link https://api.vultrinference.com/}
|
|
@@ -392,7 +394,7 @@ declare const venice: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_share
|
|
|
392
394
|
* - baseURL - `https://api.vultrinference.com/v1`
|
|
393
395
|
* - apiKey - `VULTR_API_KEY`
|
|
394
396
|
*/
|
|
395
|
-
declare const vultr:
|
|
397
|
+
declare const vultr: ModelProvider & ChatProvider<"deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-instruct" | "qwen2.5-coder-32b-instruct">;
|
|
396
398
|
/**
|
|
397
399
|
* Weights & Biases Provider
|
|
398
400
|
* @see {@link https://weave-docs.wandb.ai/guides/integrations/inference/}
|
|
@@ -400,7 +402,7 @@ declare const vultr: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared
|
|
|
400
402
|
* - baseURL - `https://api.inference.wandb.ai/v1`
|
|
401
403
|
* - apiKey - `WANDB_API_KEY`
|
|
402
404
|
*/
|
|
403
|
-
declare const wandb:
|
|
405
|
+
declare const wandb: ModelProvider & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2-Instruct" | "meta-llama/Llama-3.3-70B-Instruct" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
|
|
404
406
|
/**
|
|
405
407
|
* xAI Provider
|
|
406
408
|
* @see {@link https://docs.x.ai/docs/models}
|
|
@@ -408,7 +410,7 @@ declare const wandb: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared
|
|
|
408
410
|
* - baseURL - `https://api.x.ai/v1/`
|
|
409
411
|
* - apiKey - `XAI_API_KEY`
|
|
410
412
|
*/
|
|
411
|
-
declare const xai:
|
|
413
|
+
declare const xai: ModelProvider & ChatProvider<"grok-code-fast-1" | "grok-4-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-4-1-fast-non-reasoning" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
|
|
412
414
|
/**
|
|
413
415
|
* Xiaomi Provider
|
|
414
416
|
* @see {@link https://platform.xiaomimimo.com/#/docs}
|
|
@@ -416,7 +418,7 @@ declare const xai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_p
|
|
|
416
418
|
* - baseURL - `https://api.xiaomimimo.com/v1`
|
|
417
419
|
* - apiKey - `XIAOMI_API_KEY`
|
|
418
420
|
*/
|
|
419
|
-
declare const xiaomi:
|
|
421
|
+
declare const xiaomi: ModelProvider & ChatProvider<"mimo-v2-flash">;
|
|
420
422
|
/**
|
|
421
423
|
* Z.AI Provider
|
|
422
424
|
* @see {@link https://docs.z.ai/guides/overview/pricing}
|
|
@@ -424,7 +426,7 @@ declare const xiaomi: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_share
|
|
|
424
426
|
* - baseURL - `https://api.z.ai/api/paas/v4`
|
|
425
427
|
* - apiKey - `ZHIPU_API_KEY`
|
|
426
428
|
*/
|
|
427
|
-
declare const zai:
|
|
429
|
+
declare const zai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
428
430
|
/**
|
|
429
431
|
* Z.AI Coding Plan Provider
|
|
430
432
|
* @see {@link https://docs.z.ai/devpack/overview}
|
|
@@ -432,7 +434,7 @@ declare const zai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_p
|
|
|
432
434
|
* - baseURL - `https://api.z.ai/api/coding/paas/v4`
|
|
433
435
|
* - apiKey - `ZHIPU_API_KEY`
|
|
434
436
|
*/
|
|
435
|
-
declare const zaiCodingPlan:
|
|
437
|
+
declare const zaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
436
438
|
/**
|
|
437
439
|
* ZenMux Provider
|
|
438
440
|
* @see {@link https://docs.zenmux.ai}
|
|
@@ -440,7 +442,7 @@ declare const zaiCodingPlan: _xsai_ext_shared_providers.ModelProvider & _xsai_ex
|
|
|
440
442
|
* - baseURL - `https://zenmux.ai/api/v1`
|
|
441
443
|
* - apiKey - `ZENMUX_API_KEY`
|
|
442
444
|
*/
|
|
443
|
-
declare const zenmux:
|
|
445
|
+
declare const zenmux: ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "moonshotai/kimi-k2-thinking" | "openai/gpt-5-codex" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.5" | "moonshotai/kimi-k2-thinking-turbo" | "moonshotai/kimi-k2-0905" | "xiaomi/mimo-v2-flash" | "x-ai/grok-4-fast-non-reasoning" | "x-ai/grok-code-fast-1" | "x-ai/grok-4-fast" | "deepseek/deepseek-chat" | "minimax/minimax-m2" | "inclusionai/ring-1t" | "inclusionai/lint-1t" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1">;
|
|
444
446
|
/**
|
|
445
447
|
* Zhipu AI Provider
|
|
446
448
|
* @see {@link https://docs.z.ai/guides/overview/pricing}
|
|
@@ -448,7 +450,7 @@ declare const zenmux: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_share
|
|
|
448
450
|
* - baseURL - `https://open.bigmodel.cn/api/paas/v4`
|
|
449
451
|
* - apiKey - `ZHIPU_API_KEY`
|
|
450
452
|
*/
|
|
451
|
-
declare const zhipuai:
|
|
453
|
+
declare const zhipuai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
452
454
|
/**
|
|
453
455
|
* Zhipu AI Coding Plan Provider
|
|
454
456
|
* @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
|
|
@@ -456,7 +458,7 @@ declare const zhipuai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
456
458
|
* - baseURL - `https://open.bigmodel.cn/api/coding/paas/v4`
|
|
457
459
|
* - apiKey - `ZHIPU_API_KEY`
|
|
458
460
|
*/
|
|
459
|
-
declare const zhipuaiCodingPlan:
|
|
461
|
+
declare const zhipuaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
460
462
|
/**
|
|
461
463
|
* Novita AI Provider
|
|
462
464
|
* @see {@link https://novita.ai/docs/guides/llm-api#api-integration}
|
|
@@ -464,7 +466,7 @@ declare const zhipuaiCodingPlan: _xsai_ext_shared_providers.ModelProvider & _xsa
|
|
|
464
466
|
* - baseURL - `https://api.novita.ai/v3/openai/`
|
|
465
467
|
* - apiKey - `NOVITA_API_KEY`
|
|
466
468
|
*/
|
|
467
|
-
declare const novita:
|
|
469
|
+
declare const novita: ModelProvider & ChatProvider<string>;
|
|
468
470
|
/**
|
|
469
471
|
* StepFun Provider
|
|
470
472
|
* @see {@link https://www.stepfun.com}
|
|
@@ -472,7 +474,7 @@ declare const novita: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_share
|
|
|
472
474
|
* - baseURL - `https://api.stepfun.com/v1/`
|
|
473
475
|
* - apiKey - `STEPFUN_API_KEY`
|
|
474
476
|
*/
|
|
475
|
-
declare const stepfun:
|
|
477
|
+
declare const stepfun: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
|
|
476
478
|
/**
|
|
477
479
|
* Tencent Hunyuan Provider
|
|
478
480
|
* @see {@link https://cloud.tencent.com/document/product/1729}
|
|
@@ -480,7 +482,7 @@ declare const stepfun: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
480
482
|
* - baseURL - `https://api.hunyuan.cloud.tencent.com/v1/`
|
|
481
483
|
* - apiKey - `TENCENT_HUNYUAN_API_KEY`
|
|
482
484
|
*/
|
|
483
|
-
declare const tencentHunyuan:
|
|
485
|
+
declare const tencentHunyuan: ModelProvider & EmbedProvider<string> & ChatProvider<string>;
|
|
484
486
|
/**
|
|
485
487
|
* Ollama Provider
|
|
486
488
|
* @see {@link https://docs.ollama.com}
|
|
@@ -488,7 +490,7 @@ declare const tencentHunyuan: _xsai_ext_shared_providers.ModelProvider & _xsai_e
|
|
|
488
490
|
* - baseURL - `http://localhost:11434/v1/`
|
|
489
491
|
* - apiKey - `OLLAMA_API_KEY`
|
|
490
492
|
*/
|
|
491
|
-
declare const ollama:
|
|
493
|
+
declare const ollama: ModelProvider & EmbedProvider<string> & ChatProvider<string>;
|
|
492
494
|
/**
|
|
493
495
|
* LiteLLM Provider
|
|
494
496
|
* @see {@link https://docs.litellm.ai}
|
|
@@ -496,7 +498,7 @@ declare const ollama: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_share
|
|
|
496
498
|
* - baseURL - `http://localhost:4000/v1/`
|
|
497
499
|
* - apiKey - `LITELLM_API_KEY`
|
|
498
500
|
*/
|
|
499
|
-
declare const litellm:
|
|
501
|
+
declare const litellm: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
|
|
500
502
|
|
|
501
503
|
/**
|
|
502
504
|
* Anthropic Provider
|
|
@@ -505,7 +507,7 @@ declare const litellm: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
505
507
|
* - baseURL - `https://api.anthropic.com/v1/`
|
|
506
508
|
* - apiKey - `ANTHROPIC_API_KEY`
|
|
507
509
|
*/
|
|
508
|
-
declare const anthropic:
|
|
510
|
+
declare const anthropic: ModelProvider & ChatProviderWithExtraOptions<AnthropicModels, AnthropicOptions>;
|
|
509
511
|
/**
|
|
510
512
|
* Featherless AI Provider
|
|
511
513
|
* @see {@link https://featherless.ai/models}
|
|
@@ -513,7 +515,7 @@ declare const anthropic: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sh
|
|
|
513
515
|
* - baseURL - `https://api.featherless.ai/v1/`
|
|
514
516
|
* - apiKey - `FEATHERLESS_API_KEY`
|
|
515
517
|
*/
|
|
516
|
-
declare const featherless:
|
|
518
|
+
declare const featherless: ModelProvider & ChatProvider<string>;
|
|
517
519
|
/**
|
|
518
520
|
* OpenRouter Provider
|
|
519
521
|
* @see {@link https://openrouter.ai/models}
|
|
@@ -521,7 +523,7 @@ declare const featherless: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_
|
|
|
521
523
|
* - baseURL - `https://openrouter.ai/api/v1/`
|
|
522
524
|
* - apiKey - `OPENROUTER_API_KEY`
|
|
523
525
|
*/
|
|
524
|
-
declare const openrouter:
|
|
526
|
+
declare const openrouter: ModelProvider & ChatProviderWithExtraOptions<OpenrouterModels, OpenRouterOptions>;
|
|
525
527
|
/**
|
|
526
528
|
* Together AI Provider
|
|
527
529
|
* @see {@link https://docs.together.ai/docs/serverless-models}
|
|
@@ -529,10 +531,10 @@ declare const openrouter: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_s
|
|
|
529
531
|
* - baseURL - `https://api.together.xyz/v1/`
|
|
530
532
|
* - apiKey - `TOGETHER_API_KEY`
|
|
531
533
|
*/
|
|
532
|
-
declare const togetherai:
|
|
534
|
+
declare const togetherai: ModelProvider & EmbedProvider<string> & ChatProvider<TogetheraiModels>;
|
|
533
535
|
/** @deprecated use `siliconflow` instead. */
|
|
534
|
-
declare const siliconFlow:
|
|
536
|
+
declare const siliconFlow: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
|
|
535
537
|
/** @deprecated use `minimaxCn` instead. */
|
|
536
|
-
declare const minimaxi:
|
|
538
|
+
declare const minimaxi: ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
537
539
|
|
|
538
540
|
export { alibaba, alibabaCn, anthropic, bailing, baseten, cerebras, chutes, cortecs, deepinfra, deepseek, fastrouter, featherless, fireworks, githubCopilot, githubModels, google, groq, helicone, huggingface, iflowcn, inception, inference, ioNet, kimiForCoding, litellm, llama, lmstudio, lucidquery, minimax, minimaxCn, minimaxi, mistral, modelscope, moonshotai, moonshotaiCn, morph, nebius, novita, nvidia, ollama, ollamaCloud, openai, opencode, openrouter, ovhcloud, perplexity, poe, requesty, scaleway, siliconFlow, siliconflow, siliconflowCn, stepfun, submodel, synthetic, tencentHunyuan, togetherai, upstage, venice, vultr, wandb, xai, xiaomi, zai, zaiCodingPlan, zenmux, zhipuai, zhipuaiCodingPlan };
|