@xsai-ext/providers 0.4.0-beta.9 → 0.4.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/anthropic-Bel15KQV.d.ts +111 -0
- package/dist/create.d.ts +201 -208
- package/dist/create.js +2 -42
- package/dist/index.d.ts +327 -237
- package/dist/index.js +51 -40
- package/dist/player2-D788vRq_.js +47 -0
- package/dist/special/create.d.ts +66 -0
- package/dist/special/create.js +37 -0
- package/dist/special/index.d.ts +12 -0
- package/dist/special/index.js +6 -0
- package/dist/{together-ai-DXREpfi4.js → together-ai-CzFuvpt-.js} +106 -59
- package/dist/types-Czo87gCh.d.ts +7 -0
- package/dist/types-DW1hvH0W.d.ts +38 -0
- package/dist/utils/index.d.ts +52 -0
- package/dist/utils/index.js +37 -0
- package/package.json +16 -4
- package/dist/openrouter-DCc0QhRt.d.ts +0 -104
package/dist/index.d.ts
CHANGED
|
@@ -1,86 +1,96 @@
|
|
|
1
|
-
import
|
|
2
|
-
import { A as
|
|
1
|
+
import { C as ChatProvider, M as ModelProvider, E as EmbedProvider, I as ImageProvider, S as SpeechProvider, T as TranscriptionProvider, a as ChatProviderWithExtraOptions } from './types-DW1hvH0W.js';
|
|
2
|
+
import { A as AnthropicOptions, O as OpenRouterOptions } from './anthropic-Bel15KQV.js';
|
|
3
|
+
import { A as AnthropicModels, O as OpenrouterModels, T as TogetheraiModels } from './types-Czo87gCh.js';
|
|
4
|
+
import '@xsai/shared';
|
|
3
5
|
|
|
4
6
|
/**
|
|
5
|
-
*
|
|
6
|
-
* @see {@link https://
|
|
7
|
+
* Alibaba Provider
|
|
8
|
+
* @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
|
|
7
9
|
* @remarks
|
|
8
|
-
* - baseURL - `https://
|
|
9
|
-
* - apiKey - `
|
|
10
|
+
* - baseURL - `https://dashscope-intl.aliyuncs.com/compatible-mode/v1`
|
|
11
|
+
* - apiKey - `DASHSCOPE_API_KEY`
|
|
10
12
|
*/
|
|
11
|
-
declare const
|
|
13
|
+
declare const alibaba: ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus"> & ModelProvider;
|
|
12
14
|
/**
|
|
13
|
-
*
|
|
14
|
-
* @see {@link https://
|
|
15
|
+
* Alibaba (China) Provider
|
|
16
|
+
* @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
|
|
15
17
|
* @remarks
|
|
16
|
-
* - baseURL - `https://
|
|
17
|
-
* - apiKey - `
|
|
18
|
+
* - baseURL - `https://dashscope.aliyuncs.com/compatible-mode/v1`
|
|
19
|
+
* - apiKey - `DASHSCOPE_API_KEY`
|
|
18
20
|
*/
|
|
19
|
-
declare const
|
|
21
|
+
declare const alibabaCn: ModelProvider & ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
|
|
20
22
|
/**
|
|
21
|
-
*
|
|
22
|
-
* @see {@link https://
|
|
23
|
+
* Bailing Provider
|
|
24
|
+
* @see {@link https://alipaytbox.yuque.com/sxs0ba/ling/intro}
|
|
23
25
|
* @remarks
|
|
24
|
-
* - baseURL - `https://api.
|
|
25
|
-
* - apiKey - `
|
|
26
|
+
* - baseURL - `https://api.tbox.cn/api/llm/v1/chat/completions`
|
|
27
|
+
* - apiKey - `BAILING_API_TOKEN`
|
|
26
28
|
*/
|
|
27
|
-
declare const
|
|
29
|
+
declare const bailing: ModelProvider & ChatProvider<"Ling-1T" | "Ring-1T">;
|
|
28
30
|
/**
|
|
29
|
-
*
|
|
30
|
-
* @see {@link https://docs.
|
|
31
|
+
* Baseten Provider
|
|
32
|
+
* @see {@link https://docs.baseten.co/development/model-apis/overview}
|
|
31
33
|
* @remarks
|
|
32
|
-
* - baseURL - `https://
|
|
33
|
-
* - apiKey - `
|
|
34
|
+
* - baseURL - `https://inference.baseten.co/v1`
|
|
35
|
+
* - apiKey - `BASETEN_API_KEY`
|
|
34
36
|
*/
|
|
35
|
-
declare const
|
|
37
|
+
declare const baseten: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2">;
|
|
36
38
|
/**
|
|
37
|
-
*
|
|
38
|
-
* @see {@link https://
|
|
39
|
+
* Cerebras Provider
|
|
40
|
+
* @see {@link https://inference-docs.cerebras.ai/models/overview}
|
|
39
41
|
* @remarks
|
|
40
|
-
* - baseURL - `https://
|
|
41
|
-
* - apiKey - `
|
|
42
|
+
* - baseURL - `https://api.cerebras.ai/v1/`
|
|
43
|
+
* - apiKey - `CEREBRAS_API_KEY`
|
|
42
44
|
*/
|
|
43
|
-
declare const
|
|
45
|
+
declare const cerebras: ModelProvider & ChatProvider<"qwen-3-235b-a22b-instruct-2507" | "zai-glm-4.6" | "gpt-oss-120b"> & EmbedProvider<string>;
|
|
44
46
|
/**
|
|
45
|
-
*
|
|
46
|
-
* @see {@link https://
|
|
47
|
+
* Chutes Provider
|
|
48
|
+
* @see {@link https://llm.chutes.ai/v1/models}
|
|
47
49
|
* @remarks
|
|
48
|
-
* - baseURL - `https://
|
|
49
|
-
* - apiKey - `
|
|
50
|
+
* - baseURL - `https://llm.chutes.ai/v1`
|
|
51
|
+
* - apiKey - `CHUTES_API_KEY`
|
|
50
52
|
*/
|
|
51
|
-
declare const
|
|
53
|
+
declare const chutes: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2" | "NousResearch/Hermes-4.3-36B" | "NousResearch/Hermes-4-70B" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4-405B-FP8" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "rednote-hilab/dots.ocr" | "MiniMaxAI/MiniMax-M2" | "ArliAI/QwQ-32B-ArliAI-RpR-v1" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "tngtech/TNG-R1T-Chimera-TEE" | "OpenGVLab/InternVL3-78B" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "Alibaba-NLP/Tongyi-DeepResearch-30B-A3B" | "mistralai/Devstral-2-123B-Instruct-2512" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "unsloth/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen3-32B" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-V3" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
|
|
52
54
|
/**
|
|
53
|
-
*
|
|
54
|
-
* @see {@link https://api.
|
|
55
|
+
* Cortecs Provider
|
|
56
|
+
* @see {@link https://api.cortecs.ai/v1/models}
|
|
55
57
|
* @remarks
|
|
56
|
-
* - baseURL - `https://api.
|
|
57
|
-
* - apiKey - `
|
|
58
|
+
* - baseURL - `https://api.cortecs.ai/v1`
|
|
59
|
+
* - apiKey - `CORTECS_API_KEY`
|
|
58
60
|
*/
|
|
59
|
-
declare const
|
|
61
|
+
declare const cortecs: ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "gpt-4.1" | "gemini-2.5-pro" | "devstral-small-2512" | "claude-sonnet-4" | "llama-3.1-405b-instruct">;
|
|
60
62
|
/**
|
|
61
|
-
*
|
|
62
|
-
* @see {@link https://
|
|
63
|
+
* Deep Infra Provider
|
|
64
|
+
* @see {@link https://deepinfra.com/models}
|
|
63
65
|
* @remarks
|
|
64
|
-
* - baseURL - `https://
|
|
65
|
-
* - apiKey - `
|
|
66
|
+
* - baseURL - `https://api.deepinfra.com/v1/openai/`
|
|
67
|
+
* - apiKey - `DEEPINFRA_API_KEY`
|
|
66
68
|
*/
|
|
67
|
-
declare const
|
|
69
|
+
declare const deepinfra: ModelProvider & EmbedProvider<string> & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "zai-org/GLM-4.5" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo">;
|
|
68
70
|
/**
|
|
69
|
-
*
|
|
70
|
-
* @see {@link https://
|
|
71
|
+
* DeepSeek Provider
|
|
72
|
+
* @see {@link https://platform.deepseek.com/api-docs/pricing}
|
|
71
73
|
* @remarks
|
|
72
|
-
* - baseURL - `https://api.
|
|
73
|
-
* - apiKey - `
|
|
74
|
+
* - baseURL - `https://api.deepseek.com`
|
|
75
|
+
* - apiKey - `DEEPSEEK_API_KEY`
|
|
74
76
|
*/
|
|
75
|
-
declare const
|
|
77
|
+
declare const deepseek: ModelProvider & ChatProvider<"deepseek-chat" | "deepseek-reasoner">;
|
|
76
78
|
/**
|
|
77
|
-
*
|
|
78
|
-
* @see {@link https://
|
|
79
|
+
* FastRouter Provider
|
|
80
|
+
* @see {@link https://fastrouter.ai/models}
|
|
79
81
|
* @remarks
|
|
80
|
-
* - baseURL - `https://
|
|
81
|
-
* - apiKey - `
|
|
82
|
+
* - baseURL - `https://go.fastrouter.ai/api/v1`
|
|
83
|
+
* - apiKey - `FASTROUTER_API_KEY`
|
|
82
84
|
*/
|
|
83
|
-
declare const
|
|
85
|
+
declare const fastrouter: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
|
|
86
|
+
/**
|
|
87
|
+
* Fireworks AI Provider
|
|
88
|
+
* @see {@link https://fireworks.ai/docs/}
|
|
89
|
+
* @remarks
|
|
90
|
+
* - baseURL - `https://api.fireworks.ai/inference/v1/`
|
|
91
|
+
* - apiKey - `FIREWORKS_API_KEY`
|
|
92
|
+
*/
|
|
93
|
+
declare const fireworks: ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
|
|
84
94
|
/**
|
|
85
95
|
* GitHub Copilot Provider
|
|
86
96
|
* @see {@link https://docs.github.com/en/copilot}
|
|
@@ -88,127 +98,191 @@ declare const groq: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_
|
|
|
88
98
|
* - baseURL - `https://api.githubcopilot.com`
|
|
89
99
|
* - apiKey - `GITHUB_TOKEN`
|
|
90
100
|
*/
|
|
91
|
-
declare const githubCopilot:
|
|
101
|
+
declare const githubCopilot: ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "gemini-2.0-flash-001" | "claude-opus-4" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gpt-5.1-codex" | "claude-haiku-4.5" | "gemini-3-pro-preview" | "oswe-vscode-prime" | "claude-3.5-sonnet" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "claude-opus-41" | "gpt-5-mini" | "claude-3.7-sonnet" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "claude-3.7-sonnet-thought" | "claude-opus-4.5" | "gpt-5.2" | "claude-sonnet-4.5">;
|
|
92
102
|
/**
|
|
93
|
-
*
|
|
94
|
-
* @see {@link https://docs.
|
|
103
|
+
* GitHub Models Provider
|
|
104
|
+
* @see {@link https://docs.github.com/en/github-models}
|
|
95
105
|
* @remarks
|
|
96
|
-
* - baseURL - `https://
|
|
97
|
-
* - apiKey - `
|
|
106
|
+
* - baseURL - `https://models.github.ai/inference`
|
|
107
|
+
* - apiKey - `GITHUB_TOKEN`
|
|
98
108
|
*/
|
|
99
|
-
declare const
|
|
109
|
+
declare const githubModels: ModelProvider & ChatProvider<"openai/gpt-4.1" | "core42/jais-30b-chat" | "xai/grok-3" | "xai/grok-3-mini" | "cohere/cohere-command-r-08-2024" | "cohere/cohere-command-a" | "cohere/cohere-command-r-plus-08-2024" | "cohere/cohere-command-r" | "cohere/cohere-command-r-plus" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-r1" | "deepseek/deepseek-v3-0324" | "mistral-ai/mistral-medium-2505" | "mistral-ai/ministral-3b" | "mistral-ai/mistral-nemo" | "mistral-ai/mistral-large-2411" | "mistral-ai/codestral-2501" | "mistral-ai/mistral-small-2503" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-mini-4k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-4" | "microsoft/phi-4-mini-reasoning" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-mini-instruct" | "microsoft/phi-4-multimodal-instruct" | "microsoft/phi-3-mini-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-mini-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-4-reasoning" | "microsoft/mai-ds-r1" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o1-preview" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o1-mini" | "openai/o3" | "openai/gpt-4o-mini" | "meta/llama-3.2-11b-vision-instruct" | "meta/meta-llama-3.1-405b-instruct" | "meta/llama-4-maverick-17b-128e-instruct-fp8" | "meta/meta-llama-3-70b-instruct" | "meta/meta-llama-3.1-70b-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-3.2-90b-vision-instruct" | "meta/meta-llama-3-8b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/meta-llama-3.1-8b-instruct" | "ai21-labs/ai21-jamba-1.5-large" | "ai21-labs/ai21-jamba-1.5-mini">;
|
|
100
110
|
/**
|
|
101
|
-
*
|
|
102
|
-
* @see {@link https://
|
|
111
|
+
* Google Provider
|
|
112
|
+
* @see {@link https://ai.google.dev/gemini-api/docs/pricing}
|
|
103
113
|
* @remarks
|
|
104
|
-
* - baseURL - `https://
|
|
105
|
-
* - apiKey - `
|
|
114
|
+
* - baseURL - `https://generativelanguage.googleapis.com/v1beta/openai/`
|
|
115
|
+
* - apiKey - `GOOGLE_GENERATIVE_AI_API_KEY or GEMINI_API_KEY`
|
|
106
116
|
*/
|
|
107
|
-
declare const
|
|
117
|
+
declare const google: ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-2.5-pro" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-2.5-flash" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
|
|
108
118
|
/**
|
|
109
|
-
*
|
|
110
|
-
* @see {@link https://
|
|
119
|
+
* Groq Provider
|
|
120
|
+
* @see {@link https://console.groq.com/docs/models}
|
|
111
121
|
* @remarks
|
|
112
|
-
* - baseURL - `https://api.
|
|
113
|
-
* - apiKey - `
|
|
122
|
+
* - baseURL - `https://api.groq.com/openai/v1/`
|
|
123
|
+
* - apiKey - `GROQ_API_KEY`
|
|
114
124
|
*/
|
|
115
|
-
declare const
|
|
125
|
+
declare const groq: ModelProvider & EmbedProvider<string> & ChatProvider<"deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b">;
|
|
116
126
|
/**
|
|
117
|
-
*
|
|
118
|
-
* @see {@link https://
|
|
127
|
+
* Helicone Provider
|
|
128
|
+
* @see {@link https://helicone.ai/models}
|
|
119
129
|
* @remarks
|
|
120
|
-
* - baseURL - `https://
|
|
121
|
-
* - apiKey - `
|
|
130
|
+
* - baseURL - `https://ai-gateway.helicone.ai/v1`
|
|
131
|
+
* - apiKey - `HELICONE_API_KEY`
|
|
122
132
|
*/
|
|
123
|
-
declare const
|
|
133
|
+
declare const helicone: ModelProvider & ChatProvider<"qwen3-next-80b-a3b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-thinking" | "gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "deepseek-reasoner" | "claude-opus-4" | "grok-code-fast-1" | "gpt-5.1-codex" | "gemini-3-pro-preview" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "claude-3.7-sonnet" | "o3" | "gpt-5" | "gemini-2.5-flash" | "gemini-2.5-flash-lite" | "llama-3.1-8b-instant" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "qwen3-coder" | "grok-4-fast-reasoning" | "claude-opus-4-1" | "grok-4" | "llama-4-maverick" | "llama-prompt-guard-2-86m" | "grok-4-1-fast-reasoning" | "claude-4.5-haiku" | "llama-3.1-8b-instruct-turbo" | "gpt-4.1-mini-2025-04-14" | "llama-guard-4" | "llama-3.1-8b-instruct" | "gpt-4.1-mini" | "deepseek-v3.1-terminus" | "llama-prompt-guard-2-22m" | "claude-3.5-sonnet-v2" | "sonar-deep-research" | "claude-sonnet-4-5-20250929" | "grok-3" | "mistral-small" | "kimi-k2-0711" | "chatgpt-4o-latest" | "kimi-k2-0905" | "sonar-reasoning" | "llama-3.3-70b-instruct" | "claude-4.5-sonnet" | "codex-mini-latest" | "gpt-5-nano" | "deepseek-tng-r1t2-chimera" | "claude-4.5-opus" | "sonar" | "glm-4.6" | "qwen3-235b-a22b-thinking" | "hermes-2-pro-llama-3-8b" | "o1" | "grok-3-mini" | "sonar-pro" | "o1-mini" | "claude-3-haiku-20240307" | "o3-pro" | "qwen2.5-coder-7b-fast" | "gemma-3-12b-it" | "mistral-nemo" | "gpt-oss-20b" | "claude-3.5-haiku" | "gpt-5-chat-latest" | "gpt-4o-mini" | "sonar-reasoning-pro" | "qwen3-vl-235b-a22b-instruct" | "qwen3-30b-a3b" | "deepseek-v3.2" | "grok-4-1-fast-non-reasoning" | "gpt-5-pro" | "mistral-large-2411" | "claude-opus-4-1-20250805" | "ernie-4.5-21b-a3b-thinking" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "llama-4-scout">;
|
|
124
134
|
/**
|
|
125
|
-
*
|
|
126
|
-
* @see {@link https://docs
|
|
135
|
+
* Hugging Face Provider
|
|
136
|
+
* @see {@link https://huggingface.co/docs/inference-providers}
|
|
127
137
|
* @remarks
|
|
128
|
-
* - baseURL - `https://
|
|
129
|
-
* - apiKey - `
|
|
138
|
+
* - baseURL - `https://router.huggingface.co/v1`
|
|
139
|
+
* - apiKey - `HF_TOKEN`
|
|
130
140
|
*/
|
|
131
|
-
declare const
|
|
141
|
+
declare const huggingface: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "MiniMaxAI/MiniMax-M2" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/Deepseek-V3-0324">;
|
|
132
142
|
/**
|
|
133
|
-
*
|
|
134
|
-
* @see {@link https://
|
|
143
|
+
* iFlow Provider
|
|
144
|
+
* @see {@link https://platform.iflow.cn/en/docs}
|
|
135
145
|
* @remarks
|
|
136
|
-
* - baseURL - `https://
|
|
137
|
-
* - apiKey - `
|
|
146
|
+
* - baseURL - `https://apis.iflow.cn/v1`
|
|
147
|
+
* - apiKey - `IFLOW_API_KEY`
|
|
138
148
|
*/
|
|
139
|
-
declare const
|
|
149
|
+
declare const iflowcn: ModelProvider & ChatProvider<"qwen3-vl-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "kimi-k2-thinking" | "qwen3-coder" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "kimi-k2" | "deepseek-v3.1" | "minimax-m2" | "qwen3-235b" | "deepseek-v3.2-chat" | "qwen3-235b-a22b-thinking-2507" | "tstars2.0" | "qwen3-235b-a22b-instruct" | "qwen3-max-preview">;
|
|
140
150
|
/**
|
|
141
|
-
*
|
|
142
|
-
* @see {@link https://
|
|
151
|
+
* Inception Provider
|
|
152
|
+
* @see {@link https://platform.inceptionlabs.ai/docs}
|
|
143
153
|
* @remarks
|
|
144
|
-
* - baseURL - `https://api.
|
|
145
|
-
* - apiKey - `
|
|
154
|
+
* - baseURL - `https://api.inceptionlabs.ai/v1/`
|
|
155
|
+
* - apiKey - `INCEPTION_API_KEY`
|
|
146
156
|
*/
|
|
147
|
-
declare const
|
|
157
|
+
declare const inception: ModelProvider & ChatProvider<"mercury-coder" | "mercury">;
|
|
148
158
|
/**
|
|
149
|
-
*
|
|
150
|
-
* @see {@link https://
|
|
159
|
+
* Inference Provider
|
|
160
|
+
* @see {@link https://inference.net/models}
|
|
151
161
|
* @remarks
|
|
152
|
-
* - baseURL - `https://
|
|
153
|
-
* - apiKey - `
|
|
162
|
+
* - baseURL - `https://inference.net/v1`
|
|
163
|
+
* - apiKey - `INFERENCE_API_KEY`
|
|
154
164
|
*/
|
|
155
|
-
declare const
|
|
165
|
+
declare const inference: ModelProvider & ChatProvider<"meta/llama-3.2-11b-vision-instruct" | "mistral/mistral-nemo-12b-instruct" | "google/gemma-3" | "osmosis/osmosis-structure-0.6b" | "qwen/qwen3-embedding-4b" | "qwen/qwen-2.5-7b-vision-instruct" | "meta/llama-3.1-8b-instruct" | "meta/llama-3.2-3b-instruct" | "meta/llama-3.2-1b-instruct">;
|
|
156
166
|
/**
|
|
157
|
-
*
|
|
158
|
-
* @see {@link https://
|
|
167
|
+
* IO.NET Provider
|
|
168
|
+
* @see {@link https://io.net/docs/guides/intelligence/io-intelligence}
|
|
159
169
|
* @remarks
|
|
160
|
-
* - baseURL - `https://
|
|
161
|
-
* - apiKey - `
|
|
170
|
+
* - baseURL - `https://api.intelligence.io.solutions/api/v1`
|
|
171
|
+
* - apiKey - `IOINTELLIGENCE_API_KEY`
|
|
162
172
|
*/
|
|
163
|
-
declare const
|
|
173
|
+
declare const ioNet: ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506" | "mistralai/Mistral-Large-Instruct-2411" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar">;
|
|
164
174
|
/**
|
|
165
|
-
*
|
|
166
|
-
* @see {@link https://
|
|
175
|
+
* Kimi For Coding Provider
|
|
176
|
+
* @see {@link https://www.kimi.com/coding/docs/en/third-party-agents.html}
|
|
167
177
|
* @remarks
|
|
168
|
-
* - baseURL - `https://
|
|
169
|
-
* - apiKey - `
|
|
178
|
+
* - baseURL - `https://api.kimi.com/coding/v1`
|
|
179
|
+
* - apiKey - `KIMI_API_KEY`
|
|
170
180
|
*/
|
|
171
|
-
declare const
|
|
181
|
+
declare const kimiForCoding: ModelProvider & ChatProvider<"kimi-k2-thinking">;
|
|
172
182
|
/**
|
|
173
|
-
*
|
|
174
|
-
* @see {@link https://
|
|
183
|
+
* Llama Provider
|
|
184
|
+
* @see {@link https://llama.developer.meta.com/docs/models}
|
|
175
185
|
* @remarks
|
|
176
|
-
* - baseURL - `https://
|
|
177
|
-
* - apiKey - `
|
|
186
|
+
* - baseURL - `https://api.llama.com/compat/v1/`
|
|
187
|
+
* - apiKey - `LLAMA_API_KEY`
|
|
178
188
|
*/
|
|
179
|
-
declare const
|
|
189
|
+
declare const llama: ModelProvider & ChatProvider<"llama-3.3-70b-instruct" | "llama-3.3-8b-instruct" | "llama-4-maverick-17b-128e-instruct-fp8" | "llama-4-scout-17b-16e-instruct-fp8" | "groq-llama-4-maverick-17b-128e-instruct" | "cerebras-llama-4-scout-17b-16e-instruct" | "cerebras-llama-4-maverick-17b-128e-instruct">;
|
|
180
190
|
/**
|
|
181
|
-
*
|
|
182
|
-
* @see {@link https://
|
|
191
|
+
* LMStudio Provider
|
|
192
|
+
* @see {@link https://lmstudio.ai/models}
|
|
183
193
|
* @remarks
|
|
184
|
-
* - baseURL - `
|
|
185
|
-
* - apiKey - `
|
|
194
|
+
* - baseURL - `http://127.0.0.1:1234/v1`
|
|
195
|
+
* - apiKey - `LMSTUDIO_API_KEY`
|
|
186
196
|
*/
|
|
187
|
-
declare const
|
|
197
|
+
declare const lmstudio: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "qwen/qwen3-30b-a3b-2507" | "qwen/qwen3-coder-30b">;
|
|
188
198
|
/**
|
|
189
|
-
*
|
|
190
|
-
* @see {@link https://
|
|
199
|
+
* LucidQuery AI Provider
|
|
200
|
+
* @see {@link https://lucidquery.com/api/docs}
|
|
191
201
|
* @remarks
|
|
192
|
-
* - baseURL - `https://
|
|
193
|
-
* - apiKey - `
|
|
202
|
+
* - baseURL - `https://lucidquery.com/api/v1`
|
|
203
|
+
* - apiKey - `LUCIDQUERY_API_KEY`
|
|
194
204
|
*/
|
|
195
|
-
declare const
|
|
205
|
+
declare const lucidquery: ModelProvider & ChatProvider<"lucidquery-nexus-coder" | "lucidnova-rf1-100b">;
|
|
196
206
|
/**
|
|
197
|
-
*
|
|
198
|
-
* @see {@link https://platform.
|
|
207
|
+
* MiniMax Provider
|
|
208
|
+
* @see {@link https://platform.minimax.io/docs/guides/quickstart}
|
|
199
209
|
* @remarks
|
|
200
|
-
* - baseURL - `https://api.
|
|
201
|
-
* - apiKey - `
|
|
210
|
+
* - baseURL - `https://api.minimax.io/v1/`
|
|
211
|
+
* - apiKey - `MINIMAX_API_KEY`
|
|
202
212
|
*/
|
|
203
|
-
declare const
|
|
213
|
+
declare const minimax: ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
204
214
|
/**
|
|
205
|
-
*
|
|
206
|
-
* @see {@link https://
|
|
215
|
+
* MiniMax (China) Provider
|
|
216
|
+
* @see {@link https://platform.minimaxi.com/docs/guides/quickstart}
|
|
207
217
|
* @remarks
|
|
208
|
-
* - baseURL - `https://api.
|
|
209
|
-
* - apiKey - `
|
|
218
|
+
* - baseURL - `https://api.minimaxi.com/v1/`
|
|
219
|
+
* - apiKey - `MINIMAX_API_KEY`
|
|
210
220
|
*/
|
|
211
|
-
declare const
|
|
221
|
+
declare const minimaxCn: ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
222
|
+
/**
|
|
223
|
+
* Mistral Provider
|
|
224
|
+
* @see {@link https://docs.mistral.ai/getting-started/models/}
|
|
225
|
+
* @remarks
|
|
226
|
+
* - baseURL - `https://api.mistral.ai/v1/`
|
|
227
|
+
* - apiKey - `MISTRAL_API_KEY`
|
|
228
|
+
*/
|
|
229
|
+
declare const mistral: ModelProvider & EmbedProvider<string> & ChatProvider<"devstral-2512" | "mistral-nemo" | "mistral-large-2411" | "devstral-medium-2507" | "mistral-large-2512" | "open-mixtral-8x22b" | "ministral-8b-latest" | "pixtral-large-latest" | "mistral-small-2506" | "ministral-3b-latest" | "pixtral-12b" | "mistral-medium-2505" | "labs-devstral-small-2512" | "devstral-medium-latest" | "devstral-small-2505" | "mistral-medium-2508" | "mistral-embed" | "mistral-small-latest" | "magistral-small" | "devstral-small-2507" | "codestral-latest" | "open-mixtral-8x7b" | "open-mistral-7b" | "mistral-large-latest" | "mistral-medium-latest" | "magistral-medium-latest">;
|
|
230
|
+
/**
|
|
231
|
+
* ModelScope Provider
|
|
232
|
+
* @see {@link https://modelscope.cn/docs/model-service/API-Inference/intro}
|
|
233
|
+
* @remarks
|
|
234
|
+
* - baseURL - `https://api-inference.modelscope.cn/v1`
|
|
235
|
+
* - apiKey - `MODELSCOPE_API_KEY`
|
|
236
|
+
*/
|
|
237
|
+
declare const modelscope: ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507">;
|
|
238
|
+
/**
|
|
239
|
+
* Moonshot AI Provider
|
|
240
|
+
* @see {@link https://platform.moonshot.ai/docs/api/chat}
|
|
241
|
+
* @remarks
|
|
242
|
+
* - baseURL - `https://api.moonshot.ai/v1`
|
|
243
|
+
* - apiKey - `MOONSHOT_API_KEY`
|
|
244
|
+
*/
|
|
245
|
+
declare const moonshotai: ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
|
|
246
|
+
/**
|
|
247
|
+
* Moonshot AI (China) Provider
|
|
248
|
+
* @see {@link https://platform.moonshot.cn/docs/api/chat}
|
|
249
|
+
* @remarks
|
|
250
|
+
* - baseURL - `https://api.moonshot.cn/v1`
|
|
251
|
+
* - apiKey - `MOONSHOT_API_KEY`
|
|
252
|
+
*/
|
|
253
|
+
declare const moonshotaiCn: ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
|
|
254
|
+
/**
|
|
255
|
+
* Morph Provider
|
|
256
|
+
* @see {@link https://docs.morphllm.com/api-reference/introduction}
|
|
257
|
+
* @remarks
|
|
258
|
+
* - baseURL - `https://api.morphllm.com/v1`
|
|
259
|
+
* - apiKey - `MORPH_API_KEY`
|
|
260
|
+
*/
|
|
261
|
+
declare const morph: ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
|
|
262
|
+
/**
|
|
263
|
+
* Nebius Token Factory Provider
|
|
264
|
+
* @see {@link https://docs.tokenfactory.nebius.com/}
|
|
265
|
+
* @remarks
|
|
266
|
+
* - baseURL - `https://api.tokenfactory.nebius.com/v1`
|
|
267
|
+
* - apiKey - `NEBIUS_API_KEY`
|
|
268
|
+
*/
|
|
269
|
+
declare const nebius: ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2-instruct" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "zai-org/glm-4.5" | "zai-org/glm-4.5-air" | "deepseek-ai/deepseek-v3">;
|
|
270
|
+
/**
|
|
271
|
+
* Nvidia Provider
|
|
272
|
+
* @see {@link https://docs.api.nvidia.com/nim/}
|
|
273
|
+
* @remarks
|
|
274
|
+
* - baseURL - `https://integrate.api.nvidia.com/v1`
|
|
275
|
+
* - apiKey - `NVIDIA_API_KEY`
|
|
276
|
+
*/
|
|
277
|
+
declare const nvidia: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-4-mini-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-coder-480b-a35b-instruct" | "moonshotai/kimi-k2-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "minimaxai/minimax-m2" | "google/gemma-3-27b-it" | "openai/whisper-large-v3" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwen3-next-80b-a3b-thinking" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "black-forest-labs/flux.1-dev">;
|
|
278
|
+
/**
|
|
279
|
+
* Ollama Cloud Provider
|
|
280
|
+
* @see {@link https://docs.ollama.com/cloud}
|
|
281
|
+
* @remarks
|
|
282
|
+
* - baseURL - `https://ollama.com/v1`
|
|
283
|
+
* - apiKey - `OLLAMA_API_KEY`
|
|
284
|
+
*/
|
|
285
|
+
declare const ollamaCloud: ModelProvider & ChatProvider<"kimi-k2-thinking:cloud" | "qwen3-vl-235b-cloud" | "qwen3-coder:480b-cloud" | "gpt-oss:120b-cloud" | "deepseek-v3.1:671b-cloud" | "glm-4.6:cloud" | "cogito-2.1:671b-cloud" | "gpt-oss:20b-cloud" | "qwen3-vl-235b-instruct-cloud" | "kimi-k2:1t-cloud" | "minimax-m2:cloud" | "gemini-3-pro-preview:latest">;
|
|
212
286
|
/**
|
|
213
287
|
* OpenAI Provider
|
|
214
288
|
* @see {@link https://platform.openai.com/docs/models}
|
|
@@ -216,15 +290,23 @@ declare const wandb: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared
|
|
|
216
290
|
* - baseURL - `https://api.openai.com/v1/`
|
|
217
291
|
* - apiKey - `OPENAI_API_KEY`
|
|
218
292
|
*/
|
|
219
|
-
declare const openai:
|
|
293
|
+
declare const openai: ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "gpt-5.2" | "gpt-4.1-nano" | "gpt-4.1-mini" | "codex-mini-latest" | "gpt-5-nano" | "o1" | "o1-mini" | "o3-pro" | "gpt-5-chat-latest" | "gpt-4o-mini" | "gpt-5-pro" | "gpt-5.1-chat-latest" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "o3-deep-research" | "gpt-3.5-turbo" | "gpt-5.2-pro" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "gpt-5.2-chat-latest" | "text-embedding-ada-002" | "gpt-4o-2024-11-20" | "o4-mini-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
|
|
220
294
|
/**
|
|
221
|
-
*
|
|
222
|
-
* @see {@link https://
|
|
295
|
+
* OpenCode Zen Provider
|
|
296
|
+
* @see {@link https://opencode.ai/docs/zen}
|
|
223
297
|
* @remarks
|
|
224
|
-
* - baseURL - `https://
|
|
225
|
-
* - apiKey - `
|
|
298
|
+
* - baseURL - `https://opencode.ai/zen/v1`
|
|
299
|
+
* - apiKey - `OPENCODE_API_KEY`
|
|
300
|
+
*/
|
|
301
|
+
declare const opencode: ModelProvider & ChatProvider<"kimi-k2-thinking" | "claude-sonnet-4" | "gpt-5.1-codex" | "gpt-5.1" | "gpt-5-codex" | "gpt-5.1-codex-max" | "gpt-5" | "gpt-5.2" | "qwen3-coder" | "claude-opus-4-1" | "gpt-5-nano" | "glm-4.6" | "kimi-k2" | "claude-haiku-4-5" | "claude-opus-4-5" | "gemini-3-pro" | "claude-sonnet-4-5" | "alpha-gd4" | "big-pickle" | "claude-3-5-haiku" | "grok-code" | "gemini-3-flash" | "alpha-doubao-seed-code" | "minimax-m2.1">;
|
|
302
|
+
/**
|
|
303
|
+
* OVHcloud AI Endpoints Provider
|
|
304
|
+
* @see {@link https://www.ovhcloud.com/en/public-cloud/ai-endpoints/catalog//}
|
|
305
|
+
* @remarks
|
|
306
|
+
* - baseURL - `https://oai.endpoints.kepler.ai.cloud.ovh.net/v1`
|
|
307
|
+
* - apiKey - `OVHCLOUD_API_KEY`
|
|
226
308
|
*/
|
|
227
|
-
declare const
|
|
309
|
+
declare const ovhcloud: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "gpt-oss-20b" | "mixtral-8x7b-instruct-v0.1" | "mistral-7b-instruct-v0.3" | "qwen2.5-vl-72b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen2.5-coder-32b-instruct" | "llava-next-mistral-7b" | "meta-llama-3_1-70b-instruct" | "meta-llama-3_3-70b-instruct">;
|
|
228
310
|
/**
|
|
229
311
|
* Perplexity Provider
|
|
230
312
|
* @see {@link https://docs.perplexity.ai}
|
|
@@ -232,47 +314,47 @@ declare const zhipuaiCodingPlan: _xsai_ext_shared_providers.ModelProvider & _xsa
|
|
|
232
314
|
* - baseURL - `https://api.perplexity.ai/`
|
|
233
315
|
* - apiKey - `PERPLEXITY_API_KEY`
|
|
234
316
|
*/
|
|
235
|
-
declare const perplexity:
|
|
317
|
+
declare const perplexity: ModelProvider & ChatProvider<"sonar-reasoning" | "sonar" | "sonar-pro" | "sonar-reasoning-pro">;
|
|
236
318
|
/**
|
|
237
|
-
*
|
|
238
|
-
* @see {@link https://
|
|
319
|
+
* Poe Provider
|
|
320
|
+
* @see {@link https://creator.poe.com/docs/external-applications/openai-compatible-api}
|
|
239
321
|
* @remarks
|
|
240
|
-
* - baseURL - `https://
|
|
241
|
-
* - apiKey - `
|
|
322
|
+
* - baseURL - `https://api.poe.com/v1`
|
|
323
|
+
* - apiKey - `POE_API_KEY`
|
|
242
324
|
*/
|
|
243
|
-
declare const
|
|
325
|
+
declare const poe: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-2" | "xai/grok-4.1-fast-non-reasoning" | "ideogramai/ideogram" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram-v2" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "poetools/claude-code" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "google/gemini-deep-research" | "google/nano-banana" | "google/imagen-4" | "google/imagen-3" | "google/imagen-4-ultra" | "google/gemini-2.0-flash-lite" | "google/gemini-3-pro" | "google/veo-3.1" | "google/imagen-3-fast" | "google/lyria" | "google/gemini-2.0-flash" | "google/gemini-2.5-flash-lite" | "google/veo-3" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-2" | "google/nano-banana-pro" | "google/veo-3.1-fast" | "openai/gpt-5.2-instant" | "openai/sora-2" | "openai/o1-pro" | "openai/gpt-5.1-codex" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-4-classic" | "openai/gpt-5-chat" | "openai/o3-deep-research" | "openai/gpt-4o-search" | "openai/gpt-image-1-mini" | "openai/gpt-3.5-turbo" | "openai/gpt-5.2-pro" | "openai/o3-mini-high" | "openai/chatgpt-4o-latest" | "openai/gpt-4-turbo" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1-instant" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-4o-aug" | "openai/o3-pro" | "openai/gpt-image-1" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo-instruct" | "openai/o4-mini-deep-research" | "openai/gpt-4-classic-0314" | "openai/dall-e-3" | "openai/sora-2-pro" | "openai/gpt-5-pro" | "openai/gpt-5.2" | "openai/gpt-4o-mini-search" | "stabilityai/stablediffusionxl" | "topazlabs-co/topazlabs" | "lumalabs/ray2" | "lumalabs/dream-machine" | "anthropic/claude-opus-3" | "anthropic/claude-opus-4" | "anthropic/claude-sonnet-3.7-reasoning" | "anthropic/claude-opus-4-search" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-haiku-3.5-search" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4-reasoning" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7-search" | "anthropic/claude-opus-4-reasoning" | "anthropic/claude-sonnet-3.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-haiku-3.5" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-sonnet-4-search" | "trytako/tako" | "novita/kimi-k2-thinking" | "novita/glm-4.6">;
|
|
244
326
|
/**
|
|
245
|
-
*
|
|
246
|
-
* @see {@link https://
|
|
327
|
+
* Requesty Provider
|
|
328
|
+
* @see {@link https://requesty.ai/solution/llm-routing/models}
|
|
247
329
|
* @remarks
|
|
248
|
-
* - baseURL - `https://
|
|
249
|
-
* - apiKey - `
|
|
330
|
+
* - baseURL - `https://router.requesty.ai/v1`
|
|
331
|
+
* - apiKey - `REQUESTY_API_KEY`
|
|
250
332
|
*/
|
|
251
|
-
declare const
|
|
333
|
+
declare const requesty: ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-sonnet-4" | "openai/gpt-4.1-mini" | "openai/o4-mini" | "openai/gpt-4o-mini" | "xai/grok-4" | "anthropic/claude-opus-4" | "xai/grok-4-fast" | "google/gemini-3-flash-preview" | "google/gemini-3-pro-preview" | "anthropic/claude-opus-4-1" | "anthropic/claude-haiku-4-5" | "anthropic/claude-opus-4-5" | "anthropic/claude-sonnet-4-5" | "anthropic/claude-3-7-sonnet">;
|
|
252
334
|
/**
|
|
253
|
-
*
|
|
254
|
-
* @see {@link https://
|
|
335
|
+
* Scaleway Provider
|
|
336
|
+
* @see {@link https://www.scaleway.com/en/docs/generative-apis/}
|
|
255
337
|
* @remarks
|
|
256
|
-
* - baseURL - `https://api.
|
|
257
|
-
* - apiKey - `
|
|
338
|
+
* - baseURL - `https://api.scaleway.ai/v1`
|
|
339
|
+
* - apiKey - `SCALEWAY_API_KEY`
|
|
258
340
|
*/
|
|
259
|
-
declare const
|
|
341
|
+
declare const scaleway: ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen3-235b-a22b-instruct-2507" | "pixtral-12b-2409" | "whisper-large-v3" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
|
|
260
342
|
/**
|
|
261
|
-
*
|
|
262
|
-
* @see {@link https://
|
|
343
|
+
* SiliconFlow Provider
|
|
344
|
+
* @see {@link https://cloud.siliconflow.com/models}
|
|
263
345
|
* @remarks
|
|
264
|
-
* - baseURL - `https://api.
|
|
265
|
-
* - apiKey - `
|
|
346
|
+
* - baseURL - `https://api.siliconflow.com/v1`
|
|
347
|
+
* - apiKey - `SILICONFLOW_API_KEY`
|
|
266
348
|
*/
|
|
267
|
-
declare const
|
|
349
|
+
declare const siliconflow: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
|
|
268
350
|
/**
|
|
269
|
-
*
|
|
270
|
-
* @see {@link https://
|
|
351
|
+
* SiliconFlow (China) Provider
|
|
352
|
+
* @see {@link https://cloud.siliconflow.com/models}
|
|
271
353
|
* @remarks
|
|
272
|
-
* - baseURL - `https://
|
|
273
|
-
* - apiKey - `
|
|
354
|
+
* - baseURL - `https://api.siliconflow.cn/v1`
|
|
355
|
+
* - apiKey - `SILICONFLOW_API_KEY`
|
|
274
356
|
*/
|
|
275
|
-
declare const
|
|
357
|
+
declare const siliconflowCn: ModelProvider & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
|
|
276
358
|
/**
|
|
277
359
|
* submodel Provider
|
|
278
360
|
* @see {@link https://submodel.gitbook.io}
|
|
@@ -280,103 +362,103 @@ declare const zhipuai: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
280
362
|
* - baseURL - `https://llm.submodel.ai/v1`
|
|
281
363
|
* - apiKey - `SUBMODEL_INSTAGEN_ACCESS_KEY`
|
|
282
364
|
*/
|
|
283
|
-
declare const submodel:
|
|
365
|
+
declare const submodel: ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324" | "zai-org/GLM-4.5-FP8">;
|
|
284
366
|
/**
|
|
285
|
-
*
|
|
286
|
-
* @see {@link https://
|
|
367
|
+
* Synthetic Provider
|
|
368
|
+
* @see {@link https://synthetic.new/pricing}
|
|
287
369
|
* @remarks
|
|
288
|
-
* - baseURL - `https://api.
|
|
289
|
-
* - apiKey - `
|
|
370
|
+
* - baseURL - `https://api.synthetic.new/v1`
|
|
371
|
+
* - apiKey - `SYNTHETIC_API_KEY`
|
|
290
372
|
*/
|
|
291
|
-
declare const
|
|
373
|
+
declare const synthetic: ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
|
|
292
374
|
/**
|
|
293
|
-
*
|
|
294
|
-
* @see {@link https://
|
|
375
|
+
* Upstage Provider
|
|
376
|
+
* @see {@link https://developers.upstage.ai/docs/apis/chat}
|
|
295
377
|
* @remarks
|
|
296
|
-
* - baseURL - `https://
|
|
297
|
-
* - apiKey - `
|
|
378
|
+
* - baseURL - `https://api.upstage.ai`
|
|
379
|
+
* - apiKey - `UPSTAGE_API_KEY`
|
|
298
380
|
*/
|
|
299
|
-
declare const
|
|
381
|
+
declare const upstage: ModelProvider & ChatProvider<"solar-mini" | "solar-pro2">;
|
|
300
382
|
/**
|
|
301
|
-
*
|
|
302
|
-
* @see {@link https://
|
|
383
|
+
* Venice AI Provider
|
|
384
|
+
* @see {@link https://docs.venice.ai}
|
|
303
385
|
* @remarks
|
|
304
|
-
* - baseURL - `https://
|
|
305
|
-
* - apiKey - `
|
|
386
|
+
* - baseURL - `https://api.venice.ai/api/v1`
|
|
387
|
+
* - apiKey - `VENICE_API_KEY`
|
|
306
388
|
*/
|
|
307
|
-
declare const
|
|
389
|
+
declare const venice: ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "kimi-k2-thinking" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "deepseek-v3.2" | "qwen3-235b" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "openai-gpt-oss-120b" | "zai-org-glm-4.6" | "grok-41-fast" | "claude-opus-45" | "mistral-31-24b" | "venice-uncensored" | "openai-gpt-52" | "qwen3-4b" | "llama-3.3-70b" | "devstral-2-2512" | "llama-3.2-3b" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "zai-org-glm-4.6v" | "qwen3-next-80b">;
|
|
308
390
|
/**
|
|
309
|
-
*
|
|
310
|
-
* @see {@link https://
|
|
391
|
+
* Vultr Provider
|
|
392
|
+
* @see {@link https://api.vultrinference.com/}
|
|
311
393
|
* @remarks
|
|
312
|
-
* - baseURL - `https://api.
|
|
313
|
-
* - apiKey - `
|
|
394
|
+
* - baseURL - `https://api.vultrinference.com/v1`
|
|
395
|
+
* - apiKey - `VULTR_API_KEY`
|
|
314
396
|
*/
|
|
315
|
-
declare const
|
|
397
|
+
declare const vultr: ModelProvider & ChatProvider<"deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-instruct" | "qwen2.5-coder-32b-instruct">;
|
|
316
398
|
/**
|
|
317
|
-
*
|
|
318
|
-
* @see {@link https://
|
|
399
|
+
* Weights & Biases Provider
|
|
400
|
+
* @see {@link https://weave-docs.wandb.ai/guides/integrations/inference/}
|
|
319
401
|
* @remarks
|
|
320
|
-
* - baseURL - `
|
|
321
|
-
* - apiKey - `
|
|
402
|
+
* - baseURL - `https://api.inference.wandb.ai/v1`
|
|
403
|
+
* - apiKey - `WANDB_API_KEY`
|
|
322
404
|
*/
|
|
323
|
-
declare const
|
|
405
|
+
declare const wandb: ModelProvider & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2-Instruct" | "meta-llama/Llama-3.3-70B-Instruct" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
|
|
324
406
|
/**
|
|
325
|
-
*
|
|
326
|
-
* @see {@link https://
|
|
407
|
+
* xAI Provider
|
|
408
|
+
* @see {@link https://docs.x.ai/docs/models}
|
|
327
409
|
* @remarks
|
|
328
|
-
* - baseURL - `https://api.
|
|
329
|
-
* - apiKey - `
|
|
410
|
+
* - baseURL - `https://api.x.ai/v1/`
|
|
411
|
+
* - apiKey - `XAI_API_KEY`
|
|
330
412
|
*/
|
|
331
|
-
declare const
|
|
413
|
+
declare const xai: ModelProvider & ChatProvider<"grok-code-fast-1" | "grok-4-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-4-1-fast-non-reasoning" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
|
|
332
414
|
/**
|
|
333
|
-
*
|
|
334
|
-
* @see {@link https://
|
|
415
|
+
* Xiaomi Provider
|
|
416
|
+
* @see {@link https://platform.xiaomimimo.com/#/docs}
|
|
335
417
|
* @remarks
|
|
336
|
-
* - baseURL - `https://api
|
|
337
|
-
* - apiKey - `
|
|
418
|
+
* - baseURL - `https://api.xiaomimimo.com/v1`
|
|
419
|
+
* - apiKey - `XIAOMI_API_KEY`
|
|
338
420
|
*/
|
|
339
|
-
declare const
|
|
421
|
+
declare const xiaomi: ModelProvider & ChatProvider<"mimo-v2-flash">;
|
|
340
422
|
/**
|
|
341
|
-
*
|
|
342
|
-
* @see {@link https://
|
|
423
|
+
* Z.AI Provider
|
|
424
|
+
* @see {@link https://docs.z.ai/guides/overview/pricing}
|
|
343
425
|
* @remarks
|
|
344
|
-
* - baseURL - `https://api.
|
|
345
|
-
* - apiKey - `
|
|
426
|
+
* - baseURL - `https://api.z.ai/api/paas/v4`
|
|
427
|
+
* - apiKey - `ZHIPU_API_KEY`
|
|
346
428
|
*/
|
|
347
|
-
declare const
|
|
429
|
+
declare const zai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
348
430
|
/**
|
|
349
|
-
*
|
|
350
|
-
* @see {@link https://
|
|
431
|
+
* Z.AI Coding Plan Provider
|
|
432
|
+
* @see {@link https://docs.z.ai/devpack/overview}
|
|
351
433
|
* @remarks
|
|
352
|
-
* - baseURL - `https://api.
|
|
353
|
-
* - apiKey - `
|
|
434
|
+
* - baseURL - `https://api.z.ai/api/coding/paas/v4`
|
|
435
|
+
* - apiKey - `ZHIPU_API_KEY`
|
|
354
436
|
*/
|
|
355
|
-
declare const
|
|
437
|
+
declare const zaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
356
438
|
/**
|
|
357
|
-
*
|
|
358
|
-
* @see {@link https://
|
|
439
|
+
* ZenMux Provider
|
|
440
|
+
* @see {@link https://docs.zenmux.ai}
|
|
359
441
|
* @remarks
|
|
360
|
-
* - baseURL - `https://
|
|
361
|
-
* - apiKey - `
|
|
442
|
+
* - baseURL - `https://zenmux.ai/api/v1`
|
|
443
|
+
* - apiKey - `ZENMUX_API_KEY`
|
|
362
444
|
*/
|
|
363
|
-
declare const
|
|
445
|
+
declare const zenmux: ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "moonshotai/kimi-k2-thinking" | "openai/gpt-5-codex" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.5" | "moonshotai/kimi-k2-thinking-turbo" | "moonshotai/kimi-k2-0905" | "xiaomi/mimo-v2-flash" | "x-ai/grok-4-fast-non-reasoning" | "x-ai/grok-code-fast-1" | "x-ai/grok-4-fast" | "deepseek/deepseek-chat" | "minimax/minimax-m2" | "inclusionai/ring-1t" | "inclusionai/lint-1t" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1">;
|
|
364
446
|
/**
|
|
365
|
-
*
|
|
366
|
-
* @see {@link https://
|
|
447
|
+
* Zhipu AI Provider
|
|
448
|
+
* @see {@link https://docs.z.ai/guides/overview/pricing}
|
|
367
449
|
* @remarks
|
|
368
|
-
* - baseURL - `https://
|
|
369
|
-
* - apiKey - `
|
|
450
|
+
* - baseURL - `https://open.bigmodel.cn/api/paas/v4`
|
|
451
|
+
* - apiKey - `ZHIPU_API_KEY`
|
|
370
452
|
*/
|
|
371
|
-
declare const
|
|
453
|
+
declare const zhipuai: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
372
454
|
/**
|
|
373
|
-
*
|
|
374
|
-
* @see {@link https://
|
|
455
|
+
* Zhipu AI Coding Plan Provider
|
|
456
|
+
* @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
|
|
375
457
|
* @remarks
|
|
376
|
-
* - baseURL - `https://
|
|
377
|
-
* - apiKey - `
|
|
458
|
+
* - baseURL - `https://open.bigmodel.cn/api/coding/paas/v4`
|
|
459
|
+
* - apiKey - `ZHIPU_API_KEY`
|
|
378
460
|
*/
|
|
379
|
-
declare const
|
|
461
|
+
declare const zhipuaiCodingPlan: ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
380
462
|
/**
|
|
381
463
|
* Novita AI Provider
|
|
382
464
|
* @see {@link https://novita.ai/docs/guides/llm-api#api-integration}
|
|
@@ -384,15 +466,7 @@ declare const minimaxi: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sha
|
|
|
384
466
|
* - baseURL - `https://api.novita.ai/v3/openai/`
|
|
385
467
|
* - apiKey - `NOVITA_API_KEY`
|
|
386
468
|
*/
|
|
387
|
-
declare const novita:
|
|
388
|
-
/**
|
|
389
|
-
* SiliconFlow Provider
|
|
390
|
-
* @see {@link https://docs.siliconflow.com/en/userguide/quickstart#4-3-call-via-openai-interface}
|
|
391
|
-
* @remarks
|
|
392
|
-
* - baseURL - `https://api.siliconflow.cn/v1/`
|
|
393
|
-
* - apiKey - `SILICON_FLOW_API_KEY`
|
|
394
|
-
*/
|
|
395
|
-
declare const siliconFlow: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string> & _xsai_ext_shared_providers.ChatProvider<string>;
|
|
469
|
+
declare const novita: ModelProvider & ChatProvider<string>;
|
|
396
470
|
/**
|
|
397
471
|
* StepFun Provider
|
|
398
472
|
* @see {@link https://www.stepfun.com}
|
|
@@ -400,7 +474,7 @@ declare const siliconFlow: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_
|
|
|
400
474
|
* - baseURL - `https://api.stepfun.com/v1/`
|
|
401
475
|
* - apiKey - `STEPFUN_API_KEY`
|
|
402
476
|
*/
|
|
403
|
-
declare const stepfun:
|
|
477
|
+
declare const stepfun: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
|
|
404
478
|
/**
|
|
405
479
|
* Tencent Hunyuan Provider
|
|
406
480
|
* @see {@link https://cloud.tencent.com/document/product/1729}
|
|
@@ -408,7 +482,23 @@ declare const stepfun: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shar
|
|
|
408
482
|
* - baseURL - `https://api.hunyuan.cloud.tencent.com/v1/`
|
|
409
483
|
* - apiKey - `TENCENT_HUNYUAN_API_KEY`
|
|
410
484
|
*/
|
|
411
|
-
declare const tencentHunyuan:
|
|
485
|
+
declare const tencentHunyuan: ModelProvider & EmbedProvider<string> & ChatProvider<string>;
|
|
486
|
+
/**
|
|
487
|
+
* Ollama Provider
|
|
488
|
+
* @see {@link https://docs.ollama.com}
|
|
489
|
+
* @remarks
|
|
490
|
+
* - baseURL - `http://localhost:11434/v1/`
|
|
491
|
+
* - apiKey - `OLLAMA_API_KEY`
|
|
492
|
+
*/
|
|
493
|
+
declare const ollama: ModelProvider & EmbedProvider<string> & ChatProvider<string>;
|
|
494
|
+
/**
|
|
495
|
+
* LiteLLM Provider
|
|
496
|
+
* @see {@link https://docs.litellm.ai}
|
|
497
|
+
* @remarks
|
|
498
|
+
* - baseURL - `http://localhost:4000/v1/`
|
|
499
|
+
* - apiKey - `LITELLM_API_KEY`
|
|
500
|
+
*/
|
|
501
|
+
declare const litellm: ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
|
|
412
502
|
|
|
413
503
|
/**
|
|
414
504
|
* Anthropic Provider
|
|
@@ -417,7 +507,7 @@ declare const tencentHunyuan: _xsai_ext_shared_providers.ModelProvider & _xsai_e
|
|
|
417
507
|
* - baseURL - `https://api.anthropic.com/v1/`
|
|
418
508
|
* - apiKey - `ANTHROPIC_API_KEY`
|
|
419
509
|
*/
|
|
420
|
-
declare const anthropic:
|
|
510
|
+
declare const anthropic: ModelProvider & ChatProviderWithExtraOptions<AnthropicModels, AnthropicOptions>;
|
|
421
511
|
/**
|
|
422
512
|
* Featherless AI Provider
|
|
423
513
|
* @see {@link https://featherless.ai/models}
|
|
@@ -425,7 +515,7 @@ declare const anthropic: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_sh
|
|
|
425
515
|
* - baseURL - `https://api.featherless.ai/v1/`
|
|
426
516
|
* - apiKey - `FEATHERLESS_API_KEY`
|
|
427
517
|
*/
|
|
428
|
-
declare const featherless:
|
|
518
|
+
declare const featherless: ModelProvider & ChatProvider<string>;
|
|
429
519
|
/**
|
|
430
520
|
* OpenRouter Provider
|
|
431
521
|
* @see {@link https://openrouter.ai/models}
|
|
@@ -433,7 +523,7 @@ declare const featherless: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_
|
|
|
433
523
|
* - baseURL - `https://openrouter.ai/api/v1/`
|
|
434
524
|
* - apiKey - `OPENROUTER_API_KEY`
|
|
435
525
|
*/
|
|
436
|
-
declare const openrouter:
|
|
526
|
+
declare const openrouter: ModelProvider & ChatProviderWithExtraOptions<OpenrouterModels, OpenRouterOptions>;
|
|
437
527
|
/**
|
|
438
528
|
* Together AI Provider
|
|
439
529
|
* @see {@link https://docs.together.ai/docs/serverless-models}
|
|
@@ -441,6 +531,6 @@ declare const openrouter: _xsai_ext_shared_providers.ModelProvider & _xsai_ext_s
|
|
|
441
531
|
* - baseURL - `https://api.together.xyz/v1/`
|
|
442
532
|
* - apiKey - `TOGETHER_API_KEY`
|
|
443
533
|
*/
|
|
444
|
-
declare const togetherai:
|
|
534
|
+
declare const togetherai: ModelProvider & EmbedProvider<string> & ChatProvider<TogetheraiModels>;
|
|
445
535
|
|
|
446
|
-
export { alibaba, alibabaCn, anthropic, baseten, cerebras, chutes, cortecs, deepinfra, deepseek, fastrouter, featherless, fireworks, githubCopilot, githubModels, google, groq, huggingface, iflowcn, inception, inference, llama, lmstudio, lucidquery, minimax,
|
|
536
|
+
export { alibaba, alibabaCn, anthropic, bailing, baseten, cerebras, chutes, cortecs, deepinfra, deepseek, fastrouter, featherless, fireworks, githubCopilot, githubModels, google, groq, helicone, huggingface, iflowcn, inception, inference, ioNet, kimiForCoding, litellm, llama, lmstudio, lucidquery, minimax, minimaxCn, mistral, modelscope, moonshotai, moonshotaiCn, morph, nebius, novita, nvidia, ollama, ollamaCloud, openai, opencode, openrouter, ovhcloud, perplexity, poe, requesty, scaleway, siliconflow, siliconflowCn, stepfun, submodel, synthetic, tencentHunyuan, togetherai, upstage, venice, vultr, wandb, xai, xiaomi, zai, zaiCodingPlan, zenmux, zhipuai, zhipuaiCodingPlan };
|