@xsai-ext/providers 0.4.0-beta.9 → 0.4.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/anthropic-Bel15KQV.d.ts +111 -0
- package/dist/create.d.ts +201 -208
- package/dist/create.js +2 -42
- package/dist/index.d.ts +327 -237
- package/dist/index.js +51 -40
- package/dist/player2-D788vRq_.js +47 -0
- package/dist/special/create.d.ts +66 -0
- package/dist/special/create.js +37 -0
- package/dist/special/index.d.ts +12 -0
- package/dist/special/index.js +6 -0
- package/dist/{together-ai-DXREpfi4.js → together-ai-CzFuvpt-.js} +106 -59
- package/dist/types-Czo87gCh.d.ts +7 -0
- package/dist/types-DW1hvH0W.d.ts +38 -0
- package/dist/utils/index.d.ts +52 -0
- package/dist/utils/index.js +37 -0
- package/package.json +16 -4
- package/dist/openrouter-DCc0QhRt.d.ts +0 -104
package/dist/create.d.ts
CHANGED
|
@@ -1,336 +1,329 @@
|
|
|
1
|
-
import
|
|
2
|
-
|
|
3
|
-
|
|
1
|
+
import { C as ChatProvider, M as ModelProvider, E as EmbedProvider, I as ImageProvider, S as SpeechProvider, T as TranscriptionProvider } from './types-DW1hvH0W.js';
|
|
2
|
+
export { c as createAnthropic, a as createOpenRouter } from './anthropic-Bel15KQV.js';
|
|
3
|
+
import { T as TogetheraiModels } from './types-Czo87gCh.js';
|
|
4
|
+
import '@xsai/shared';
|
|
4
5
|
|
|
5
6
|
/**
|
|
6
|
-
* Create a
|
|
7
|
-
* @see {@link https://
|
|
7
|
+
* Create a Alibaba Provider
|
|
8
|
+
* @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
|
|
8
9
|
*/
|
|
9
|
-
declare const
|
|
10
|
+
declare const createAlibaba: (apiKey: string, baseURL?: string) => ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus"> & ModelProvider;
|
|
10
11
|
/**
|
|
11
|
-
* Create a
|
|
12
|
-
* @see {@link https://
|
|
12
|
+
* Create a Alibaba (China) Provider
|
|
13
|
+
* @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
|
|
13
14
|
*/
|
|
14
|
-
declare const
|
|
15
|
+
declare const createAlibabaCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
|
|
15
16
|
/**
|
|
16
|
-
* Create a
|
|
17
|
-
* @see {@link https://
|
|
17
|
+
* Create a Bailing Provider
|
|
18
|
+
* @see {@link https://alipaytbox.yuque.com/sxs0ba/ling/intro}
|
|
18
19
|
*/
|
|
19
|
-
declare const
|
|
20
|
+
declare const createBailing: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Ling-1T" | "Ring-1T">;
|
|
20
21
|
/**
|
|
21
|
-
* Create a
|
|
22
|
-
* @see {@link https://docs.
|
|
22
|
+
* Create a Baseten Provider
|
|
23
|
+
* @see {@link https://docs.baseten.co/development/model-apis/overview}
|
|
23
24
|
*/
|
|
24
|
-
declare const
|
|
25
|
+
declare const createBaseten: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2">;
|
|
25
26
|
/**
|
|
26
|
-
* Create a
|
|
27
|
-
* @see {@link https://
|
|
27
|
+
* Create a Cerebras Provider
|
|
28
|
+
* @see {@link https://inference-docs.cerebras.ai/models/overview}
|
|
28
29
|
*/
|
|
29
|
-
declare const
|
|
30
|
+
declare const createCerebras: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen-3-235b-a22b-instruct-2507" | "zai-glm-4.6" | "gpt-oss-120b"> & EmbedProvider<string>;
|
|
30
31
|
/**
|
|
31
|
-
* Create a
|
|
32
|
-
* @see {@link https://
|
|
32
|
+
* Create a Chutes Provider
|
|
33
|
+
* @see {@link https://llm.chutes.ai/v1/models}
|
|
33
34
|
*/
|
|
34
|
-
declare const
|
|
35
|
+
declare const createChutes: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2" | "NousResearch/Hermes-4.3-36B" | "NousResearch/Hermes-4-70B" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4-405B-FP8" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "rednote-hilab/dots.ocr" | "MiniMaxAI/MiniMax-M2" | "ArliAI/QwQ-32B-ArliAI-RpR-v1" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "tngtech/TNG-R1T-Chimera-TEE" | "OpenGVLab/InternVL3-78B" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "Alibaba-NLP/Tongyi-DeepResearch-30B-A3B" | "mistralai/Devstral-2-123B-Instruct-2512" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "unsloth/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen3-32B" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-V3" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
|
|
35
36
|
/**
|
|
36
|
-
* Create a
|
|
37
|
-
* @see {@link https://api.
|
|
37
|
+
* Create a Cortecs Provider
|
|
38
|
+
* @see {@link https://api.cortecs.ai/v1/models}
|
|
38
39
|
*/
|
|
39
|
-
declare const
|
|
40
|
+
declare const createCortecs: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "gpt-4.1" | "gemini-2.5-pro" | "devstral-small-2512" | "claude-sonnet-4" | "llama-3.1-405b-instruct">;
|
|
40
41
|
/**
|
|
41
|
-
* Create a
|
|
42
|
-
* @see {@link https://
|
|
42
|
+
* Create a Deep Infra Provider
|
|
43
|
+
* @see {@link https://deepinfra.com/models}
|
|
43
44
|
*/
|
|
44
|
-
declare const
|
|
45
|
+
declare const createDeepinfra: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "zai-org/GLM-4.5" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo">;
|
|
45
46
|
/**
|
|
46
|
-
* Create a
|
|
47
|
-
* @see {@link https://
|
|
47
|
+
* Create a DeepSeek Provider
|
|
48
|
+
* @see {@link https://platform.deepseek.com/api-docs/pricing}
|
|
48
49
|
*/
|
|
49
|
-
declare const
|
|
50
|
+
declare const createDeepSeek: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"deepseek-chat" | "deepseek-reasoner">;
|
|
50
51
|
/**
|
|
51
|
-
* Create a
|
|
52
|
-
* @see {@link https://
|
|
52
|
+
* Create a FastRouter Provider
|
|
53
|
+
* @see {@link https://fastrouter.ai/models}
|
|
54
|
+
*/
|
|
55
|
+
declare const createFastrouter: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
|
|
56
|
+
/**
|
|
57
|
+
* Create a Fireworks AI Provider
|
|
58
|
+
* @see {@link https://fireworks.ai/docs/}
|
|
53
59
|
*/
|
|
54
|
-
declare const
|
|
60
|
+
declare const createFireworks: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
|
|
55
61
|
/**
|
|
56
62
|
* Create a GitHub Copilot Provider
|
|
57
63
|
* @see {@link https://docs.github.com/en/copilot}
|
|
58
64
|
*/
|
|
59
|
-
declare const createGithubCopilot: (apiKey: string, baseURL?: string) =>
|
|
65
|
+
declare const createGithubCopilot: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "gemini-2.0-flash-001" | "claude-opus-4" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gpt-5.1-codex" | "claude-haiku-4.5" | "gemini-3-pro-preview" | "oswe-vscode-prime" | "claude-3.5-sonnet" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "claude-opus-41" | "gpt-5-mini" | "claude-3.7-sonnet" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "claude-3.7-sonnet-thought" | "claude-opus-4.5" | "gpt-5.2" | "claude-sonnet-4.5">;
|
|
60
66
|
/**
|
|
61
|
-
* Create a
|
|
62
|
-
* @see {@link https://docs.
|
|
67
|
+
* Create a GitHub Models Provider
|
|
68
|
+
* @see {@link https://docs.github.com/en/github-models}
|
|
69
|
+
*/
|
|
70
|
+
declare const createGithubModels: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-4.1" | "core42/jais-30b-chat" | "xai/grok-3" | "xai/grok-3-mini" | "cohere/cohere-command-r-08-2024" | "cohere/cohere-command-a" | "cohere/cohere-command-r-plus-08-2024" | "cohere/cohere-command-r" | "cohere/cohere-command-r-plus" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-r1" | "deepseek/deepseek-v3-0324" | "mistral-ai/mistral-medium-2505" | "mistral-ai/ministral-3b" | "mistral-ai/mistral-nemo" | "mistral-ai/mistral-large-2411" | "mistral-ai/codestral-2501" | "mistral-ai/mistral-small-2503" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-mini-4k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-4" | "microsoft/phi-4-mini-reasoning" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-mini-instruct" | "microsoft/phi-4-multimodal-instruct" | "microsoft/phi-3-mini-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-mini-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-4-reasoning" | "microsoft/mai-ds-r1" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o1-preview" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o1-mini" | "openai/o3" | "openai/gpt-4o-mini" | "meta/llama-3.2-11b-vision-instruct" | "meta/meta-llama-3.1-405b-instruct" | "meta/llama-4-maverick-17b-128e-instruct-fp8" | "meta/meta-llama-3-70b-instruct" | "meta/meta-llama-3.1-70b-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-3.2-90b-vision-instruct" | "meta/meta-llama-3-8b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/meta-llama-3.1-8b-instruct" | "ai21-labs/ai21-jamba-1.5-large" | "ai21-labs/ai21-jamba-1.5-mini">;
|
|
71
|
+
/**
|
|
72
|
+
* Create a Google Provider
|
|
73
|
+
* @see {@link https://ai.google.dev/gemini-api/docs/pricing}
|
|
63
74
|
*/
|
|
64
|
-
declare const
|
|
75
|
+
declare const createGoogleGenerativeAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-2.5-pro" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-2.5-flash" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
|
|
65
76
|
/**
|
|
66
|
-
* Create a
|
|
67
|
-
* @see {@link https://
|
|
77
|
+
* Create a Groq Provider
|
|
78
|
+
* @see {@link https://console.groq.com/docs/models}
|
|
68
79
|
*/
|
|
69
|
-
declare const
|
|
80
|
+
declare const createGroq: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b">;
|
|
70
81
|
/**
|
|
71
|
-
* Create a
|
|
72
|
-
* @see {@link https://
|
|
82
|
+
* Create a Helicone Provider
|
|
83
|
+
* @see {@link https://helicone.ai/models}
|
|
73
84
|
*/
|
|
74
|
-
declare const
|
|
85
|
+
declare const createHelicone: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-next-80b-a3b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-thinking" | "gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "deepseek-reasoner" | "claude-opus-4" | "grok-code-fast-1" | "gpt-5.1-codex" | "gemini-3-pro-preview" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "claude-3.7-sonnet" | "o3" | "gpt-5" | "gemini-2.5-flash" | "gemini-2.5-flash-lite" | "llama-3.1-8b-instant" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "qwen3-coder" | "grok-4-fast-reasoning" | "claude-opus-4-1" | "grok-4" | "llama-4-maverick" | "llama-prompt-guard-2-86m" | "grok-4-1-fast-reasoning" | "claude-4.5-haiku" | "llama-3.1-8b-instruct-turbo" | "gpt-4.1-mini-2025-04-14" | "llama-guard-4" | "llama-3.1-8b-instruct" | "gpt-4.1-mini" | "deepseek-v3.1-terminus" | "llama-prompt-guard-2-22m" | "claude-3.5-sonnet-v2" | "sonar-deep-research" | "claude-sonnet-4-5-20250929" | "grok-3" | "mistral-small" | "kimi-k2-0711" | "chatgpt-4o-latest" | "kimi-k2-0905" | "sonar-reasoning" | "llama-3.3-70b-instruct" | "claude-4.5-sonnet" | "codex-mini-latest" | "gpt-5-nano" | "deepseek-tng-r1t2-chimera" | "claude-4.5-opus" | "sonar" | "glm-4.6" | "qwen3-235b-a22b-thinking" | "hermes-2-pro-llama-3-8b" | "o1" | "grok-3-mini" | "sonar-pro" | "o1-mini" | "claude-3-haiku-20240307" | "o3-pro" | "qwen2.5-coder-7b-fast" | "gemma-3-12b-it" | "mistral-nemo" | "gpt-oss-20b" | "claude-3.5-haiku" | "gpt-5-chat-latest" | "gpt-4o-mini" | "sonar-reasoning-pro" | "qwen3-vl-235b-a22b-instruct" | "qwen3-30b-a3b" | "deepseek-v3.2" | "grok-4-1-fast-non-reasoning" | "gpt-5-pro" | "mistral-large-2411" | "claude-opus-4-1-20250805" | "ernie-4.5-21b-a3b-thinking" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "llama-4-scout">;
|
|
75
86
|
/**
|
|
76
|
-
* Create a
|
|
77
|
-
* @see {@link https://
|
|
87
|
+
* Create a Hugging Face Provider
|
|
88
|
+
* @see {@link https://huggingface.co/docs/inference-providers}
|
|
78
89
|
*/
|
|
79
|
-
declare const
|
|
90
|
+
declare const createHuggingface: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "MiniMaxAI/MiniMax-M2" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/Deepseek-V3-0324">;
|
|
80
91
|
/**
|
|
81
|
-
* Create a
|
|
82
|
-
* @see {@link https://
|
|
92
|
+
* Create a iFlow Provider
|
|
93
|
+
* @see {@link https://platform.iflow.cn/en/docs}
|
|
83
94
|
*/
|
|
84
|
-
declare const
|
|
95
|
+
declare const createIflowcn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-vl-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "kimi-k2-thinking" | "qwen3-coder" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "kimi-k2" | "deepseek-v3.1" | "minimax-m2" | "qwen3-235b" | "deepseek-v3.2-chat" | "qwen3-235b-a22b-thinking-2507" | "tstars2.0" | "qwen3-235b-a22b-instruct" | "qwen3-max-preview">;
|
|
85
96
|
/**
|
|
86
|
-
* Create a
|
|
87
|
-
* @see {@link https://
|
|
97
|
+
* Create a Inception Provider
|
|
98
|
+
* @see {@link https://platform.inceptionlabs.ai/docs}
|
|
88
99
|
*/
|
|
89
|
-
declare const
|
|
100
|
+
declare const createInception: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"mercury-coder" | "mercury">;
|
|
90
101
|
/**
|
|
91
|
-
* Create a
|
|
92
|
-
* @see {@link https://
|
|
102
|
+
* Create a Inference Provider
|
|
103
|
+
* @see {@link https://inference.net/models}
|
|
93
104
|
*/
|
|
94
|
-
declare const
|
|
105
|
+
declare const createInference: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"meta/llama-3.2-11b-vision-instruct" | "mistral/mistral-nemo-12b-instruct" | "google/gemma-3" | "osmosis/osmosis-structure-0.6b" | "qwen/qwen3-embedding-4b" | "qwen/qwen-2.5-7b-vision-instruct" | "meta/llama-3.1-8b-instruct" | "meta/llama-3.2-3b-instruct" | "meta/llama-3.2-1b-instruct">;
|
|
95
106
|
/**
|
|
96
|
-
* Create a
|
|
97
|
-
* @see {@link https://
|
|
107
|
+
* Create a IO.NET Provider
|
|
108
|
+
* @see {@link https://io.net/docs/guides/intelligence/io-intelligence}
|
|
98
109
|
*/
|
|
99
|
-
declare const
|
|
110
|
+
declare const createIoNet: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506" | "mistralai/Mistral-Large-Instruct-2411" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar">;
|
|
100
111
|
/**
|
|
101
|
-
* Create a
|
|
102
|
-
* @see {@link https://
|
|
112
|
+
* Create a Kimi For Coding Provider
|
|
113
|
+
* @see {@link https://www.kimi.com/coding/docs/en/third-party-agents.html}
|
|
103
114
|
*/
|
|
104
|
-
declare const
|
|
115
|
+
declare const createKimiForCoding: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking">;
|
|
105
116
|
/**
|
|
106
|
-
* Create a
|
|
107
|
-
* @see {@link https://
|
|
117
|
+
* Create a Llama Provider
|
|
118
|
+
* @see {@link https://llama.developer.meta.com/docs/models}
|
|
108
119
|
*/
|
|
109
|
-
declare const
|
|
120
|
+
declare const createLlama: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"llama-3.3-70b-instruct" | "llama-3.3-8b-instruct" | "llama-4-maverick-17b-128e-instruct-fp8" | "llama-4-scout-17b-16e-instruct-fp8" | "groq-llama-4-maverick-17b-128e-instruct" | "cerebras-llama-4-scout-17b-16e-instruct" | "cerebras-llama-4-maverick-17b-128e-instruct">;
|
|
110
121
|
/**
|
|
111
|
-
* Create a
|
|
112
|
-
* @see {@link https://
|
|
122
|
+
* Create a LMStudio Provider
|
|
123
|
+
* @see {@link https://lmstudio.ai/models}
|
|
113
124
|
*/
|
|
114
|
-
declare const
|
|
125
|
+
declare const createLmstudio: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "qwen/qwen3-30b-a3b-2507" | "qwen/qwen3-coder-30b">;
|
|
115
126
|
/**
|
|
116
|
-
* Create a
|
|
117
|
-
* @see {@link https://
|
|
127
|
+
* Create a LucidQuery AI Provider
|
|
128
|
+
* @see {@link https://lucidquery.com/api/docs}
|
|
118
129
|
*/
|
|
119
|
-
declare const
|
|
130
|
+
declare const createLucidquery: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"lucidquery-nexus-coder" | "lucidnova-rf1-100b">;
|
|
120
131
|
/**
|
|
121
|
-
* Create a
|
|
122
|
-
* @see {@link https://
|
|
132
|
+
* Create a MiniMax Provider
|
|
133
|
+
* @see {@link https://platform.minimax.io/docs/guides/quickstart}
|
|
123
134
|
*/
|
|
124
|
-
declare const
|
|
135
|
+
declare const createMinimax: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
125
136
|
/**
|
|
126
|
-
* Create a
|
|
127
|
-
* @see {@link https://platform.
|
|
137
|
+
* Create a MiniMax (China) Provider
|
|
138
|
+
* @see {@link https://platform.minimaxi.com/docs/guides/quickstart}
|
|
128
139
|
*/
|
|
129
|
-
declare const
|
|
140
|
+
declare const createMinimaxCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
130
141
|
/**
|
|
131
|
-
* Create a
|
|
132
|
-
* @see {@link https://
|
|
142
|
+
* Create a Mistral Provider
|
|
143
|
+
* @see {@link https://docs.mistral.ai/getting-started/models/}
|
|
144
|
+
*/
|
|
145
|
+
declare const createMistral: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"devstral-2512" | "mistral-nemo" | "mistral-large-2411" | "devstral-medium-2507" | "mistral-large-2512" | "open-mixtral-8x22b" | "ministral-8b-latest" | "pixtral-large-latest" | "mistral-small-2506" | "ministral-3b-latest" | "pixtral-12b" | "mistral-medium-2505" | "labs-devstral-small-2512" | "devstral-medium-latest" | "devstral-small-2505" | "mistral-medium-2508" | "mistral-embed" | "mistral-small-latest" | "magistral-small" | "devstral-small-2507" | "codestral-latest" | "open-mixtral-8x7b" | "open-mistral-7b" | "mistral-large-latest" | "mistral-medium-latest" | "magistral-medium-latest">;
|
|
146
|
+
/**
|
|
147
|
+
* Create a ModelScope Provider
|
|
148
|
+
* @see {@link https://modelscope.cn/docs/model-service/API-Inference/intro}
|
|
149
|
+
*/
|
|
150
|
+
declare const createModelscope: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507">;
|
|
151
|
+
/**
|
|
152
|
+
* Create a Moonshot AI Provider
|
|
153
|
+
* @see {@link https://platform.moonshot.ai/docs/api/chat}
|
|
154
|
+
*/
|
|
155
|
+
declare const createMoonshotai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
|
|
156
|
+
/**
|
|
157
|
+
* Create a Moonshot AI (China) Provider
|
|
158
|
+
* @see {@link https://platform.moonshot.cn/docs/api/chat}
|
|
159
|
+
*/
|
|
160
|
+
declare const createMoonshotaiCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
|
|
161
|
+
/**
|
|
162
|
+
* Create a Morph Provider
|
|
163
|
+
* @see {@link https://docs.morphllm.com/api-reference/introduction}
|
|
164
|
+
*/
|
|
165
|
+
declare const createMorph: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
|
|
166
|
+
/**
|
|
167
|
+
* Create a Nebius Token Factory Provider
|
|
168
|
+
* @see {@link https://docs.tokenfactory.nebius.com/}
|
|
169
|
+
*/
|
|
170
|
+
declare const createNebius: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2-instruct" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "zai-org/glm-4.5" | "zai-org/glm-4.5-air" | "deepseek-ai/deepseek-v3">;
|
|
171
|
+
/**
|
|
172
|
+
* Create a Nvidia Provider
|
|
173
|
+
* @see {@link https://docs.api.nvidia.com/nim/}
|
|
174
|
+
*/
|
|
175
|
+
declare const createNvidia: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-4-mini-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-coder-480b-a35b-instruct" | "moonshotai/kimi-k2-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "minimaxai/minimax-m2" | "google/gemma-3-27b-it" | "openai/whisper-large-v3" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwen3-next-80b-a3b-thinking" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "black-forest-labs/flux.1-dev">;
|
|
176
|
+
/**
|
|
177
|
+
* Create a Ollama Cloud Provider
|
|
178
|
+
* @see {@link https://docs.ollama.com/cloud}
|
|
133
179
|
*/
|
|
134
|
-
declare const
|
|
180
|
+
declare const createOllamaCloud: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking:cloud" | "qwen3-vl-235b-cloud" | "qwen3-coder:480b-cloud" | "gpt-oss:120b-cloud" | "deepseek-v3.1:671b-cloud" | "glm-4.6:cloud" | "cogito-2.1:671b-cloud" | "gpt-oss:20b-cloud" | "qwen3-vl-235b-instruct-cloud" | "kimi-k2:1t-cloud" | "minimax-m2:cloud" | "gemini-3-pro-preview:latest">;
|
|
135
181
|
/**
|
|
136
182
|
* Create a OpenAI Provider
|
|
137
183
|
* @see {@link https://platform.openai.com/docs/models}
|
|
138
184
|
*/
|
|
139
|
-
declare const createOpenAI: (apiKey: string, baseURL?: string) =>
|
|
185
|
+
declare const createOpenAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "gpt-5.2" | "gpt-4.1-nano" | "gpt-4.1-mini" | "codex-mini-latest" | "gpt-5-nano" | "o1" | "o1-mini" | "o3-pro" | "gpt-5-chat-latest" | "gpt-4o-mini" | "gpt-5-pro" | "gpt-5.1-chat-latest" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "o3-deep-research" | "gpt-3.5-turbo" | "gpt-5.2-pro" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "gpt-5.2-chat-latest" | "text-embedding-ada-002" | "gpt-4o-2024-11-20" | "o4-mini-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
|
|
140
186
|
/**
|
|
141
|
-
* Create a
|
|
142
|
-
* @see {@link https://
|
|
187
|
+
* Create a OpenCode Zen Provider
|
|
188
|
+
* @see {@link https://opencode.ai/docs/zen}
|
|
189
|
+
*/
|
|
190
|
+
declare const createOpencode: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking" | "claude-sonnet-4" | "gpt-5.1-codex" | "gpt-5.1" | "gpt-5-codex" | "gpt-5.1-codex-max" | "gpt-5" | "gpt-5.2" | "qwen3-coder" | "claude-opus-4-1" | "gpt-5-nano" | "glm-4.6" | "kimi-k2" | "claude-haiku-4-5" | "claude-opus-4-5" | "gemini-3-pro" | "claude-sonnet-4-5" | "alpha-gd4" | "big-pickle" | "claude-3-5-haiku" | "grok-code" | "gemini-3-flash" | "alpha-doubao-seed-code" | "minimax-m2.1">;
|
|
191
|
+
/**
|
|
192
|
+
* Create a OVHcloud AI Endpoints Provider
|
|
193
|
+
* @see {@link https://www.ovhcloud.com/en/public-cloud/ai-endpoints/catalog//}
|
|
143
194
|
*/
|
|
144
|
-
declare const
|
|
195
|
+
declare const createOvhcloud: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "gpt-oss-20b" | "mixtral-8x7b-instruct-v0.1" | "mistral-7b-instruct-v0.3" | "qwen2.5-vl-72b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen2.5-coder-32b-instruct" | "llava-next-mistral-7b" | "meta-llama-3_1-70b-instruct" | "meta-llama-3_3-70b-instruct">;
|
|
145
196
|
/**
|
|
146
197
|
* Create a Perplexity Provider
|
|
147
198
|
* @see {@link https://docs.perplexity.ai}
|
|
148
199
|
*/
|
|
149
|
-
declare const createPerplexity: (apiKey: string, baseURL?: string) =>
|
|
200
|
+
declare const createPerplexity: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"sonar-reasoning" | "sonar" | "sonar-pro" | "sonar-reasoning-pro">;
|
|
150
201
|
/**
|
|
151
|
-
* Create a
|
|
152
|
-
* @see {@link https://
|
|
202
|
+
* Create a Poe Provider
|
|
203
|
+
* @see {@link https://creator.poe.com/docs/external-applications/openai-compatible-api}
|
|
153
204
|
*/
|
|
154
|
-
declare const
|
|
205
|
+
declare const createPoe: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-2" | "xai/grok-4.1-fast-non-reasoning" | "ideogramai/ideogram" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram-v2" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "poetools/claude-code" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "google/gemini-deep-research" | "google/nano-banana" | "google/imagen-4" | "google/imagen-3" | "google/imagen-4-ultra" | "google/gemini-2.0-flash-lite" | "google/gemini-3-pro" | "google/veo-3.1" | "google/imagen-3-fast" | "google/lyria" | "google/gemini-2.0-flash" | "google/gemini-2.5-flash-lite" | "google/veo-3" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-2" | "google/nano-banana-pro" | "google/veo-3.1-fast" | "openai/gpt-5.2-instant" | "openai/sora-2" | "openai/o1-pro" | "openai/gpt-5.1-codex" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-4-classic" | "openai/gpt-5-chat" | "openai/o3-deep-research" | "openai/gpt-4o-search" | "openai/gpt-image-1-mini" | "openai/gpt-3.5-turbo" | "openai/gpt-5.2-pro" | "openai/o3-mini-high" | "openai/chatgpt-4o-latest" | "openai/gpt-4-turbo" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1-instant" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-4o-aug" | "openai/o3-pro" | "openai/gpt-image-1" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo-instruct" | "openai/o4-mini-deep-research" | "openai/gpt-4-classic-0314" | "openai/dall-e-3" | "openai/sora-2-pro" | "openai/gpt-5-pro" | "openai/gpt-5.2" | "openai/gpt-4o-mini-search" | "stabilityai/stablediffusionxl" | "topazlabs-co/topazlabs" | "lumalabs/ray2" | "lumalabs/dream-machine" | "anthropic/claude-opus-3" | "anthropic/claude-opus-4" | "anthropic/claude-sonnet-3.7-reasoning" | "anthropic/claude-opus-4-search" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-haiku-3.5-search" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4-reasoning" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7-search" | "anthropic/claude-opus-4-reasoning" | "anthropic/claude-sonnet-3.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-haiku-3.5" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-sonnet-4-search" | "trytako/tako" | "novita/kimi-k2-thinking" | "novita/glm-4.6">;
|
|
155
206
|
/**
|
|
156
|
-
* Create a
|
|
157
|
-
* @see {@link https://
|
|
207
|
+
* Create a Requesty Provider
|
|
208
|
+
* @see {@link https://requesty.ai/solution/llm-routing/models}
|
|
158
209
|
*/
|
|
159
|
-
declare const
|
|
210
|
+
declare const createRequesty: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-sonnet-4" | "openai/gpt-4.1-mini" | "openai/o4-mini" | "openai/gpt-4o-mini" | "xai/grok-4" | "anthropic/claude-opus-4" | "xai/grok-4-fast" | "google/gemini-3-flash-preview" | "google/gemini-3-pro-preview" | "anthropic/claude-opus-4-1" | "anthropic/claude-haiku-4-5" | "anthropic/claude-opus-4-5" | "anthropic/claude-sonnet-4-5" | "anthropic/claude-3-7-sonnet">;
|
|
160
211
|
/**
|
|
161
|
-
* Create a
|
|
162
|
-
* @see {@link https://
|
|
212
|
+
* Create a Scaleway Provider
|
|
213
|
+
* @see {@link https://www.scaleway.com/en/docs/generative-apis/}
|
|
163
214
|
*/
|
|
164
|
-
declare const
|
|
215
|
+
declare const createScaleway: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen3-235b-a22b-instruct-2507" | "pixtral-12b-2409" | "whisper-large-v3" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
|
|
165
216
|
/**
|
|
166
|
-
* Create a
|
|
167
|
-
* @see {@link https://
|
|
217
|
+
* Create a SiliconFlow Provider
|
|
218
|
+
* @see {@link https://cloud.siliconflow.com/models}
|
|
168
219
|
*/
|
|
169
|
-
declare const
|
|
220
|
+
declare const createSiliconFlow: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
|
|
170
221
|
/**
|
|
171
|
-
* Create a
|
|
172
|
-
* @see {@link https://
|
|
222
|
+
* Create a SiliconFlow (China) Provider
|
|
223
|
+
* @see {@link https://cloud.siliconflow.com/models}
|
|
173
224
|
*/
|
|
174
|
-
declare const
|
|
225
|
+
declare const createSiliconflowCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
|
|
175
226
|
/**
|
|
176
227
|
* Create a submodel Provider
|
|
177
228
|
* @see {@link https://submodel.gitbook.io}
|
|
178
229
|
*/
|
|
179
|
-
declare const createSubmodel: (apiKey: string, baseURL?: string) =>
|
|
230
|
+
declare const createSubmodel: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324" | "zai-org/GLM-4.5-FP8">;
|
|
180
231
|
/**
|
|
181
|
-
* Create a
|
|
182
|
-
* @see {@link https://
|
|
232
|
+
* Create a Synthetic Provider
|
|
233
|
+
* @see {@link https://synthetic.new/pricing}
|
|
183
234
|
*/
|
|
184
|
-
declare const
|
|
235
|
+
declare const createSynthetic: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
|
|
185
236
|
/**
|
|
186
|
-
* Create a
|
|
187
|
-
* @see {@link https://
|
|
237
|
+
* Create a Upstage Provider
|
|
238
|
+
* @see {@link https://developers.upstage.ai/docs/apis/chat}
|
|
188
239
|
*/
|
|
189
|
-
declare const
|
|
240
|
+
declare const createUpstage: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"solar-mini" | "solar-pro2">;
|
|
190
241
|
/**
|
|
191
|
-
* Create a
|
|
192
|
-
* @see {@link https://
|
|
242
|
+
* Create a Venice AI Provider
|
|
243
|
+
* @see {@link https://docs.venice.ai}
|
|
193
244
|
*/
|
|
194
|
-
declare const
|
|
245
|
+
declare const createVenice: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "kimi-k2-thinking" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "deepseek-v3.2" | "qwen3-235b" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "openai-gpt-oss-120b" | "zai-org-glm-4.6" | "grok-41-fast" | "claude-opus-45" | "mistral-31-24b" | "venice-uncensored" | "openai-gpt-52" | "qwen3-4b" | "llama-3.3-70b" | "devstral-2-2512" | "llama-3.2-3b" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "zai-org-glm-4.6v" | "qwen3-next-80b">;
|
|
195
246
|
/**
|
|
196
|
-
* Create a
|
|
197
|
-
* @see {@link https://
|
|
247
|
+
* Create a Vultr Provider
|
|
248
|
+
* @see {@link https://api.vultrinference.com/}
|
|
198
249
|
*/
|
|
199
|
-
declare const
|
|
250
|
+
declare const createVultr: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-instruct" | "qwen2.5-coder-32b-instruct">;
|
|
200
251
|
/**
|
|
201
|
-
* Create a
|
|
202
|
-
* @see {@link https://
|
|
252
|
+
* Create a Weights & Biases Provider
|
|
253
|
+
* @see {@link https://weave-docs.wandb.ai/guides/integrations/inference/}
|
|
203
254
|
*/
|
|
204
|
-
declare const
|
|
255
|
+
declare const createWandb: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2-Instruct" | "meta-llama/Llama-3.3-70B-Instruct" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
|
|
205
256
|
/**
|
|
206
|
-
* Create a
|
|
207
|
-
* @see {@link https://
|
|
257
|
+
* Create a xAI Provider
|
|
258
|
+
* @see {@link https://docs.x.ai/docs/models}
|
|
208
259
|
*/
|
|
209
|
-
declare const
|
|
260
|
+
declare const createXai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"grok-code-fast-1" | "grok-4-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-4-1-fast-non-reasoning" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
|
|
210
261
|
/**
|
|
211
|
-
* Create a
|
|
212
|
-
* @see {@link https://
|
|
262
|
+
* Create a Xiaomi Provider
|
|
263
|
+
* @see {@link https://platform.xiaomimimo.com/#/docs}
|
|
213
264
|
*/
|
|
214
|
-
declare const
|
|
265
|
+
declare const createXiaomi: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"mimo-v2-flash">;
|
|
215
266
|
/**
|
|
216
|
-
* Create a
|
|
217
|
-
* @see {@link https://
|
|
267
|
+
* Create a Z.AI Provider
|
|
268
|
+
* @see {@link https://docs.z.ai/guides/overview/pricing}
|
|
218
269
|
*/
|
|
219
|
-
declare const
|
|
270
|
+
declare const createZai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
220
271
|
/**
|
|
221
|
-
* Create a
|
|
222
|
-
* @see {@link https://
|
|
272
|
+
* Create a Z.AI Coding Plan Provider
|
|
273
|
+
* @see {@link https://docs.z.ai/devpack/overview}
|
|
223
274
|
*/
|
|
224
|
-
declare const
|
|
275
|
+
declare const createZaiCodingPlan: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
225
276
|
/**
|
|
226
|
-
* Create a
|
|
227
|
-
* @see {@link https://
|
|
277
|
+
* Create a ZenMux Provider
|
|
278
|
+
* @see {@link https://docs.zenmux.ai}
|
|
228
279
|
*/
|
|
229
|
-
declare const
|
|
280
|
+
declare const createZenmux: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "moonshotai/kimi-k2-thinking" | "openai/gpt-5-codex" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.5" | "moonshotai/kimi-k2-thinking-turbo" | "moonshotai/kimi-k2-0905" | "xiaomi/mimo-v2-flash" | "x-ai/grok-4-fast-non-reasoning" | "x-ai/grok-code-fast-1" | "x-ai/grok-4-fast" | "deepseek/deepseek-chat" | "minimax/minimax-m2" | "inclusionai/ring-1t" | "inclusionai/lint-1t" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1">;
|
|
230
281
|
/**
|
|
231
|
-
* Create a
|
|
232
|
-
* @see {@link https://
|
|
282
|
+
* Create a Zhipu AI Provider
|
|
283
|
+
* @see {@link https://docs.z.ai/guides/overview/pricing}
|
|
233
284
|
*/
|
|
234
|
-
declare const
|
|
285
|
+
declare const createZhipuai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
235
286
|
/**
|
|
236
|
-
* Create a
|
|
237
|
-
* @see {@link https://
|
|
287
|
+
* Create a Zhipu AI Coding Plan Provider
|
|
288
|
+
* @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
|
|
238
289
|
*/
|
|
239
|
-
declare const
|
|
290
|
+
declare const createZhipuaiCodingPlan: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
240
291
|
/**
|
|
241
292
|
* Create a Novita AI Provider
|
|
242
293
|
* @see {@link https://novita.ai/docs/guides/llm-api#api-integration}
|
|
243
294
|
*/
|
|
244
|
-
declare const createNovita: (apiKey: string, baseURL?: string) =>
|
|
245
|
-
/**
|
|
246
|
-
* Create a SiliconFlow Provider
|
|
247
|
-
* @see {@link https://docs.siliconflow.com/en/userguide/quickstart#4-3-call-via-openai-interface}
|
|
248
|
-
*/
|
|
249
|
-
declare const createSiliconFlow: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string> & _xsai_ext_shared_providers.ChatProvider<string>;
|
|
295
|
+
declare const createNovita: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<string>;
|
|
250
296
|
/**
|
|
251
297
|
* Create a StepFun Provider
|
|
252
298
|
* @see {@link https://www.stepfun.com}
|
|
253
299
|
*/
|
|
254
|
-
declare const createStepfun: (apiKey: string, baseURL?: string) =>
|
|
300
|
+
declare const createStepfun: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
|
|
255
301
|
/**
|
|
256
302
|
* Create a Tencent Hunyuan Provider
|
|
257
303
|
* @see {@link https://cloud.tencent.com/document/product/1729}
|
|
258
304
|
*/
|
|
259
|
-
declare const createTencentHunyuan: (apiKey: string, baseURL?: string) =>
|
|
260
|
-
|
|
305
|
+
declare const createTencentHunyuan: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<string>;
|
|
261
306
|
/**
|
|
262
|
-
* Create a
|
|
263
|
-
* @see {@link https://docs.
|
|
307
|
+
* Create a Ollama Provider
|
|
308
|
+
* @see {@link https://docs.ollama.com}
|
|
264
309
|
*/
|
|
265
|
-
declare const
|
|
266
|
-
|
|
267
|
-
|
|
268
|
-
|
|
269
|
-
|
|
270
|
-
|
|
271
|
-
* If passed in as a function, it is treated as an accessTokenFetcher.
|
|
272
|
-
*
|
|
273
|
-
* @see {@link https://learn.microsoft.com/en-us/azure/api-management/api-management-authenticate-authorize-azure-openai}
|
|
274
|
-
*/
|
|
275
|
-
apiKey: (() => Promise<string> | string) | string;
|
|
276
|
-
/**
|
|
277
|
-
* The Azure API version to use (`api-version` param).
|
|
278
|
-
*
|
|
279
|
-
* Notice: Different deployment over different time may have different API versions, please
|
|
280
|
-
* follow the exact prompt from either [Azure AI Foundry](https://ai.azure.com/) or Azure OpenAI service
|
|
281
|
-
* to get the correct API version from the Azure OpenAI Service endpoint.
|
|
282
|
-
*
|
|
283
|
-
* On Azure AI Foundry portal, you can go to https://ai.azure.com/build/overview > Choose the project >
|
|
284
|
-
* Overview > Endpoints and keys > Included capabilities > Azure OpenAI Service to get the correct endpoint.
|
|
285
|
-
*
|
|
286
|
-
* @see {@link https://learn.microsoft.com/en-us/azure/ai-services/openai/reference#rest-api-versioning}
|
|
287
|
-
*/
|
|
288
|
-
apiVersion?: string;
|
|
289
|
-
/**
|
|
290
|
-
* Azure resource name.
|
|
291
|
-
*
|
|
292
|
-
* On Azure AI Foundry portal, you can go to https://ai.azure.com/build/overview > Choose the project >
|
|
293
|
-
* Overview > Endpoints and keys > Included capabilities > Azure OpenAI Service to get the correct endpoint.
|
|
294
|
-
*
|
|
295
|
-
* @see {@link https://learn.microsoft.com/en-us/azure/ai-services/openai/reference#uri-parameters}
|
|
296
|
-
*/
|
|
297
|
-
resourceName: string;
|
|
298
|
-
}
|
|
299
|
-
/**
|
|
300
|
-
* Create a Azure Provider
|
|
301
|
-
* @see {@link https://ai.azure.com/explore/models}
|
|
302
|
-
* @see {@link https://learn.microsoft.com/en-us/azure/ai-foundry/model-inference/concepts/endpoints?tabs=rest#routing}
|
|
303
|
-
* @remarks
|
|
304
|
-
* For Azure AI services, you can have multiple deployments of the same model with different names.
|
|
305
|
-
*
|
|
306
|
-
* Please pass your deployment name as the `model` parameter. By default, Azure will use the model name
|
|
307
|
-
* as the deployment name when deploying a model.
|
|
308
|
-
*/
|
|
309
|
-
declare const createAzure: (options: CreateAzureOptions) => Promise<_xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string> & _xsai_ext_shared_providers.ChatProvider<AzureModels>>;
|
|
310
|
+
declare const createOllama: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<string>;
|
|
311
|
+
/**
|
|
312
|
+
* Create a LiteLLM Provider
|
|
313
|
+
* @see {@link https://docs.litellm.ai}
|
|
314
|
+
*/
|
|
315
|
+
declare const createLitellm: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
|
|
310
316
|
|
|
311
317
|
/**
|
|
312
318
|
* Create a Featherless AI Provider
|
|
313
319
|
* @see {@link https://featherless.ai/models}
|
|
314
320
|
*/
|
|
315
|
-
declare const createFeatherless: (apiKey: string, baseURL?: string) =>
|
|
316
|
-
|
|
317
|
-
/** @deprecated use `createFeatherless` instead. */
|
|
318
|
-
declare const createFatherless: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<string>;
|
|
319
|
-
/** @deprecated use `createAlibabaCn` instead. */
|
|
320
|
-
declare const createQwen: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
|
|
321
|
-
/** @deprecated use `createXai` instead. */
|
|
322
|
-
declare const createXAI: (apiKey: string, baseURL?: string) => _xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.ChatProvider<"grok-4-fast-non-reasoning" | "grok-3-fast" | "grok-4" | "grok-2-vision" | "grok-code-fast-1" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-3" | "grok-4-fast" | "grok-2-latest" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-3-mini" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
|
|
321
|
+
declare const createFeatherless: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<string>;
|
|
323
322
|
|
|
324
323
|
/**
|
|
325
324
|
* Create a Together AI Provider
|
|
326
325
|
* @see {@link https://docs.together.ai/docs/serverless-models}
|
|
327
326
|
*/
|
|
328
|
-
declare const createTogetherAI: (apiKey: string, baseURL?: string) =>
|
|
329
|
-
|
|
330
|
-
/**
|
|
331
|
-
* Create a Workers AI Provider
|
|
332
|
-
* @see {@link https://developers.cloudflare.com/workers-ai}
|
|
333
|
-
*/
|
|
334
|
-
declare const createWorkersAI: (apiKey: string, accountId: string) => _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<CloudflareWorkersAiModels>;
|
|
327
|
+
declare const createTogetherAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<TogetheraiModels>;
|
|
335
328
|
|
|
336
|
-
export { createAlibaba, createAlibabaCn,
|
|
329
|
+
export { createAlibaba, createAlibabaCn, createBailing, createBaseten, createCerebras, createChutes, createCortecs, createDeepSeek, createDeepinfra, createFastrouter, createFeatherless, createFireworks, createGithubCopilot, createGithubModels, createGoogleGenerativeAI, createGroq, createHelicone, createHuggingface, createIflowcn, createInception, createInference, createIoNet, createKimiForCoding, createLitellm, createLlama, createLmstudio, createLucidquery, createMinimax, createMinimaxCn, createMistral, createModelscope, createMoonshotai, createMoonshotaiCn, createMorph, createNebius, createNovita, createNvidia, createOllama, createOllamaCloud, createOpenAI, createOpencode, createOvhcloud, createPerplexity, createPoe, createRequesty, createScaleway, createSiliconFlow, createSiliconflowCn, createStepfun, createSubmodel, createSynthetic, createTencentHunyuan, createTogetherAI, createUpstage, createVenice, createVultr, createWandb, createXai, createXiaomi, createZai, createZaiCodingPlan, createZenmux, createZhipuai, createZhipuaiCodingPlan };
|