@xsai-ext/providers 0.4.0-beta.11 → 0.4.0-beta.13
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/anthropic-Bel15KQV.d.ts +111 -0
- package/dist/create.d.ts +73 -122
- package/dist/create.js +4 -37
- package/dist/index.d.ts +72 -70
- package/dist/index.js +2 -2
- package/dist/player2-D788vRq_.js +47 -0
- package/dist/special/create.d.ts +66 -0
- package/dist/special/create.js +37 -0
- package/dist/special/index.d.ts +12 -0
- package/dist/special/index.js +6 -0
- package/dist/{together-ai-BNlmR9ei.js → together-ai-D5anbRie.js} +2 -2
- package/dist/{anthropic-B6NN_uBS.d.ts → types-Czo87gCh.d.ts} +1 -111
- package/dist/types-DW1hvH0W.d.ts +38 -0
- package/dist/utils/index.d.ts +52 -0
- package/dist/utils/index.js +37 -0
- package/package.json +16 -4
package/dist/create.d.ts
CHANGED
|
@@ -1,387 +1,338 @@
|
|
|
1
|
-
import
|
|
2
|
-
|
|
3
|
-
|
|
1
|
+
import { C as ChatProvider, M as ModelProvider, E as EmbedProvider, I as ImageProvider, S as SpeechProvider, T as TranscriptionProvider } from './types-DW1hvH0W.js';
|
|
2
|
+
export { c as createAnthropic, a as createOpenRouter } from './anthropic-Bel15KQV.js';
|
|
3
|
+
import { T as TogetheraiModels } from './types-Czo87gCh.js';
|
|
4
|
+
import '@xsai/shared';
|
|
4
5
|
|
|
5
6
|
/**
|
|
6
7
|
* Create a Alibaba Provider
|
|
7
8
|
* @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
|
|
8
9
|
*/
|
|
9
|
-
declare const createAlibaba: (apiKey: string, baseURL?: string) =>
|
|
10
|
+
declare const createAlibaba: (apiKey: string, baseURL?: string) => ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus"> & ModelProvider;
|
|
10
11
|
/**
|
|
11
12
|
* Create a Alibaba (China) Provider
|
|
12
13
|
* @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
|
|
13
14
|
*/
|
|
14
|
-
declare const createAlibabaCn: (apiKey: string, baseURL?: string) =>
|
|
15
|
+
declare const createAlibabaCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
|
|
15
16
|
/**
|
|
16
17
|
* Create a Bailing Provider
|
|
17
18
|
* @see {@link https://alipaytbox.yuque.com/sxs0ba/ling/intro}
|
|
18
19
|
*/
|
|
19
|
-
declare const createBailing: (apiKey: string, baseURL?: string) =>
|
|
20
|
+
declare const createBailing: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Ling-1T" | "Ring-1T">;
|
|
20
21
|
/**
|
|
21
22
|
* Create a Baseten Provider
|
|
22
23
|
* @see {@link https://docs.baseten.co/development/model-apis/overview}
|
|
23
24
|
*/
|
|
24
|
-
declare const createBaseten: (apiKey: string, baseURL?: string) =>
|
|
25
|
+
declare const createBaseten: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2">;
|
|
25
26
|
/**
|
|
26
27
|
* Create a Cerebras Provider
|
|
27
28
|
* @see {@link https://inference-docs.cerebras.ai/models/overview}
|
|
28
29
|
*/
|
|
29
|
-
declare const createCerebras: (apiKey: string, baseURL?: string) =>
|
|
30
|
+
declare const createCerebras: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen-3-235b-a22b-instruct-2507" | "zai-glm-4.6" | "gpt-oss-120b"> & EmbedProvider<string>;
|
|
30
31
|
/**
|
|
31
32
|
* Create a Chutes Provider
|
|
32
33
|
* @see {@link https://llm.chutes.ai/v1/models}
|
|
33
34
|
*/
|
|
34
|
-
declare const createChutes: (apiKey: string, baseURL?: string) =>
|
|
35
|
+
declare const createChutes: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "deepseek-ai/DeepSeek-V3.2" | "NousResearch/Hermes-4.3-36B" | "NousResearch/Hermes-4-70B" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4-405B-FP8" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "rednote-hilab/dots.ocr" | "MiniMaxAI/MiniMax-M2" | "ArliAI/QwQ-32B-ArliAI-RpR-v1" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "tngtech/TNG-R1T-Chimera-TEE" | "OpenGVLab/InternVL3-78B" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "Alibaba-NLP/Tongyi-DeepResearch-30B-A3B" | "mistralai/Devstral-2-123B-Instruct-2512" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "unsloth/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen3-32B" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-R1-0528-Qwen3-8B" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-V3" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324">;
|
|
35
36
|
/**
|
|
36
37
|
* Create a Cortecs Provider
|
|
37
38
|
* @see {@link https://api.cortecs.ai/v1/models}
|
|
38
39
|
*/
|
|
39
|
-
declare const createCortecs: (apiKey: string, baseURL?: string) =>
|
|
40
|
+
declare const createCortecs: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "gpt-4.1" | "gemini-2.5-pro" | "devstral-small-2512" | "claude-sonnet-4" | "llama-3.1-405b-instruct">;
|
|
40
41
|
/**
|
|
41
42
|
* Create a Deep Infra Provider
|
|
42
43
|
* @see {@link https://deepinfra.com/models}
|
|
43
44
|
*/
|
|
44
|
-
declare const createDeepinfra: (apiKey: string, baseURL?: string) =>
|
|
45
|
+
declare const createDeepinfra: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "zai-org/GLM-4.5" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo">;
|
|
45
46
|
/**
|
|
46
47
|
* Create a DeepSeek Provider
|
|
47
48
|
* @see {@link https://platform.deepseek.com/api-docs/pricing}
|
|
48
49
|
*/
|
|
49
|
-
declare const createDeepSeek: (apiKey: string, baseURL?: string) =>
|
|
50
|
+
declare const createDeepSeek: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"deepseek-chat" | "deepseek-reasoner">;
|
|
50
51
|
/**
|
|
51
52
|
* Create a FastRouter Provider
|
|
52
53
|
* @see {@link https://fastrouter.ai/models}
|
|
53
54
|
*/
|
|
54
|
-
declare const createFastrouter: (apiKey: string, baseURL?: string) =>
|
|
55
|
+
declare const createFastrouter: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
|
|
55
56
|
/**
|
|
56
57
|
* Create a Fireworks AI Provider
|
|
57
58
|
* @see {@link https://fireworks.ai/docs/}
|
|
58
59
|
*/
|
|
59
|
-
declare const createFireworks: (apiKey: string, baseURL?: string) =>
|
|
60
|
+
declare const createFireworks: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
|
|
60
61
|
/**
|
|
61
62
|
* Create a GitHub Copilot Provider
|
|
62
63
|
* @see {@link https://docs.github.com/en/copilot}
|
|
63
64
|
*/
|
|
64
|
-
declare const createGithubCopilot: (apiKey: string, baseURL?: string) =>
|
|
65
|
+
declare const createGithubCopilot: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "gemini-2.0-flash-001" | "claude-opus-4" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gpt-5.1-codex" | "claude-haiku-4.5" | "gemini-3-pro-preview" | "oswe-vscode-prime" | "claude-3.5-sonnet" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "claude-opus-41" | "gpt-5-mini" | "claude-3.7-sonnet" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "claude-3.7-sonnet-thought" | "claude-opus-4.5" | "gpt-5.2" | "claude-sonnet-4.5">;
|
|
65
66
|
/**
|
|
66
67
|
* Create a GitHub Models Provider
|
|
67
68
|
* @see {@link https://docs.github.com/en/github-models}
|
|
68
69
|
*/
|
|
69
|
-
declare const createGithubModels: (apiKey: string, baseURL?: string) =>
|
|
70
|
+
declare const createGithubModels: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-4.1" | "core42/jais-30b-chat" | "xai/grok-3" | "xai/grok-3-mini" | "cohere/cohere-command-r-08-2024" | "cohere/cohere-command-a" | "cohere/cohere-command-r-plus-08-2024" | "cohere/cohere-command-r" | "cohere/cohere-command-r-plus" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-r1" | "deepseek/deepseek-v3-0324" | "mistral-ai/mistral-medium-2505" | "mistral-ai/ministral-3b" | "mistral-ai/mistral-nemo" | "mistral-ai/mistral-large-2411" | "mistral-ai/codestral-2501" | "mistral-ai/mistral-small-2503" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-mini-4k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-4" | "microsoft/phi-4-mini-reasoning" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-mini-instruct" | "microsoft/phi-4-multimodal-instruct" | "microsoft/phi-3-mini-128k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-mini-instruct" | "microsoft/phi-3-medium-4k-instruct" | "microsoft/phi-4-reasoning" | "microsoft/mai-ds-r1" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o1-preview" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o1-mini" | "openai/o3" | "openai/gpt-4o-mini" | "meta/llama-3.2-11b-vision-instruct" | "meta/meta-llama-3.1-405b-instruct" | "meta/llama-4-maverick-17b-128e-instruct-fp8" | "meta/meta-llama-3-70b-instruct" | "meta/meta-llama-3.1-70b-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-3.2-90b-vision-instruct" | "meta/meta-llama-3-8b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "meta/meta-llama-3.1-8b-instruct" | "ai21-labs/ai21-jamba-1.5-large" | "ai21-labs/ai21-jamba-1.5-mini">;
|
|
70
71
|
/**
|
|
71
72
|
* Create a Google Provider
|
|
72
73
|
* @see {@link https://ai.google.dev/gemini-api/docs/pricing}
|
|
73
74
|
*/
|
|
74
|
-
declare const createGoogleGenerativeAI: (apiKey: string, baseURL?: string) =>
|
|
75
|
+
declare const createGoogleGenerativeAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-2.5-pro" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-2.5-flash" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
|
|
75
76
|
/**
|
|
76
77
|
* Create a Groq Provider
|
|
77
78
|
* @see {@link https://console.groq.com/docs/models}
|
|
78
79
|
*/
|
|
79
|
-
declare const createGroq: (apiKey: string, baseURL?: string) =>
|
|
80
|
+
declare const createGroq: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b">;
|
|
80
81
|
/**
|
|
81
82
|
* Create a Helicone Provider
|
|
82
83
|
* @see {@link https://helicone.ai/models}
|
|
83
84
|
*/
|
|
84
|
-
declare const createHelicone: (apiKey: string, baseURL?: string) =>
|
|
85
|
+
declare const createHelicone: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-next-80b-a3b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-thinking" | "gpt-4.1" | "gemini-2.5-pro" | "claude-sonnet-4" | "deepseek-reasoner" | "claude-opus-4" | "grok-code-fast-1" | "gpt-5.1-codex" | "gemini-3-pro-preview" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "claude-3.7-sonnet" | "o3" | "gpt-5" | "gemini-2.5-flash" | "gemini-2.5-flash-lite" | "llama-3.1-8b-instant" | "gemma2-9b-it" | "llama-3.3-70b-versatile" | "gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "qwen3-coder" | "grok-4-fast-reasoning" | "claude-opus-4-1" | "grok-4" | "llama-4-maverick" | "llama-prompt-guard-2-86m" | "grok-4-1-fast-reasoning" | "claude-4.5-haiku" | "llama-3.1-8b-instruct-turbo" | "gpt-4.1-mini-2025-04-14" | "llama-guard-4" | "llama-3.1-8b-instruct" | "gpt-4.1-mini" | "deepseek-v3.1-terminus" | "llama-prompt-guard-2-22m" | "claude-3.5-sonnet-v2" | "sonar-deep-research" | "claude-sonnet-4-5-20250929" | "grok-3" | "mistral-small" | "kimi-k2-0711" | "chatgpt-4o-latest" | "kimi-k2-0905" | "sonar-reasoning" | "llama-3.3-70b-instruct" | "claude-4.5-sonnet" | "codex-mini-latest" | "gpt-5-nano" | "deepseek-tng-r1t2-chimera" | "claude-4.5-opus" | "sonar" | "glm-4.6" | "qwen3-235b-a22b-thinking" | "hermes-2-pro-llama-3-8b" | "o1" | "grok-3-mini" | "sonar-pro" | "o1-mini" | "claude-3-haiku-20240307" | "o3-pro" | "qwen2.5-coder-7b-fast" | "gemma-3-12b-it" | "mistral-nemo" | "gpt-oss-20b" | "claude-3.5-haiku" | "gpt-5-chat-latest" | "gpt-4o-mini" | "sonar-reasoning-pro" | "qwen3-vl-235b-a22b-instruct" | "qwen3-30b-a3b" | "deepseek-v3.2" | "grok-4-1-fast-non-reasoning" | "gpt-5-pro" | "mistral-large-2411" | "claude-opus-4-1-20250805" | "ernie-4.5-21b-a3b-thinking" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "llama-4-scout">;
|
|
85
86
|
/**
|
|
86
87
|
* Create a Hugging Face Provider
|
|
87
88
|
* @see {@link https://huggingface.co/docs/inference-providers}
|
|
88
89
|
*/
|
|
89
|
-
declare const createHuggingface: (apiKey: string, baseURL?: string) =>
|
|
90
|
+
declare const createHuggingface: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "MiniMaxAI/MiniMax-M2" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.5" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "moonshotai/Kimi-K2-Instruct" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/Deepseek-V3-0324">;
|
|
90
91
|
/**
|
|
91
92
|
* Create a iFlow Provider
|
|
92
93
|
* @see {@link https://platform.iflow.cn/en/docs}
|
|
93
94
|
*/
|
|
94
|
-
declare const createIflowcn: (apiKey: string, baseURL?: string) =>
|
|
95
|
+
declare const createIflowcn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-vl-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "kimi-k2-thinking" | "qwen3-coder" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "kimi-k2" | "deepseek-v3.1" | "minimax-m2" | "qwen3-235b" | "deepseek-v3.2-chat" | "qwen3-235b-a22b-thinking-2507" | "tstars2.0" | "qwen3-235b-a22b-instruct" | "qwen3-max-preview">;
|
|
95
96
|
/**
|
|
96
97
|
* Create a Inception Provider
|
|
97
98
|
* @see {@link https://platform.inceptionlabs.ai/docs}
|
|
98
99
|
*/
|
|
99
|
-
declare const createInception: (apiKey: string, baseURL?: string) =>
|
|
100
|
+
declare const createInception: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"mercury-coder" | "mercury">;
|
|
100
101
|
/**
|
|
101
102
|
* Create a Inference Provider
|
|
102
103
|
* @see {@link https://inference.net/models}
|
|
103
104
|
*/
|
|
104
|
-
declare const createInference: (apiKey: string, baseURL?: string) =>
|
|
105
|
+
declare const createInference: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"meta/llama-3.2-11b-vision-instruct" | "mistral/mistral-nemo-12b-instruct" | "google/gemma-3" | "osmosis/osmosis-structure-0.6b" | "qwen/qwen3-embedding-4b" | "qwen/qwen-2.5-7b-vision-instruct" | "meta/llama-3.1-8b-instruct" | "meta/llama-3.2-3b-instruct" | "meta/llama-3.2-1b-instruct">;
|
|
105
106
|
/**
|
|
106
107
|
* Create a IO.NET Provider
|
|
107
108
|
* @see {@link https://io.net/docs/guides/intelligence/io-intelligence}
|
|
108
109
|
*/
|
|
109
|
-
declare const createIoNet: (apiKey: string, baseURL?: string) =>
|
|
110
|
+
declare const createIoNet: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506" | "mistralai/Mistral-Large-Instruct-2411" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar">;
|
|
110
111
|
/**
|
|
111
112
|
* Create a Kimi For Coding Provider
|
|
112
113
|
* @see {@link https://www.kimi.com/coding/docs/en/third-party-agents.html}
|
|
113
114
|
*/
|
|
114
|
-
declare const createKimiForCoding: (apiKey: string, baseURL?: string) =>
|
|
115
|
+
declare const createKimiForCoding: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking">;
|
|
115
116
|
/**
|
|
116
117
|
* Create a Llama Provider
|
|
117
118
|
* @see {@link https://llama.developer.meta.com/docs/models}
|
|
118
119
|
*/
|
|
119
|
-
declare const createLlama: (apiKey: string, baseURL?: string) =>
|
|
120
|
+
declare const createLlama: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"llama-3.3-70b-instruct" | "llama-3.3-8b-instruct" | "llama-4-maverick-17b-128e-instruct-fp8" | "llama-4-scout-17b-16e-instruct-fp8" | "groq-llama-4-maverick-17b-128e-instruct" | "cerebras-llama-4-scout-17b-16e-instruct" | "cerebras-llama-4-maverick-17b-128e-instruct">;
|
|
120
121
|
/**
|
|
121
122
|
* Create a LMStudio Provider
|
|
122
123
|
* @see {@link https://lmstudio.ai/models}
|
|
123
124
|
*/
|
|
124
|
-
declare const createLmstudio: (apiKey: string, baseURL?: string) =>
|
|
125
|
+
declare const createLmstudio: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "qwen/qwen3-30b-a3b-2507" | "qwen/qwen3-coder-30b">;
|
|
125
126
|
/**
|
|
126
127
|
* Create a LucidQuery AI Provider
|
|
127
128
|
* @see {@link https://lucidquery.com/api/docs}
|
|
128
129
|
*/
|
|
129
|
-
declare const createLucidquery: (apiKey: string, baseURL?: string) =>
|
|
130
|
+
declare const createLucidquery: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"lucidquery-nexus-coder" | "lucidnova-rf1-100b">;
|
|
130
131
|
/**
|
|
131
132
|
* Create a MiniMax Provider
|
|
132
133
|
* @see {@link https://platform.minimax.io/docs/guides/quickstart}
|
|
133
134
|
*/
|
|
134
|
-
declare const createMinimax: (apiKey: string, baseURL?: string) =>
|
|
135
|
+
declare const createMinimax: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
135
136
|
/**
|
|
136
137
|
* Create a MiniMax (China) Provider
|
|
137
138
|
* @see {@link https://platform.minimaxi.com/docs/guides/quickstart}
|
|
138
139
|
*/
|
|
139
|
-
declare const createMinimaxCn: (apiKey: string, baseURL?: string) =>
|
|
140
|
+
declare const createMinimaxCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
140
141
|
/**
|
|
141
142
|
* Create a Mistral Provider
|
|
142
143
|
* @see {@link https://docs.mistral.ai/getting-started/models/}
|
|
143
144
|
*/
|
|
144
|
-
declare const createMistral: (apiKey: string, baseURL?: string) =>
|
|
145
|
+
declare const createMistral: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"devstral-2512" | "mistral-nemo" | "mistral-large-2411" | "devstral-medium-2507" | "mistral-large-2512" | "open-mixtral-8x22b" | "ministral-8b-latest" | "pixtral-large-latest" | "mistral-small-2506" | "ministral-3b-latest" | "pixtral-12b" | "mistral-medium-2505" | "labs-devstral-small-2512" | "devstral-medium-latest" | "devstral-small-2505" | "mistral-medium-2508" | "mistral-embed" | "mistral-small-latest" | "magistral-small" | "devstral-small-2507" | "codestral-latest" | "open-mixtral-8x7b" | "open-mistral-7b" | "mistral-large-latest" | "mistral-medium-latest" | "magistral-medium-latest">;
|
|
145
146
|
/**
|
|
146
147
|
* Create a ModelScope Provider
|
|
147
148
|
* @see {@link https://modelscope.cn/docs/model-service/API-Inference/intro}
|
|
148
149
|
*/
|
|
149
|
-
declare const createModelscope: (apiKey: string, baseURL?: string) =>
|
|
150
|
+
declare const createModelscope: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507">;
|
|
150
151
|
/**
|
|
151
152
|
* Create a Moonshot AI Provider
|
|
152
153
|
* @see {@link https://platform.moonshot.ai/docs/api/chat}
|
|
153
154
|
*/
|
|
154
|
-
declare const createMoonshotai: (apiKey: string, baseURL?: string) =>
|
|
155
|
+
declare const createMoonshotai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
|
|
155
156
|
/**
|
|
156
157
|
* Create a Moonshot AI (China) Provider
|
|
157
158
|
* @see {@link https://platform.moonshot.cn/docs/api/chat}
|
|
158
159
|
*/
|
|
159
|
-
declare const createMoonshotaiCn: (apiKey: string, baseURL?: string) =>
|
|
160
|
+
declare const createMoonshotaiCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-turbo-preview" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
|
|
160
161
|
/**
|
|
161
162
|
* Create a Morph Provider
|
|
162
163
|
* @see {@link https://docs.morphllm.com/api-reference/introduction}
|
|
163
164
|
*/
|
|
164
|
-
declare const createMorph: (apiKey: string, baseURL?: string) =>
|
|
165
|
+
declare const createMorph: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
|
|
165
166
|
/**
|
|
166
167
|
* Create a Nebius Token Factory Provider
|
|
167
168
|
* @see {@link https://docs.tokenfactory.nebius.com/}
|
|
168
169
|
*/
|
|
169
|
-
declare const createNebius: (apiKey: string, baseURL?: string) =>
|
|
170
|
+
declare const createNebius: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-20b" | "openai/gpt-oss-120b" | "moonshotai/kimi-k2-instruct" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-235b-a22b-thinking-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "zai-org/glm-4.5" | "zai-org/glm-4.5-air" | "deepseek-ai/deepseek-v3">;
|
|
170
171
|
/**
|
|
171
172
|
* Create a Nvidia Provider
|
|
172
173
|
* @see {@link https://docs.api.nvidia.com/nim/}
|
|
173
174
|
*/
|
|
174
|
-
declare const createNvidia: (apiKey: string, baseURL?: string) =>
|
|
175
|
+
declare const createNvidia: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-4-mini-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-coder-480b-a35b-instruct" | "moonshotai/kimi-k2-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "minimaxai/minimax-m2" | "google/gemma-3-27b-it" | "openai/whisper-large-v3" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwen3-next-80b-a3b-thinking" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "black-forest-labs/flux.1-dev">;
|
|
175
176
|
/**
|
|
176
177
|
* Create a Ollama Cloud Provider
|
|
177
178
|
* @see {@link https://docs.ollama.com/cloud}
|
|
178
179
|
*/
|
|
179
|
-
declare const createOllamaCloud: (apiKey: string, baseURL?: string) =>
|
|
180
|
+
declare const createOllamaCloud: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking:cloud" | "qwen3-vl-235b-cloud" | "qwen3-coder:480b-cloud" | "gpt-oss:120b-cloud" | "deepseek-v3.1:671b-cloud" | "glm-4.6:cloud" | "cogito-2.1:671b-cloud" | "gpt-oss:20b-cloud" | "qwen3-vl-235b-instruct-cloud" | "kimi-k2:1t-cloud" | "minimax-m2:cloud" | "gemini-3-pro-preview:latest">;
|
|
180
181
|
/**
|
|
181
182
|
* Create a OpenAI Provider
|
|
182
183
|
* @see {@link https://platform.openai.com/docs/models}
|
|
183
184
|
*/
|
|
184
|
-
declare const createOpenAI: (apiKey: string, baseURL?: string) =>
|
|
185
|
+
declare const createOpenAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "o3-mini" | "gpt-5.1" | "gpt-5-codex" | "gpt-4o" | "o4-mini" | "gpt-5-mini" | "gpt-5.1-codex-max" | "o3" | "gpt-5" | "gpt-5.2" | "gpt-4.1-nano" | "gpt-4.1-mini" | "codex-mini-latest" | "gpt-5-nano" | "o1" | "o1-mini" | "o3-pro" | "gpt-5-chat-latest" | "gpt-4o-mini" | "gpt-5-pro" | "gpt-5.1-chat-latest" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "o3-deep-research" | "gpt-3.5-turbo" | "gpt-5.2-pro" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "gpt-5.2-chat-latest" | "text-embedding-ada-002" | "gpt-4o-2024-11-20" | "o4-mini-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
|
|
185
186
|
/**
|
|
186
187
|
* Create a OpenCode Zen Provider
|
|
187
188
|
* @see {@link https://opencode.ai/docs/zen}
|
|
188
189
|
*/
|
|
189
|
-
declare const createOpencode: (apiKey: string, baseURL?: string) =>
|
|
190
|
+
declare const createOpencode: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking" | "claude-sonnet-4" | "gpt-5.1-codex" | "gpt-5.1" | "gpt-5-codex" | "gpt-5.1-codex-max" | "gpt-5" | "gpt-5.2" | "qwen3-coder" | "claude-opus-4-1" | "gpt-5-nano" | "glm-4.6" | "kimi-k2" | "claude-haiku-4-5" | "claude-opus-4-5" | "gemini-3-pro" | "claude-sonnet-4-5" | "alpha-gd4" | "big-pickle" | "claude-3-5-haiku" | "grok-code" | "gemini-3-flash" | "alpha-doubao-seed-code" | "minimax-m2.1">;
|
|
190
191
|
/**
|
|
191
192
|
* Create a OVHcloud AI Endpoints Provider
|
|
192
193
|
* @see {@link https://www.ovhcloud.com/en/public-cloud/ai-endpoints/catalog//}
|
|
193
194
|
*/
|
|
194
|
-
declare const createOvhcloud: (apiKey: string, baseURL?: string) =>
|
|
195
|
+
declare const createOvhcloud: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "gpt-oss-20b" | "mixtral-8x7b-instruct-v0.1" | "mistral-7b-instruct-v0.3" | "qwen2.5-vl-72b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen2.5-coder-32b-instruct" | "llava-next-mistral-7b" | "meta-llama-3_1-70b-instruct" | "meta-llama-3_3-70b-instruct">;
|
|
195
196
|
/**
|
|
196
197
|
* Create a Perplexity Provider
|
|
197
198
|
* @see {@link https://docs.perplexity.ai}
|
|
198
199
|
*/
|
|
199
|
-
declare const createPerplexity: (apiKey: string, baseURL?: string) =>
|
|
200
|
+
declare const createPerplexity: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"sonar-reasoning" | "sonar" | "sonar-pro" | "sonar-reasoning-pro">;
|
|
200
201
|
/**
|
|
201
202
|
* Create a Poe Provider
|
|
202
203
|
* @see {@link https://creator.poe.com/docs/external-applications/openai-compatible-api}
|
|
203
204
|
*/
|
|
204
|
-
declare const createPoe: (apiKey: string, baseURL?: string) =>
|
|
205
|
+
declare const createPoe: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-2" | "xai/grok-4.1-fast-non-reasoning" | "ideogramai/ideogram" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram-v2" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "poetools/claude-code" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "google/gemini-deep-research" | "google/nano-banana" | "google/imagen-4" | "google/imagen-3" | "google/imagen-4-ultra" | "google/gemini-2.0-flash-lite" | "google/gemini-3-pro" | "google/veo-3.1" | "google/imagen-3-fast" | "google/lyria" | "google/gemini-2.0-flash" | "google/gemini-2.5-flash-lite" | "google/veo-3" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-2" | "google/nano-banana-pro" | "google/veo-3.1-fast" | "openai/gpt-5.2-instant" | "openai/sora-2" | "openai/o1-pro" | "openai/gpt-5.1-codex" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-4-classic" | "openai/gpt-5-chat" | "openai/o3-deep-research" | "openai/gpt-4o-search" | "openai/gpt-image-1-mini" | "openai/gpt-3.5-turbo" | "openai/gpt-5.2-pro" | "openai/o3-mini-high" | "openai/chatgpt-4o-latest" | "openai/gpt-4-turbo" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1-instant" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-4o-aug" | "openai/o3-pro" | "openai/gpt-image-1" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo-instruct" | "openai/o4-mini-deep-research" | "openai/gpt-4-classic-0314" | "openai/dall-e-3" | "openai/sora-2-pro" | "openai/gpt-5-pro" | "openai/gpt-5.2" | "openai/gpt-4o-mini-search" | "stabilityai/stablediffusionxl" | "topazlabs-co/topazlabs" | "lumalabs/ray2" | "lumalabs/dream-machine" | "anthropic/claude-opus-3" | "anthropic/claude-opus-4" | "anthropic/claude-sonnet-3.7-reasoning" | "anthropic/claude-opus-4-search" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-haiku-3.5-search" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4-reasoning" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7-search" | "anthropic/claude-opus-4-reasoning" | "anthropic/claude-sonnet-3.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-haiku-3.5" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-sonnet-4-search" | "trytako/tako" | "novita/kimi-k2-thinking" | "novita/glm-4.6">;
|
|
205
206
|
/**
|
|
206
207
|
* Create a Requesty Provider
|
|
207
208
|
* @see {@link https://requesty.ai/solution/llm-routing/models}
|
|
208
209
|
*/
|
|
209
|
-
declare const createRequesty: (apiKey: string, baseURL?: string) =>
|
|
210
|
+
declare const createRequesty: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-sonnet-4" | "openai/gpt-4.1-mini" | "openai/o4-mini" | "openai/gpt-4o-mini" | "xai/grok-4" | "anthropic/claude-opus-4" | "xai/grok-4-fast" | "google/gemini-3-flash-preview" | "google/gemini-3-pro-preview" | "anthropic/claude-opus-4-1" | "anthropic/claude-haiku-4-5" | "anthropic/claude-opus-4-5" | "anthropic/claude-sonnet-4-5" | "anthropic/claude-3-7-sonnet">;
|
|
210
211
|
/**
|
|
211
212
|
* Create a Scaleway Provider
|
|
212
213
|
* @see {@link https://www.scaleway.com/en/docs/generative-apis/}
|
|
213
214
|
*/
|
|
214
|
-
declare const createScaleway: (apiKey: string, baseURL?: string) =>
|
|
215
|
+
declare const createScaleway: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen3-235b-a22b-instruct-2507" | "pixtral-12b-2409" | "whisper-large-v3" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
|
|
215
216
|
/**
|
|
216
217
|
* Create a SiliconFlow Provider
|
|
217
218
|
* @see {@link https://cloud.siliconflow.com/models}
|
|
218
219
|
*/
|
|
219
|
-
declare const createSiliconFlow: (apiKey: string, baseURL?: string) =>
|
|
220
|
+
declare const createSiliconFlow: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
|
|
220
221
|
/**
|
|
221
222
|
* Create a SiliconFlow (China) Provider
|
|
222
223
|
* @see {@link https://cloud.siliconflow.com/models}
|
|
223
224
|
*/
|
|
224
|
-
declare const createSiliconflowCn: (apiKey: string, baseURL?: string) =>
|
|
225
|
+
declare const createSiliconflowCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen-qwq-32b" | "deepseek-ai-deepseek-r1-distill-qwen-7b" | "z-ai-glm-4.5-air" | "qwen-qwen2.5-72b-instruct-128k" | "deepseek-ai-deepseek-vl2" | "moonshotai-kimi-dev-72b" | "qwen-qwen2.5-coder-32b-instruct" | "qwen-qwen3-omni-30b-a3b-captioner" | "qwen-qwen3-vl-235b-a22b-thinking" | "thudm-glm-z1-9b-0414" | "qwen-qwen3-vl-30b-a3b-thinking" | "deepseek-ai-deepseek-v3.2-exp" | "qwen-qwen2.5-vl-32b-instruct" | "qwen-qwen3-235b-a22b-thinking-2507" | "qwen-qwen3-vl-32b-instruct" | "inclusionai-ling-flash-2.0" | "moonshotai-kimi-k2-instruct" | "inclusionai-ling-mini-2.0" | "qwen-qwen3-coder-480b-a35b-instruct" | "qwen-qwen3-omni-30b-a3b-instruct" | "moonshotai-kimi-k2-instruct-0905" | "qwen-qwen3-30b-a3b-thinking-2507" | "qwen-qwen3-14b" | "deepseek-ai-deepseek-r1" | "deepseek-ai-deepseek-v3.1" | "z-ai-glm-4.5" | "qwen-qwen3-30b-a3b-instruct-2507" | "zai-org-glm-4.5v" | "inclusionai-ring-flash-2.0" | "thudm-glm-z1-32b-0414" | "qwen-qwen2.5-vl-72b-instruct" | "qwen-qwen3-vl-32b-thinking" | "tencent-hunyuan-mt-7b" | "qwen-qwen3-30b-a3b" | "openai-gpt-oss-120b" | "minimaxai-minimax-m1-80k" | "deepseek-ai-deepseek-v3.1-terminus" | "zai-org-glm-4.5-air" | "thudm-glm-4-9b-0414" | "qwen-qwen3-coder-30b-a3b-instruct" | "stepfun-ai-step3" | "thudm-glm-4.1v-9b-thinking" | "qwen-qwen3-next-80b-a3b-thinking" | "qwen-qwen3-vl-235b-a22b-instruct" | "zai-org-glm-4.5" | "deepseek-ai-deepseek-r1-distill-qwen-14b" | "deepseek-ai-deepseek-v3" | "openai-gpt-oss-20b" | "qwen-qwen2.5-7b-instruct" | "qwen-qwen2.5-32b-instruct" | "minimaxai-minimax-m2" | "bytedance-seed-seed-oss-36b-instruct" | "qwen-qwen2.5-vl-7b-instruct" | "qwen-qwen3-vl-8b-thinking" | "qwen-qwen3-vl-8b-instruct" | "nex-agi-deepseek-v3.1-nex-n1" | "qwen-qwen3-8b" | "qwen-qwen2.5-72b-instruct" | "qwen-qwen3-235b-a22b" | "meta-llama-meta-llama-3.1-8b-instruct" | "qwen-qwen3-235b-a22b-instruct-2507" | "baidu-ernie-4.5-300b-a47b" | "qwen-qwen3-omni-30b-a3b-thinking" | "zai-org-glm-4.6" | "qwen-qwen3-32b" | "tencent-hunyuan-a13b-instruct" | "thudm-glm-4-32b-0414" | "deepseek-ai-deepseek-r1-distill-qwen-32b" | "qwen-qwen3-next-80b-a3b-instruct" | "qwen-qwen3-vl-30b-a3b-instruct" | "moonshotai-kimi-k2-thinking" | "qwen-qwen2.5-14b-instruct">;
|
|
225
226
|
/**
|
|
226
227
|
* Create a submodel Provider
|
|
227
228
|
* @see {@link https://submodel.gitbook.io}
|
|
228
229
|
*/
|
|
229
|
-
declare const createSubmodel: (apiKey: string, baseURL?: string) =>
|
|
230
|
+
declare const createSubmodel: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "deepseek-ai/DeepSeek-V3-0324" | "zai-org/GLM-4.5-FP8">;
|
|
230
231
|
/**
|
|
231
232
|
* Create a Synthetic Provider
|
|
232
233
|
* @see {@link https://synthetic.new/pricing}
|
|
233
234
|
*/
|
|
234
|
-
declare const createSynthetic: (apiKey: string, baseURL?: string) =>
|
|
235
|
+
declare const createSynthetic: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
|
|
235
236
|
/**
|
|
236
237
|
* Create a Upstage Provider
|
|
237
238
|
* @see {@link https://developers.upstage.ai/docs/apis/chat}
|
|
238
239
|
*/
|
|
239
|
-
declare const createUpstage: (apiKey: string, baseURL?: string) =>
|
|
240
|
+
declare const createUpstage: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"solar-mini" | "solar-pro2">;
|
|
240
241
|
/**
|
|
241
242
|
* Create a Venice AI Provider
|
|
242
243
|
* @see {@link https://docs.venice.ai}
|
|
243
244
|
*/
|
|
244
|
-
declare const createVenice: (apiKey: string, baseURL?: string) =>
|
|
245
|
+
declare const createVenice: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "kimi-k2-thinking" | "gemini-3-flash-preview" | "gemini-3-pro-preview" | "deepseek-v3.2" | "qwen3-235b" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "openai-gpt-oss-120b" | "zai-org-glm-4.6" | "grok-41-fast" | "claude-opus-45" | "mistral-31-24b" | "venice-uncensored" | "openai-gpt-52" | "qwen3-4b" | "llama-3.3-70b" | "devstral-2-2512" | "llama-3.2-3b" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "zai-org-glm-4.6v" | "qwen3-next-80b">;
|
|
245
246
|
/**
|
|
246
247
|
* Create a Vultr Provider
|
|
247
248
|
* @see {@link https://api.vultrinference.com/}
|
|
248
249
|
*/
|
|
249
|
-
declare const createVultr: (apiKey: string, baseURL?: string) =>
|
|
250
|
+
declare const createVultr: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"deepseek-r1-distill-qwen-32b" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-instruct" | "qwen2.5-coder-32b-instruct">;
|
|
250
251
|
/**
|
|
251
252
|
* Create a Weights & Biases Provider
|
|
252
253
|
* @see {@link https://weave-docs.wandb.ai/guides/integrations/inference/}
|
|
253
254
|
*/
|
|
254
|
-
declare const createWandb: (apiKey: string, baseURL?: string) =>
|
|
255
|
+
declare const createWandb: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3-0324" | "moonshotai/Kimi-K2-Instruct" | "meta-llama/Llama-3.3-70B-Instruct" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
|
|
255
256
|
/**
|
|
256
257
|
* Create a xAI Provider
|
|
257
258
|
* @see {@link https://docs.x.ai/docs/models}
|
|
258
259
|
*/
|
|
259
|
-
declare const createXai: (apiKey: string, baseURL?: string) =>
|
|
260
|
+
declare const createXai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"grok-code-fast-1" | "grok-4-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-4-1-fast-non-reasoning" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
|
|
260
261
|
/**
|
|
261
262
|
* Create a Xiaomi Provider
|
|
262
263
|
* @see {@link https://platform.xiaomimimo.com/#/docs}
|
|
263
264
|
*/
|
|
264
|
-
declare const createXiaomi: (apiKey: string, baseURL?: string) =>
|
|
265
|
+
declare const createXiaomi: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"mimo-v2-flash">;
|
|
265
266
|
/**
|
|
266
267
|
* Create a Z.AI Provider
|
|
267
268
|
* @see {@link https://docs.z.ai/guides/overview/pricing}
|
|
268
269
|
*/
|
|
269
|
-
declare const createZai: (apiKey: string, baseURL?: string) =>
|
|
270
|
+
declare const createZai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
270
271
|
/**
|
|
271
272
|
* Create a Z.AI Coding Plan Provider
|
|
272
273
|
* @see {@link https://docs.z.ai/devpack/overview}
|
|
273
274
|
*/
|
|
274
|
-
declare const createZaiCodingPlan: (apiKey: string, baseURL?: string) =>
|
|
275
|
+
declare const createZaiCodingPlan: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
275
276
|
/**
|
|
276
277
|
* Create a ZenMux Provider
|
|
277
278
|
* @see {@link https://docs.zenmux.ai}
|
|
278
279
|
*/
|
|
279
|
-
declare const createZenmux: (apiKey: string, baseURL?: string) =>
|
|
280
|
+
declare const createZenmux: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "moonshotai/kimi-k2-thinking" | "openai/gpt-5-codex" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.5" | "moonshotai/kimi-k2-thinking-turbo" | "moonshotai/kimi-k2-0905" | "xiaomi/mimo-v2-flash" | "x-ai/grok-4-fast-non-reasoning" | "x-ai/grok-code-fast-1" | "x-ai/grok-4-fast" | "deepseek/deepseek-chat" | "minimax/minimax-m2" | "inclusionai/ring-1t" | "inclusionai/lint-1t" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1">;
|
|
280
281
|
/**
|
|
281
282
|
* Create a Zhipu AI Provider
|
|
282
283
|
* @see {@link https://docs.z.ai/guides/overview/pricing}
|
|
283
284
|
*/
|
|
284
|
-
declare const createZhipuai: (apiKey: string, baseURL?: string) =>
|
|
285
|
+
declare const createZhipuai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
285
286
|
/**
|
|
286
287
|
* Create a Zhipu AI Coding Plan Provider
|
|
287
288
|
* @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
|
|
288
289
|
*/
|
|
289
|
-
declare const createZhipuaiCodingPlan: (apiKey: string, baseURL?: string) =>
|
|
290
|
+
declare const createZhipuaiCodingPlan: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
290
291
|
/**
|
|
291
292
|
* Create a Novita AI Provider
|
|
292
293
|
* @see {@link https://novita.ai/docs/guides/llm-api#api-integration}
|
|
293
294
|
*/
|
|
294
|
-
declare const createNovita: (apiKey: string, baseURL?: string) =>
|
|
295
|
+
declare const createNovita: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<string>;
|
|
295
296
|
/**
|
|
296
297
|
* Create a StepFun Provider
|
|
297
298
|
* @see {@link https://www.stepfun.com}
|
|
298
299
|
*/
|
|
299
|
-
declare const createStepfun: (apiKey: string, baseURL?: string) =>
|
|
300
|
+
declare const createStepfun: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
|
|
300
301
|
/**
|
|
301
302
|
* Create a Tencent Hunyuan Provider
|
|
302
303
|
* @see {@link https://cloud.tencent.com/document/product/1729}
|
|
303
304
|
*/
|
|
304
|
-
declare const createTencentHunyuan: (apiKey: string, baseURL?: string) =>
|
|
305
|
+
declare const createTencentHunyuan: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<string>;
|
|
305
306
|
/**
|
|
306
307
|
* Create a Ollama Provider
|
|
307
308
|
* @see {@link https://docs.ollama.com}
|
|
308
309
|
*/
|
|
309
|
-
declare const createOllama: (apiKey: string, baseURL?: string) =>
|
|
310
|
+
declare const createOllama: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<string>;
|
|
310
311
|
/**
|
|
311
312
|
* Create a LiteLLM Provider
|
|
312
313
|
* @see {@link https://docs.litellm.ai}
|
|
313
314
|
*/
|
|
314
|
-
declare const createLitellm: (apiKey: string, baseURL?: string) =>
|
|
315
|
-
|
|
316
|
-
interface CreateAzureOptions {
|
|
317
|
-
/**
|
|
318
|
-
* The static API key or AD access token fetcher for authorization.
|
|
319
|
-
*
|
|
320
|
-
* If passed in as a function, it is treated as an accessTokenFetcher.
|
|
321
|
-
*
|
|
322
|
-
* @see {@link https://learn.microsoft.com/en-us/azure/api-management/api-management-authenticate-authorize-azure-openai}
|
|
323
|
-
*/
|
|
324
|
-
apiKey: (() => Promise<string> | string) | string;
|
|
325
|
-
/**
|
|
326
|
-
* The Azure API version to use (`api-version` param).
|
|
327
|
-
*
|
|
328
|
-
* Notice: Different deployment over different time may have different API versions, please
|
|
329
|
-
* follow the exact prompt from either [Azure AI Foundry](https://ai.azure.com/) or Azure OpenAI service
|
|
330
|
-
* to get the correct API version from the Azure OpenAI Service endpoint.
|
|
331
|
-
*
|
|
332
|
-
* On Azure AI Foundry portal, you can go to https://ai.azure.com/build/overview > Choose the project >
|
|
333
|
-
* Overview > Endpoints and keys > Included capabilities > Azure OpenAI Service to get the correct endpoint.
|
|
334
|
-
*
|
|
335
|
-
* @see {@link https://learn.microsoft.com/en-us/azure/ai-services/openai/reference#rest-api-versioning}
|
|
336
|
-
*/
|
|
337
|
-
apiVersion?: string;
|
|
338
|
-
/**
|
|
339
|
-
* Azure resource name.
|
|
340
|
-
*
|
|
341
|
-
* On Azure AI Foundry portal, you can go to https://ai.azure.com/build/overview > Choose the project >
|
|
342
|
-
* Overview > Endpoints and keys > Included capabilities > Azure OpenAI Service to get the correct endpoint.
|
|
343
|
-
*
|
|
344
|
-
* @see {@link https://learn.microsoft.com/en-us/azure/ai-services/openai/reference#uri-parameters}
|
|
345
|
-
*/
|
|
346
|
-
resourceName: string;
|
|
347
|
-
}
|
|
348
|
-
/**
|
|
349
|
-
* Create a Azure Provider
|
|
350
|
-
* @see {@link https://ai.azure.com/explore/models}
|
|
351
|
-
* @see {@link https://learn.microsoft.com/en-us/azure/ai-foundry/model-inference/concepts/endpoints?tabs=rest#routing}
|
|
352
|
-
* @remarks
|
|
353
|
-
* For Azure AI services, you can have multiple deployments of the same model with different names.
|
|
354
|
-
*
|
|
355
|
-
* Please pass your deployment name as the `model` parameter. By default, Azure will use the model name
|
|
356
|
-
* as the deployment name when deploying a model.
|
|
357
|
-
*/
|
|
358
|
-
declare const createAzure: (options: CreateAzureOptions) => Promise<_xsai_ext_shared_providers.ModelProvider & _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.SpeechProvider<string> & _xsai_ext_shared_providers.TranscriptionProvider<string> & _xsai_ext_shared_providers.ChatProvider<AzureModels>>;
|
|
315
|
+
declare const createLitellm: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<string>;
|
|
359
316
|
|
|
360
317
|
/**
|
|
361
318
|
* Create a Featherless AI Provider
|
|
362
319
|
* @see {@link https://featherless.ai/models}
|
|
363
320
|
*/
|
|
364
|
-
declare const createFeatherless: (apiKey: string, baseURL?: string) =>
|
|
321
|
+
declare const createFeatherless: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<string>;
|
|
365
322
|
|
|
366
323
|
/** @deprecated use `createFeatherless` instead. */
|
|
367
|
-
declare const createFatherless: (apiKey: string, baseURL?: string) =>
|
|
324
|
+
declare const createFatherless: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<string>;
|
|
368
325
|
/** @deprecated use `createAlibabaCn` instead. */
|
|
369
|
-
declare const createQwen: (apiKey: string, baseURL?: string) =>
|
|
326
|
+
declare const createQwen: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-max" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
|
|
370
327
|
/** @deprecated use `createXai` instead. */
|
|
371
|
-
declare const createXAI: (apiKey: string, baseURL?: string) =>
|
|
328
|
+
declare const createXAI: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"grok-code-fast-1" | "grok-4-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-4-1-fast-non-reasoning" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
|
|
372
329
|
/** @deprecated use `createMinimaxCn` instead. */
|
|
373
|
-
declare const createMinimaxi: (apiKey: string, baseURL?: string) =>
|
|
330
|
+
declare const createMinimaxi: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
374
331
|
|
|
375
332
|
/**
|
|
376
333
|
* Create a Together AI Provider
|
|
377
334
|
* @see {@link https://docs.together.ai/docs/serverless-models}
|
|
378
335
|
*/
|
|
379
|
-
declare const createTogetherAI: (apiKey: string, baseURL?: string) =>
|
|
380
|
-
|
|
381
|
-
/**
|
|
382
|
-
* Create a Workers AI Provider
|
|
383
|
-
* @see {@link https://developers.cloudflare.com/workers-ai}
|
|
384
|
-
*/
|
|
385
|
-
declare const createWorkersAI: (apiKey: string, accountId: string) => _xsai_ext_shared_providers.EmbedProvider<string> & _xsai_ext_shared_providers.ChatProvider<CloudflareWorkersAiModels>;
|
|
336
|
+
declare const createTogetherAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<TogetheraiModels>;
|
|
386
337
|
|
|
387
|
-
export { createAlibaba, createAlibabaCn,
|
|
338
|
+
export { createAlibaba, createAlibabaCn, createBailing, createBaseten, createCerebras, createChutes, createCortecs, createDeepSeek, createDeepinfra, createFastrouter, createFatherless, createFeatherless, createFireworks, createGithubCopilot, createGithubModels, createGoogleGenerativeAI, createGroq, createHelicone, createHuggingface, createIflowcn, createInception, createInference, createIoNet, createKimiForCoding, createLitellm, createLlama, createLmstudio, createLucidquery, createMinimax, createMinimaxCn, createMinimaxi, createMistral, createModelscope, createMoonshotai, createMoonshotaiCn, createMorph, createNebius, createNovita, createNvidia, createOllama, createOllamaCloud, createOpenAI, createOpencode, createOvhcloud, createPerplexity, createPoe, createQwen, createRequesty, createScaleway, createSiliconFlow, createSiliconflowCn, createStepfun, createSubmodel, createSynthetic, createTencentHunyuan, createTogetherAI, createUpstage, createVenice, createVultr, createWandb, createXAI, createXai, createXiaomi, createZai, createZaiCodingPlan, createZenmux, createZhipuai, createZhipuaiCodingPlan };
|