@xsai-ext/providers 0.4.1 → 0.4.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/{anthropic-Bel15KQV.d.ts → anthropic-C6HzSBof.d.ts} +1 -1
- package/dist/create.d.ts +87 -49
- package/dist/create.js +6 -1
- package/dist/index.d.ts +112 -54
- package/dist/index.js +12 -4
- package/dist/special/create.d.ts +1 -1
- package/dist/{together-ai-CzFuvpt-.js → together-ai-SWwaUOgP.js} +35 -6
- package/dist/types-B2HZyaCH.d.ts +7 -0
- package/package.json +2 -2
- package/dist/types-Czo87gCh.d.ts +0 -7
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { M as ModelProvider, a as ChatProviderWithExtraOptions } from './types-DW1hvH0W.js';
|
|
2
|
-
import { O as OpenrouterModels, A as AnthropicModels } from './types-
|
|
2
|
+
import { O as OpenrouterModels, A as AnthropicModels } from './types-B2HZyaCH.js';
|
|
3
3
|
|
|
4
4
|
interface OpenRouterOptions {
|
|
5
5
|
extraHeaders?: (Headers | Record<string, string>) & {
|
package/dist/create.d.ts
CHANGED
|
@@ -1,18 +1,23 @@
|
|
|
1
1
|
import { C as ChatProvider, M as ModelProvider, E as EmbedProvider, I as ImageProvider, S as SpeechProvider, T as TranscriptionProvider } from './types-DW1hvH0W.js';
|
|
2
|
-
export { c as createAnthropic, a as createOpenRouter } from './anthropic-
|
|
3
|
-
import { T as TogetheraiModels } from './types-
|
|
2
|
+
export { c as createAnthropic, a as createOpenRouter } from './anthropic-C6HzSBof.js';
|
|
3
|
+
import { T as TogetheraiModels } from './types-B2HZyaCH.js';
|
|
4
4
|
import '@xsai/shared';
|
|
5
5
|
|
|
6
|
+
/**
|
|
7
|
+
* Create a Abacus Provider
|
|
8
|
+
* @see {@link https://abacus.ai/help/api}
|
|
9
|
+
*/
|
|
10
|
+
declare const createAbacus: (apiKey: string, baseURL?: string) => ChatProvider<"gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "gemini-2.0-flash-001" | "gemini-3-flash-preview" | "route-llm" | "grok-code-fast-1" | "kimi-k2-turbo-preview" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "gpt-4.1-mini" | "claude-opus-4-5-20251101" | "qwen-2.5-coder-32b" | "claude-sonnet-4-5-20250929" | "grok-4-0709" | "o3-mini" | "gpt-5.2-chat-latest" | "gemini-2.0-pro-exp-02-05" | "gpt-5.1" | "gpt-5-nano" | "claude-sonnet-4-20250514" | "gpt-4.1" | "o4-mini" | "claude-opus-4-20250514" | "gpt-5-mini" | "o3-pro" | "claude-3-7-sonnet-20250219" | "gemini-2.5-pro" | "gpt-4o-2024-11-20" | "o3" | "gpt-4o-mini" | "qwen3-max" | "gpt-5" | "grok-4-1-fast-non-reasoning" | "llama-3.3-70b-versatile" | "claude-opus-4-1-20250805" | "gpt-5.2" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "deepseek/deepseek-v3.1" | "openai/gpt-oss-120b" | "meta-llama/Meta-Llama-3.1-8B-Instruct" | "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "meta-llama/Meta-Llama-3.1-70B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/QwQ-32B" | "Qwen/Qwen3-32B" | "Qwen/qwen3-coder-480b-a35b-instruct" | "zai-org/glm-4.7" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2"> & ModelProvider;
|
|
6
11
|
/**
|
|
7
12
|
* Create a Alibaba Provider
|
|
8
13
|
* @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
|
|
9
14
|
*/
|
|
10
|
-
declare const createAlibaba: (apiKey: string, baseURL?: string) => ChatProvider<"qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-
|
|
15
|
+
declare const createAlibaba: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-max" | "qwen3-livetranslate-flash-realtime" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen-plus-character-ja" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus">;
|
|
11
16
|
/**
|
|
12
17
|
* Create a Alibaba (China) Provider
|
|
13
18
|
* @see {@link https://www.alibabacloud.com/help/en/model-studio/models}
|
|
14
19
|
*/
|
|
15
|
-
declare const createAlibabaCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-
|
|
20
|
+
declare const createAlibabaCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-max" | "qwen3-asr-flash" | "qwen-omni-turbo" | "qwen-vl-max" | "qwen3-next-80b-a3b-instruct" | "qwen-turbo" | "qwen3-vl-235b-a22b" | "qwen3-coder-flash" | "qwen3-vl-30b-a3b" | "qwen3-14b" | "qvq-max" | "qwen2-5-14b-instruct" | "qwq-plus" | "qwen3-coder-30b-a3b-instruct" | "qwen-vl-ocr" | "qwen2-5-72b-instruct" | "qwen3-omni-flash" | "qwen-flash" | "qwen3-8b" | "qwen3-omni-flash-realtime" | "qwen2-5-vl-72b-instruct" | "qwen3-vl-plus" | "qwen-plus" | "qwen2-5-32b-instruct" | "qwen2-5-omni-7b" | "qwen-max" | "qwen2-5-7b-instruct" | "qwen2-5-vl-7b-instruct" | "qwen3-235b-a22b" | "qwen-omni-turbo-realtime" | "qwen-mt-turbo" | "qwen3-coder-480b-a35b-instruct" | "qwen-mt-plus" | "qwen3-coder-plus" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "qwen-vl-plus" | "deepseek-r1-distill-qwen-7b" | "deepseek-r1-0528" | "deepseek-v3" | "deepseek-v3-2-exp" | "deepseek-r1" | "deepseek-r1-distill-qwen-32b" | "qwen-plus-character" | "qwen2-5-coder-32b-instruct" | "qwen-math-plus" | "qwen-doc-turbo" | "qwen-deep-research" | "qwen-long" | "qwen2-5-math-72b-instruct" | "moonshot-kimi-k2-instruct" | "tongyi-intent-detect-v3" | "deepseek-v3-1" | "deepseek-r1-distill-llama-70b" | "qwen2-5-coder-7b-instruct" | "deepseek-r1-distill-qwen-14b" | "qwen-math-turbo" | "deepseek-r1-distill-llama-8b" | "qwq-32b" | "qwen2-5-math-7b-instruct" | "deepseek-r1-distill-qwen-1-5b">;
|
|
16
21
|
/**
|
|
17
22
|
* Create a Bailing Provider
|
|
18
23
|
* @see {@link https://alipaytbox.yuque.com/sxs0ba/ling/intro}
|
|
@@ -22,27 +27,32 @@ declare const createBailing: (apiKey: string, baseURL?: string) => ModelProvider
|
|
|
22
27
|
* Create a Baseten Provider
|
|
23
28
|
* @see {@link https://docs.baseten.co/development/model-apis/overview}
|
|
24
29
|
*/
|
|
25
|
-
declare const createBaseten: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.
|
|
30
|
+
declare const createBaseten: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "zai-org/GLM-4.6">;
|
|
26
31
|
/**
|
|
27
32
|
* Create a Cerebras Provider
|
|
28
33
|
* @see {@link https://inference-docs.cerebras.ai/models/overview}
|
|
29
34
|
*/
|
|
30
|
-
declare const createCerebras: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen-3-235b-a22b-instruct-2507" | "
|
|
35
|
+
declare const createCerebras: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"zai-glm-4.7" | "qwen-3-235b-a22b-instruct-2507" | "gpt-oss-120b"> & EmbedProvider<string>;
|
|
31
36
|
/**
|
|
32
37
|
* Create a Chutes Provider
|
|
33
38
|
* @see {@link https://llm.chutes.ai/v1/models}
|
|
34
39
|
*/
|
|
35
|
-
declare const createChutes: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"
|
|
40
|
+
declare const createChutes: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen2.5-72B-Instruct" | "Qwen/Qwen3-32B" | "moonshotai/Kimi-K2-Instruct-0905" | "NousResearch/Hermes-4.3-36B" | "NousResearch/Hermes-4-70B" | "NousResearch/Hermes-4-14B" | "NousResearch/Hermes-4-405B-FP8-TEE" | "NousResearch/DeepHermes-3-Mistral-24B-Preview" | "rednote-hilab/dots.ocr" | "moonshotai/Kimi-K2-Thinking-TEE" | "MiniMaxAI/MiniMax-M2.1-TEE" | "nvidia/NVIDIA-Nemotron-3-Nano-30B-A3B-BF16" | "tngtech/DeepSeek-R1T-Chimera" | "tngtech/DeepSeek-TNG-R1T2-Chimera" | "tngtech/TNG-R1T-Chimera-TEE" | "XiaomiMiMo/MiMo-V2-Flash" | "OpenGVLab/InternVL3-78B-TEE" | "openai/gpt-oss-120b-TEE" | "openai/gpt-oss-20b" | "chutesai/Mistral-Small-3.1-24B-Instruct-2503" | "chutesai/Mistral-Small-3.2-24B-Instruct-2506" | "mistralai/Devstral-2-123B-Instruct-2512" | "mistralai/Devstral-2-123B-Instruct-2512-TEE" | "unsloth/Mistral-Nemo-Instruct-2407" | "unsloth/gemma-3-4b-it" | "unsloth/Mistral-Small-24B-Instruct-2501" | "unsloth/gemma-3-12b-it" | "unsloth/gemma-3-27b-it" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3Guard-Gen-0.6B" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen2.5-VL-72B-Instruct-TEE" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-235B-A22B-Instruct-2507-TEE" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8-TEE" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "miromind-ai/MiroThinker-v1.5-235B" | "zai-org/GLM-4.6-TEE" | "zai-org/GLM-4.5-TEE" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.7-TEE" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-V3-0324-TEE" | "deepseek-ai/DeepSeek-V3.2-Speciale-TEE" | "deepseek-ai/DeepSeek-V3.1-Terminus-TEE" | "deepseek-ai/DeepSeek-V3" | "deepseek-ai/DeepSeek-R1-TEE" | "deepseek-ai/DeepSeek-R1-Distill-Llama-70B" | "deepseek-ai/DeepSeek-R1-0528-TEE" | "deepseek-ai/DeepSeek-V3.2-TEE" | "deepseek-ai/DeepSeek-V3.1-TEE">;
|
|
41
|
+
/**
|
|
42
|
+
* Create a Cohere Provider
|
|
43
|
+
* @see {@link https://docs.cohere.com/docs/models}
|
|
44
|
+
*/
|
|
45
|
+
declare const createCohere: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"command-a-translate-08-2025" | "command-a-03-2025" | "command-r-08-2024" | "command-r-plus-08-2024" | "command-r7b-12-2024" | "command-a-reasoning-08-2025" | "command-a-vision-07-2025">;
|
|
36
46
|
/**
|
|
37
47
|
* Create a Cortecs Provider
|
|
38
48
|
* @see {@link https://api.cortecs.ai/v1/models}
|
|
39
49
|
*/
|
|
40
|
-
declare const createCortecs: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "
|
|
50
|
+
declare const createCortecs: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"gpt-4.1" | "gemini-2.5-pro" | "qwen3-coder-480b-a35b-instruct" | "qwen3-next-80b-a3b-thinking" | "qwen3-32b" | "gpt-oss-120b" | "nova-pro-v1" | "devstral-2512" | "intellect-3" | "claude-4-5-sonnet" | "deepseek-v3-0324" | "kimi-k2-thinking" | "kimi-k2-instruct" | "devstral-small-2512" | "claude-sonnet-4" | "llama-3.1-405b-instruct">;
|
|
41
51
|
/**
|
|
42
52
|
* Create a Deep Infra Provider
|
|
43
53
|
* @see {@link https://deepinfra.com/models}
|
|
44
54
|
*/
|
|
45
|
-
declare const createDeepinfra: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"Qwen/Qwen3-Coder-480B-A35B-Instruct" | "openai/gpt-oss-20b" | "
|
|
55
|
+
declare const createDeepinfra: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"openai/gpt-oss-120b" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "openai/gpt-oss-20b" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2" | "MiniMaxAI/MiniMax-M2.1" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo" | "zai-org/GLM-4.5">;
|
|
46
56
|
/**
|
|
47
57
|
* Create a DeepSeek Provider
|
|
48
58
|
* @see {@link https://platform.deepseek.com/api-docs/pricing}
|
|
@@ -52,17 +62,27 @@ declare const createDeepSeek: (apiKey: string, baseURL?: string) => ModelProvide
|
|
|
52
62
|
* Create a FastRouter Provider
|
|
53
63
|
* @see {@link https://fastrouter.ai/models}
|
|
54
64
|
*/
|
|
55
|
-
declare const createFastrouter: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-
|
|
65
|
+
declare const createFastrouter: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "openai/gpt-oss-20b" | "moonshotai/kimi-k2" | "x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "qwen/qwen3-coder" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "deepseek-ai/deepseek-r1-distill-llama-70b">;
|
|
56
66
|
/**
|
|
57
67
|
* Create a Fireworks AI Provider
|
|
58
68
|
* @see {@link https://fireworks.ai/docs/}
|
|
59
69
|
*/
|
|
60
|
-
declare const createFireworks: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
|
|
70
|
+
declare const createFireworks: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"accounts/fireworks/models/deepseek-r1-0528" | "accounts/fireworks/models/deepseek-v3p1" | "accounts/fireworks/models/deepseek-v3p2" | "accounts/fireworks/models/minimax-m2" | "accounts/fireworks/models/minimax-m2p1" | "accounts/fireworks/models/glm-4p7" | "accounts/fireworks/models/deepseek-v3-0324" | "accounts/fireworks/models/glm-4p6" | "accounts/fireworks/models/kimi-k2-thinking" | "accounts/fireworks/models/kimi-k2-instruct" | "accounts/fireworks/models/qwen3-235b-a22b" | "accounts/fireworks/models/gpt-oss-20b" | "accounts/fireworks/models/gpt-oss-120b" | "accounts/fireworks/models/glm-4p5-air" | "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct" | "accounts/fireworks/models/glm-4p5">;
|
|
71
|
+
/**
|
|
72
|
+
* Create a Firmware Provider
|
|
73
|
+
* @see {@link https://docs.firmware.ai}
|
|
74
|
+
*/
|
|
75
|
+
declare const createFirmware: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"grok-4-fast-non-reasoning" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "claude-sonnet-4-5-20250929" | "gpt-5-nano" | "gpt-5-mini" | "gemini-2.5-pro" | "gpt-5" | "gpt-5.2" | "claude-haiku-4-5-20251001" | "deepseek-chat" | "deepseek-reasoner" | "grok-4-fast-reasoning" | "deepseek-coder" | "claude-opus-4-5" | "gpt-4o">;
|
|
76
|
+
/**
|
|
77
|
+
* Create a Friendli Provider
|
|
78
|
+
* @see {@link https://friendli.ai/docs/guides/serverless_endpoints/introduction}
|
|
79
|
+
*/
|
|
80
|
+
declare const createFriendli: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-32B" | "zai-org/GLM-4.6" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "meta-llama-3.3-70b-instruct" | "meta-llama-3.1-8b-instruct" | "LGAI-EXAONE/K-EXAONE-236B-A23B" | "LGAI-EXAONE/EXAONE-4.0.1-32B" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct">;
|
|
61
81
|
/**
|
|
62
82
|
* Create a GitHub Copilot Provider
|
|
63
83
|
* @see {@link https://docs.github.com/en/copilot}
|
|
64
84
|
*/
|
|
65
|
-
declare const createGithubCopilot: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"
|
|
85
|
+
declare const createGithubCopilot: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"gemini-2.0-flash-001" | "gemini-3-flash-preview" | "grok-code-fast-1" | "gemini-3-pro-preview" | "o3-mini" | "gpt-5.1" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "gemini-2.5-pro" | "o3" | "gpt-5" | "gpt-5.2" | "claude-sonnet-4" | "gpt-4o" | "claude-opus-4" | "gpt-5.2-codex" | "gpt-5.1-codex" | "claude-haiku-4.5" | "claude-3.5-sonnet" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "claude-opus-41" | "claude-3.7-sonnet" | "gpt-5.1-codex-max" | "claude-3.7-sonnet-thought" | "claude-opus-4.5" | "claude-sonnet-4.5">;
|
|
66
86
|
/**
|
|
67
87
|
* Create a GitHub Models Provider
|
|
68
88
|
* @see {@link https://docs.github.com/en/github-models}
|
|
@@ -72,27 +92,27 @@ declare const createGithubModels: (apiKey: string, baseURL?: string) => ModelPro
|
|
|
72
92
|
* Create a Google Provider
|
|
73
93
|
* @see {@link https://ai.google.dev/gemini-api/docs/pricing}
|
|
74
94
|
*/
|
|
75
|
-
declare const createGoogleGenerativeAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-
|
|
95
|
+
declare const createGoogleGenerativeAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"gemini-3-flash-preview" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "gemini-2.5-pro" | "gemini-embedding-001" | "gemini-2.5-flash-image" | "gemini-2.5-flash-preview-05-20" | "gemini-flash-lite-latest" | "gemini-flash-latest" | "gemini-2.5-pro-preview-05-06" | "gemini-2.5-flash-preview-tts" | "gemini-2.0-flash-lite" | "gemini-live-2.5-flash-preview-native-audio" | "gemini-2.0-flash" | "gemini-2.5-flash-lite" | "gemini-2.5-pro-preview-06-05" | "gemini-live-2.5-flash" | "gemini-2.5-flash-lite-preview-06-17" | "gemini-2.5-flash-image-preview" | "gemini-2.5-flash-preview-09-2025" | "gemini-2.5-flash-preview-04-17" | "gemini-2.5-pro-preview-tts" | "gemini-1.5-flash" | "gemini-1.5-flash-8b" | "gemini-2.5-flash-lite-preview-09-2025" | "gemini-1.5-pro">;
|
|
76
96
|
/**
|
|
77
97
|
* Create a Groq Provider
|
|
78
98
|
* @see {@link https://console.groq.com/docs/models}
|
|
79
99
|
*/
|
|
80
|
-
declare const createGroq: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"
|
|
100
|
+
declare const createGroq: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"llama-3.3-70b-versatile" | "openai/gpt-oss-120b" | "deepseek-r1-distill-llama-70b" | "openai/gpt-oss-20b" | "llama-3.1-8b-instant" | "mistral-saba-24b" | "llama3-8b-8192" | "qwen-qwq-32b" | "llama3-70b-8192" | "llama-guard-3-8b" | "gemma2-9b-it" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-32b" | "meta-llama/llama-4-scout-17b-16e-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct" | "meta-llama/llama-guard-4-12b">;
|
|
81
101
|
/**
|
|
82
102
|
* Create a Helicone Provider
|
|
83
103
|
* @see {@link https://helicone.ai/models}
|
|
84
104
|
*/
|
|
85
|
-
declare const createHelicone: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"
|
|
105
|
+
declare const createHelicone: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"gpt-4.1-nano" | "grok-4-fast-non-reasoning" | "grok-code-fast-1" | "gemini-3-pro-preview" | "gemini-2.5-flash" | "gpt-4.1-mini" | "claude-sonnet-4-5-20250929" | "o3-mini" | "gpt-5.1" | "gpt-5-nano" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "o3-pro" | "gemini-2.5-pro" | "o3" | "gpt-4o-mini" | "gpt-5" | "grok-4-1-fast-non-reasoning" | "llama-3.3-70b-versatile" | "claude-opus-4-1-20250805" | "gpt-5.1-chat-latest" | "claude-haiku-4-5-20251001" | "qwen3-next-80b-a3b-instruct" | "qwen3-coder-30b-a3b-instruct" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "kimi-k2-thinking" | "claude-sonnet-4" | "deepseek-reasoner" | "grok-4-fast-reasoning" | "gpt-4o" | "claude-opus-4" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "claude-3.7-sonnet" | "gemini-2.5-flash-lite" | "llama-3.1-8b-instant" | "gemma2-9b-it" | "qwen3-coder" | "claude-opus-4-1" | "grok-4" | "llama-4-maverick" | "llama-prompt-guard-2-86m" | "grok-4-1-fast-reasoning" | "claude-4.5-haiku" | "llama-3.1-8b-instruct-turbo" | "gpt-4.1-mini-2025-04-14" | "llama-guard-4" | "llama-3.1-8b-instruct" | "deepseek-v3.1-terminus" | "llama-prompt-guard-2-22m" | "claude-3.5-sonnet-v2" | "sonar-deep-research" | "grok-3" | "mistral-small" | "kimi-k2-0711" | "chatgpt-4o-latest" | "kimi-k2-0905" | "sonar-reasoning" | "llama-3.3-70b-instruct" | "claude-4.5-sonnet" | "codex-mini-latest" | "deepseek-tng-r1t2-chimera" | "claude-4.5-opus" | "sonar" | "glm-4.6" | "qwen3-235b-a22b-thinking" | "hermes-2-pro-llama-3-8b" | "o1" | "grok-3-mini" | "sonar-pro" | "o1-mini" | "claude-3-haiku-20240307" | "qwen2.5-coder-7b-fast" | "gemma-3-12b-it" | "mistral-nemo" | "gpt-oss-20b" | "claude-3.5-haiku" | "gpt-5-chat-latest" | "sonar-reasoning-pro" | "qwen3-vl-235b-a22b-instruct" | "qwen3-30b-a3b" | "deepseek-v3.2" | "gpt-5-pro" | "mistral-large-2411" | "ernie-4.5-21b-a3b-thinking" | "llama-4-scout">;
|
|
86
106
|
/**
|
|
87
107
|
* Create a Hugging Face Provider
|
|
88
108
|
* @see {@link https://huggingface.co/docs/inference-providers}
|
|
89
109
|
*/
|
|
90
|
-
declare const createHuggingface: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.
|
|
110
|
+
declare const createHuggingface: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "XiaomiMiMo/MiMo-V2-Flash" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2.1" | "Qwen/Qwen3-Embedding-8B" | "Qwen/Qwen3-Embedding-4B" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "deepseek-ai/DeepSeek-R1-0528">;
|
|
91
111
|
/**
|
|
92
112
|
* Create a iFlow Provider
|
|
93
113
|
* @see {@link https://platform.iflow.cn/en/docs}
|
|
94
114
|
*/
|
|
95
|
-
declare const createIflowcn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-
|
|
115
|
+
declare const createIflowcn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-max" | "qwen3-vl-plus" | "qwen3-coder-plus" | "qwen3-32b" | "deepseek-v3" | "deepseek-r1" | "kimi-k2-0905" | "glm-4.6" | "deepseek-v3.2" | "kimi-k2" | "qwen3-235b" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct" | "qwen3-max-preview">;
|
|
96
116
|
/**
|
|
97
117
|
* Create a Inception Provider
|
|
98
118
|
* @see {@link https://platform.inceptionlabs.ai/docs}
|
|
@@ -107,7 +127,7 @@ declare const createInference: (apiKey: string, baseURL?: string) => ModelProvid
|
|
|
107
127
|
* Create a IO.NET Provider
|
|
108
128
|
* @see {@link https://io.net/docs/guides/intelligence/io-intelligence}
|
|
109
129
|
*/
|
|
110
|
-
declare const createIoNet: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "
|
|
130
|
+
declare const createIoNet: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "mistralai/Devstral-Small-2505" | "mistralai/Mistral-Nemo-Instruct-2407" | "mistralai/Magistral-Small-2506" | "mistralai/Mistral-Large-Instruct-2411" | "meta-llama/Llama-3.3-70B-Instruct" | "meta-llama/Llama-3.2-90B-Vision-Instruct" | "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar">;
|
|
111
131
|
/**
|
|
112
132
|
* Create a Kimi For Coding Provider
|
|
113
133
|
* @see {@link https://www.kimi.com/coding/docs/en/third-party-agents.html}
|
|
@@ -132,12 +152,12 @@ declare const createLucidquery: (apiKey: string, baseURL?: string) => ModelProvi
|
|
|
132
152
|
* Create a MiniMax Provider
|
|
133
153
|
* @see {@link https://platform.minimax.io/docs/guides/quickstart}
|
|
134
154
|
*/
|
|
135
|
-
declare const createMinimax: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
155
|
+
declare const createMinimax: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2" | "MiniMax-M2.1">;
|
|
136
156
|
/**
|
|
137
157
|
* Create a MiniMax (China) Provider
|
|
138
158
|
* @see {@link https://platform.minimaxi.com/docs/guides/quickstart}
|
|
139
159
|
*/
|
|
140
|
-
declare const createMinimaxCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2">;
|
|
160
|
+
declare const createMinimaxCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"MiniMax-M2" | "MiniMax-M2.1">;
|
|
141
161
|
/**
|
|
142
162
|
* Create a Mistral Provider
|
|
143
163
|
* @see {@link https://docs.mistral.ai/getting-started/models/}
|
|
@@ -147,32 +167,42 @@ declare const createMistral: (apiKey: string, baseURL?: string) => ModelProvider
|
|
|
147
167
|
* Create a ModelScope Provider
|
|
148
168
|
* @see {@link https://modelscope.cn/docs/model-service/API-Inference/intro}
|
|
149
169
|
*/
|
|
150
|
-
declare const createModelscope: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-
|
|
170
|
+
declare const createModelscope: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "ZhipuAI/GLM-4.5" | "ZhipuAI/GLM-4.6" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct">;
|
|
151
171
|
/**
|
|
152
172
|
* Create a Moonshot AI Provider
|
|
153
173
|
* @see {@link https://platform.moonshot.ai/docs/api/chat}
|
|
154
174
|
*/
|
|
155
|
-
declare const createMoonshotai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-
|
|
175
|
+
declare const createMoonshotai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-turbo-preview" | "kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
|
|
156
176
|
/**
|
|
157
177
|
* Create a Moonshot AI (China) Provider
|
|
158
178
|
* @see {@link https://platform.moonshot.cn/docs/api/chat}
|
|
159
179
|
*/
|
|
160
|
-
declare const createMoonshotaiCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-
|
|
180
|
+
declare const createMoonshotaiCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-turbo-preview" | "kimi-k2-thinking" | "kimi-k2-thinking-turbo" | "kimi-k2-0711-preview" | "kimi-k2-0905-preview">;
|
|
161
181
|
/**
|
|
162
182
|
* Create a Morph Provider
|
|
163
183
|
* @see {@link https://docs.morphllm.com/api-reference/introduction}
|
|
164
184
|
*/
|
|
165
185
|
declare const createMorph: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"morph-v3-large" | "auto" | "morph-v3-fast">;
|
|
186
|
+
/**
|
|
187
|
+
* Create a NanoGPT Provider
|
|
188
|
+
* @see {@link https://docs.nano-gpt.com}
|
|
189
|
+
*/
|
|
190
|
+
declare const createNanoGpt: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "zai-org/glm-4.7" | "qwen/qwen3-coder" | "deepseek/deepseek-r1" | "moonshotai/kimi-k2-instruct" | "moonshotai/kimi-k2-thinking" | "nousresearch/hermes-4-405b:thinking" | "nvidia/llama-3_3-nemotron-super-49b-v1_5" | "deepseek/deepseek-v3.2:thinking" | "minimax/minimax-m2.1" | "z-ai/glm-4.6:thinking" | "z-ai/glm-4.6" | "qwen/qwen3-235b-a22b-thinking-2507" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "meta-llama/llama-4-maverick" | "meta-llama/llama-3.3-70b-instruct" | "zai-org/glm-4.5-air" | "zai-org/glm-4.7:thinking" | "zai-org/glm-4.5-air:thinking">;
|
|
166
191
|
/**
|
|
167
192
|
* Create a Nebius Token Factory Provider
|
|
168
193
|
* @see {@link https://docs.tokenfactory.nebius.com/}
|
|
169
194
|
*/
|
|
170
|
-
declare const createNebius: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-
|
|
195
|
+
declare const createNebius: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "zai-org/glm-4.5" | "openai/gpt-oss-20b" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-235b-a22b-thinking-2507" | "zai-org/glm-4.5-air" | "NousResearch/hermes-4-70b" | "NousResearch/hermes-4-405b" | "nvidia/llama-3_1-nemotron-ultra-253b-v1" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "meta-llama/llama-3_1-405b-instruct" | "meta-llama/llama-3.3-70b-instruct-fast" | "meta-llama/llama-3.3-70b-instruct-base" | "deepseek-ai/deepseek-v3">;
|
|
196
|
+
/**
|
|
197
|
+
* Create a NovitaAI Provider
|
|
198
|
+
* @see {@link https://novita.ai/docs/guides/introduction}
|
|
199
|
+
*/
|
|
200
|
+
declare const createNovitaAi: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"deepseek/deepseek-v3.1" | "openai/gpt-oss-120b" | "zai-org/glm-4.7" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "openai/gpt-oss-20b" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-v3-0324" | "moonshotai/kimi-k2-instruct" | "meta-llama/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-thinking" | "minimax/minimax-m2.1" | "qwen/qwen3-235b-a22b-thinking-2507" | "meta-llama/llama-3.3-70b-instruct" | "zai-org/glm-4.5-air" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "baichuan/baichuan-m2-32b" | "moonshotai/kimi-k2-0905" | "nousresearch/hermes-2-pro-llama-3-8b" | "paddlepaddle/paddleocr-vl" | "kwaipilot/kat-coder" | "kwaipilot/kat-coder-pro" | "xiaomimimo/mimo-v2-flash" | "deepseek/deepseek-prover-v2-671b" | "deepseek/deepseek-r1-0528-qwen3-8b" | "deepseek/deepseek-v3.1-terminus" | "deepseek/deepseek-r1-turbo" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-r1-distill-llama-70b" | "deepseek/deepseek-ocr" | "deepseek/deepseek-v3.2" | "deepseek/deepseek-v3-turbo" | "sao10k/l3-8b-lunaris" | "sao10k/L3-8B-Stheno-v3.2" | "sao10k/l31-70b-euryale-v2.2" | "sao10k/l3-70b-euryale-v2.1" | "skywork/r1v4-lite" | "minimaxai/minimax-m1-80k" | "minimax/minimax-m2" | "google/gemma-3-27b-it" | "microsoft/wizardlm-2-8x22b" | "gryphe/mythomax-l2-13b" | "baidu/ernie-4.5-vl-28b-a3b-thinking" | "baidu/ernie-4.5-300b-a47b-paddle" | "baidu/ernie-4.5-21B-a3b" | "baidu/ernie-4.5-21B-a3b-thinking" | "baidu/ernie-4.5-vl-424b-a47b" | "baidu/ernie-4.5-vl-28b-a3b" | "qwen/qwen3-vl-30b-a3b-thinking" | "qwen/qwen3-omni-30b-a3b-thinking" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-8b-fp8" | "qwen/qwen2.5-vl-72b-instruct" | "qwen/qwen3-4b-fp8" | "qwen/qwen3-coder-30b-a3b-instruct" | "qwen/qwen3-vl-8b-instruct" | "qwen/qwen2.5-7b-instruct" | "qwen/qwen3-30b-a3b-fp8" | "qwen/qwen3-32b-fp8" | "qwen/qwen3-omni-30b-a3b-instruct" | "qwen/qwen-2.5-72b-instruct" | "qwen/qwen3-vl-235b-a22b-thinking" | "qwen/qwen-mt-plus" | "qwen/qwen3-max" | "qwen/qwen3-vl-235b-a22b-instruct" | "qwen/qwen3-vl-30b-a3b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-235b-a22b-fp8" | "mistralai/mistral-nemo" | "meta-llama/llama-3-70b-instruct" | "meta-llama/llama-3-8b-instruct" | "meta-llama/llama-3.1-8b-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct-fp8" | "zai-org/glm-4.5v" | "zai-org/glm-4.6v" | "zai-org/autoglm-phone-9b-multilingual">;
|
|
171
201
|
/**
|
|
172
202
|
* Create a Nvidia Provider
|
|
173
203
|
* @see {@link https://docs.api.nvidia.com/nim/}
|
|
174
204
|
*/
|
|
175
|
-
declare const createNvidia: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-4-mini-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "qwen/qwen3-coder-480b-a35b-instruct" | "
|
|
205
|
+
declare const createNvidia: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "microsoft/phi-3-medium-128k-instruct" | "microsoft/phi-3-small-128k-instruct" | "microsoft/phi-3.5-vision-instruct" | "microsoft/phi-3-small-8k-instruct" | "microsoft/phi-3.5-moe-instruct" | "microsoft/phi-4-mini-instruct" | "microsoft/phi-3-medium-4k-instruct" | "meta/llama-3.2-11b-vision-instruct" | "meta/llama-3.3-70b-instruct" | "meta/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-instruct-0905" | "moonshotai/kimi-k2-instruct" | "meta/llama-3.2-1b-instruct" | "moonshotai/kimi-k2-thinking" | "mistralai/devstral-2-123b-instruct-2512" | "mistralai/mistral-large-3-675b-instruct-2512" | "mistralai/ministral-14b-instruct-2512" | "qwen/qwen3-coder-480b-a35b-instruct" | "google/gemma-3-27b-it" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "nvidia/nvidia-nemotron-nano-9b-v2" | "nvidia/cosmos-nemotron-34b" | "nvidia/llama-embed-nemotron-8b" | "nvidia/nemotron-3-nano-30b-a3b" | "nvidia/parakeet-tdt-0.6b-v2" | "nvidia/nemoretriever-ocr-v1" | "nvidia/llama-3.3-nemotron-super-49b-v1" | "nvidia/llama-3.1-nemotron-51b-instruct" | "nvidia/llama3-chatqa-1.5-70b" | "nvidia/llama-3.1-nemotron-ultra-253b-v1" | "nvidia/llama-3.1-nemotron-70b-instruct" | "nvidia/nemotron-4-340b-instruct" | "nvidia/llama-3.3-nemotron-super-49b-v1.5" | "minimaxai/minimax-m2" | "google/gemma-3n-e2b-it" | "google/codegemma-1.1-7b" | "google/gemma-3n-e4b-it" | "google/gemma-2-2b-it" | "google/gemma-3-12b-it" | "google/codegemma-7b" | "google/gemma-3-1b-it" | "google/gemma-2-27b-it" | "microsoft/phi-3-vision-128k-instruct" | "openai/whisper-large-v3" | "qwen/qwen2.5-coder-32b-instruct" | "qwen/qwen2.5-coder-7b-instruct" | "qwen/qwen3-235b-a22b" | "qwen/qwq-32b" | "mistralai/mamba-codestral-7b-v0.1" | "mistralai/mistral-large-2-instruct" | "mistralai/codestral-22b-instruct-v0.1" | "mistralai/mistral-small-3.1-24b-instruct-2503" | "meta/llama3-70b-instruct" | "meta/llama-4-maverick-17b-128e-instruct" | "meta/codellama-70b" | "meta/llama-3.1-405b-instruct" | "meta/llama3-8b-instruct" | "meta/llama-3.1-70b-instruct" | "deepseek-ai/deepseek-r1-0528" | "deepseek-ai/deepseek-r1" | "deepseek-ai/deepseek-v3.1-terminus" | "deepseek-ai/deepseek-v3.1" | "deepseek-ai/deepseek-coder-6.7b-instruct" | "black-forest-labs/flux.1-dev">;
|
|
176
206
|
/**
|
|
177
207
|
* Create a Ollama Cloud Provider
|
|
178
208
|
* @see {@link https://docs.ollama.com/cloud}
|
|
@@ -182,12 +212,12 @@ declare const createOllamaCloud: (apiKey: string, baseURL?: string) => ModelProv
|
|
|
182
212
|
* Create a OpenAI Provider
|
|
183
213
|
* @see {@link https://platform.openai.com/docs/models}
|
|
184
214
|
*/
|
|
185
|
-
declare const createOpenAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1" | "gpt-
|
|
215
|
+
declare const createOpenAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<"gpt-4.1-nano" | "gpt-4.1-mini" | "o3-mini" | "gpt-5.2-chat-latest" | "gpt-5.1" | "gpt-5-nano" | "gpt-4.1" | "o4-mini" | "gpt-5-mini" | "o3-pro" | "gpt-4o-2024-11-20" | "o3" | "gpt-4o-mini" | "gpt-5" | "gpt-5.2" | "gpt-5.1-chat-latest" | "gpt-4o" | "gpt-5.2-codex" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "gpt-5.1-codex-max" | "codex-mini-latest" | "o1" | "o1-mini" | "gpt-5-chat-latest" | "gpt-5-pro" | "text-embedding-3-small" | "gpt-4" | "o1-pro" | "gpt-4o-2024-05-13" | "gpt-4o-2024-08-06" | "o3-deep-research" | "gpt-3.5-turbo" | "gpt-5.2-pro" | "text-embedding-3-large" | "gpt-4-turbo" | "o1-preview" | "text-embedding-ada-002" | "o4-mini-deep-research"> & ImageProvider<string> & SpeechProvider<string> & TranscriptionProvider<string>;
|
|
186
216
|
/**
|
|
187
217
|
* Create a OpenCode Zen Provider
|
|
188
218
|
* @see {@link https://opencode.ai/docs/zen}
|
|
189
219
|
*/
|
|
190
|
-
declare const createOpencode: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"kimi-k2-thinking" | "claude-sonnet-4" | "
|
|
220
|
+
declare const createOpencode: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"gpt-5.1" | "gpt-5-nano" | "gpt-5" | "gpt-5.2" | "kimi-k2-thinking" | "claude-sonnet-4" | "claude-opus-4-5" | "gpt-5.2-codex" | "gpt-5.1-codex" | "gpt-5.1-codex-mini" | "gpt-5-codex" | "gpt-5.1-codex-max" | "qwen3-coder" | "claude-opus-4-1" | "glm-4.6" | "kimi-k2" | "claude-haiku-4-5" | "gemini-3-pro" | "alpha-glm-4.7" | "claude-sonnet-4-5" | "alpha-gd4" | "big-pickle" | "claude-3-5-haiku" | "glm-4.7-free" | "grok-code" | "gemini-3-flash" | "minimax-m2.1-free">;
|
|
191
221
|
/**
|
|
192
222
|
* Create a OVHcloud AI Endpoints Provider
|
|
193
223
|
* @see {@link https://www.ovhcloud.com/en/public-cloud/ai-endpoints/catalog//}
|
|
@@ -197,12 +227,17 @@ declare const createOvhcloud: (apiKey: string, baseURL?: string) => ModelProvide
|
|
|
197
227
|
* Create a Perplexity Provider
|
|
198
228
|
* @see {@link https://docs.perplexity.ai}
|
|
199
229
|
*/
|
|
200
|
-
declare const createPerplexity: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"sonar
|
|
230
|
+
declare const createPerplexity: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"sonar" | "sonar-pro" | "sonar-reasoning-pro">;
|
|
201
231
|
/**
|
|
202
232
|
* Create a Poe Provider
|
|
203
233
|
* @see {@link https://creator.poe.com/docs/external-applications/openai-compatible-api}
|
|
204
234
|
*/
|
|
205
|
-
declare const createPoe: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-
|
|
235
|
+
declare const createPoe: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5-nano" | "openai/gpt-4.1" | "openai/gpt-5-mini" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "xai/grok-3" | "xai/grok-3-mini" | "openai/gpt-4.1-nano" | "openai/gpt-4.1-mini" | "openai/o3-mini" | "openai/gpt-4o" | "openai/o4-mini" | "openai/o1" | "openai/o3" | "openai/gpt-4o-mini" | "xai/grok-4-fast-non-reasoning" | "xai/grok-4-fast-reasoning" | "xai/grok-4.1-fast-reasoning" | "xai/grok-4" | "xai/grok-code-fast-1" | "xai/grok-4.1-fast-non-reasoning" | "ideogramai/ideogram" | "ideogramai/ideogram-v2a" | "ideogramai/ideogram-v2a-turbo" | "ideogramai/ideogram-v2" | "runwayml/runway" | "runwayml/runway-gen-4-turbo" | "poetools/claude-code" | "elevenlabs/elevenlabs-v3" | "elevenlabs/elevenlabs-music" | "elevenlabs/elevenlabs-v2.5-turbo" | "google/gemini-deep-research" | "google/nano-banana" | "google/imagen-4" | "google/imagen-3" | "google/imagen-4-ultra" | "google/gemini-2.0-flash-lite" | "google/gemini-3-pro" | "google/veo-3.1" | "google/imagen-3-fast" | "google/lyria" | "google/gemini-2.0-flash" | "google/gemini-2.5-flash-lite" | "google/veo-3" | "google/veo-3-fast" | "google/imagen-4-fast" | "google/veo-2" | "google/gemini-3-flash" | "google/nano-banana-pro" | "google/veo-3.1-fast" | "openai/gpt-5.2-instant" | "openai/sora-2" | "openai/o1-pro" | "openai/gpt-5.1-codex" | "openai/gpt-3.5-turbo-raw" | "openai/gpt-4-classic" | "openai/gpt-5-chat" | "openai/o3-deep-research" | "openai/gpt-4o-search" | "openai/gpt-image-1.5" | "openai/gpt-image-1-mini" | "openai/gpt-3.5-turbo" | "openai/gpt-5.2-pro" | "openai/o3-mini-high" | "openai/chatgpt-4o-latest" | "openai/gpt-4-turbo" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1-instant" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-4o-aug" | "openai/o3-pro" | "openai/gpt-image-1" | "openai/gpt-5.1-codex-max" | "openai/gpt-3.5-turbo-instruct" | "openai/o4-mini-deep-research" | "openai/gpt-4-classic-0314" | "openai/dall-e-3" | "openai/sora-2-pro" | "openai/gpt-5-pro" | "openai/gpt-5.2" | "openai/gpt-4o-mini-search" | "stabilityai/stablediffusionxl" | "topazlabs-co/topazlabs" | "lumalabs/ray2" | "lumalabs/dream-machine" | "anthropic/claude-opus-3" | "anthropic/claude-opus-4" | "anthropic/claude-sonnet-3.7-reasoning" | "anthropic/claude-opus-4-search" | "anthropic/claude-sonnet-3.7" | "anthropic/claude-haiku-3.5-search" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4-reasoning" | "anthropic/claude-haiku-3" | "anthropic/claude-sonnet-3.7-search" | "anthropic/claude-opus-4-reasoning" | "anthropic/claude-sonnet-3.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-haiku-3.5" | "anthropic/claude-sonnet-3.5-june" | "anthropic/claude-sonnet-4.5" | "anthropic/claude-sonnet-4-search" | "trytako/tako" | "novita/glm-4.7" | "novita/kimi-k2-thinking" | "novita/kat-coder-pro" | "novita/glm-4.6" | "novita/minimax-m2.1" | "novita/glm-4.6v" | "cerebras/gpt-oss-120b-cs" | "cerebras/zai-glm-4.6-cs">;
|
|
236
|
+
/**
|
|
237
|
+
* Create a Privatemode AI Provider
|
|
238
|
+
* @see {@link https://docs.privatemode.ai/api/overview}
|
|
239
|
+
*/
|
|
240
|
+
declare const createPrivatemodeAi: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"gpt-oss-120b" | "whisper-large-v3" | "qwen3-embedding-4b" | "gemma-3-27b" | "qwen3-coder-30b-a3b">;
|
|
206
241
|
/**
|
|
207
242
|
* Create a Requesty Provider
|
|
208
243
|
* @see {@link https://requesty.ai/solution/llm-routing/models}
|
|
@@ -212,37 +247,42 @@ declare const createRequesty: (apiKey: string, baseURL?: string) => ModelProvide
|
|
|
212
247
|
* Create a Scaleway Provider
|
|
213
248
|
* @see {@link https://www.scaleway.com/en/docs/generative-apis/}
|
|
214
249
|
*/
|
|
215
|
-
declare const createScaleway: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "qwen3-235b-a22b-instruct-2507" | "
|
|
250
|
+
declare const createScaleway: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"qwen3-coder-30b-a3b-instruct" | "deepseek-r1-distill-llama-70b" | "gpt-oss-120b" | "llama-3.1-8b-instruct" | "llama-3.3-70b-instruct" | "mistral-nemo-instruct-2407" | "mistral-small-3.2-24b-instruct-2506" | "whisper-large-v3" | "qwen3-235b-a22b-instruct-2507" | "devstral-2-123b-instruct-2512" | "pixtral-12b-2409" | "voxtral-small-24b-2507" | "bge-multilingual-gemma2" | "gemma-3-27b-it">;
|
|
216
251
|
/**
|
|
217
252
|
* Create a SiliconFlow Provider
|
|
218
253
|
* @see {@link https://cloud.siliconflow.com/models}
|
|
219
254
|
*/
|
|
220
|
-
declare const createSiliconFlow: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"
|
|
255
|
+
declare const createSiliconFlow: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & SpeechProvider<string> & TranscriptionProvider<string> & ChatProvider<"openai/gpt-oss-120b" | "meta-llama/Meta-Llama-3.1-8B-Instruct" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/QwQ-32B" | "Qwen/Qwen3-32B" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.7" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-V3" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2" | "zai-org/GLM-4.5" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "inclusionAI/Ling-mini-2.0" | "inclusionAI/Ling-flash-2.0" | "inclusionAI/Ring-flash-2.0" | "moonshotai/Kimi-Dev-72B" | "tencent/Hunyuan-MT-7B" | "tencent/Hunyuan-A13B-Instruct" | "MiniMaxAI/MiniMax-M1-80k" | "THUDM/GLM-4-32B-0414" | "THUDM/GLM-4.1V-9B-Thinking" | "THUDM/GLM-Z1-9B-0414" | "THUDM/GLM-4-9B-0414" | "THUDM/GLM-Z1-32B-0414" | "stepfun-ai/step3" | "nex-agi/DeepSeek-V3.1-Nex-N1" | "baidu/ERNIE-4.5-300B-A47B" | "ByteDance-Seed/Seed-OSS-36B-Instruct" | "Qwen/Qwen3-VL-30B-A3B-Instruct" | "Qwen/Qwen3-Omni-30B-A3B-Captioner" | "Qwen/Qwen3-8B" | "Qwen/Qwen3-Omni-30B-A3B-Instruct" | "Qwen/Qwen3-VL-8B-Thinking" | "Qwen/Qwen2.5-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct-128K" | "Qwen/Qwen2.5-7B-Instruct" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen2.5-VL-7B-Instruct" | "Qwen/Qwen3-VL-8B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-VL-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Thinking" | "Qwen/Qwen3-Omni-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Instruct" | "Qwen/Qwen2.5-14B-Instruct" | "zai-org/GLM-4.5V" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" | "deepseek-ai/deepseek-vl2" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B" | "deepseek-ai/DeepSeek-V3.2-Exp" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B" | "deepseek-ai/DeepSeek-V3.1">;
|
|
221
256
|
/**
|
|
222
257
|
* Create a SiliconFlow (China) Provider
|
|
223
258
|
* @see {@link https://cloud.siliconflow.com/models}
|
|
224
259
|
*/
|
|
225
|
-
declare const createSiliconflowCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"
|
|
260
|
+
declare const createSiliconflowCn: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen2.5-72B-Instruct" | "Qwen/QwQ-32B" | "Qwen/Qwen3-32B" | "deepseek-ai/DeepSeek-R1" | "deepseek-ai/DeepSeek-V3.1-Terminus" | "deepseek-ai/DeepSeek-V3.2" | "moonshotai/Kimi-K2-Instruct-0905" | "moonshotai/Kimi-K2-Thinking" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "zai-org/GLM-4.6" | "openai/gpt-oss-20b" | "Qwen/Qwen3-30B-A3B" | "Qwen/Qwen3-14B" | "Qwen/Qwen2.5-VL-32B-Instruct" | "Qwen/Qwen2.5-Coder-32B-Instruct" | "Qwen/Qwen3-235B-A22B" | "Qwen/Qwen3-VL-235B-A22B-Instruct" | "Qwen/Qwen3-30B-A3B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "Qwen/Qwen3-Next-80B-A3B-Instruct" | "zai-org/GLM-4.6V" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-V3" | "moonshotai/Kimi-K2-Instruct" | "MiniMaxAI/MiniMax-M2" | "zai-org/GLM-4.5" | "Qwen/Qwen3-Next-80B-A3B-Thinking" | "Qwen/Qwen3-30B-A3B-Thinking-2507" | "Qwen/Qwen3-Coder-30B-A3B-Instruct" | "inclusionAI/Ling-mini-2.0" | "inclusionAI/Ling-flash-2.0" | "inclusionAI/Ring-flash-2.0" | "moonshotai/Kimi-Dev-72B" | "tencent/Hunyuan-MT-7B" | "tencent/Hunyuan-A13B-Instruct" | "MiniMaxAI/MiniMax-M1-80k" | "THUDM/GLM-4-32B-0414" | "THUDM/GLM-4.1V-9B-Thinking" | "THUDM/GLM-Z1-9B-0414" | "THUDM/GLM-4-9B-0414" | "THUDM/GLM-Z1-32B-0414" | "stepfun-ai/step3" | "nex-agi/DeepSeek-V3.1-Nex-N1" | "baidu/ERNIE-4.5-300B-A47B" | "ByteDance-Seed/Seed-OSS-36B-Instruct" | "Qwen/Qwen3-VL-30B-A3B-Instruct" | "Qwen/Qwen3-Omni-30B-A3B-Captioner" | "Qwen/Qwen3-8B" | "Qwen/Qwen3-Omni-30B-A3B-Instruct" | "Qwen/Qwen3-VL-8B-Thinking" | "Qwen/Qwen2.5-32B-Instruct" | "Qwen/Qwen2.5-72B-Instruct-128K" | "Qwen/Qwen2.5-7B-Instruct" | "Qwen/Qwen2.5-VL-72B-Instruct" | "Qwen/Qwen2.5-VL-7B-Instruct" | "Qwen/Qwen3-VL-8B-Instruct" | "Qwen/Qwen3-VL-235B-A22B-Thinking" | "Qwen/Qwen3-VL-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Thinking" | "Qwen/Qwen3-Omni-30B-A3B-Thinking" | "Qwen/Qwen3-VL-32B-Instruct" | "Qwen/Qwen2.5-14B-Instruct" | "zai-org/GLM-4.5V" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" | "deepseek-ai/deepseek-vl2" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B" | "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B" | "Kwaipilot/KAT-Dev" | "ascend-tribe/pangu-pro-moe">;
|
|
226
261
|
/**
|
|
227
262
|
* Create a submodel Provider
|
|
228
263
|
* @see {@link https://submodel.gitbook.io}
|
|
229
264
|
*/
|
|
230
|
-
declare const createSubmodel: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-
|
|
265
|
+
declare const createSubmodel: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"openai/gpt-oss-120b" | "Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "zai-org/GLM-4.5-Air" | "deepseek-ai/DeepSeek-R1-0528" | "deepseek-ai/DeepSeek-V3.1" | "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8" | "zai-org/GLM-4.5-FP8" | "deepseek-ai/DeepSeek-V3-0324">;
|
|
231
266
|
/**
|
|
232
267
|
* Create a Synthetic Provider
|
|
233
268
|
* @see {@link https://synthetic.new/pricing}
|
|
234
269
|
*/
|
|
235
|
-
declare const createSynthetic: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct" | "hf:moonshotai/Kimi-K2-
|
|
270
|
+
declare const createSynthetic: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"hf:Qwen/Qwen3-235B-A22B-Instruct-2507" | "hf:Qwen/Qwen2.5-Coder-32B-Instruct" | "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct" | "hf:Qwen/Qwen3-235B-A22B-Thinking-2507" | "hf:MiniMaxAI/MiniMax-M2" | "hf:MiniMaxAI/MiniMax-M2.1" | "hf:meta-llama/Llama-3.1-70B-Instruct" | "hf:meta-llama/Llama-3.1-8B-Instruct" | "hf:meta-llama/Llama-3.3-70B-Instruct" | "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct" | "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8" | "hf:meta-llama/Llama-3.1-405B-Instruct" | "hf:moonshotai/Kimi-K2-Instruct-0905" | "hf:moonshotai/Kimi-K2-Thinking" | "hf:zai-org/GLM-4.5" | "hf:zai-org/GLM-4.7" | "hf:zai-org/GLM-4.6" | "hf:deepseek-ai/DeepSeek-R1" | "hf:deepseek-ai/DeepSeek-R1-0528" | "hf:deepseek-ai/DeepSeek-V3.1-Terminus" | "hf:deepseek-ai/DeepSeek-V3.2" | "hf:deepseek-ai/DeepSeek-V3" | "hf:deepseek-ai/DeepSeek-V3.1" | "hf:deepseek-ai/DeepSeek-V3-0324" | "hf:openai/gpt-oss-120b">;
|
|
236
271
|
/**
|
|
237
272
|
* Create a Upstage Provider
|
|
238
273
|
* @see {@link https://developers.upstage.ai/docs/apis/chat}
|
|
239
274
|
*/
|
|
240
|
-
declare const createUpstage: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"solar-mini" | "solar-pro2">;
|
|
275
|
+
declare const createUpstage: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"solar-mini" | "solar-pro3" | "solar-pro2">;
|
|
241
276
|
/**
|
|
242
277
|
* Create a Venice AI Provider
|
|
243
278
|
* @see {@link https://docs.venice.ai}
|
|
244
279
|
*/
|
|
245
|
-
declare const createVenice: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"
|
|
280
|
+
declare const createVenice: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"gemini-3-flash-preview" | "grok-code-fast-1" | "gemini-3-pro-preview" | "qwen3-coder-480b-a35b-instruct" | "kimi-k2-thinking" | "deepseek-v3.2" | "qwen3-235b-a22b-thinking-2507" | "qwen3-235b-a22b-instruct-2507" | "grok-41-fast" | "claude-opus-45" | "mistral-31-24b" | "zai-org-glm-4.7" | "venice-uncensored" | "openai-gpt-52" | "qwen3-4b" | "llama-3.3-70b" | "claude-sonnet-45" | "openai-gpt-oss-120b" | "llama-3.2-3b" | "google-gemma-3-27b-it" | "hermes-3-llama-3.1-405b" | "zai-org-glm-4.6v" | "minimax-m21" | "qwen3-next-80b" | "openai-gpt-52-codex">;
|
|
281
|
+
/**
|
|
282
|
+
* Create a Vivgrid Provider
|
|
283
|
+
* @see {@link https://docs.vivgrid.com/models}
|
|
284
|
+
*/
|
|
285
|
+
declare const createVivgrid: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"gpt-5.1-codex">;
|
|
246
286
|
/**
|
|
247
287
|
* Create a Vultr Provider
|
|
248
288
|
* @see {@link https://api.vultrinference.com/}
|
|
@@ -252,12 +292,12 @@ declare const createVultr: (apiKey: string, baseURL?: string) => ModelProvider &
|
|
|
252
292
|
* Create a Weights & Biases Provider
|
|
253
293
|
* @see {@link https://weave-docs.wandb.ai/guides/integrations/inference/}
|
|
254
294
|
*/
|
|
255
|
-
declare const createWandb: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-
|
|
295
|
+
declare const createWandb: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"Qwen/Qwen3-235B-A22B-Instruct-2507" | "Qwen/Qwen3-Coder-480B-A35B-Instruct" | "Qwen/Qwen3-235B-A22B-Thinking-2507" | "moonshotai/Kimi-K2-Instruct" | "meta-llama/Llama-4-Scout-17B-16E-Instruct" | "deepseek-ai/DeepSeek-R1-0528" | "meta-llama/Llama-3.3-70B-Instruct" | "deepseek-ai/DeepSeek-V3-0324" | "microsoft/Phi-4-mini-instruct" | "meta-llama/Llama-3.1-8B-Instruct">;
|
|
256
296
|
/**
|
|
257
297
|
* Create a xAI Provider
|
|
258
298
|
* @see {@link https://docs.x.ai/docs/models}
|
|
259
299
|
*/
|
|
260
|
-
declare const createXai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"grok-code-fast-1" | "grok-4-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-
|
|
300
|
+
declare const createXai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"grok-4-fast-non-reasoning" | "grok-code-fast-1" | "grok-4-1-fast-non-reasoning" | "grok-4" | "grok-3" | "grok-3-mini" | "grok-3-fast" | "grok-2-vision" | "grok-2" | "grok-3-mini-fast-latest" | "grok-2-vision-1212" | "grok-4-fast" | "grok-2-latest" | "grok-4-1-fast" | "grok-2-1212" | "grok-3-fast-latest" | "grok-3-latest" | "grok-2-vision-latest" | "grok-vision-beta" | "grok-beta" | "grok-3-mini-latest" | "grok-3-mini-fast">;
|
|
261
301
|
/**
|
|
262
302
|
* Create a Xiaomi Provider
|
|
263
303
|
* @see {@link https://platform.xiaomimimo.com/#/docs}
|
|
@@ -267,32 +307,27 @@ declare const createXiaomi: (apiKey: string, baseURL?: string) => ModelProvider
|
|
|
267
307
|
* Create a Z.AI Provider
|
|
268
308
|
* @see {@link https://docs.z.ai/guides/overview/pricing}
|
|
269
309
|
*/
|
|
270
|
-
declare const createZai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
310
|
+
declare const createZai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
271
311
|
/**
|
|
272
312
|
* Create a Z.AI Coding Plan Provider
|
|
273
313
|
* @see {@link https://docs.z.ai/devpack/overview}
|
|
274
314
|
*/
|
|
275
|
-
declare const createZaiCodingPlan: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
315
|
+
declare const createZaiCodingPlan: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v">;
|
|
276
316
|
/**
|
|
277
317
|
* Create a ZenMux Provider
|
|
278
318
|
* @see {@link https://docs.zenmux.ai}
|
|
279
319
|
*/
|
|
280
|
-
declare const createZenmux: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "moonshotai/kimi-k2-thinking" | "openai/gpt-5-codex" | "anthropic/claude-haiku-4.5" | "anthropic/claude-sonnet-4.5" | "moonshotai/kimi-k2-thinking-turbo" | "
|
|
320
|
+
declare const createZenmux: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"x-ai/grok-4" | "google/gemini-2.5-flash" | "google/gemini-2.5-pro" | "openai/gpt-5" | "anthropic/claude-opus-4.1" | "anthropic/claude-sonnet-4" | "moonshotai/kimi-k2-thinking" | "minimax/minimax-m2.1" | "z-ai/glm-4.6" | "moonshotai/kimi-k2-0905" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-v3.2" | "minimax/minimax-m2" | "google/gemini-2.5-flash-lite" | "openai/gpt-5.1-codex" | "openai/gpt-5.1-codex-mini" | "openai/gpt-5.1" | "openai/gpt-5-codex" | "openai/gpt-5.2" | "anthropic/claude-opus-4" | "anthropic/claude-haiku-4.5" | "anthropic/claude-opus-4.5" | "anthropic/claude-sonnet-4.5" | "google/gemini-3-flash-preview" | "google/gemini-3-pro-preview" | "stepfun/step-3" | "moonshotai/kimi-k2-thinking-turbo" | "xiaomi/mimo-v2-flash-free" | "xiaomi/mimo-v2-flash" | "x-ai/grok-code-fast-1" | "x-ai/grok-4.1-fast-non-reasoning" | "x-ai/grok-4-fast" | "x-ai/grok-4.1-fast" | "deepseek/deepseek-chat" | "deepseek/deepseek-reasoner" | "google/gemini-3-flash-preview-free" | "volcengine/doubao-seed-code" | "volcengine/doubao-seed-1.8" | "openai/gpt-5.1-chat" | "baidu/ernie-5.0-thinking-preview" | "inclusionai/ring-1t" | "inclusionai/ling-1t" | "z-ai/glm-4.7" | "z-ai/glm-4.6v-flash-free" | "z-ai/glm-4.6v-flash" | "z-ai/glm-4.5" | "z-ai/glm-4.5-air" | "z-ai/glm-4.6v" | "qwen/qwen3-coder-plus" | "kuaishou/kat-coder-pro-v1-free" | "kuaishou/kat-coder-pro-v1">;
|
|
281
321
|
/**
|
|
282
322
|
* Create a Zhipu AI Provider
|
|
283
323
|
* @see {@link https://docs.z.ai/guides/overview/pricing}
|
|
284
324
|
*/
|
|
285
|
-
declare const createZhipuai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
325
|
+
declare const createZhipuai: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
286
326
|
/**
|
|
287
327
|
* Create a Zhipu AI Coding Plan Provider
|
|
288
328
|
* @see {@link https://docs.bigmodel.cn/cn/coding-plan/overview}
|
|
289
329
|
*/
|
|
290
|
-
declare const createZhipuaiCodingPlan: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
291
|
-
/**
|
|
292
|
-
* Create a Novita AI Provider
|
|
293
|
-
* @see {@link https://novita.ai/docs/guides/llm-api#api-integration}
|
|
294
|
-
*/
|
|
295
|
-
declare const createNovita: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<string>;
|
|
330
|
+
declare const createZhipuaiCodingPlan: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"glm-4.6" | "glm-4.7" | "glm-4.5-flash" | "glm-4.5" | "glm-4.5-air" | "glm-4.5v" | "glm-4.6v" | "glm-4.6v-flash">;
|
|
296
331
|
/**
|
|
297
332
|
* Create a StepFun Provider
|
|
298
333
|
* @see {@link https://www.stepfun.com}
|
|
@@ -326,4 +361,7 @@ declare const createFeatherless: (apiKey: string, baseURL?: string) => ModelProv
|
|
|
326
361
|
*/
|
|
327
362
|
declare const createTogetherAI: (apiKey: string, baseURL?: string) => ModelProvider & EmbedProvider<string> & ChatProvider<TogetheraiModels>;
|
|
328
363
|
|
|
329
|
-
|
|
364
|
+
/** @deprecated use `createNovitaAi` instead. */
|
|
365
|
+
declare const createNovita: (apiKey: string, baseURL?: string) => ModelProvider & ChatProvider<"deepseek/deepseek-v3.1" | "openai/gpt-oss-120b" | "zai-org/glm-4.7" | "zai-org/glm-4.5" | "zai-org/glm-4.6" | "openai/gpt-oss-20b" | "deepseek/deepseek-r1-0528" | "deepseek/deepseek-v3-0324" | "moonshotai/kimi-k2-instruct" | "meta-llama/llama-4-scout-17b-16e-instruct" | "moonshotai/kimi-k2-thinking" | "minimax/minimax-m2.1" | "qwen/qwen3-235b-a22b-thinking-2507" | "meta-llama/llama-3.3-70b-instruct" | "zai-org/glm-4.5-air" | "qwen/qwen3-235b-a22b-instruct-2507" | "qwen/qwen3-coder-480b-a35b-instruct" | "baichuan/baichuan-m2-32b" | "moonshotai/kimi-k2-0905" | "nousresearch/hermes-2-pro-llama-3-8b" | "paddlepaddle/paddleocr-vl" | "kwaipilot/kat-coder" | "kwaipilot/kat-coder-pro" | "xiaomimimo/mimo-v2-flash" | "deepseek/deepseek-prover-v2-671b" | "deepseek/deepseek-r1-0528-qwen3-8b" | "deepseek/deepseek-v3.1-terminus" | "deepseek/deepseek-r1-turbo" | "deepseek/deepseek-v3.2-exp" | "deepseek/deepseek-r1-distill-llama-70b" | "deepseek/deepseek-ocr" | "deepseek/deepseek-v3.2" | "deepseek/deepseek-v3-turbo" | "sao10k/l3-8b-lunaris" | "sao10k/L3-8B-Stheno-v3.2" | "sao10k/l31-70b-euryale-v2.2" | "sao10k/l3-70b-euryale-v2.1" | "skywork/r1v4-lite" | "minimaxai/minimax-m1-80k" | "minimax/minimax-m2" | "google/gemma-3-27b-it" | "microsoft/wizardlm-2-8x22b" | "gryphe/mythomax-l2-13b" | "baidu/ernie-4.5-vl-28b-a3b-thinking" | "baidu/ernie-4.5-300b-a47b-paddle" | "baidu/ernie-4.5-21B-a3b" | "baidu/ernie-4.5-21B-a3b-thinking" | "baidu/ernie-4.5-vl-424b-a47b" | "baidu/ernie-4.5-vl-28b-a3b" | "qwen/qwen3-vl-30b-a3b-thinking" | "qwen/qwen3-omni-30b-a3b-thinking" | "qwen/qwen3-next-80b-a3b-instruct" | "qwen/qwen3-8b-fp8" | "qwen/qwen2.5-vl-72b-instruct" | "qwen/qwen3-4b-fp8" | "qwen/qwen3-coder-30b-a3b-instruct" | "qwen/qwen3-vl-8b-instruct" | "qwen/qwen2.5-7b-instruct" | "qwen/qwen3-30b-a3b-fp8" | "qwen/qwen3-32b-fp8" | "qwen/qwen3-omni-30b-a3b-instruct" | "qwen/qwen-2.5-72b-instruct" | "qwen/qwen3-vl-235b-a22b-thinking" | "qwen/qwen-mt-plus" | "qwen/qwen3-max" | "qwen/qwen3-vl-235b-a22b-instruct" | "qwen/qwen3-vl-30b-a3b-instruct" | "qwen/qwen3-next-80b-a3b-thinking" | "qwen/qwen3-235b-a22b-fp8" | "mistralai/mistral-nemo" | "meta-llama/llama-3-70b-instruct" | "meta-llama/llama-3-8b-instruct" | "meta-llama/llama-3.1-8b-instruct" | "meta-llama/llama-4-maverick-17b-128e-instruct-fp8" | "zai-org/glm-4.5v" | "zai-org/glm-4.6v" | "zai-org/autoglm-phone-9b-multilingual">;
|
|
366
|
+
|
|
367
|
+
export { createAbacus, createAlibaba, createAlibabaCn, createBailing, createBaseten, createCerebras, createChutes, createCohere, createCortecs, createDeepSeek, createDeepinfra, createFastrouter, createFeatherless, createFireworks, createFirmware, createFriendli, createGithubCopilot, createGithubModels, createGoogleGenerativeAI, createGroq, createHelicone, createHuggingface, createIflowcn, createInception, createInference, createIoNet, createKimiForCoding, createLitellm, createLlama, createLmstudio, createLucidquery, createMinimax, createMinimaxCn, createMistral, createModelscope, createMoonshotai, createMoonshotaiCn, createMorph, createNanoGpt, createNebius, createNovita, createNovitaAi, createNvidia, createOllama, createOllamaCloud, createOpenAI, createOpencode, createOvhcloud, createPerplexity, createPoe, createPrivatemodeAi, createRequesty, createScaleway, createSiliconFlow, createSiliconflowCn, createStepfun, createSubmodel, createSynthetic, createTencentHunyuan, createTogetherAI, createUpstage, createVenice, createVivgrid, createVultr, createWandb, createXai, createXiaomi, createZai, createZaiCodingPlan, createZenmux, createZhipuai, createZhipuaiCodingPlan };
|
package/dist/create.js
CHANGED
|
@@ -1,3 +1,8 @@
|
|
|
1
|
-
|
|
1
|
+
import { M as createNovitaAi } from './together-ai-SWwaUOgP.js';
|
|
2
|
+
export { c as createAbacus, a as createAlibaba, b as createAlibabaCn, af as createAnthropic, d as createBailing, e as createBaseten, f as createCerebras, g as createChutes, h as createCohere, i as createCortecs, k as createDeepSeek, j as createDeepinfra, l as createFastrouter, ag as createFeatherless, m as createFireworks, n as createFirmware, o as createFriendli, p as createGithubCopilot, q as createGithubModels, r as createGoogleGenerativeAI, s as createGroq, t as createHelicone, u as createHuggingface, v as createIflowcn, w as createInception, x as createInference, y as createIoNet, z as createKimiForCoding, ae as createLitellm, A as createLlama, B as createLmstudio, C as createLucidquery, D as createMinimax, E as createMinimaxCn, F as createMistral, G as createModelscope, H as createMoonshotai, I as createMoonshotaiCn, J as createMorph, K as createNanoGpt, L as createNebius, N as createNvidia, ad as createOllama, O as createOllamaCloud, P as createOpenAI, ah as createOpenRouter, Q as createOpencode, R as createOvhcloud, S as createPerplexity, T as createPoe, U as createPrivatemodeAi, V as createRequesty, W as createScaleway, X as createSiliconFlow, Y as createSiliconflowCn, ab as createStepfun, Z as createSubmodel, _ as createSynthetic, ac as createTencentHunyuan, ai as createTogetherAI, $ as createUpstage, a0 as createVenice, a1 as createVivgrid, a2 as createVultr, a3 as createWandb, a4 as createXai, a5 as createXiaomi, a6 as createZai, a7 as createZaiCodingPlan, a8 as createZenmux, a9 as createZhipuai, aa as createZhipuaiCodingPlan } from './together-ai-SWwaUOgP.js';
|
|
2
3
|
import './utils/index.js';
|
|
3
4
|
import '@xsai/shared';
|
|
5
|
+
|
|
6
|
+
const createNovita = createNovitaAi;
|
|
7
|
+
|
|
8
|
+
export { createNovita, createNovitaAi };
|