@juspay/neurolink 8.8.0 → 8.10.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +17 -0
- package/dist/adapters/providerImageAdapter.js +131 -14
- package/dist/cli/factories/commandFactory.d.ts +1 -0
- package/dist/cli/factories/commandFactory.js +52 -2
- package/dist/constants/enums.d.ts +346 -15
- package/dist/constants/enums.js +476 -19
- package/dist/factories/providerRegistry.js +8 -7
- package/dist/lib/adapters/providerImageAdapter.js +131 -14
- package/dist/lib/constants/enums.d.ts +346 -15
- package/dist/lib/constants/enums.js +476 -19
- package/dist/lib/factories/providerRegistry.js +8 -7
- package/dist/lib/models/modelRegistry.js +1962 -210
- package/dist/lib/types/fileTypes.d.ts +10 -1
- package/dist/lib/types/generateTypes.d.ts +7 -0
- package/dist/lib/types/streamTypes.d.ts +7 -0
- package/dist/lib/utils/csvProcessor.d.ts +17 -0
- package/dist/lib/utils/csvProcessor.js +56 -4
- package/dist/models/modelRegistry.js +1962 -210
- package/dist/types/fileTypes.d.ts +10 -1
- package/dist/types/generateTypes.d.ts +7 -0
- package/dist/types/streamTypes.d.ts +7 -0
- package/dist/utils/csvProcessor.d.ts +17 -0
- package/dist/utils/csvProcessor.js +56 -4
- package/package.json +1 -1
|
@@ -20,39 +20,152 @@ export declare enum AIProviderName {
|
|
|
20
20
|
* Supported Models for Amazon Bedrock
|
|
21
21
|
*/
|
|
22
22
|
export declare enum BedrockModels {
|
|
23
|
-
CLAUDE_4_5_SONNET = "anthropic.claude-sonnet-4-5-20250929-v1:0",
|
|
24
23
|
CLAUDE_4_5_OPUS = "anthropic.claude-opus-4-5-20251124-v1:0",
|
|
24
|
+
CLAUDE_4_5_SONNET = "anthropic.claude-sonnet-4-5-20250929-v1:0",
|
|
25
25
|
CLAUDE_4_5_HAIKU = "anthropic.claude-haiku-4-5-20251001-v1:0",
|
|
26
|
+
CLAUDE_4_1_OPUS = "anthropic.claude-opus-4-1-20250805-v1:0",
|
|
27
|
+
CLAUDE_4_SONNET = "anthropic.claude-sonnet-4-20250514-v1:0",
|
|
26
28
|
CLAUDE_3_7_SONNET = "anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
27
29
|
CLAUDE_3_5_SONNET = "anthropic.claude-3-5-sonnet-20241022-v1:0",
|
|
30
|
+
CLAUDE_3_5_HAIKU = "anthropic.claude-3-5-haiku-20241022-v1:0",
|
|
28
31
|
CLAUDE_3_SONNET = "anthropic.claude-3-sonnet-20240229-v1:0",
|
|
29
|
-
CLAUDE_3_HAIKU = "anthropic.claude-3-haiku-20240307-v1:0"
|
|
32
|
+
CLAUDE_3_HAIKU = "anthropic.claude-3-haiku-20240307-v1:0",
|
|
33
|
+
NOVA_PREMIER = "amazon.nova-premier-v1:0",
|
|
34
|
+
NOVA_PRO = "amazon.nova-pro-v1:0",
|
|
35
|
+
NOVA_LITE = "amazon.nova-lite-v1:0",
|
|
36
|
+
NOVA_MICRO = "amazon.nova-micro-v1:0",
|
|
37
|
+
NOVA_2_LITE = "amazon.nova-2-lite-v1:0",
|
|
38
|
+
NOVA_2_SONIC = "amazon.nova-2-sonic-v1:0",
|
|
39
|
+
NOVA_SONIC = "amazon.nova-sonic-v1:0",
|
|
40
|
+
NOVA_CANVAS = "amazon.nova-canvas-v1:0",
|
|
41
|
+
NOVA_REEL = "amazon.nova-reel-v1:0",
|
|
42
|
+
NOVA_REEL_V1_1 = "amazon.nova-reel-v1:1",
|
|
43
|
+
NOVA_MULTIMODAL_EMBEDDINGS = "amazon.nova-2-multimodal-embeddings-v1:0",
|
|
44
|
+
TITAN_TEXT_LARGE = "amazon.titan-tg1-large",
|
|
45
|
+
TITAN_EMBED_TEXT_V2 = "amazon.titan-embed-text-v2:0",
|
|
46
|
+
TITAN_EMBED_TEXT_V1 = "amazon.titan-embed-text-v1",
|
|
47
|
+
TITAN_EMBED_G1_TEXT_02 = "amazon.titan-embed-g1-text-02",
|
|
48
|
+
TITAN_EMBED_IMAGE_V1 = "amazon.titan-embed-image-v1",
|
|
49
|
+
TITAN_IMAGE_GENERATOR_V2 = "amazon.titan-image-generator-v2:0",
|
|
50
|
+
LLAMA_4_MAVERICK_17B = "meta.llama4-maverick-17b-instruct-v1:0",
|
|
51
|
+
LLAMA_4_SCOUT_17B = "meta.llama4-scout-17b-instruct-v1:0",
|
|
52
|
+
LLAMA_3_3_70B = "meta.llama3-3-70b-instruct-v1:0",
|
|
53
|
+
LLAMA_3_2_90B = "meta.llama3-2-90b-instruct-v1:0",
|
|
54
|
+
LLAMA_3_2_11B = "meta.llama3-2-11b-instruct-v1:0",
|
|
55
|
+
LLAMA_3_2_3B = "meta.llama3-2-3b-instruct-v1:0",
|
|
56
|
+
LLAMA_3_2_1B = "meta.llama3-2-1b-instruct-v1:0",
|
|
57
|
+
LLAMA_3_1_405B = "meta.llama3-1-405b-instruct-v1:0",
|
|
58
|
+
LLAMA_3_1_70B = "meta.llama3-1-70b-instruct-v1:0",
|
|
59
|
+
LLAMA_3_1_8B = "meta.llama3-1-8b-instruct-v1:0",
|
|
60
|
+
LLAMA_3_70B = "meta.llama3-70b-instruct-v1:0",
|
|
61
|
+
LLAMA_3_8B = "meta.llama3-8b-instruct-v1:0",
|
|
62
|
+
MISTRAL_LARGE_3 = "mistral.mistral-large-3-675b-instruct",
|
|
63
|
+
MISTRAL_LARGE_2407 = "mistral.mistral-large-2407-v1:0",
|
|
64
|
+
MISTRAL_LARGE_2402 = "mistral.mistral-large-2402-v1:0",
|
|
65
|
+
MAGISTRAL_SMALL_2509 = "mistral.magistral-small-2509",
|
|
66
|
+
MINISTRAL_3_14B = "mistral.ministral-3-14b-instruct",
|
|
67
|
+
MINISTRAL_3_8B = "mistral.ministral-3-8b-instruct",
|
|
68
|
+
MINISTRAL_3_3B = "mistral.ministral-3-3b-instruct",
|
|
69
|
+
MISTRAL_7B = "mistral.mistral-7b-instruct-v0:2",
|
|
70
|
+
MIXTRAL_8x7B = "mistral.mixtral-8x7b-instruct-v0:1",
|
|
71
|
+
PIXTRAL_LARGE_2502 = "mistral.pixtral-large-2502-v1:0",
|
|
72
|
+
VOXTRAL_SMALL_24B = "mistral.voxtral-small-24b-2507",
|
|
73
|
+
VOXTRAL_MINI_3B = "mistral.voxtral-mini-3b-2507",
|
|
74
|
+
COHERE_COMMAND_R_PLUS = "cohere.command-r-plus-v1:0",
|
|
75
|
+
COHERE_COMMAND_R = "cohere.command-r-v1:0",
|
|
76
|
+
DEEPSEEK_R1 = "deepseek.r1-v1:0",
|
|
77
|
+
DEEPSEEK_V3 = "deepseek.v3-v1:0",
|
|
78
|
+
QWEN_3_235B_A22B = "qwen.qwen3-235b-a22b-2507-v1:0",
|
|
79
|
+
QWEN_3_CODER_480B_A35B = "qwen.qwen3-coder-480b-a35b-v1:0",
|
|
80
|
+
QWEN_3_CODER_30B_A3B = "qwen.qwen3-coder-30b-a3b-v1:0",
|
|
81
|
+
QWEN_3_32B = "qwen.qwen3-32b-v1:0",
|
|
82
|
+
QWEN_3_NEXT_80B_A3B = "qwen.qwen3-next-80b-a3b",
|
|
83
|
+
QWEN_3_VL_235B_A22B = "qwen.qwen3-vl-235b-a22b",
|
|
84
|
+
GEMMA_3_27B_IT = "google.gemma-3-27b-it",
|
|
85
|
+
GEMMA_3_12B_IT = "google.gemma-3-12b-it",
|
|
86
|
+
GEMMA_3_4B_IT = "google.gemma-3-4b-it",
|
|
87
|
+
JAMBA_1_5_LARGE = "ai21.jamba-1-5-large-v1:0",
|
|
88
|
+
JAMBA_1_5_MINI = "ai21.jamba-1-5-mini-v1:0"
|
|
30
89
|
}
|
|
31
90
|
/**
|
|
32
91
|
* Supported Models for OpenAI
|
|
33
92
|
*/
|
|
34
93
|
export declare enum OpenAIModels {
|
|
94
|
+
GPT_5_2 = "gpt-5.2",
|
|
95
|
+
GPT_5_2_CHAT_LATEST = "gpt-5.2-chat-latest",
|
|
96
|
+
GPT_5_2_PRO = "gpt-5.2-pro",
|
|
97
|
+
GPT_5 = "gpt-5",
|
|
98
|
+
GPT_5_MINI = "gpt-5-mini",
|
|
99
|
+
GPT_5_NANO = "gpt-5-nano",
|
|
100
|
+
GPT_4_1 = "gpt-4.1",
|
|
101
|
+
GPT_4_1_MINI = "gpt-4.1-mini",
|
|
102
|
+
GPT_4_1_NANO = "gpt-4.1-nano",
|
|
103
|
+
GPT_4O = "gpt-4o",
|
|
104
|
+
GPT_4O_MINI = "gpt-4o-mini",
|
|
105
|
+
O3 = "o3",
|
|
106
|
+
O3_MINI = "o3-mini",
|
|
107
|
+
O3_PRO = "o3-pro",
|
|
108
|
+
O4_MINI = "o4-mini",
|
|
109
|
+
O1 = "o1",
|
|
110
|
+
O1_PREVIEW = "o1-preview",
|
|
111
|
+
O1_MINI = "o1-mini",
|
|
35
112
|
GPT_4 = "gpt-4",
|
|
36
113
|
GPT_4_TURBO = "gpt-4-turbo",
|
|
114
|
+
GPT_3_5_TURBO = "gpt-3.5-turbo"
|
|
115
|
+
}
|
|
116
|
+
/**
|
|
117
|
+
* Supported Models for Azure OpenAI
|
|
118
|
+
* Note: Azure uses deployment names, these are model identifiers
|
|
119
|
+
*/
|
|
120
|
+
export declare enum AzureOpenAIModels {
|
|
121
|
+
GPT_5_1 = "gpt-5.1",
|
|
122
|
+
GPT_5_1_CHAT = "gpt-5.1-chat",
|
|
123
|
+
GPT_5_1_CODEX = "gpt-5.1-codex",
|
|
124
|
+
GPT_5_1_CODEX_MINI = "gpt-5.1-codex-mini",
|
|
125
|
+
GPT_5_1_CODEX_MAX = "gpt-5.1-codex-max",
|
|
126
|
+
GPT_5 = "gpt-5",
|
|
127
|
+
GPT_5_MINI = "gpt-5-mini",
|
|
128
|
+
GPT_5_NANO = "gpt-5-nano",
|
|
129
|
+
GPT_5_CHAT = "gpt-5-chat",
|
|
130
|
+
GPT_5_CODEX = "gpt-5-codex",
|
|
131
|
+
GPT_5_PRO = "gpt-5-pro",
|
|
132
|
+
GPT_5_TURBO = "gpt-5-turbo",
|
|
133
|
+
O4_MINI = "o4-mini",
|
|
134
|
+
O3 = "o3",
|
|
135
|
+
O3_MINI = "o3-mini",
|
|
136
|
+
O3_PRO = "o3-pro",
|
|
137
|
+
O1 = "o1",
|
|
138
|
+
O1_MINI = "o1-mini",
|
|
139
|
+
O1_PREVIEW = "o1-preview",
|
|
140
|
+
CODEX_MINI = "codex-mini",
|
|
141
|
+
GPT_4_1 = "gpt-4.1",
|
|
142
|
+
GPT_4_1_NANO = "gpt-4.1-nano",
|
|
143
|
+
GPT_4_1_MINI = "gpt-4.1-mini",
|
|
37
144
|
GPT_4O = "gpt-4o",
|
|
38
145
|
GPT_4O_MINI = "gpt-4o-mini",
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
|
|
146
|
+
GPT_4_TURBO = "gpt-4-turbo",
|
|
147
|
+
GPT_4 = "gpt-4",
|
|
148
|
+
GPT_4_32K = "gpt-4-32k",
|
|
149
|
+
GPT_3_5_TURBO = "gpt-35-turbo",
|
|
150
|
+
GPT_3_5_TURBO_INSTRUCT = "gpt-35-turbo-instruct"
|
|
42
151
|
}
|
|
43
152
|
/**
|
|
44
153
|
* Supported Models for Google Vertex AI
|
|
45
154
|
*/
|
|
46
155
|
export declare enum VertexModels {
|
|
47
|
-
CLAUDE_4_5_SONNET = "claude-sonnet-4-5@20250929",
|
|
48
156
|
CLAUDE_4_5_OPUS = "claude-opus-4-5@20251124",
|
|
157
|
+
CLAUDE_4_5_SONNET = "claude-sonnet-4-5@20250929",
|
|
158
|
+
CLAUDE_4_5_HAIKU = "claude-haiku-4-5@20251001",
|
|
49
159
|
CLAUDE_4_0_SONNET = "claude-sonnet-4@20250514",
|
|
50
160
|
CLAUDE_4_0_OPUS = "claude-opus-4@20250514",
|
|
161
|
+
CLAUDE_3_7_SONNET = "claude-3-7-sonnet@20250219",
|
|
51
162
|
CLAUDE_3_5_SONNET = "claude-3-5-sonnet-20241022",
|
|
52
163
|
CLAUDE_3_5_HAIKU = "claude-3-5-haiku-20241022",
|
|
53
164
|
CLAUDE_3_SONNET = "claude-3-sonnet-20240229",
|
|
54
165
|
CLAUDE_3_OPUS = "claude-3-opus-20240229",
|
|
55
166
|
CLAUDE_3_HAIKU = "claude-3-haiku-20240307",
|
|
167
|
+
/** Gemini 3 Pro - Base model with adaptive thinking */
|
|
168
|
+
GEMINI_3_PRO = "gemini-3-pro",
|
|
56
169
|
/** Gemini 3 Pro Preview - Versioned preview (November 2025) */
|
|
57
170
|
GEMINI_3_PRO_PREVIEW_11_2025 = "gemini-3-pro-preview-11-2025",
|
|
58
171
|
/** Gemini 3 Pro Latest - Auto-updated alias (always points to latest preview) */
|
|
@@ -62,43 +175,261 @@ export declare enum VertexModels {
|
|
|
62
175
|
GEMINI_2_5_PRO = "gemini-2.5-pro",
|
|
63
176
|
GEMINI_2_5_FLASH = "gemini-2.5-flash",
|
|
64
177
|
GEMINI_2_5_FLASH_LITE = "gemini-2.5-flash-lite",
|
|
178
|
+
GEMINI_2_5_FLASH_IMAGE = "gemini-2.5-flash-image",
|
|
179
|
+
GEMINI_2_0_FLASH = "gemini-2.0-flash",
|
|
65
180
|
GEMINI_2_0_FLASH_001 = "gemini-2.0-flash-001",
|
|
66
181
|
/** Gemini 2.0 Flash Lite - GA, production-ready, cost-optimized */
|
|
67
182
|
GEMINI_2_0_FLASH_LITE = "gemini-2.0-flash-lite",
|
|
68
|
-
GEMINI_1_5_PRO = "gemini-1.5-pro",
|
|
69
|
-
GEMINI_1_5_FLASH = "gemini-1.5-flash"
|
|
183
|
+
GEMINI_1_5_PRO = "gemini-1.5-pro-002",
|
|
184
|
+
GEMINI_1_5_FLASH = "gemini-1.5-flash-002"
|
|
70
185
|
}
|
|
71
186
|
/**
|
|
72
187
|
* Supported Models for Google AI Studio
|
|
73
188
|
*/
|
|
74
189
|
export declare enum GoogleAIModels {
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
/** Gemini 3 Pro Latest - Auto-updated alias (always points to latest preview) */
|
|
78
|
-
GEMINI_3_PRO_LATEST = "gemini-3-pro-latest",
|
|
190
|
+
GEMINI_3_PRO_PREVIEW = "gemini-3-pro-preview",
|
|
191
|
+
GEMINI_3_PRO_IMAGE_PREVIEW = "gemini-3-pro-image-preview",
|
|
79
192
|
GEMINI_2_5_PRO = "gemini-2.5-pro",
|
|
80
193
|
GEMINI_2_5_FLASH = "gemini-2.5-flash",
|
|
81
194
|
GEMINI_2_5_FLASH_LITE = "gemini-2.5-flash-lite",
|
|
195
|
+
GEMINI_2_5_FLASH_IMAGE = "gemini-2.5-flash-image",
|
|
196
|
+
GEMINI_2_5_FLASH_LIVE = "gemini-2.5-flash-native-audio-preview-09-2025",
|
|
197
|
+
GEMINI_2_0_FLASH = "gemini-2.0-flash",
|
|
82
198
|
GEMINI_2_0_FLASH_001 = "gemini-2.0-flash-001",
|
|
83
|
-
/** Gemini 2.0 Flash Lite - GA, production-ready, cost-optimized */
|
|
84
199
|
GEMINI_2_0_FLASH_LITE = "gemini-2.0-flash-lite",
|
|
200
|
+
GEMINI_2_0_FLASH_IMAGE = "gemini-2.0-flash-preview-image-generation",
|
|
85
201
|
GEMINI_1_5_PRO = "gemini-1.5-pro",
|
|
86
202
|
GEMINI_1_5_FLASH = "gemini-1.5-flash",
|
|
87
|
-
|
|
203
|
+
GEMINI_EMBEDDING = "gemini-embedding-001",
|
|
204
|
+
TEXT_EMBEDDING_004 = "text-embedding-004"
|
|
88
205
|
}
|
|
89
206
|
/**
|
|
90
207
|
* Supported Models for Anthropic (Direct API)
|
|
91
208
|
*/
|
|
92
209
|
export declare enum AnthropicModels {
|
|
210
|
+
CLAUDE_OPUS_4_5 = "claude-opus-4-5-20251101",
|
|
93
211
|
CLAUDE_SONNET_4_5 = "claude-sonnet-4-5-20250929",
|
|
94
|
-
CLAUDE_OPUS_4_5 = "claude-opus-4-5-20251124",
|
|
95
212
|
CLAUDE_4_5_HAIKU = "claude-haiku-4-5-20251001",
|
|
213
|
+
CLAUDE_OPUS_4_1 = "claude-opus-4-1-20250805",
|
|
214
|
+
CLAUDE_OPUS_4_0 = "claude-opus-4-20250514",
|
|
215
|
+
CLAUDE_SONNET_4_0 = "claude-sonnet-4-20250514",
|
|
216
|
+
CLAUDE_SONNET_3_7 = "claude-3-7-sonnet-20250219",
|
|
96
217
|
CLAUDE_3_5_SONNET = "claude-3-5-sonnet-20241022",
|
|
97
218
|
CLAUDE_3_5_HAIKU = "claude-3-5-haiku-20241022",
|
|
98
219
|
CLAUDE_3_SONNET = "claude-3-sonnet-20240229",
|
|
99
220
|
CLAUDE_3_OPUS = "claude-3-opus-20240229",
|
|
100
221
|
CLAUDE_3_HAIKU = "claude-3-haiku-20240307"
|
|
101
222
|
}
|
|
223
|
+
/**
|
|
224
|
+
* Supported Models for Mistral AI
|
|
225
|
+
*/
|
|
226
|
+
export declare enum MistralModels {
|
|
227
|
+
MISTRAL_LARGE_LATEST = "mistral-large-latest",
|
|
228
|
+
MISTRAL_LARGE_2512 = "mistral-large-2512",
|
|
229
|
+
MISTRAL_MEDIUM_LATEST = "mistral-medium-latest",
|
|
230
|
+
MISTRAL_MEDIUM_2508 = "mistral-medium-2508",
|
|
231
|
+
MISTRAL_SMALL_LATEST = "mistral-small-latest",
|
|
232
|
+
MISTRAL_SMALL_2506 = "mistral-small-2506",
|
|
233
|
+
MAGISTRAL_MEDIUM_LATEST = "magistral-medium-latest",
|
|
234
|
+
MAGISTRAL_SMALL_LATEST = "magistral-small-latest",
|
|
235
|
+
MINISTRAL_14B_2512 = "ministral-14b-2512",
|
|
236
|
+
MINISTRAL_8B_2512 = "ministral-8b-2512",
|
|
237
|
+
MINISTRAL_3B_2512 = "ministral-3b-2512",
|
|
238
|
+
CODESTRAL_LATEST = "codestral-latest",
|
|
239
|
+
CODESTRAL_2508 = "codestral-2508",
|
|
240
|
+
CODESTRAL_EMBED = "codestral-embed",
|
|
241
|
+
DEVSTRAL_MEDIUM_LATEST = "devstral-medium-latest",
|
|
242
|
+
DEVSTRAL_SMALL_LATEST = "devstral-small-latest",
|
|
243
|
+
PIXTRAL_LARGE = "pixtral-large",
|
|
244
|
+
PIXTRAL_12B = "pixtral-12b",
|
|
245
|
+
VOXTRAL_SMALL_LATEST = "voxtral-small-latest",
|
|
246
|
+
VOXTRAL_MINI_LATEST = "voxtral-mini-latest",
|
|
247
|
+
MISTRAL_NEMO = "mistral-nemo",
|
|
248
|
+
MISTRAL_EMBED = "mistral-embed",
|
|
249
|
+
MISTRAL_MODERATION_LATEST = "mistral-moderation-latest"
|
|
250
|
+
}
|
|
251
|
+
/**
|
|
252
|
+
* Supported Models for Ollama (Local)
|
|
253
|
+
* All models can be run locally without requiring API keys or cloud services
|
|
254
|
+
*/
|
|
255
|
+
export declare enum OllamaModels {
|
|
256
|
+
LLAMA4_SCOUT = "llama4:scout",
|
|
257
|
+
LLAMA4_MAVERICK = "llama4:maverick",
|
|
258
|
+
LLAMA4_LATEST = "llama4:latest",
|
|
259
|
+
LLAMA3_3_LATEST = "llama3.3:latest",
|
|
260
|
+
LLAMA3_3_70B = "llama3.3:70b",
|
|
261
|
+
LLAMA3_2_LATEST = "llama3.2:latest",
|
|
262
|
+
LLAMA3_2_3B = "llama3.2:3b",
|
|
263
|
+
LLAMA3_2_1B = "llama3.2:1b",
|
|
264
|
+
LLAMA3_1_8B = "llama3.1:8b",
|
|
265
|
+
LLAMA3_1_70B = "llama3.1:70b",
|
|
266
|
+
LLAMA3_1_405B = "llama3.1:405b",
|
|
267
|
+
QWEN3_4B = "qwen3:4b",
|
|
268
|
+
QWEN3_8B = "qwen3:8b",
|
|
269
|
+
QWEN3_14B = "qwen3:14b",
|
|
270
|
+
QWEN3_32B = "qwen3:32b",
|
|
271
|
+
QWEN3_72B = "qwen3:72b",
|
|
272
|
+
QWEN2_5_3B = "qwen2.5:3b",
|
|
273
|
+
QWEN2_5_7B = "qwen2.5:7b",
|
|
274
|
+
QWEN2_5_14B = "qwen2.5:14b",
|
|
275
|
+
QWEN2_5_32B = "qwen2.5:32b",
|
|
276
|
+
QWEN2_5_72B = "qwen2.5:72b",
|
|
277
|
+
QWQ_32B = "qwq:32b",
|
|
278
|
+
QWQ_LATEST = "qwq:latest",
|
|
279
|
+
DEEPSEEK_R1_1_5B = "deepseek-r1:1.5b",
|
|
280
|
+
DEEPSEEK_R1_7B = "deepseek-r1:7b",
|
|
281
|
+
DEEPSEEK_R1_8B = "deepseek-r1:8b",
|
|
282
|
+
DEEPSEEK_R1_14B = "deepseek-r1:14b",
|
|
283
|
+
DEEPSEEK_R1_32B = "deepseek-r1:32b",
|
|
284
|
+
DEEPSEEK_R1_70B = "deepseek-r1:70b",
|
|
285
|
+
DEEPSEEK_V3_671B = "deepseek-v3:671b",
|
|
286
|
+
DEEPSEEK_V3_LATEST = "deepseek-v3:latest",
|
|
287
|
+
MISTRAL_LATEST = "mistral:latest",
|
|
288
|
+
MISTRAL_7B = "mistral:7b",
|
|
289
|
+
MISTRAL_SMALL_LATEST = "mistral-small:latest",
|
|
290
|
+
MISTRAL_NEMO_LATEST = "mistral-nemo:latest",
|
|
291
|
+
MISTRAL_LARGE_LATEST = "mistral-large:latest",
|
|
292
|
+
GEMMA3_LATEST = "gemma3:latest",
|
|
293
|
+
GEMMA2_2B = "gemma2:2b",
|
|
294
|
+
GEMMA2_9B = "gemma2:9b",
|
|
295
|
+
GEMMA2_27B = "gemma2:27b",
|
|
296
|
+
PHI4_LATEST = "phi4:latest",
|
|
297
|
+
PHI4_14B = "phi4:14b",
|
|
298
|
+
PHI3_MINI = "phi3:mini",
|
|
299
|
+
PHI3_3_8B = "phi3:3.8b",
|
|
300
|
+
PHI3_MEDIUM = "phi3:medium",
|
|
301
|
+
PHI3_14B = "phi3:14b",
|
|
302
|
+
LLAVA_7B = "llava:7b",
|
|
303
|
+
LLAVA_13B = "llava:13b",
|
|
304
|
+
LLAVA_34B = "llava:34b",
|
|
305
|
+
LLAVA_LLAMA3_8B = "llava-llama3:8b",
|
|
306
|
+
CODELLAMA_7B = "codellama:7b",
|
|
307
|
+
CODELLAMA_13B = "codellama:13b",
|
|
308
|
+
CODELLAMA_34B = "codellama:34b",
|
|
309
|
+
CODELLAMA_70B = "codellama:70b",
|
|
310
|
+
QWEN2_5_CODER_7B = "qwen2.5-coder:7b",
|
|
311
|
+
QWEN2_5_CODER_32B = "qwen2.5-coder:32b",
|
|
312
|
+
STARCODER2_3B = "starcoder2:3b",
|
|
313
|
+
STARCODER2_7B = "starcoder2:7b",
|
|
314
|
+
STARCODER2_15B = "starcoder2:15b",
|
|
315
|
+
MIXTRAL_8X7B = "mixtral:8x7b",
|
|
316
|
+
MIXTRAL_8X22B = "mixtral:8x22b",
|
|
317
|
+
COMMAND_R_PLUS = "command-r-plus:104b"
|
|
318
|
+
}
|
|
319
|
+
/**
|
|
320
|
+
* Common Models for LiteLLM Proxy
|
|
321
|
+
* LiteLLM supports 100+ models through unified proxy interface
|
|
322
|
+
* Models use provider-specific prefixes (e.g., "openai/", "anthropic/")
|
|
323
|
+
*/
|
|
324
|
+
export declare enum LiteLLMModels {
|
|
325
|
+
OPENAI_GPT_5 = "openai/gpt-5",
|
|
326
|
+
OPENAI_GPT_4O = "openai/gpt-4o",
|
|
327
|
+
OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
|
328
|
+
OPENAI_GPT_4_TURBO = "openai/gpt-4-turbo",
|
|
329
|
+
OPENAI_GPT_4 = "openai/gpt-4",
|
|
330
|
+
OPENAI_GPT_3_5_TURBO = "openai/gpt-3.5-turbo",
|
|
331
|
+
ANTHROPIC_CLAUDE_SONNET_4_5 = "anthropic/claude-sonnet-4-5-20250929",
|
|
332
|
+
ANTHROPIC_CLAUDE_OPUS_4_1 = "anthropic/claude-opus-4-1-20250805",
|
|
333
|
+
ANTHROPIC_CLAUDE_3_5_SONNET = "anthropic/claude-3-5-sonnet-20240620",
|
|
334
|
+
ANTHROPIC_CLAUDE_3_HAIKU = "anthropic/claude-3-haiku-20240307",
|
|
335
|
+
VERTEX_GEMINI_2_5_PRO = "vertex_ai/gemini-2.5-pro",
|
|
336
|
+
VERTEX_GEMINI_1_5_PRO = "vertex_ai/gemini-1.5-pro",
|
|
337
|
+
VERTEX_GEMINI_1_5_FLASH = "vertex_ai/gemini-1.5-flash",
|
|
338
|
+
GEMINI_2_5_PRO = "gemini/gemini-2.5-pro",
|
|
339
|
+
GEMINI_2_0_FLASH = "gemini/gemini-2.0-flash",
|
|
340
|
+
GEMINI_1_5_PRO = "gemini/gemini-1.5-pro",
|
|
341
|
+
GEMINI_1_5_FLASH = "gemini/gemini-1.5-flash",
|
|
342
|
+
GROQ_LLAMA_3_1_70B_VERSATILE = "groq/llama-3.1-70b-versatile",
|
|
343
|
+
GROQ_LLAMA_3_1_8B_INSTANT = "groq/llama-3.1-8b-instant",
|
|
344
|
+
GROQ_LLAMA_3_2_11B_VISION = "groq/llama-3.2-11b-vision-preview",
|
|
345
|
+
GROQ_MIXTRAL_8X7B = "groq/mixtral-8x7b-32768",
|
|
346
|
+
TOGETHER_LLAMA_2_70B_CHAT = "together_ai/togethercomputer/llama-2-70b-chat",
|
|
347
|
+
TOGETHER_MIXTRAL_8X7B = "together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1",
|
|
348
|
+
TOGETHER_CODELLAMA_34B = "together_ai/codellama/CodeLlama-34b-Instruct-hf",
|
|
349
|
+
DEEPINFRA_LLAMA_3_70B = "deepinfra/meta-llama/Meta-Llama-3-70B-Instruct",
|
|
350
|
+
DEEPINFRA_LLAMA_2_70B = "deepinfra/meta-llama/Llama-2-70b-chat-hf",
|
|
351
|
+
DEEPINFRA_MISTRAL_7B = "deepinfra/mistralai/Mistral-7B-Instruct-v0.1",
|
|
352
|
+
MISTRAL_LARGE = "mistral/mistral-large-latest",
|
|
353
|
+
MISTRAL_SMALL = "mistral/mistral-small-latest",
|
|
354
|
+
MISTRAL_MAGISTRAL_MEDIUM = "mistral/magistral-medium-2506",
|
|
355
|
+
BEDROCK_CLAUDE_3_5_SONNET = "bedrock/anthropic.claude-3-5-sonnet-20240620-v1:0",
|
|
356
|
+
BEDROCK_CLAUDE_3_HAIKU = "bedrock/anthropic.claude-3-haiku-20240307-v1:0",
|
|
357
|
+
PERPLEXITY_SONAR_PRO = "perplexity/sonar-pro",
|
|
358
|
+
PERPLEXITY_SONAR_REASONING_PRO = "perplexity/sonar-reasoning-pro"
|
|
359
|
+
}
|
|
360
|
+
/**
|
|
361
|
+
* Supported Models for Hugging Face Inference API
|
|
362
|
+
*/
|
|
363
|
+
export declare enum HuggingFaceModels {
|
|
364
|
+
LLAMA_3_3_70B_INSTRUCT = "meta-llama/Llama-3.3-70B-Instruct",
|
|
365
|
+
LLAMA_3_2_1B = "meta-llama/Llama-3.2-1B",
|
|
366
|
+
LLAMA_3_2_3B_INSTRUCT = "meta-llama/Llama-3.2-3B-Instruct",
|
|
367
|
+
LLAMA_3_1_8B = "meta-llama/Llama-3.1-8B",
|
|
368
|
+
LLAMA_3_1_70B_INSTRUCT = "meta-llama/Llama-3.1-70B-Instruct",
|
|
369
|
+
LLAMA_3_1_405B_INSTRUCT = "meta-llama/Llama-3.1-405B-Instruct",
|
|
370
|
+
LLAMA_3_8B_INSTRUCT = "meta-llama/Meta-Llama-3-8B-Instruct",
|
|
371
|
+
LLAMA_3_70B_INSTRUCT = "meta-llama/Meta-Llama-3-70B-Instruct",
|
|
372
|
+
MISTRAL_LARGE_3_675B = "mistralai/Mistral-Large-3-675B-Instruct-2512",
|
|
373
|
+
MISTRAL_SMALL_3_1_24B = "mistralai/Mistral-Small-3.1-24B-Instruct-2503",
|
|
374
|
+
MISTRAL_SMALL_24B = "mistralai/Mistral-Small-24B-Instruct-2501",
|
|
375
|
+
MISTRAL_7B_INSTRUCT = "mistralai/Mistral-7B-Instruct-v0.2",
|
|
376
|
+
MIXTRAL_8X7B_INSTRUCT = "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
|
377
|
+
DEVSTRAL_2 = "mistralai/Devstral-2",
|
|
378
|
+
QWEN_2_5_7B = "Qwen/Qwen2.5-7B",
|
|
379
|
+
QWEN_2_5_32B = "Qwen/Qwen2.5-32B",
|
|
380
|
+
QWEN_2_5_72B_INSTRUCT = "Qwen/Qwen2.5-72B-Instruct",
|
|
381
|
+
QWEN_2_5_CODER_7B = "Qwen/Qwen2.5-Coder-7B",
|
|
382
|
+
QWEN_2_5_CODER_32B_INSTRUCT = "Qwen/Qwen2.5-Coder-32B-Instruct",
|
|
383
|
+
QWQ_32B = "Qwen/QwQ-32B",
|
|
384
|
+
QWEN_2_5_VL_32B = "Qwen/Qwen2.5-VL-32B-Instruct",
|
|
385
|
+
DEEPSEEK_R1 = "deepseek-ai/DeepSeek-R1",
|
|
386
|
+
DEEPSEEK_V3 = "deepseek-ai/DeepSeek-V3",
|
|
387
|
+
DEEPSEEK_V3_1 = "deepseek-ai/DeepSeek-V3.1",
|
|
388
|
+
DEEPSEEK_V3_2_EXP = "deepseek-ai/DeepSeek-V3.2-Exp",
|
|
389
|
+
PHI_4 = "microsoft/phi-4",
|
|
390
|
+
PHI_4_REASONING = "microsoft/Phi-4-reasoning",
|
|
391
|
+
PHI_4_MINI_INSTRUCT = "microsoft/Phi-4-mini-instruct",
|
|
392
|
+
PHI_4_MINI_REASONING = "microsoft/Phi-4-mini-reasoning",
|
|
393
|
+
PHI_3_MINI_128K_INSTRUCT = "microsoft/Phi-3-mini-128k-instruct",
|
|
394
|
+
PHI_3_VISION_128K_INSTRUCT = "microsoft/Phi-3-vision-128k-instruct",
|
|
395
|
+
GEMMA_3_270M = "google/gemma-3-270m",
|
|
396
|
+
GEMMA_3_1B_IT = "google/gemma-3-1b-it",
|
|
397
|
+
GEMMA_3_4B_IT = "google/gemma-3-4b-it",
|
|
398
|
+
GEMMA_3_12B_IT = "google/gemma-3-12b-it",
|
|
399
|
+
GEMMA_3_27B_IT = "google/gemma-3-27b-it",
|
|
400
|
+
GEMMA_2_9B = "google/gemma-2-9b",
|
|
401
|
+
GEMMA_2_27B = "google/gemma-2-27b",
|
|
402
|
+
GEMMA_2B = "google/gemma-2b",
|
|
403
|
+
GEMMA_7B = "google/gemma-7b",
|
|
404
|
+
FALCON_40B_INSTRUCT = "tiiuae/falcon-40b-instruct",
|
|
405
|
+
FALCON_180B_CHAT = "tiiuae/falcon-180B-chat",
|
|
406
|
+
STARCODER2_15B = "bigcode/starcoder2-15b",
|
|
407
|
+
CODELLAMA_34B_INSTRUCT = "codellama/CodeLlama-34b-Instruct-hf",
|
|
408
|
+
BLOOM_7B1 = "bigscience/bloom-7b1",
|
|
409
|
+
BLOOM_1B3 = "bigscience/bloom-1b3"
|
|
410
|
+
}
|
|
411
|
+
/**
|
|
412
|
+
* Supported Models for AWS SageMaker JumpStart
|
|
413
|
+
* https://docs.aws.amazon.com/sagemaker/latest/dg/jumpstart-foundation-models-latest.html
|
|
414
|
+
*/
|
|
415
|
+
export declare enum SageMakerModels {
|
|
416
|
+
LLAMA_4_SCOUT_17B_16E = "meta-llama-4-scout-17b-16e-instruct",
|
|
417
|
+
LLAMA_4_MAVERICK_17B_128E = "meta-llama-4-maverick-17b-128e-instruct",
|
|
418
|
+
LLAMA_4_MAVERICK_17B_128E_FP8 = "meta-llama-4-maverick-17b-128e-instruct-fp8",
|
|
419
|
+
LLAMA_3_8B = "meta-llama-3-8b-instruct",
|
|
420
|
+
LLAMA_3_70B = "meta-llama-3-70b-instruct",
|
|
421
|
+
CODE_LLAMA_7B = "meta-code-llama-7b",
|
|
422
|
+
CODE_LLAMA_13B = "meta-code-llama-13b",
|
|
423
|
+
CODE_LLAMA_34B = "meta-code-llama-34b",
|
|
424
|
+
MISTRAL_SMALL_24B = "mistral-small-24b-instruct-2501",
|
|
425
|
+
MISTRAL_7B_INSTRUCT = "mistral-7b-instruct-v0.3",
|
|
426
|
+
MIXTRAL_8X7B = "mistral-mixtral-8x7b-instruct-v0.1",
|
|
427
|
+
MIXTRAL_8X22B = "mistral-mixtral-8x22b-instruct-v0.1",
|
|
428
|
+
FALCON_3_7B = "tii-falcon-3-7b-instruct",
|
|
429
|
+
FALCON_3_10B = "tii-falcon-3-10b-instruct",
|
|
430
|
+
FALCON_40B = "tii-falcon-40b-instruct",
|
|
431
|
+
FALCON_180B = "tii-falcon-180b"
|
|
432
|
+
}
|
|
102
433
|
/**
|
|
103
434
|
* API Versions for various providers
|
|
104
435
|
*/
|