@juspay/neurolink 8.8.0 → 8.10.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -20,39 +20,152 @@ export declare enum AIProviderName {
20
20
  * Supported Models for Amazon Bedrock
21
21
  */
22
22
  export declare enum BedrockModels {
23
- CLAUDE_4_5_SONNET = "anthropic.claude-sonnet-4-5-20250929-v1:0",
24
23
  CLAUDE_4_5_OPUS = "anthropic.claude-opus-4-5-20251124-v1:0",
24
+ CLAUDE_4_5_SONNET = "anthropic.claude-sonnet-4-5-20250929-v1:0",
25
25
  CLAUDE_4_5_HAIKU = "anthropic.claude-haiku-4-5-20251001-v1:0",
26
+ CLAUDE_4_1_OPUS = "anthropic.claude-opus-4-1-20250805-v1:0",
27
+ CLAUDE_4_SONNET = "anthropic.claude-sonnet-4-20250514-v1:0",
26
28
  CLAUDE_3_7_SONNET = "anthropic.claude-3-7-sonnet-20250219-v1:0",
27
29
  CLAUDE_3_5_SONNET = "anthropic.claude-3-5-sonnet-20241022-v1:0",
30
+ CLAUDE_3_5_HAIKU = "anthropic.claude-3-5-haiku-20241022-v1:0",
28
31
  CLAUDE_3_SONNET = "anthropic.claude-3-sonnet-20240229-v1:0",
29
- CLAUDE_3_HAIKU = "anthropic.claude-3-haiku-20240307-v1:0"
32
+ CLAUDE_3_HAIKU = "anthropic.claude-3-haiku-20240307-v1:0",
33
+ NOVA_PREMIER = "amazon.nova-premier-v1:0",
34
+ NOVA_PRO = "amazon.nova-pro-v1:0",
35
+ NOVA_LITE = "amazon.nova-lite-v1:0",
36
+ NOVA_MICRO = "amazon.nova-micro-v1:0",
37
+ NOVA_2_LITE = "amazon.nova-2-lite-v1:0",
38
+ NOVA_2_SONIC = "amazon.nova-2-sonic-v1:0",
39
+ NOVA_SONIC = "amazon.nova-sonic-v1:0",
40
+ NOVA_CANVAS = "amazon.nova-canvas-v1:0",
41
+ NOVA_REEL = "amazon.nova-reel-v1:0",
42
+ NOVA_REEL_V1_1 = "amazon.nova-reel-v1:1",
43
+ NOVA_MULTIMODAL_EMBEDDINGS = "amazon.nova-2-multimodal-embeddings-v1:0",
44
+ TITAN_TEXT_LARGE = "amazon.titan-tg1-large",
45
+ TITAN_EMBED_TEXT_V2 = "amazon.titan-embed-text-v2:0",
46
+ TITAN_EMBED_TEXT_V1 = "amazon.titan-embed-text-v1",
47
+ TITAN_EMBED_G1_TEXT_02 = "amazon.titan-embed-g1-text-02",
48
+ TITAN_EMBED_IMAGE_V1 = "amazon.titan-embed-image-v1",
49
+ TITAN_IMAGE_GENERATOR_V2 = "amazon.titan-image-generator-v2:0",
50
+ LLAMA_4_MAVERICK_17B = "meta.llama4-maverick-17b-instruct-v1:0",
51
+ LLAMA_4_SCOUT_17B = "meta.llama4-scout-17b-instruct-v1:0",
52
+ LLAMA_3_3_70B = "meta.llama3-3-70b-instruct-v1:0",
53
+ LLAMA_3_2_90B = "meta.llama3-2-90b-instruct-v1:0",
54
+ LLAMA_3_2_11B = "meta.llama3-2-11b-instruct-v1:0",
55
+ LLAMA_3_2_3B = "meta.llama3-2-3b-instruct-v1:0",
56
+ LLAMA_3_2_1B = "meta.llama3-2-1b-instruct-v1:0",
57
+ LLAMA_3_1_405B = "meta.llama3-1-405b-instruct-v1:0",
58
+ LLAMA_3_1_70B = "meta.llama3-1-70b-instruct-v1:0",
59
+ LLAMA_3_1_8B = "meta.llama3-1-8b-instruct-v1:0",
60
+ LLAMA_3_70B = "meta.llama3-70b-instruct-v1:0",
61
+ LLAMA_3_8B = "meta.llama3-8b-instruct-v1:0",
62
+ MISTRAL_LARGE_3 = "mistral.mistral-large-3-675b-instruct",
63
+ MISTRAL_LARGE_2407 = "mistral.mistral-large-2407-v1:0",
64
+ MISTRAL_LARGE_2402 = "mistral.mistral-large-2402-v1:0",
65
+ MAGISTRAL_SMALL_2509 = "mistral.magistral-small-2509",
66
+ MINISTRAL_3_14B = "mistral.ministral-3-14b-instruct",
67
+ MINISTRAL_3_8B = "mistral.ministral-3-8b-instruct",
68
+ MINISTRAL_3_3B = "mistral.ministral-3-3b-instruct",
69
+ MISTRAL_7B = "mistral.mistral-7b-instruct-v0:2",
70
+ MIXTRAL_8x7B = "mistral.mixtral-8x7b-instruct-v0:1",
71
+ PIXTRAL_LARGE_2502 = "mistral.pixtral-large-2502-v1:0",
72
+ VOXTRAL_SMALL_24B = "mistral.voxtral-small-24b-2507",
73
+ VOXTRAL_MINI_3B = "mistral.voxtral-mini-3b-2507",
74
+ COHERE_COMMAND_R_PLUS = "cohere.command-r-plus-v1:0",
75
+ COHERE_COMMAND_R = "cohere.command-r-v1:0",
76
+ DEEPSEEK_R1 = "deepseek.r1-v1:0",
77
+ DEEPSEEK_V3 = "deepseek.v3-v1:0",
78
+ QWEN_3_235B_A22B = "qwen.qwen3-235b-a22b-2507-v1:0",
79
+ QWEN_3_CODER_480B_A35B = "qwen.qwen3-coder-480b-a35b-v1:0",
80
+ QWEN_3_CODER_30B_A3B = "qwen.qwen3-coder-30b-a3b-v1:0",
81
+ QWEN_3_32B = "qwen.qwen3-32b-v1:0",
82
+ QWEN_3_NEXT_80B_A3B = "qwen.qwen3-next-80b-a3b",
83
+ QWEN_3_VL_235B_A22B = "qwen.qwen3-vl-235b-a22b",
84
+ GEMMA_3_27B_IT = "google.gemma-3-27b-it",
85
+ GEMMA_3_12B_IT = "google.gemma-3-12b-it",
86
+ GEMMA_3_4B_IT = "google.gemma-3-4b-it",
87
+ JAMBA_1_5_LARGE = "ai21.jamba-1-5-large-v1:0",
88
+ JAMBA_1_5_MINI = "ai21.jamba-1-5-mini-v1:0"
30
89
  }
31
90
  /**
32
91
  * Supported Models for OpenAI
33
92
  */
34
93
  export declare enum OpenAIModels {
94
+ GPT_5_2 = "gpt-5.2",
95
+ GPT_5_2_CHAT_LATEST = "gpt-5.2-chat-latest",
96
+ GPT_5_2_PRO = "gpt-5.2-pro",
97
+ GPT_5 = "gpt-5",
98
+ GPT_5_MINI = "gpt-5-mini",
99
+ GPT_5_NANO = "gpt-5-nano",
100
+ GPT_4_1 = "gpt-4.1",
101
+ GPT_4_1_MINI = "gpt-4.1-mini",
102
+ GPT_4_1_NANO = "gpt-4.1-nano",
103
+ GPT_4O = "gpt-4o",
104
+ GPT_4O_MINI = "gpt-4o-mini",
105
+ O3 = "o3",
106
+ O3_MINI = "o3-mini",
107
+ O3_PRO = "o3-pro",
108
+ O4_MINI = "o4-mini",
109
+ O1 = "o1",
110
+ O1_PREVIEW = "o1-preview",
111
+ O1_MINI = "o1-mini",
35
112
  GPT_4 = "gpt-4",
36
113
  GPT_4_TURBO = "gpt-4-turbo",
114
+ GPT_3_5_TURBO = "gpt-3.5-turbo"
115
+ }
116
+ /**
117
+ * Supported Models for Azure OpenAI
118
+ * Note: Azure uses deployment names, these are model identifiers
119
+ */
120
+ export declare enum AzureOpenAIModels {
121
+ GPT_5_1 = "gpt-5.1",
122
+ GPT_5_1_CHAT = "gpt-5.1-chat",
123
+ GPT_5_1_CODEX = "gpt-5.1-codex",
124
+ GPT_5_1_CODEX_MINI = "gpt-5.1-codex-mini",
125
+ GPT_5_1_CODEX_MAX = "gpt-5.1-codex-max",
126
+ GPT_5 = "gpt-5",
127
+ GPT_5_MINI = "gpt-5-mini",
128
+ GPT_5_NANO = "gpt-5-nano",
129
+ GPT_5_CHAT = "gpt-5-chat",
130
+ GPT_5_CODEX = "gpt-5-codex",
131
+ GPT_5_PRO = "gpt-5-pro",
132
+ GPT_5_TURBO = "gpt-5-turbo",
133
+ O4_MINI = "o4-mini",
134
+ O3 = "o3",
135
+ O3_MINI = "o3-mini",
136
+ O3_PRO = "o3-pro",
137
+ O1 = "o1",
138
+ O1_MINI = "o1-mini",
139
+ O1_PREVIEW = "o1-preview",
140
+ CODEX_MINI = "codex-mini",
141
+ GPT_4_1 = "gpt-4.1",
142
+ GPT_4_1_NANO = "gpt-4.1-nano",
143
+ GPT_4_1_MINI = "gpt-4.1-mini",
37
144
  GPT_4O = "gpt-4o",
38
145
  GPT_4O_MINI = "gpt-4o-mini",
39
- GPT_3_5_TURBO = "gpt-3.5-turbo",
40
- O1_PREVIEW = "o1-preview",
41
- O1_MINI = "o1-mini"
146
+ GPT_4_TURBO = "gpt-4-turbo",
147
+ GPT_4 = "gpt-4",
148
+ GPT_4_32K = "gpt-4-32k",
149
+ GPT_3_5_TURBO = "gpt-35-turbo",
150
+ GPT_3_5_TURBO_INSTRUCT = "gpt-35-turbo-instruct"
42
151
  }
43
152
  /**
44
153
  * Supported Models for Google Vertex AI
45
154
  */
46
155
  export declare enum VertexModels {
47
- CLAUDE_4_5_SONNET = "claude-sonnet-4-5@20250929",
48
156
  CLAUDE_4_5_OPUS = "claude-opus-4-5@20251124",
157
+ CLAUDE_4_5_SONNET = "claude-sonnet-4-5@20250929",
158
+ CLAUDE_4_5_HAIKU = "claude-haiku-4-5@20251001",
49
159
  CLAUDE_4_0_SONNET = "claude-sonnet-4@20250514",
50
160
  CLAUDE_4_0_OPUS = "claude-opus-4@20250514",
161
+ CLAUDE_3_7_SONNET = "claude-3-7-sonnet@20250219",
51
162
  CLAUDE_3_5_SONNET = "claude-3-5-sonnet-20241022",
52
163
  CLAUDE_3_5_HAIKU = "claude-3-5-haiku-20241022",
53
164
  CLAUDE_3_SONNET = "claude-3-sonnet-20240229",
54
165
  CLAUDE_3_OPUS = "claude-3-opus-20240229",
55
166
  CLAUDE_3_HAIKU = "claude-3-haiku-20240307",
167
+ /** Gemini 3 Pro - Base model with adaptive thinking */
168
+ GEMINI_3_PRO = "gemini-3-pro",
56
169
  /** Gemini 3 Pro Preview - Versioned preview (November 2025) */
57
170
  GEMINI_3_PRO_PREVIEW_11_2025 = "gemini-3-pro-preview-11-2025",
58
171
  /** Gemini 3 Pro Latest - Auto-updated alias (always points to latest preview) */
@@ -62,43 +175,261 @@ export declare enum VertexModels {
62
175
  GEMINI_2_5_PRO = "gemini-2.5-pro",
63
176
  GEMINI_2_5_FLASH = "gemini-2.5-flash",
64
177
  GEMINI_2_5_FLASH_LITE = "gemini-2.5-flash-lite",
178
+ GEMINI_2_5_FLASH_IMAGE = "gemini-2.5-flash-image",
179
+ GEMINI_2_0_FLASH = "gemini-2.0-flash",
65
180
  GEMINI_2_0_FLASH_001 = "gemini-2.0-flash-001",
66
181
  /** Gemini 2.0 Flash Lite - GA, production-ready, cost-optimized */
67
182
  GEMINI_2_0_FLASH_LITE = "gemini-2.0-flash-lite",
68
- GEMINI_1_5_PRO = "gemini-1.5-pro",
69
- GEMINI_1_5_FLASH = "gemini-1.5-flash"
183
+ GEMINI_1_5_PRO = "gemini-1.5-pro-002",
184
+ GEMINI_1_5_FLASH = "gemini-1.5-flash-002"
70
185
  }
71
186
  /**
72
187
  * Supported Models for Google AI Studio
73
188
  */
74
189
  export declare enum GoogleAIModels {
75
- /** Gemini 3 Pro Preview - Versioned preview (November 2025) */
76
- GEMINI_3_PRO_PREVIEW_11_2025 = "gemini-3-pro-preview-11-2025",
77
- /** Gemini 3 Pro Latest - Auto-updated alias (always points to latest preview) */
78
- GEMINI_3_PRO_LATEST = "gemini-3-pro-latest",
190
+ GEMINI_3_PRO_PREVIEW = "gemini-3-pro-preview",
191
+ GEMINI_3_PRO_IMAGE_PREVIEW = "gemini-3-pro-image-preview",
79
192
  GEMINI_2_5_PRO = "gemini-2.5-pro",
80
193
  GEMINI_2_5_FLASH = "gemini-2.5-flash",
81
194
  GEMINI_2_5_FLASH_LITE = "gemini-2.5-flash-lite",
195
+ GEMINI_2_5_FLASH_IMAGE = "gemini-2.5-flash-image",
196
+ GEMINI_2_5_FLASH_LIVE = "gemini-2.5-flash-native-audio-preview-09-2025",
197
+ GEMINI_2_0_FLASH = "gemini-2.0-flash",
82
198
  GEMINI_2_0_FLASH_001 = "gemini-2.0-flash-001",
83
- /** Gemini 2.0 Flash Lite - GA, production-ready, cost-optimized */
84
199
  GEMINI_2_0_FLASH_LITE = "gemini-2.0-flash-lite",
200
+ GEMINI_2_0_FLASH_IMAGE = "gemini-2.0-flash-preview-image-generation",
85
201
  GEMINI_1_5_PRO = "gemini-1.5-pro",
86
202
  GEMINI_1_5_FLASH = "gemini-1.5-flash",
87
- GEMINI_1_5_FLASH_LITE = "gemini-1.5-flash-lite"
203
+ GEMINI_EMBEDDING = "gemini-embedding-001",
204
+ TEXT_EMBEDDING_004 = "text-embedding-004"
88
205
  }
89
206
  /**
90
207
  * Supported Models for Anthropic (Direct API)
91
208
  */
92
209
  export declare enum AnthropicModels {
210
+ CLAUDE_OPUS_4_5 = "claude-opus-4-5-20251101",
93
211
  CLAUDE_SONNET_4_5 = "claude-sonnet-4-5-20250929",
94
- CLAUDE_OPUS_4_5 = "claude-opus-4-5-20251124",
95
212
  CLAUDE_4_5_HAIKU = "claude-haiku-4-5-20251001",
213
+ CLAUDE_OPUS_4_1 = "claude-opus-4-1-20250805",
214
+ CLAUDE_OPUS_4_0 = "claude-opus-4-20250514",
215
+ CLAUDE_SONNET_4_0 = "claude-sonnet-4-20250514",
216
+ CLAUDE_SONNET_3_7 = "claude-3-7-sonnet-20250219",
96
217
  CLAUDE_3_5_SONNET = "claude-3-5-sonnet-20241022",
97
218
  CLAUDE_3_5_HAIKU = "claude-3-5-haiku-20241022",
98
219
  CLAUDE_3_SONNET = "claude-3-sonnet-20240229",
99
220
  CLAUDE_3_OPUS = "claude-3-opus-20240229",
100
221
  CLAUDE_3_HAIKU = "claude-3-haiku-20240307"
101
222
  }
223
+ /**
224
+ * Supported Models for Mistral AI
225
+ */
226
+ export declare enum MistralModels {
227
+ MISTRAL_LARGE_LATEST = "mistral-large-latest",
228
+ MISTRAL_LARGE_2512 = "mistral-large-2512",
229
+ MISTRAL_MEDIUM_LATEST = "mistral-medium-latest",
230
+ MISTRAL_MEDIUM_2508 = "mistral-medium-2508",
231
+ MISTRAL_SMALL_LATEST = "mistral-small-latest",
232
+ MISTRAL_SMALL_2506 = "mistral-small-2506",
233
+ MAGISTRAL_MEDIUM_LATEST = "magistral-medium-latest",
234
+ MAGISTRAL_SMALL_LATEST = "magistral-small-latest",
235
+ MINISTRAL_14B_2512 = "ministral-14b-2512",
236
+ MINISTRAL_8B_2512 = "ministral-8b-2512",
237
+ MINISTRAL_3B_2512 = "ministral-3b-2512",
238
+ CODESTRAL_LATEST = "codestral-latest",
239
+ CODESTRAL_2508 = "codestral-2508",
240
+ CODESTRAL_EMBED = "codestral-embed",
241
+ DEVSTRAL_MEDIUM_LATEST = "devstral-medium-latest",
242
+ DEVSTRAL_SMALL_LATEST = "devstral-small-latest",
243
+ PIXTRAL_LARGE = "pixtral-large",
244
+ PIXTRAL_12B = "pixtral-12b",
245
+ VOXTRAL_SMALL_LATEST = "voxtral-small-latest",
246
+ VOXTRAL_MINI_LATEST = "voxtral-mini-latest",
247
+ MISTRAL_NEMO = "mistral-nemo",
248
+ MISTRAL_EMBED = "mistral-embed",
249
+ MISTRAL_MODERATION_LATEST = "mistral-moderation-latest"
250
+ }
251
+ /**
252
+ * Supported Models for Ollama (Local)
253
+ * All models can be run locally without requiring API keys or cloud services
254
+ */
255
+ export declare enum OllamaModels {
256
+ LLAMA4_SCOUT = "llama4:scout",
257
+ LLAMA4_MAVERICK = "llama4:maverick",
258
+ LLAMA4_LATEST = "llama4:latest",
259
+ LLAMA3_3_LATEST = "llama3.3:latest",
260
+ LLAMA3_3_70B = "llama3.3:70b",
261
+ LLAMA3_2_LATEST = "llama3.2:latest",
262
+ LLAMA3_2_3B = "llama3.2:3b",
263
+ LLAMA3_2_1B = "llama3.2:1b",
264
+ LLAMA3_1_8B = "llama3.1:8b",
265
+ LLAMA3_1_70B = "llama3.1:70b",
266
+ LLAMA3_1_405B = "llama3.1:405b",
267
+ QWEN3_4B = "qwen3:4b",
268
+ QWEN3_8B = "qwen3:8b",
269
+ QWEN3_14B = "qwen3:14b",
270
+ QWEN3_32B = "qwen3:32b",
271
+ QWEN3_72B = "qwen3:72b",
272
+ QWEN2_5_3B = "qwen2.5:3b",
273
+ QWEN2_5_7B = "qwen2.5:7b",
274
+ QWEN2_5_14B = "qwen2.5:14b",
275
+ QWEN2_5_32B = "qwen2.5:32b",
276
+ QWEN2_5_72B = "qwen2.5:72b",
277
+ QWQ_32B = "qwq:32b",
278
+ QWQ_LATEST = "qwq:latest",
279
+ DEEPSEEK_R1_1_5B = "deepseek-r1:1.5b",
280
+ DEEPSEEK_R1_7B = "deepseek-r1:7b",
281
+ DEEPSEEK_R1_8B = "deepseek-r1:8b",
282
+ DEEPSEEK_R1_14B = "deepseek-r1:14b",
283
+ DEEPSEEK_R1_32B = "deepseek-r1:32b",
284
+ DEEPSEEK_R1_70B = "deepseek-r1:70b",
285
+ DEEPSEEK_V3_671B = "deepseek-v3:671b",
286
+ DEEPSEEK_V3_LATEST = "deepseek-v3:latest",
287
+ MISTRAL_LATEST = "mistral:latest",
288
+ MISTRAL_7B = "mistral:7b",
289
+ MISTRAL_SMALL_LATEST = "mistral-small:latest",
290
+ MISTRAL_NEMO_LATEST = "mistral-nemo:latest",
291
+ MISTRAL_LARGE_LATEST = "mistral-large:latest",
292
+ GEMMA3_LATEST = "gemma3:latest",
293
+ GEMMA2_2B = "gemma2:2b",
294
+ GEMMA2_9B = "gemma2:9b",
295
+ GEMMA2_27B = "gemma2:27b",
296
+ PHI4_LATEST = "phi4:latest",
297
+ PHI4_14B = "phi4:14b",
298
+ PHI3_MINI = "phi3:mini",
299
+ PHI3_3_8B = "phi3:3.8b",
300
+ PHI3_MEDIUM = "phi3:medium",
301
+ PHI3_14B = "phi3:14b",
302
+ LLAVA_7B = "llava:7b",
303
+ LLAVA_13B = "llava:13b",
304
+ LLAVA_34B = "llava:34b",
305
+ LLAVA_LLAMA3_8B = "llava-llama3:8b",
306
+ CODELLAMA_7B = "codellama:7b",
307
+ CODELLAMA_13B = "codellama:13b",
308
+ CODELLAMA_34B = "codellama:34b",
309
+ CODELLAMA_70B = "codellama:70b",
310
+ QWEN2_5_CODER_7B = "qwen2.5-coder:7b",
311
+ QWEN2_5_CODER_32B = "qwen2.5-coder:32b",
312
+ STARCODER2_3B = "starcoder2:3b",
313
+ STARCODER2_7B = "starcoder2:7b",
314
+ STARCODER2_15B = "starcoder2:15b",
315
+ MIXTRAL_8X7B = "mixtral:8x7b",
316
+ MIXTRAL_8X22B = "mixtral:8x22b",
317
+ COMMAND_R_PLUS = "command-r-plus:104b"
318
+ }
319
+ /**
320
+ * Common Models for LiteLLM Proxy
321
+ * LiteLLM supports 100+ models through unified proxy interface
322
+ * Models use provider-specific prefixes (e.g., "openai/", "anthropic/")
323
+ */
324
+ export declare enum LiteLLMModels {
325
+ OPENAI_GPT_5 = "openai/gpt-5",
326
+ OPENAI_GPT_4O = "openai/gpt-4o",
327
+ OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
328
+ OPENAI_GPT_4_TURBO = "openai/gpt-4-turbo",
329
+ OPENAI_GPT_4 = "openai/gpt-4",
330
+ OPENAI_GPT_3_5_TURBO = "openai/gpt-3.5-turbo",
331
+ ANTHROPIC_CLAUDE_SONNET_4_5 = "anthropic/claude-sonnet-4-5-20250929",
332
+ ANTHROPIC_CLAUDE_OPUS_4_1 = "anthropic/claude-opus-4-1-20250805",
333
+ ANTHROPIC_CLAUDE_3_5_SONNET = "anthropic/claude-3-5-sonnet-20240620",
334
+ ANTHROPIC_CLAUDE_3_HAIKU = "anthropic/claude-3-haiku-20240307",
335
+ VERTEX_GEMINI_2_5_PRO = "vertex_ai/gemini-2.5-pro",
336
+ VERTEX_GEMINI_1_5_PRO = "vertex_ai/gemini-1.5-pro",
337
+ VERTEX_GEMINI_1_5_FLASH = "vertex_ai/gemini-1.5-flash",
338
+ GEMINI_2_5_PRO = "gemini/gemini-2.5-pro",
339
+ GEMINI_2_0_FLASH = "gemini/gemini-2.0-flash",
340
+ GEMINI_1_5_PRO = "gemini/gemini-1.5-pro",
341
+ GEMINI_1_5_FLASH = "gemini/gemini-1.5-flash",
342
+ GROQ_LLAMA_3_1_70B_VERSATILE = "groq/llama-3.1-70b-versatile",
343
+ GROQ_LLAMA_3_1_8B_INSTANT = "groq/llama-3.1-8b-instant",
344
+ GROQ_LLAMA_3_2_11B_VISION = "groq/llama-3.2-11b-vision-preview",
345
+ GROQ_MIXTRAL_8X7B = "groq/mixtral-8x7b-32768",
346
+ TOGETHER_LLAMA_2_70B_CHAT = "together_ai/togethercomputer/llama-2-70b-chat",
347
+ TOGETHER_MIXTRAL_8X7B = "together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1",
348
+ TOGETHER_CODELLAMA_34B = "together_ai/codellama/CodeLlama-34b-Instruct-hf",
349
+ DEEPINFRA_LLAMA_3_70B = "deepinfra/meta-llama/Meta-Llama-3-70B-Instruct",
350
+ DEEPINFRA_LLAMA_2_70B = "deepinfra/meta-llama/Llama-2-70b-chat-hf",
351
+ DEEPINFRA_MISTRAL_7B = "deepinfra/mistralai/Mistral-7B-Instruct-v0.1",
352
+ MISTRAL_LARGE = "mistral/mistral-large-latest",
353
+ MISTRAL_SMALL = "mistral/mistral-small-latest",
354
+ MISTRAL_MAGISTRAL_MEDIUM = "mistral/magistral-medium-2506",
355
+ BEDROCK_CLAUDE_3_5_SONNET = "bedrock/anthropic.claude-3-5-sonnet-20240620-v1:0",
356
+ BEDROCK_CLAUDE_3_HAIKU = "bedrock/anthropic.claude-3-haiku-20240307-v1:0",
357
+ PERPLEXITY_SONAR_PRO = "perplexity/sonar-pro",
358
+ PERPLEXITY_SONAR_REASONING_PRO = "perplexity/sonar-reasoning-pro"
359
+ }
360
+ /**
361
+ * Supported Models for Hugging Face Inference API
362
+ */
363
+ export declare enum HuggingFaceModels {
364
+ LLAMA_3_3_70B_INSTRUCT = "meta-llama/Llama-3.3-70B-Instruct",
365
+ LLAMA_3_2_1B = "meta-llama/Llama-3.2-1B",
366
+ LLAMA_3_2_3B_INSTRUCT = "meta-llama/Llama-3.2-3B-Instruct",
367
+ LLAMA_3_1_8B = "meta-llama/Llama-3.1-8B",
368
+ LLAMA_3_1_70B_INSTRUCT = "meta-llama/Llama-3.1-70B-Instruct",
369
+ LLAMA_3_1_405B_INSTRUCT = "meta-llama/Llama-3.1-405B-Instruct",
370
+ LLAMA_3_8B_INSTRUCT = "meta-llama/Meta-Llama-3-8B-Instruct",
371
+ LLAMA_3_70B_INSTRUCT = "meta-llama/Meta-Llama-3-70B-Instruct",
372
+ MISTRAL_LARGE_3_675B = "mistralai/Mistral-Large-3-675B-Instruct-2512",
373
+ MISTRAL_SMALL_3_1_24B = "mistralai/Mistral-Small-3.1-24B-Instruct-2503",
374
+ MISTRAL_SMALL_24B = "mistralai/Mistral-Small-24B-Instruct-2501",
375
+ MISTRAL_7B_INSTRUCT = "mistralai/Mistral-7B-Instruct-v0.2",
376
+ MIXTRAL_8X7B_INSTRUCT = "mistralai/Mixtral-8x7B-Instruct-v0.1",
377
+ DEVSTRAL_2 = "mistralai/Devstral-2",
378
+ QWEN_2_5_7B = "Qwen/Qwen2.5-7B",
379
+ QWEN_2_5_32B = "Qwen/Qwen2.5-32B",
380
+ QWEN_2_5_72B_INSTRUCT = "Qwen/Qwen2.5-72B-Instruct",
381
+ QWEN_2_5_CODER_7B = "Qwen/Qwen2.5-Coder-7B",
382
+ QWEN_2_5_CODER_32B_INSTRUCT = "Qwen/Qwen2.5-Coder-32B-Instruct",
383
+ QWQ_32B = "Qwen/QwQ-32B",
384
+ QWEN_2_5_VL_32B = "Qwen/Qwen2.5-VL-32B-Instruct",
385
+ DEEPSEEK_R1 = "deepseek-ai/DeepSeek-R1",
386
+ DEEPSEEK_V3 = "deepseek-ai/DeepSeek-V3",
387
+ DEEPSEEK_V3_1 = "deepseek-ai/DeepSeek-V3.1",
388
+ DEEPSEEK_V3_2_EXP = "deepseek-ai/DeepSeek-V3.2-Exp",
389
+ PHI_4 = "microsoft/phi-4",
390
+ PHI_4_REASONING = "microsoft/Phi-4-reasoning",
391
+ PHI_4_MINI_INSTRUCT = "microsoft/Phi-4-mini-instruct",
392
+ PHI_4_MINI_REASONING = "microsoft/Phi-4-mini-reasoning",
393
+ PHI_3_MINI_128K_INSTRUCT = "microsoft/Phi-3-mini-128k-instruct",
394
+ PHI_3_VISION_128K_INSTRUCT = "microsoft/Phi-3-vision-128k-instruct",
395
+ GEMMA_3_270M = "google/gemma-3-270m",
396
+ GEMMA_3_1B_IT = "google/gemma-3-1b-it",
397
+ GEMMA_3_4B_IT = "google/gemma-3-4b-it",
398
+ GEMMA_3_12B_IT = "google/gemma-3-12b-it",
399
+ GEMMA_3_27B_IT = "google/gemma-3-27b-it",
400
+ GEMMA_2_9B = "google/gemma-2-9b",
401
+ GEMMA_2_27B = "google/gemma-2-27b",
402
+ GEMMA_2B = "google/gemma-2b",
403
+ GEMMA_7B = "google/gemma-7b",
404
+ FALCON_40B_INSTRUCT = "tiiuae/falcon-40b-instruct",
405
+ FALCON_180B_CHAT = "tiiuae/falcon-180B-chat",
406
+ STARCODER2_15B = "bigcode/starcoder2-15b",
407
+ CODELLAMA_34B_INSTRUCT = "codellama/CodeLlama-34b-Instruct-hf",
408
+ BLOOM_7B1 = "bigscience/bloom-7b1",
409
+ BLOOM_1B3 = "bigscience/bloom-1b3"
410
+ }
411
+ /**
412
+ * Supported Models for AWS SageMaker JumpStart
413
+ * https://docs.aws.amazon.com/sagemaker/latest/dg/jumpstart-foundation-models-latest.html
414
+ */
415
+ export declare enum SageMakerModels {
416
+ LLAMA_4_SCOUT_17B_16E = "meta-llama-4-scout-17b-16e-instruct",
417
+ LLAMA_4_MAVERICK_17B_128E = "meta-llama-4-maverick-17b-128e-instruct",
418
+ LLAMA_4_MAVERICK_17B_128E_FP8 = "meta-llama-4-maverick-17b-128e-instruct-fp8",
419
+ LLAMA_3_8B = "meta-llama-3-8b-instruct",
420
+ LLAMA_3_70B = "meta-llama-3-70b-instruct",
421
+ CODE_LLAMA_7B = "meta-code-llama-7b",
422
+ CODE_LLAMA_13B = "meta-code-llama-13b",
423
+ CODE_LLAMA_34B = "meta-code-llama-34b",
424
+ MISTRAL_SMALL_24B = "mistral-small-24b-instruct-2501",
425
+ MISTRAL_7B_INSTRUCT = "mistral-7b-instruct-v0.3",
426
+ MIXTRAL_8X7B = "mistral-mixtral-8x7b-instruct-v0.1",
427
+ MIXTRAL_8X22B = "mistral-mixtral-8x22b-instruct-v0.1",
428
+ FALCON_3_7B = "tii-falcon-3-7b-instruct",
429
+ FALCON_3_10B = "tii-falcon-3-10b-instruct",
430
+ FALCON_40B = "tii-falcon-40b-instruct",
431
+ FALCON_180B = "tii-falcon-180b"
432
+ }
102
433
  /**
103
434
  * API Versions for various providers
104
435
  */