@juspay/neurolink 8.9.0 → 8.10.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -25,42 +25,209 @@ export var AIProviderName;
25
25
  */
26
26
  export var BedrockModels;
27
27
  (function (BedrockModels) {
28
+ // ============================================================================
29
+ // ANTHROPIC CLAUDE MODELS
30
+ // ============================================================================
28
31
  // Claude 4.5 Series (Latest - September-November 2025)
29
- BedrockModels["CLAUDE_4_5_SONNET"] = "anthropic.claude-sonnet-4-5-20250929-v1:0";
30
32
  BedrockModels["CLAUDE_4_5_OPUS"] = "anthropic.claude-opus-4-5-20251124-v1:0";
33
+ BedrockModels["CLAUDE_4_5_SONNET"] = "anthropic.claude-sonnet-4-5-20250929-v1:0";
31
34
  BedrockModels["CLAUDE_4_5_HAIKU"] = "anthropic.claude-haiku-4-5-20251001-v1:0";
35
+ // Claude 4 Series (May-August 2025)
36
+ BedrockModels["CLAUDE_4_1_OPUS"] = "anthropic.claude-opus-4-1-20250805-v1:0";
37
+ BedrockModels["CLAUDE_4_SONNET"] = "anthropic.claude-sonnet-4-20250514-v1:0";
32
38
  // Claude 3.7 Series
33
39
  BedrockModels["CLAUDE_3_7_SONNET"] = "anthropic.claude-3-7-sonnet-20250219-v1:0";
34
40
  // Claude 3.5 Series
35
41
  BedrockModels["CLAUDE_3_5_SONNET"] = "anthropic.claude-3-5-sonnet-20241022-v1:0";
42
+ BedrockModels["CLAUDE_3_5_HAIKU"] = "anthropic.claude-3-5-haiku-20241022-v1:0";
36
43
  // Claude 3 Series (Legacy support)
37
44
  BedrockModels["CLAUDE_3_SONNET"] = "anthropic.claude-3-sonnet-20240229-v1:0";
38
45
  BedrockModels["CLAUDE_3_HAIKU"] = "anthropic.claude-3-haiku-20240307-v1:0";
46
+ // ============================================================================
47
+ // AMAZON NOVA MODELS
48
+ // ============================================================================
49
+ // Nova Generation 1
50
+ BedrockModels["NOVA_PREMIER"] = "amazon.nova-premier-v1:0";
51
+ BedrockModels["NOVA_PRO"] = "amazon.nova-pro-v1:0";
52
+ BedrockModels["NOVA_LITE"] = "amazon.nova-lite-v1:0";
53
+ BedrockModels["NOVA_MICRO"] = "amazon.nova-micro-v1:0";
54
+ // Nova Generation 2 (December 2025)
55
+ BedrockModels["NOVA_2_LITE"] = "amazon.nova-2-lite-v1:0";
56
+ BedrockModels["NOVA_2_SONIC"] = "amazon.nova-2-sonic-v1:0";
57
+ // Nova Specialized Models
58
+ BedrockModels["NOVA_SONIC"] = "amazon.nova-sonic-v1:0";
59
+ BedrockModels["NOVA_CANVAS"] = "amazon.nova-canvas-v1:0";
60
+ BedrockModels["NOVA_REEL"] = "amazon.nova-reel-v1:0";
61
+ BedrockModels["NOVA_REEL_V1_1"] = "amazon.nova-reel-v1:1";
62
+ BedrockModels["NOVA_MULTIMODAL_EMBEDDINGS"] = "amazon.nova-2-multimodal-embeddings-v1:0";
63
+ // ============================================================================
64
+ // AMAZON TITAN MODELS
65
+ // ============================================================================
66
+ // Titan Text Generation
67
+ BedrockModels["TITAN_TEXT_LARGE"] = "amazon.titan-tg1-large";
68
+ // Titan Text Embeddings
69
+ BedrockModels["TITAN_EMBED_TEXT_V2"] = "amazon.titan-embed-text-v2:0";
70
+ BedrockModels["TITAN_EMBED_TEXT_V1"] = "amazon.titan-embed-text-v1";
71
+ BedrockModels["TITAN_EMBED_G1_TEXT_02"] = "amazon.titan-embed-g1-text-02";
72
+ // Titan Multimodal Embeddings
73
+ BedrockModels["TITAN_EMBED_IMAGE_V1"] = "amazon.titan-embed-image-v1";
74
+ // Titan Image Generation
75
+ BedrockModels["TITAN_IMAGE_GENERATOR_V2"] = "amazon.titan-image-generator-v2:0";
76
+ // ============================================================================
77
+ // META LLAMA MODELS
78
+ // ============================================================================
79
+ // Llama 4 Series (2025)
80
+ BedrockModels["LLAMA_4_MAVERICK_17B"] = "meta.llama4-maverick-17b-instruct-v1:0";
81
+ BedrockModels["LLAMA_4_SCOUT_17B"] = "meta.llama4-scout-17b-instruct-v1:0";
82
+ // Llama 3.3 Series
83
+ BedrockModels["LLAMA_3_3_70B"] = "meta.llama3-3-70b-instruct-v1:0";
84
+ // Llama 3.2 Series (Multimodal)
85
+ BedrockModels["LLAMA_3_2_90B"] = "meta.llama3-2-90b-instruct-v1:0";
86
+ BedrockModels["LLAMA_3_2_11B"] = "meta.llama3-2-11b-instruct-v1:0";
87
+ BedrockModels["LLAMA_3_2_3B"] = "meta.llama3-2-3b-instruct-v1:0";
88
+ BedrockModels["LLAMA_3_2_1B"] = "meta.llama3-2-1b-instruct-v1:0";
89
+ // Llama 3.1 Series
90
+ BedrockModels["LLAMA_3_1_405B"] = "meta.llama3-1-405b-instruct-v1:0";
91
+ BedrockModels["LLAMA_3_1_70B"] = "meta.llama3-1-70b-instruct-v1:0";
92
+ BedrockModels["LLAMA_3_1_8B"] = "meta.llama3-1-8b-instruct-v1:0";
93
+ // Llama 3 Series (Legacy)
94
+ BedrockModels["LLAMA_3_70B"] = "meta.llama3-70b-instruct-v1:0";
95
+ BedrockModels["LLAMA_3_8B"] = "meta.llama3-8b-instruct-v1:0";
96
+ // ============================================================================
97
+ // MISTRAL AI MODELS
98
+ // ============================================================================
99
+ // Mistral Large Series
100
+ BedrockModels["MISTRAL_LARGE_3"] = "mistral.mistral-large-3-675b-instruct";
101
+ BedrockModels["MISTRAL_LARGE_2407"] = "mistral.mistral-large-2407-v1:0";
102
+ BedrockModels["MISTRAL_LARGE_2402"] = "mistral.mistral-large-2402-v1:0";
103
+ // Magistral & Ministral Series
104
+ BedrockModels["MAGISTRAL_SMALL_2509"] = "mistral.magistral-small-2509";
105
+ BedrockModels["MINISTRAL_3_14B"] = "mistral.ministral-3-14b-instruct";
106
+ BedrockModels["MINISTRAL_3_8B"] = "mistral.ministral-3-8b-instruct";
107
+ BedrockModels["MINISTRAL_3_3B"] = "mistral.ministral-3-3b-instruct";
108
+ // Mistral Base Series
109
+ BedrockModels["MISTRAL_7B"] = "mistral.mistral-7b-instruct-v0:2";
110
+ BedrockModels["MIXTRAL_8x7B"] = "mistral.mixtral-8x7b-instruct-v0:1";
111
+ // Mistral Multimodal & Audio
112
+ BedrockModels["PIXTRAL_LARGE_2502"] = "mistral.pixtral-large-2502-v1:0";
113
+ BedrockModels["VOXTRAL_SMALL_24B"] = "mistral.voxtral-small-24b-2507";
114
+ BedrockModels["VOXTRAL_MINI_3B"] = "mistral.voxtral-mini-3b-2507";
115
+ // ============================================================================
116
+ // OTHER MODELS
117
+ // ============================================================================
118
+ // Cohere Models
119
+ BedrockModels["COHERE_COMMAND_R_PLUS"] = "cohere.command-r-plus-v1:0";
120
+ BedrockModels["COHERE_COMMAND_R"] = "cohere.command-r-v1:0";
121
+ // DeepSeek Models
122
+ BedrockModels["DEEPSEEK_R1"] = "deepseek.r1-v1:0";
123
+ BedrockModels["DEEPSEEK_V3"] = "deepseek.v3-v1:0";
124
+ // Qwen Models
125
+ BedrockModels["QWEN_3_235B_A22B"] = "qwen.qwen3-235b-a22b-2507-v1:0";
126
+ BedrockModels["QWEN_3_CODER_480B_A35B"] = "qwen.qwen3-coder-480b-a35b-v1:0";
127
+ BedrockModels["QWEN_3_CODER_30B_A3B"] = "qwen.qwen3-coder-30b-a3b-v1:0";
128
+ BedrockModels["QWEN_3_32B"] = "qwen.qwen3-32b-v1:0";
129
+ BedrockModels["QWEN_3_NEXT_80B_A3B"] = "qwen.qwen3-next-80b-a3b";
130
+ BedrockModels["QWEN_3_VL_235B_A22B"] = "qwen.qwen3-vl-235b-a22b";
131
+ // Google Gemma
132
+ BedrockModels["GEMMA_3_27B_IT"] = "google.gemma-3-27b-it";
133
+ BedrockModels["GEMMA_3_12B_IT"] = "google.gemma-3-12b-it";
134
+ BedrockModels["GEMMA_3_4B_IT"] = "google.gemma-3-4b-it";
135
+ // AI21 Labs Models
136
+ BedrockModels["JAMBA_1_5_LARGE"] = "ai21.jamba-1-5-large-v1:0";
137
+ BedrockModels["JAMBA_1_5_MINI"] = "ai21.jamba-1-5-mini-v1:0";
39
138
  })(BedrockModels || (BedrockModels = {}));
40
139
  /**
41
140
  * Supported Models for OpenAI
42
141
  */
43
142
  export var OpenAIModels;
44
143
  (function (OpenAIModels) {
45
- OpenAIModels["GPT_4"] = "gpt-4";
46
- OpenAIModels["GPT_4_TURBO"] = "gpt-4-turbo";
144
+ // GPT-5.2 Series (Released December 11, 2025) - Latest flagship models
145
+ OpenAIModels["GPT_5_2"] = "gpt-5.2";
146
+ OpenAIModels["GPT_5_2_CHAT_LATEST"] = "gpt-5.2-chat-latest";
147
+ OpenAIModels["GPT_5_2_PRO"] = "gpt-5.2-pro";
148
+ // GPT-5 Series (Released August 7, 2025)
149
+ OpenAIModels["GPT_5"] = "gpt-5";
150
+ OpenAIModels["GPT_5_MINI"] = "gpt-5-mini";
151
+ OpenAIModels["GPT_5_NANO"] = "gpt-5-nano";
152
+ // GPT-4.1 Series (Released April 14, 2025)
153
+ OpenAIModels["GPT_4_1"] = "gpt-4.1";
154
+ OpenAIModels["GPT_4_1_MINI"] = "gpt-4.1-mini";
155
+ OpenAIModels["GPT_4_1_NANO"] = "gpt-4.1-nano";
156
+ // GPT-4o Series
47
157
  OpenAIModels["GPT_4O"] = "gpt-4o";
48
158
  OpenAIModels["GPT_4O_MINI"] = "gpt-4o-mini";
49
- OpenAIModels["GPT_3_5_TURBO"] = "gpt-3.5-turbo";
159
+ // O-Series Reasoning Models
160
+ OpenAIModels["O3"] = "o3";
161
+ OpenAIModels["O3_MINI"] = "o3-mini";
162
+ OpenAIModels["O3_PRO"] = "o3-pro";
163
+ OpenAIModels["O4_MINI"] = "o4-mini";
164
+ OpenAIModels["O1"] = "o1";
50
165
  OpenAIModels["O1_PREVIEW"] = "o1-preview";
51
166
  OpenAIModels["O1_MINI"] = "o1-mini";
167
+ // GPT-4 Series (Legacy)
168
+ OpenAIModels["GPT_4"] = "gpt-4";
169
+ OpenAIModels["GPT_4_TURBO"] = "gpt-4-turbo";
170
+ // Legacy Models
171
+ OpenAIModels["GPT_3_5_TURBO"] = "gpt-3.5-turbo";
52
172
  })(OpenAIModels || (OpenAIModels = {}));
173
+ /**
174
+ * Supported Models for Azure OpenAI
175
+ * Note: Azure uses deployment names, these are model identifiers
176
+ */
177
+ export var AzureOpenAIModels;
178
+ (function (AzureOpenAIModels) {
179
+ // GPT-5.1 Series (Latest - December 2025)
180
+ AzureOpenAIModels["GPT_5_1"] = "gpt-5.1";
181
+ AzureOpenAIModels["GPT_5_1_CHAT"] = "gpt-5.1-chat";
182
+ AzureOpenAIModels["GPT_5_1_CODEX"] = "gpt-5.1-codex";
183
+ AzureOpenAIModels["GPT_5_1_CODEX_MINI"] = "gpt-5.1-codex-mini";
184
+ AzureOpenAIModels["GPT_5_1_CODEX_MAX"] = "gpt-5.1-codex-max";
185
+ // GPT-5.0 Series
186
+ AzureOpenAIModels["GPT_5"] = "gpt-5";
187
+ AzureOpenAIModels["GPT_5_MINI"] = "gpt-5-mini";
188
+ AzureOpenAIModels["GPT_5_NANO"] = "gpt-5-nano";
189
+ AzureOpenAIModels["GPT_5_CHAT"] = "gpt-5-chat";
190
+ AzureOpenAIModels["GPT_5_CODEX"] = "gpt-5-codex";
191
+ AzureOpenAIModels["GPT_5_PRO"] = "gpt-5-pro";
192
+ AzureOpenAIModels["GPT_5_TURBO"] = "gpt-5-turbo";
193
+ // O-Series Reasoning Models
194
+ AzureOpenAIModels["O4_MINI"] = "o4-mini";
195
+ AzureOpenAIModels["O3"] = "o3";
196
+ AzureOpenAIModels["O3_MINI"] = "o3-mini";
197
+ AzureOpenAIModels["O3_PRO"] = "o3-pro";
198
+ AzureOpenAIModels["O1"] = "o1";
199
+ AzureOpenAIModels["O1_MINI"] = "o1-mini";
200
+ AzureOpenAIModels["O1_PREVIEW"] = "o1-preview";
201
+ AzureOpenAIModels["CODEX_MINI"] = "codex-mini";
202
+ // GPT-4.1 Series
203
+ AzureOpenAIModels["GPT_4_1"] = "gpt-4.1";
204
+ AzureOpenAIModels["GPT_4_1_NANO"] = "gpt-4.1-nano";
205
+ AzureOpenAIModels["GPT_4_1_MINI"] = "gpt-4.1-mini";
206
+ // GPT-4o Series (Multimodal)
207
+ AzureOpenAIModels["GPT_4O"] = "gpt-4o";
208
+ AzureOpenAIModels["GPT_4O_MINI"] = "gpt-4o-mini";
209
+ // GPT-4 Turbo & GPT-4
210
+ AzureOpenAIModels["GPT_4_TURBO"] = "gpt-4-turbo";
211
+ AzureOpenAIModels["GPT_4"] = "gpt-4";
212
+ AzureOpenAIModels["GPT_4_32K"] = "gpt-4-32k";
213
+ // GPT-3.5 Turbo (Legacy)
214
+ AzureOpenAIModels["GPT_3_5_TURBO"] = "gpt-35-turbo";
215
+ AzureOpenAIModels["GPT_3_5_TURBO_INSTRUCT"] = "gpt-35-turbo-instruct";
216
+ })(AzureOpenAIModels || (AzureOpenAIModels = {}));
53
217
  /**
54
218
  * Supported Models for Google Vertex AI
55
219
  */
56
220
  export var VertexModels;
57
221
  (function (VertexModels) {
58
222
  // Claude 4.5 Series (Latest - December 2025)
59
- VertexModels["CLAUDE_4_5_SONNET"] = "claude-sonnet-4-5@20250929";
60
223
  VertexModels["CLAUDE_4_5_OPUS"] = "claude-opus-4-5@20251124";
224
+ VertexModels["CLAUDE_4_5_SONNET"] = "claude-sonnet-4-5@20250929";
225
+ VertexModels["CLAUDE_4_5_HAIKU"] = "claude-haiku-4-5@20251001";
61
226
  // Claude 4 Series (May 2025)
62
227
  VertexModels["CLAUDE_4_0_SONNET"] = "claude-sonnet-4@20250514";
63
228
  VertexModels["CLAUDE_4_0_OPUS"] = "claude-opus-4@20250514";
229
+ // Claude 3.7 Series (February 2025)
230
+ VertexModels["CLAUDE_3_7_SONNET"] = "claude-3-7-sonnet@20250219";
64
231
  // Claude 3.5 Series (Still supported)
65
232
  VertexModels["CLAUDE_3_5_SONNET"] = "claude-3-5-sonnet-20241022";
66
233
  VertexModels["CLAUDE_3_5_HAIKU"] = "claude-3-5-haiku-20241022";
@@ -69,6 +236,8 @@ export var VertexModels;
69
236
  VertexModels["CLAUDE_3_OPUS"] = "claude-3-opus-20240229";
70
237
  VertexModels["CLAUDE_3_HAIKU"] = "claude-3-haiku-20240307";
71
238
  // Gemini 3 Series (Preview)
239
+ /** Gemini 3 Pro - Base model with adaptive thinking */
240
+ VertexModels["GEMINI_3_PRO"] = "gemini-3-pro";
72
241
  /** Gemini 3 Pro Preview - Versioned preview (November 2025) */
73
242
  VertexModels["GEMINI_3_PRO_PREVIEW_11_2025"] = "gemini-3-pro-preview-11-2025";
74
243
  /** Gemini 3 Pro Latest - Auto-updated alias (always points to latest preview) */
@@ -79,36 +248,41 @@ export var VertexModels;
79
248
  VertexModels["GEMINI_2_5_PRO"] = "gemini-2.5-pro";
80
249
  VertexModels["GEMINI_2_5_FLASH"] = "gemini-2.5-flash";
81
250
  VertexModels["GEMINI_2_5_FLASH_LITE"] = "gemini-2.5-flash-lite";
251
+ VertexModels["GEMINI_2_5_FLASH_IMAGE"] = "gemini-2.5-flash-image";
82
252
  // Gemini 2.0 Series
253
+ VertexModels["GEMINI_2_0_FLASH"] = "gemini-2.0-flash";
83
254
  VertexModels["GEMINI_2_0_FLASH_001"] = "gemini-2.0-flash-001";
84
255
  /** Gemini 2.0 Flash Lite - GA, production-ready, cost-optimized */
85
256
  VertexModels["GEMINI_2_0_FLASH_LITE"] = "gemini-2.0-flash-lite";
86
257
  // Gemini 1.5 Series (Legacy support)
87
- VertexModels["GEMINI_1_5_PRO"] = "gemini-1.5-pro";
88
- VertexModels["GEMINI_1_5_FLASH"] = "gemini-1.5-flash";
258
+ VertexModels["GEMINI_1_5_PRO"] = "gemini-1.5-pro-002";
259
+ VertexModels["GEMINI_1_5_FLASH"] = "gemini-1.5-flash-002";
89
260
  })(VertexModels || (VertexModels = {}));
90
261
  /**
91
262
  * Supported Models for Google AI Studio
92
263
  */
93
264
  export var GoogleAIModels;
94
265
  (function (GoogleAIModels) {
95
- // Gemini 3 Series (Preview)
96
- /** Gemini 3 Pro Preview - Versioned preview (November 2025) */
97
- GoogleAIModels["GEMINI_3_PRO_PREVIEW_11_2025"] = "gemini-3-pro-preview-11-2025";
98
- /** Gemini 3 Pro Latest - Auto-updated alias (always points to latest preview) */
99
- GoogleAIModels["GEMINI_3_PRO_LATEST"] = "gemini-3-pro-latest";
100
- // Gemini 2.5 Series (Latest - 2025)
266
+ // Gemini 3 Series
267
+ GoogleAIModels["GEMINI_3_PRO_PREVIEW"] = "gemini-3-pro-preview";
268
+ GoogleAIModels["GEMINI_3_PRO_IMAGE_PREVIEW"] = "gemini-3-pro-image-preview";
269
+ // Gemini 2.5 Series
101
270
  GoogleAIModels["GEMINI_2_5_PRO"] = "gemini-2.5-pro";
102
271
  GoogleAIModels["GEMINI_2_5_FLASH"] = "gemini-2.5-flash";
103
272
  GoogleAIModels["GEMINI_2_5_FLASH_LITE"] = "gemini-2.5-flash-lite";
273
+ GoogleAIModels["GEMINI_2_5_FLASH_IMAGE"] = "gemini-2.5-flash-image";
274
+ GoogleAIModels["GEMINI_2_5_FLASH_LIVE"] = "gemini-2.5-flash-native-audio-preview-09-2025";
104
275
  // Gemini 2.0 Series
276
+ GoogleAIModels["GEMINI_2_0_FLASH"] = "gemini-2.0-flash";
105
277
  GoogleAIModels["GEMINI_2_0_FLASH_001"] = "gemini-2.0-flash-001";
106
- /** Gemini 2.0 Flash Lite - GA, production-ready, cost-optimized */
107
278
  GoogleAIModels["GEMINI_2_0_FLASH_LITE"] = "gemini-2.0-flash-lite";
108
- // Gemini 1.5 Series (Legacy support)
279
+ GoogleAIModels["GEMINI_2_0_FLASH_IMAGE"] = "gemini-2.0-flash-preview-image-generation";
280
+ // Gemini 1.5 Series (Legacy)
109
281
  GoogleAIModels["GEMINI_1_5_PRO"] = "gemini-1.5-pro";
110
282
  GoogleAIModels["GEMINI_1_5_FLASH"] = "gemini-1.5-flash";
111
- GoogleAIModels["GEMINI_1_5_FLASH_LITE"] = "gemini-1.5-flash-lite";
283
+ // Embedding Models
284
+ GoogleAIModels["GEMINI_EMBEDDING"] = "gemini-embedding-001";
285
+ GoogleAIModels["TEXT_EMBEDDING_004"] = "text-embedding-004";
112
286
  })(GoogleAIModels || (GoogleAIModels = {}));
113
287
  /**
114
288
  * Supported Models for Anthropic (Direct API)
@@ -116,17 +290,300 @@ export var GoogleAIModels;
116
290
  export var AnthropicModels;
117
291
  (function (AnthropicModels) {
118
292
  // Claude 4.5 Series (Latest - September-November 2025)
293
+ AnthropicModels["CLAUDE_OPUS_4_5"] = "claude-opus-4-5-20251101";
119
294
  AnthropicModels["CLAUDE_SONNET_4_5"] = "claude-sonnet-4-5-20250929";
120
- AnthropicModels["CLAUDE_OPUS_4_5"] = "claude-opus-4-5-20251124";
121
295
  AnthropicModels["CLAUDE_4_5_HAIKU"] = "claude-haiku-4-5-20251001";
122
- // Claude 3.5 Series
296
+ // Claude 4.1 Series (Legacy)
297
+ AnthropicModels["CLAUDE_OPUS_4_1"] = "claude-opus-4-1-20250805";
298
+ // Claude 4.0 Series (Legacy)
299
+ AnthropicModels["CLAUDE_OPUS_4_0"] = "claude-opus-4-20250514";
300
+ AnthropicModels["CLAUDE_SONNET_4_0"] = "claude-sonnet-4-20250514";
301
+ // Claude 3.7 Series (Legacy)
302
+ AnthropicModels["CLAUDE_SONNET_3_7"] = "claude-3-7-sonnet-20250219";
303
+ // Claude 3.5 Series (Legacy)
123
304
  AnthropicModels["CLAUDE_3_5_SONNET"] = "claude-3-5-sonnet-20241022";
124
305
  AnthropicModels["CLAUDE_3_5_HAIKU"] = "claude-3-5-haiku-20241022";
125
- // Claude 3 Series (Legacy support)
306
+ // Claude 3 Series (Legacy - Deprecated)
126
307
  AnthropicModels["CLAUDE_3_SONNET"] = "claude-3-sonnet-20240229";
127
308
  AnthropicModels["CLAUDE_3_OPUS"] = "claude-3-opus-20240229";
128
309
  AnthropicModels["CLAUDE_3_HAIKU"] = "claude-3-haiku-20240307";
129
310
  })(AnthropicModels || (AnthropicModels = {}));
311
+ /**
312
+ * Supported Models for Mistral AI
313
+ */
314
+ export var MistralModels;
315
+ (function (MistralModels) {
316
+ // Mistral Large (Latest)
317
+ MistralModels["MISTRAL_LARGE_LATEST"] = "mistral-large-latest";
318
+ MistralModels["MISTRAL_LARGE_2512"] = "mistral-large-2512";
319
+ // Mistral Medium
320
+ MistralModels["MISTRAL_MEDIUM_LATEST"] = "mistral-medium-latest";
321
+ MistralModels["MISTRAL_MEDIUM_2508"] = "mistral-medium-2508";
322
+ // Mistral Small
323
+ MistralModels["MISTRAL_SMALL_LATEST"] = "mistral-small-latest";
324
+ MistralModels["MISTRAL_SMALL_2506"] = "mistral-small-2506";
325
+ // Magistral (Reasoning)
326
+ MistralModels["MAGISTRAL_MEDIUM_LATEST"] = "magistral-medium-latest";
327
+ MistralModels["MAGISTRAL_SMALL_LATEST"] = "magistral-small-latest";
328
+ // Ministral (Edge Models)
329
+ MistralModels["MINISTRAL_14B_2512"] = "ministral-14b-2512";
330
+ MistralModels["MINISTRAL_8B_2512"] = "ministral-8b-2512";
331
+ MistralModels["MINISTRAL_3B_2512"] = "ministral-3b-2512";
332
+ // Codestral (Code Generation)
333
+ MistralModels["CODESTRAL_LATEST"] = "codestral-latest";
334
+ MistralModels["CODESTRAL_2508"] = "codestral-2508";
335
+ MistralModels["CODESTRAL_EMBED"] = "codestral-embed";
336
+ // Devstral (Software Development)
337
+ MistralModels["DEVSTRAL_MEDIUM_LATEST"] = "devstral-medium-latest";
338
+ MistralModels["DEVSTRAL_SMALL_LATEST"] = "devstral-small-latest";
339
+ // Pixtral (Multimodal/Vision)
340
+ MistralModels["PIXTRAL_LARGE"] = "pixtral-large";
341
+ MistralModels["PIXTRAL_12B"] = "pixtral-12b";
342
+ // Voxtral (Audio)
343
+ MistralModels["VOXTRAL_SMALL_LATEST"] = "voxtral-small-latest";
344
+ MistralModels["VOXTRAL_MINI_LATEST"] = "voxtral-mini-latest";
345
+ // Specialized Models
346
+ MistralModels["MISTRAL_NEMO"] = "mistral-nemo";
347
+ MistralModels["MISTRAL_EMBED"] = "mistral-embed";
348
+ MistralModels["MISTRAL_MODERATION_LATEST"] = "mistral-moderation-latest";
349
+ })(MistralModels || (MistralModels = {}));
350
+ /**
351
+ * Supported Models for Ollama (Local)
352
+ * All models can be run locally without requiring API keys or cloud services
353
+ */
354
+ export var OllamaModels;
355
+ (function (OllamaModels) {
356
+ // Llama 4 Series - Multimodal with vision and tool capabilities
357
+ OllamaModels["LLAMA4_SCOUT"] = "llama4:scout";
358
+ OllamaModels["LLAMA4_MAVERICK"] = "llama4:maverick";
359
+ OllamaModels["LLAMA4_LATEST"] = "llama4:latest";
360
+ // Llama 3.3 Series - High-performance models
361
+ OllamaModels["LLAMA3_3_LATEST"] = "llama3.3:latest";
362
+ OllamaModels["LLAMA3_3_70B"] = "llama3.3:70b";
363
+ // Llama 3.2 Series - Optimized for edge and mobile deployment
364
+ OllamaModels["LLAMA3_2_LATEST"] = "llama3.2:latest";
365
+ OllamaModels["LLAMA3_2_3B"] = "llama3.2:3b";
366
+ OllamaModels["LLAMA3_2_1B"] = "llama3.2:1b";
367
+ // Llama 3.1 Series - Open models rivaling proprietary models
368
+ OllamaModels["LLAMA3_1_8B"] = "llama3.1:8b";
369
+ OllamaModels["LLAMA3_1_70B"] = "llama3.1:70b";
370
+ OllamaModels["LLAMA3_1_405B"] = "llama3.1:405b";
371
+ // Qwen 3 Series - Advanced reasoning and multilingual support
372
+ OllamaModels["QWEN3_4B"] = "qwen3:4b";
373
+ OllamaModels["QWEN3_8B"] = "qwen3:8b";
374
+ OllamaModels["QWEN3_14B"] = "qwen3:14b";
375
+ OllamaModels["QWEN3_32B"] = "qwen3:32b";
376
+ OllamaModels["QWEN3_72B"] = "qwen3:72b";
377
+ // Qwen 2.5 Series - Enhanced coding and mathematics
378
+ OllamaModels["QWEN2_5_3B"] = "qwen2.5:3b";
379
+ OllamaModels["QWEN2_5_7B"] = "qwen2.5:7b";
380
+ OllamaModels["QWEN2_5_14B"] = "qwen2.5:14b";
381
+ OllamaModels["QWEN2_5_32B"] = "qwen2.5:32b";
382
+ OllamaModels["QWEN2_5_72B"] = "qwen2.5:72b";
383
+ // Qwen Reasoning Model
384
+ OllamaModels["QWQ_32B"] = "qwq:32b";
385
+ OllamaModels["QWQ_LATEST"] = "qwq:latest";
386
+ // DeepSeek-R1 Series - State-of-the-art reasoning models
387
+ OllamaModels["DEEPSEEK_R1_1_5B"] = "deepseek-r1:1.5b";
388
+ OllamaModels["DEEPSEEK_R1_7B"] = "deepseek-r1:7b";
389
+ OllamaModels["DEEPSEEK_R1_8B"] = "deepseek-r1:8b";
390
+ OllamaModels["DEEPSEEK_R1_14B"] = "deepseek-r1:14b";
391
+ OllamaModels["DEEPSEEK_R1_32B"] = "deepseek-r1:32b";
392
+ OllamaModels["DEEPSEEK_R1_70B"] = "deepseek-r1:70b";
393
+ // DeepSeek-V3 Series - Mixture of Experts model
394
+ OllamaModels["DEEPSEEK_V3_671B"] = "deepseek-v3:671b";
395
+ OllamaModels["DEEPSEEK_V3_LATEST"] = "deepseek-v3:latest";
396
+ // Mistral AI Series - Efficient general-purpose models
397
+ OllamaModels["MISTRAL_LATEST"] = "mistral:latest";
398
+ OllamaModels["MISTRAL_7B"] = "mistral:7b";
399
+ OllamaModels["MISTRAL_SMALL_LATEST"] = "mistral-small:latest";
400
+ OllamaModels["MISTRAL_NEMO_LATEST"] = "mistral-nemo:latest";
401
+ OllamaModels["MISTRAL_LARGE_LATEST"] = "mistral-large:latest";
402
+ // Google Gemma Series - Efficient edge and cloud models
403
+ OllamaModels["GEMMA3_LATEST"] = "gemma3:latest";
404
+ OllamaModels["GEMMA2_2B"] = "gemma2:2b";
405
+ OllamaModels["GEMMA2_9B"] = "gemma2:9b";
406
+ OllamaModels["GEMMA2_27B"] = "gemma2:27b";
407
+ // Microsoft Phi Series - Compact, efficient models
408
+ OllamaModels["PHI4_LATEST"] = "phi4:latest";
409
+ OllamaModels["PHI4_14B"] = "phi4:14b";
410
+ OllamaModels["PHI3_MINI"] = "phi3:mini";
411
+ OllamaModels["PHI3_3_8B"] = "phi3:3.8b";
412
+ OllamaModels["PHI3_MEDIUM"] = "phi3:medium";
413
+ OllamaModels["PHI3_14B"] = "phi3:14b";
414
+ // Vision-Language Models
415
+ OllamaModels["LLAVA_7B"] = "llava:7b";
416
+ OllamaModels["LLAVA_13B"] = "llava:13b";
417
+ OllamaModels["LLAVA_34B"] = "llava:34b";
418
+ OllamaModels["LLAVA_LLAMA3_8B"] = "llava-llama3:8b";
419
+ // Code-Specialized Models
420
+ OllamaModels["CODELLAMA_7B"] = "codellama:7b";
421
+ OllamaModels["CODELLAMA_13B"] = "codellama:13b";
422
+ OllamaModels["CODELLAMA_34B"] = "codellama:34b";
423
+ OllamaModels["CODELLAMA_70B"] = "codellama:70b";
424
+ OllamaModels["QWEN2_5_CODER_7B"] = "qwen2.5-coder:7b";
425
+ OllamaModels["QWEN2_5_CODER_32B"] = "qwen2.5-coder:32b";
426
+ OllamaModels["STARCODER2_3B"] = "starcoder2:3b";
427
+ OllamaModels["STARCODER2_7B"] = "starcoder2:7b";
428
+ OllamaModels["STARCODER2_15B"] = "starcoder2:15b";
429
+ // Mixture of Experts Models
430
+ OllamaModels["MIXTRAL_8X7B"] = "mixtral:8x7b";
431
+ OllamaModels["MIXTRAL_8X22B"] = "mixtral:8x22b";
432
+ // Enterprise Models
433
+ OllamaModels["COMMAND_R_PLUS"] = "command-r-plus:104b";
434
+ })(OllamaModels || (OllamaModels = {}));
435
+ /**
436
+ * Common Models for LiteLLM Proxy
437
+ * LiteLLM supports 100+ models through unified proxy interface
438
+ * Models use provider-specific prefixes (e.g., "openai/", "anthropic/")
439
+ */
440
+ export var LiteLLMModels;
441
+ (function (LiteLLMModels) {
442
+ // OpenAI via LiteLLM
443
+ LiteLLMModels["OPENAI_GPT_5"] = "openai/gpt-5";
444
+ LiteLLMModels["OPENAI_GPT_4O"] = "openai/gpt-4o";
445
+ LiteLLMModels["OPENAI_GPT_4O_MINI"] = "openai/gpt-4o-mini";
446
+ LiteLLMModels["OPENAI_GPT_4_TURBO"] = "openai/gpt-4-turbo";
447
+ LiteLLMModels["OPENAI_GPT_4"] = "openai/gpt-4";
448
+ LiteLLMModels["OPENAI_GPT_3_5_TURBO"] = "openai/gpt-3.5-turbo";
449
+ // Anthropic via LiteLLM
450
+ LiteLLMModels["ANTHROPIC_CLAUDE_SONNET_4_5"] = "anthropic/claude-sonnet-4-5-20250929";
451
+ LiteLLMModels["ANTHROPIC_CLAUDE_OPUS_4_1"] = "anthropic/claude-opus-4-1-20250805";
452
+ LiteLLMModels["ANTHROPIC_CLAUDE_3_5_SONNET"] = "anthropic/claude-3-5-sonnet-20240620";
453
+ LiteLLMModels["ANTHROPIC_CLAUDE_3_HAIKU"] = "anthropic/claude-3-haiku-20240307";
454
+ // Google Vertex AI via LiteLLM
455
+ LiteLLMModels["VERTEX_GEMINI_2_5_PRO"] = "vertex_ai/gemini-2.5-pro";
456
+ LiteLLMModels["VERTEX_GEMINI_1_5_PRO"] = "vertex_ai/gemini-1.5-pro";
457
+ LiteLLMModels["VERTEX_GEMINI_1_5_FLASH"] = "vertex_ai/gemini-1.5-flash";
458
+ // Google AI Studio (Gemini) via LiteLLM
459
+ LiteLLMModels["GEMINI_2_5_PRO"] = "gemini/gemini-2.5-pro";
460
+ LiteLLMModels["GEMINI_2_0_FLASH"] = "gemini/gemini-2.0-flash";
461
+ LiteLLMModels["GEMINI_1_5_PRO"] = "gemini/gemini-1.5-pro";
462
+ LiteLLMModels["GEMINI_1_5_FLASH"] = "gemini/gemini-1.5-flash";
463
+ // Groq via LiteLLM
464
+ LiteLLMModels["GROQ_LLAMA_3_1_70B_VERSATILE"] = "groq/llama-3.1-70b-versatile";
465
+ LiteLLMModels["GROQ_LLAMA_3_1_8B_INSTANT"] = "groq/llama-3.1-8b-instant";
466
+ LiteLLMModels["GROQ_LLAMA_3_2_11B_VISION"] = "groq/llama-3.2-11b-vision-preview";
467
+ LiteLLMModels["GROQ_MIXTRAL_8X7B"] = "groq/mixtral-8x7b-32768";
468
+ // Together AI via LiteLLM
469
+ LiteLLMModels["TOGETHER_LLAMA_2_70B_CHAT"] = "together_ai/togethercomputer/llama-2-70b-chat";
470
+ LiteLLMModels["TOGETHER_MIXTRAL_8X7B"] = "together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1";
471
+ LiteLLMModels["TOGETHER_CODELLAMA_34B"] = "together_ai/codellama/CodeLlama-34b-Instruct-hf";
472
+ // DeepInfra via LiteLLM
473
+ LiteLLMModels["DEEPINFRA_LLAMA_3_70B"] = "deepinfra/meta-llama/Meta-Llama-3-70B-Instruct";
474
+ LiteLLMModels["DEEPINFRA_LLAMA_2_70B"] = "deepinfra/meta-llama/Llama-2-70b-chat-hf";
475
+ LiteLLMModels["DEEPINFRA_MISTRAL_7B"] = "deepinfra/mistralai/Mistral-7B-Instruct-v0.1";
476
+ // Mistral AI via LiteLLM
477
+ LiteLLMModels["MISTRAL_LARGE"] = "mistral/mistral-large-latest";
478
+ LiteLLMModels["MISTRAL_SMALL"] = "mistral/mistral-small-latest";
479
+ LiteLLMModels["MISTRAL_MAGISTRAL_MEDIUM"] = "mistral/magistral-medium-2506";
480
+ // AWS Bedrock via LiteLLM
481
+ LiteLLMModels["BEDROCK_CLAUDE_3_5_SONNET"] = "bedrock/anthropic.claude-3-5-sonnet-20240620-v1:0";
482
+ LiteLLMModels["BEDROCK_CLAUDE_3_HAIKU"] = "bedrock/anthropic.claude-3-haiku-20240307-v1:0";
483
+ // Perplexity AI via LiteLLM
484
+ LiteLLMModels["PERPLEXITY_SONAR_PRO"] = "perplexity/sonar-pro";
485
+ LiteLLMModels["PERPLEXITY_SONAR_REASONING_PRO"] = "perplexity/sonar-reasoning-pro";
486
+ })(LiteLLMModels || (LiteLLMModels = {}));
487
+ /**
488
+ * Supported Models for Hugging Face Inference API
489
+ */
490
+ export var HuggingFaceModels;
491
+ (function (HuggingFaceModels) {
492
+ // Meta Llama 3.3
493
+ HuggingFaceModels["LLAMA_3_3_70B_INSTRUCT"] = "meta-llama/Llama-3.3-70B-Instruct";
494
+ // Meta Llama 3.2
495
+ HuggingFaceModels["LLAMA_3_2_1B"] = "meta-llama/Llama-3.2-1B";
496
+ HuggingFaceModels["LLAMA_3_2_3B_INSTRUCT"] = "meta-llama/Llama-3.2-3B-Instruct";
497
+ // Meta Llama 3.1
498
+ HuggingFaceModels["LLAMA_3_1_8B"] = "meta-llama/Llama-3.1-8B";
499
+ HuggingFaceModels["LLAMA_3_1_70B_INSTRUCT"] = "meta-llama/Llama-3.1-70B-Instruct";
500
+ HuggingFaceModels["LLAMA_3_1_405B_INSTRUCT"] = "meta-llama/Llama-3.1-405B-Instruct";
501
+ // Meta Llama 3.0
502
+ HuggingFaceModels["LLAMA_3_8B_INSTRUCT"] = "meta-llama/Meta-Llama-3-8B-Instruct";
503
+ HuggingFaceModels["LLAMA_3_70B_INSTRUCT"] = "meta-llama/Meta-Llama-3-70B-Instruct";
504
+ // Mistral Large
505
+ HuggingFaceModels["MISTRAL_LARGE_3_675B"] = "mistralai/Mistral-Large-3-675B-Instruct-2512";
506
+ // Mistral Small
507
+ HuggingFaceModels["MISTRAL_SMALL_3_1_24B"] = "mistralai/Mistral-Small-3.1-24B-Instruct-2503";
508
+ HuggingFaceModels["MISTRAL_SMALL_24B"] = "mistralai/Mistral-Small-24B-Instruct-2501";
509
+ // Mistral
510
+ HuggingFaceModels["MISTRAL_7B_INSTRUCT"] = "mistralai/Mistral-7B-Instruct-v0.2";
511
+ HuggingFaceModels["MIXTRAL_8X7B_INSTRUCT"] = "mistralai/Mixtral-8x7B-Instruct-v0.1";
512
+ // Mistral Devstral
513
+ HuggingFaceModels["DEVSTRAL_2"] = "mistralai/Devstral-2";
514
+ // Qwen 2.5
515
+ HuggingFaceModels["QWEN_2_5_7B"] = "Qwen/Qwen2.5-7B";
516
+ HuggingFaceModels["QWEN_2_5_32B"] = "Qwen/Qwen2.5-32B";
517
+ HuggingFaceModels["QWEN_2_5_72B_INSTRUCT"] = "Qwen/Qwen2.5-72B-Instruct";
518
+ // Qwen 2.5 Coder
519
+ HuggingFaceModels["QWEN_2_5_CODER_7B"] = "Qwen/Qwen2.5-Coder-7B";
520
+ HuggingFaceModels["QWEN_2_5_CODER_32B_INSTRUCT"] = "Qwen/Qwen2.5-Coder-32B-Instruct";
521
+ // Qwen QwQ
522
+ HuggingFaceModels["QWQ_32B"] = "Qwen/QwQ-32B";
523
+ // Qwen 2.5 VL (Multimodal)
524
+ HuggingFaceModels["QWEN_2_5_VL_32B"] = "Qwen/Qwen2.5-VL-32B-Instruct";
525
+ // DeepSeek
526
+ HuggingFaceModels["DEEPSEEK_R1"] = "deepseek-ai/DeepSeek-R1";
527
+ HuggingFaceModels["DEEPSEEK_V3"] = "deepseek-ai/DeepSeek-V3";
528
+ HuggingFaceModels["DEEPSEEK_V3_1"] = "deepseek-ai/DeepSeek-V3.1";
529
+ HuggingFaceModels["DEEPSEEK_V3_2_EXP"] = "deepseek-ai/DeepSeek-V3.2-Exp";
530
+ // Microsoft Phi
531
+ HuggingFaceModels["PHI_4"] = "microsoft/phi-4";
532
+ HuggingFaceModels["PHI_4_REASONING"] = "microsoft/Phi-4-reasoning";
533
+ HuggingFaceModels["PHI_4_MINI_INSTRUCT"] = "microsoft/Phi-4-mini-instruct";
534
+ HuggingFaceModels["PHI_4_MINI_REASONING"] = "microsoft/Phi-4-mini-reasoning";
535
+ HuggingFaceModels["PHI_3_MINI_128K_INSTRUCT"] = "microsoft/Phi-3-mini-128k-instruct";
536
+ HuggingFaceModels["PHI_3_VISION_128K_INSTRUCT"] = "microsoft/Phi-3-vision-128k-instruct";
537
+ // Google Gemma 3
538
+ HuggingFaceModels["GEMMA_3_270M"] = "google/gemma-3-270m";
539
+ HuggingFaceModels["GEMMA_3_1B_IT"] = "google/gemma-3-1b-it";
540
+ HuggingFaceModels["GEMMA_3_4B_IT"] = "google/gemma-3-4b-it";
541
+ HuggingFaceModels["GEMMA_3_12B_IT"] = "google/gemma-3-12b-it";
542
+ HuggingFaceModels["GEMMA_3_27B_IT"] = "google/gemma-3-27b-it";
543
+ // Google Gemma 2
544
+ HuggingFaceModels["GEMMA_2_9B"] = "google/gemma-2-9b";
545
+ HuggingFaceModels["GEMMA_2_27B"] = "google/gemma-2-27b";
546
+ // Google Gemma 1
547
+ HuggingFaceModels["GEMMA_2B"] = "google/gemma-2b";
548
+ HuggingFaceModels["GEMMA_7B"] = "google/gemma-7b";
549
+ // Falcon
550
+ HuggingFaceModels["FALCON_40B_INSTRUCT"] = "tiiuae/falcon-40b-instruct";
551
+ HuggingFaceModels["FALCON_180B_CHAT"] = "tiiuae/falcon-180B-chat";
552
+ // Code Models
553
+ HuggingFaceModels["STARCODER2_15B"] = "bigcode/starcoder2-15b";
554
+ HuggingFaceModels["CODELLAMA_34B_INSTRUCT"] = "codellama/CodeLlama-34b-Instruct-hf";
555
+ // BLOOM
556
+ HuggingFaceModels["BLOOM_7B1"] = "bigscience/bloom-7b1";
557
+ HuggingFaceModels["BLOOM_1B3"] = "bigscience/bloom-1b3";
558
+ })(HuggingFaceModels || (HuggingFaceModels = {}));
559
+ /**
560
+ * Supported Models for AWS SageMaker JumpStart
561
+ * https://docs.aws.amazon.com/sagemaker/latest/dg/jumpstart-foundation-models-latest.html
562
+ */
563
+ export var SageMakerModels;
564
+ (function (SageMakerModels) {
565
+ // Meta Llama 4 Series (Latest - 2025)
566
+ SageMakerModels["LLAMA_4_SCOUT_17B_16E"] = "meta-llama-4-scout-17b-16e-instruct";
567
+ SageMakerModels["LLAMA_4_MAVERICK_17B_128E"] = "meta-llama-4-maverick-17b-128e-instruct";
568
+ SageMakerModels["LLAMA_4_MAVERICK_17B_128E_FP8"] = "meta-llama-4-maverick-17b-128e-instruct-fp8";
569
+ // Meta Llama 3 Series
570
+ SageMakerModels["LLAMA_3_8B"] = "meta-llama-3-8b-instruct";
571
+ SageMakerModels["LLAMA_3_70B"] = "meta-llama-3-70b-instruct";
572
+ // Meta Code Llama Series
573
+ SageMakerModels["CODE_LLAMA_7B"] = "meta-code-llama-7b";
574
+ SageMakerModels["CODE_LLAMA_13B"] = "meta-code-llama-13b";
575
+ SageMakerModels["CODE_LLAMA_34B"] = "meta-code-llama-34b";
576
+ // Mistral AI Models
577
+ SageMakerModels["MISTRAL_SMALL_24B"] = "mistral-small-24b-instruct-2501";
578
+ SageMakerModels["MISTRAL_7B_INSTRUCT"] = "mistral-7b-instruct-v0.3";
579
+ SageMakerModels["MIXTRAL_8X7B"] = "mistral-mixtral-8x7b-instruct-v0.1";
580
+ SageMakerModels["MIXTRAL_8X22B"] = "mistral-mixtral-8x22b-instruct-v0.1";
581
+ // Falcon Models
582
+ SageMakerModels["FALCON_3_7B"] = "tii-falcon-3-7b-instruct";
583
+ SageMakerModels["FALCON_3_10B"] = "tii-falcon-3-10b-instruct";
584
+ SageMakerModels["FALCON_40B"] = "tii-falcon-40b-instruct";
585
+ SageMakerModels["FALCON_180B"] = "tii-falcon-180b";
586
+ })(SageMakerModels || (SageMakerModels = {}));
130
587
  /**
131
588
  * API Versions for various providers
132
589
  */
@@ -1,6 +1,6 @@
1
1
  import { ProviderFactory } from "./providerFactory.js";
2
2
  import { logger } from "../utils/logger.js";
3
- import { AIProviderName, GoogleAIModels, OpenAIModels, } from "../constants/enums.js";
3
+ import { AIProviderName, GoogleAIModels, OpenAIModels, AnthropicModels, VertexModels, MistralModels, OllamaModels, LiteLLMModels, HuggingFaceModels, } from "../constants/enums.js";
4
4
  /**
5
5
  * Provider Registry - registers all providers with the factory
6
6
  * This is where we migrate providers one by one to the new pattern
@@ -34,7 +34,7 @@ export class ProviderRegistry {
34
34
  ProviderFactory.registerProvider(AIProviderName.ANTHROPIC, async (modelName, _providerName, sdk) => {
35
35
  const { AnthropicProvider } = await import("../providers/anthropic.js");
36
36
  return new AnthropicProvider(modelName, sdk);
37
- }, "claude-3-5-sonnet-20241022", ["claude", "anthropic"]);
37
+ }, AnthropicModels.CLAUDE_SONNET_4_0, ["claude", "anthropic"]);
38
38
  // Register Amazon Bedrock provider
39
39
  ProviderFactory.registerProvider(AIProviderName.BEDROCK, async (modelName, _providerName, sdk, region) => {
40
40
  const { AmazonBedrockProvider } = await import("../providers/amazonBedrock.js");
@@ -54,27 +54,28 @@ export class ProviderRegistry {
54
54
  ProviderFactory.registerProvider(AIProviderName.VERTEX, async (modelName, providerName, sdk, region) => {
55
55
  const { GoogleVertexProvider } = await import("../providers/googleVertex.js");
56
56
  return new GoogleVertexProvider(modelName, providerName, sdk, region);
57
- }, "claude-sonnet-4@20250514", ["vertex", "googleVertex"]);
57
+ }, VertexModels.CLAUDE_4_0_SONNET, ["vertex", "googleVertex"]);
58
58
  // Register Hugging Face provider (Unified Router implementation)
59
59
  ProviderFactory.registerProvider(AIProviderName.HUGGINGFACE, async (modelName) => {
60
60
  const { HuggingFaceProvider } = await import("../providers/huggingFace.js");
61
61
  return new HuggingFaceProvider(modelName);
62
- }, process.env.HUGGINGFACE_MODEL || "microsoft/DialoGPT-medium", ["huggingface", "hf"]);
62
+ }, process.env.HUGGINGFACE_MODEL ||
63
+ HuggingFaceModels.QWEN_2_5_72B_INSTRUCT, ["huggingface", "hf"]);
63
64
  // Register Mistral AI provider
64
65
  ProviderFactory.registerProvider(AIProviderName.MISTRAL, async (modelName, _providerName, sdk) => {
65
66
  const { MistralProvider } = await import("../providers/mistral.js");
66
67
  return new MistralProvider(modelName, sdk);
67
- }, "mistral-large-latest", ["mistral"]);
68
+ }, MistralModels.MISTRAL_LARGE_LATEST, ["mistral"]);
68
69
  // Register Ollama provider
69
70
  ProviderFactory.registerProvider(AIProviderName.OLLAMA, async (modelName) => {
70
71
  const { OllamaProvider } = await import("../providers/ollama.js");
71
72
  return new OllamaProvider(modelName);
72
- }, process.env.OLLAMA_MODEL || "llama3.1:8b", ["ollama", "local"]);
73
+ }, process.env.OLLAMA_MODEL || OllamaModels.LLAMA3_2_LATEST, ["ollama", "local"]);
73
74
  // Register LiteLLM provider
74
75
  ProviderFactory.registerProvider(AIProviderName.LITELLM, async (modelName, _providerName, sdk) => {
75
76
  const { LiteLLMProvider } = await import("../providers/litellm.js");
76
77
  return new LiteLLMProvider(modelName, sdk);
77
- }, process.env.LITELLM_MODEL || "openai/gpt-4o-mini", ["litellm"]);
78
+ }, process.env.LITELLM_MODEL || LiteLLMModels.OPENAI_GPT_4O_MINI, ["litellm"]);
78
79
  // Register OpenAI Compatible provider
79
80
  ProviderFactory.registerProvider(AIProviderName.OPENAI_COMPATIBLE, async (modelName, _providerName, sdk) => {
80
81
  const { OpenAICompatibleProvider } = await import("../providers/openaiCompatible.js");