@juspay/neurolink 8.9.0 → 8.10.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +11 -0
- package/dist/adapters/providerImageAdapter.js +131 -14
- package/dist/cli/factories/commandFactory.d.ts +1 -0
- package/dist/cli/factories/commandFactory.js +52 -2
- package/dist/constants/enums.d.ts +346 -15
- package/dist/constants/enums.js +476 -19
- package/dist/factories/providerRegistry.js +8 -7
- package/dist/lib/adapters/providerImageAdapter.js +131 -14
- package/dist/lib/constants/enums.d.ts +346 -15
- package/dist/lib/constants/enums.js +476 -19
- package/dist/lib/factories/providerRegistry.js +8 -7
- package/dist/lib/models/modelRegistry.js +1962 -210
- package/dist/lib/types/generateTypes.d.ts +7 -0
- package/dist/lib/types/streamTypes.d.ts +7 -0
- package/dist/models/modelRegistry.js +1962 -210
- package/dist/types/generateTypes.d.ts +7 -0
- package/dist/types/streamTypes.d.ts +7 -0
- package/package.json +1 -1
package/dist/constants/enums.js
CHANGED
|
@@ -25,42 +25,209 @@ export var AIProviderName;
|
|
|
25
25
|
*/
|
|
26
26
|
export var BedrockModels;
|
|
27
27
|
(function (BedrockModels) {
|
|
28
|
+
// ============================================================================
|
|
29
|
+
// ANTHROPIC CLAUDE MODELS
|
|
30
|
+
// ============================================================================
|
|
28
31
|
// Claude 4.5 Series (Latest - September-November 2025)
|
|
29
|
-
BedrockModels["CLAUDE_4_5_SONNET"] = "anthropic.claude-sonnet-4-5-20250929-v1:0";
|
|
30
32
|
BedrockModels["CLAUDE_4_5_OPUS"] = "anthropic.claude-opus-4-5-20251124-v1:0";
|
|
33
|
+
BedrockModels["CLAUDE_4_5_SONNET"] = "anthropic.claude-sonnet-4-5-20250929-v1:0";
|
|
31
34
|
BedrockModels["CLAUDE_4_5_HAIKU"] = "anthropic.claude-haiku-4-5-20251001-v1:0";
|
|
35
|
+
// Claude 4 Series (May-August 2025)
|
|
36
|
+
BedrockModels["CLAUDE_4_1_OPUS"] = "anthropic.claude-opus-4-1-20250805-v1:0";
|
|
37
|
+
BedrockModels["CLAUDE_4_SONNET"] = "anthropic.claude-sonnet-4-20250514-v1:0";
|
|
32
38
|
// Claude 3.7 Series
|
|
33
39
|
BedrockModels["CLAUDE_3_7_SONNET"] = "anthropic.claude-3-7-sonnet-20250219-v1:0";
|
|
34
40
|
// Claude 3.5 Series
|
|
35
41
|
BedrockModels["CLAUDE_3_5_SONNET"] = "anthropic.claude-3-5-sonnet-20241022-v1:0";
|
|
42
|
+
BedrockModels["CLAUDE_3_5_HAIKU"] = "anthropic.claude-3-5-haiku-20241022-v1:0";
|
|
36
43
|
// Claude 3 Series (Legacy support)
|
|
37
44
|
BedrockModels["CLAUDE_3_SONNET"] = "anthropic.claude-3-sonnet-20240229-v1:0";
|
|
38
45
|
BedrockModels["CLAUDE_3_HAIKU"] = "anthropic.claude-3-haiku-20240307-v1:0";
|
|
46
|
+
// ============================================================================
|
|
47
|
+
// AMAZON NOVA MODELS
|
|
48
|
+
// ============================================================================
|
|
49
|
+
// Nova Generation 1
|
|
50
|
+
BedrockModels["NOVA_PREMIER"] = "amazon.nova-premier-v1:0";
|
|
51
|
+
BedrockModels["NOVA_PRO"] = "amazon.nova-pro-v1:0";
|
|
52
|
+
BedrockModels["NOVA_LITE"] = "amazon.nova-lite-v1:0";
|
|
53
|
+
BedrockModels["NOVA_MICRO"] = "amazon.nova-micro-v1:0";
|
|
54
|
+
// Nova Generation 2 (December 2025)
|
|
55
|
+
BedrockModels["NOVA_2_LITE"] = "amazon.nova-2-lite-v1:0";
|
|
56
|
+
BedrockModels["NOVA_2_SONIC"] = "amazon.nova-2-sonic-v1:0";
|
|
57
|
+
// Nova Specialized Models
|
|
58
|
+
BedrockModels["NOVA_SONIC"] = "amazon.nova-sonic-v1:0";
|
|
59
|
+
BedrockModels["NOVA_CANVAS"] = "amazon.nova-canvas-v1:0";
|
|
60
|
+
BedrockModels["NOVA_REEL"] = "amazon.nova-reel-v1:0";
|
|
61
|
+
BedrockModels["NOVA_REEL_V1_1"] = "amazon.nova-reel-v1:1";
|
|
62
|
+
BedrockModels["NOVA_MULTIMODAL_EMBEDDINGS"] = "amazon.nova-2-multimodal-embeddings-v1:0";
|
|
63
|
+
// ============================================================================
|
|
64
|
+
// AMAZON TITAN MODELS
|
|
65
|
+
// ============================================================================
|
|
66
|
+
// Titan Text Generation
|
|
67
|
+
BedrockModels["TITAN_TEXT_LARGE"] = "amazon.titan-tg1-large";
|
|
68
|
+
// Titan Text Embeddings
|
|
69
|
+
BedrockModels["TITAN_EMBED_TEXT_V2"] = "amazon.titan-embed-text-v2:0";
|
|
70
|
+
BedrockModels["TITAN_EMBED_TEXT_V1"] = "amazon.titan-embed-text-v1";
|
|
71
|
+
BedrockModels["TITAN_EMBED_G1_TEXT_02"] = "amazon.titan-embed-g1-text-02";
|
|
72
|
+
// Titan Multimodal Embeddings
|
|
73
|
+
BedrockModels["TITAN_EMBED_IMAGE_V1"] = "amazon.titan-embed-image-v1";
|
|
74
|
+
// Titan Image Generation
|
|
75
|
+
BedrockModels["TITAN_IMAGE_GENERATOR_V2"] = "amazon.titan-image-generator-v2:0";
|
|
76
|
+
// ============================================================================
|
|
77
|
+
// META LLAMA MODELS
|
|
78
|
+
// ============================================================================
|
|
79
|
+
// Llama 4 Series (2025)
|
|
80
|
+
BedrockModels["LLAMA_4_MAVERICK_17B"] = "meta.llama4-maverick-17b-instruct-v1:0";
|
|
81
|
+
BedrockModels["LLAMA_4_SCOUT_17B"] = "meta.llama4-scout-17b-instruct-v1:0";
|
|
82
|
+
// Llama 3.3 Series
|
|
83
|
+
BedrockModels["LLAMA_3_3_70B"] = "meta.llama3-3-70b-instruct-v1:0";
|
|
84
|
+
// Llama 3.2 Series (Multimodal)
|
|
85
|
+
BedrockModels["LLAMA_3_2_90B"] = "meta.llama3-2-90b-instruct-v1:0";
|
|
86
|
+
BedrockModels["LLAMA_3_2_11B"] = "meta.llama3-2-11b-instruct-v1:0";
|
|
87
|
+
BedrockModels["LLAMA_3_2_3B"] = "meta.llama3-2-3b-instruct-v1:0";
|
|
88
|
+
BedrockModels["LLAMA_3_2_1B"] = "meta.llama3-2-1b-instruct-v1:0";
|
|
89
|
+
// Llama 3.1 Series
|
|
90
|
+
BedrockModels["LLAMA_3_1_405B"] = "meta.llama3-1-405b-instruct-v1:0";
|
|
91
|
+
BedrockModels["LLAMA_3_1_70B"] = "meta.llama3-1-70b-instruct-v1:0";
|
|
92
|
+
BedrockModels["LLAMA_3_1_8B"] = "meta.llama3-1-8b-instruct-v1:0";
|
|
93
|
+
// Llama 3 Series (Legacy)
|
|
94
|
+
BedrockModels["LLAMA_3_70B"] = "meta.llama3-70b-instruct-v1:0";
|
|
95
|
+
BedrockModels["LLAMA_3_8B"] = "meta.llama3-8b-instruct-v1:0";
|
|
96
|
+
// ============================================================================
|
|
97
|
+
// MISTRAL AI MODELS
|
|
98
|
+
// ============================================================================
|
|
99
|
+
// Mistral Large Series
|
|
100
|
+
BedrockModels["MISTRAL_LARGE_3"] = "mistral.mistral-large-3-675b-instruct";
|
|
101
|
+
BedrockModels["MISTRAL_LARGE_2407"] = "mistral.mistral-large-2407-v1:0";
|
|
102
|
+
BedrockModels["MISTRAL_LARGE_2402"] = "mistral.mistral-large-2402-v1:0";
|
|
103
|
+
// Magistral & Ministral Series
|
|
104
|
+
BedrockModels["MAGISTRAL_SMALL_2509"] = "mistral.magistral-small-2509";
|
|
105
|
+
BedrockModels["MINISTRAL_3_14B"] = "mistral.ministral-3-14b-instruct";
|
|
106
|
+
BedrockModels["MINISTRAL_3_8B"] = "mistral.ministral-3-8b-instruct";
|
|
107
|
+
BedrockModels["MINISTRAL_3_3B"] = "mistral.ministral-3-3b-instruct";
|
|
108
|
+
// Mistral Base Series
|
|
109
|
+
BedrockModels["MISTRAL_7B"] = "mistral.mistral-7b-instruct-v0:2";
|
|
110
|
+
BedrockModels["MIXTRAL_8x7B"] = "mistral.mixtral-8x7b-instruct-v0:1";
|
|
111
|
+
// Mistral Multimodal & Audio
|
|
112
|
+
BedrockModels["PIXTRAL_LARGE_2502"] = "mistral.pixtral-large-2502-v1:0";
|
|
113
|
+
BedrockModels["VOXTRAL_SMALL_24B"] = "mistral.voxtral-small-24b-2507";
|
|
114
|
+
BedrockModels["VOXTRAL_MINI_3B"] = "mistral.voxtral-mini-3b-2507";
|
|
115
|
+
// ============================================================================
|
|
116
|
+
// OTHER MODELS
|
|
117
|
+
// ============================================================================
|
|
118
|
+
// Cohere Models
|
|
119
|
+
BedrockModels["COHERE_COMMAND_R_PLUS"] = "cohere.command-r-plus-v1:0";
|
|
120
|
+
BedrockModels["COHERE_COMMAND_R"] = "cohere.command-r-v1:0";
|
|
121
|
+
// DeepSeek Models
|
|
122
|
+
BedrockModels["DEEPSEEK_R1"] = "deepseek.r1-v1:0";
|
|
123
|
+
BedrockModels["DEEPSEEK_V3"] = "deepseek.v3-v1:0";
|
|
124
|
+
// Qwen Models
|
|
125
|
+
BedrockModels["QWEN_3_235B_A22B"] = "qwen.qwen3-235b-a22b-2507-v1:0";
|
|
126
|
+
BedrockModels["QWEN_3_CODER_480B_A35B"] = "qwen.qwen3-coder-480b-a35b-v1:0";
|
|
127
|
+
BedrockModels["QWEN_3_CODER_30B_A3B"] = "qwen.qwen3-coder-30b-a3b-v1:0";
|
|
128
|
+
BedrockModels["QWEN_3_32B"] = "qwen.qwen3-32b-v1:0";
|
|
129
|
+
BedrockModels["QWEN_3_NEXT_80B_A3B"] = "qwen.qwen3-next-80b-a3b";
|
|
130
|
+
BedrockModels["QWEN_3_VL_235B_A22B"] = "qwen.qwen3-vl-235b-a22b";
|
|
131
|
+
// Google Gemma
|
|
132
|
+
BedrockModels["GEMMA_3_27B_IT"] = "google.gemma-3-27b-it";
|
|
133
|
+
BedrockModels["GEMMA_3_12B_IT"] = "google.gemma-3-12b-it";
|
|
134
|
+
BedrockModels["GEMMA_3_4B_IT"] = "google.gemma-3-4b-it";
|
|
135
|
+
// AI21 Labs Models
|
|
136
|
+
BedrockModels["JAMBA_1_5_LARGE"] = "ai21.jamba-1-5-large-v1:0";
|
|
137
|
+
BedrockModels["JAMBA_1_5_MINI"] = "ai21.jamba-1-5-mini-v1:0";
|
|
39
138
|
})(BedrockModels || (BedrockModels = {}));
|
|
40
139
|
/**
|
|
41
140
|
* Supported Models for OpenAI
|
|
42
141
|
*/
|
|
43
142
|
export var OpenAIModels;
|
|
44
143
|
(function (OpenAIModels) {
|
|
45
|
-
|
|
46
|
-
OpenAIModels["
|
|
144
|
+
// GPT-5.2 Series (Released December 11, 2025) - Latest flagship models
|
|
145
|
+
OpenAIModels["GPT_5_2"] = "gpt-5.2";
|
|
146
|
+
OpenAIModels["GPT_5_2_CHAT_LATEST"] = "gpt-5.2-chat-latest";
|
|
147
|
+
OpenAIModels["GPT_5_2_PRO"] = "gpt-5.2-pro";
|
|
148
|
+
// GPT-5 Series (Released August 7, 2025)
|
|
149
|
+
OpenAIModels["GPT_5"] = "gpt-5";
|
|
150
|
+
OpenAIModels["GPT_5_MINI"] = "gpt-5-mini";
|
|
151
|
+
OpenAIModels["GPT_5_NANO"] = "gpt-5-nano";
|
|
152
|
+
// GPT-4.1 Series (Released April 14, 2025)
|
|
153
|
+
OpenAIModels["GPT_4_1"] = "gpt-4.1";
|
|
154
|
+
OpenAIModels["GPT_4_1_MINI"] = "gpt-4.1-mini";
|
|
155
|
+
OpenAIModels["GPT_4_1_NANO"] = "gpt-4.1-nano";
|
|
156
|
+
// GPT-4o Series
|
|
47
157
|
OpenAIModels["GPT_4O"] = "gpt-4o";
|
|
48
158
|
OpenAIModels["GPT_4O_MINI"] = "gpt-4o-mini";
|
|
49
|
-
|
|
159
|
+
// O-Series Reasoning Models
|
|
160
|
+
OpenAIModels["O3"] = "o3";
|
|
161
|
+
OpenAIModels["O3_MINI"] = "o3-mini";
|
|
162
|
+
OpenAIModels["O3_PRO"] = "o3-pro";
|
|
163
|
+
OpenAIModels["O4_MINI"] = "o4-mini";
|
|
164
|
+
OpenAIModels["O1"] = "o1";
|
|
50
165
|
OpenAIModels["O1_PREVIEW"] = "o1-preview";
|
|
51
166
|
OpenAIModels["O1_MINI"] = "o1-mini";
|
|
167
|
+
// GPT-4 Series (Legacy)
|
|
168
|
+
OpenAIModels["GPT_4"] = "gpt-4";
|
|
169
|
+
OpenAIModels["GPT_4_TURBO"] = "gpt-4-turbo";
|
|
170
|
+
// Legacy Models
|
|
171
|
+
OpenAIModels["GPT_3_5_TURBO"] = "gpt-3.5-turbo";
|
|
52
172
|
})(OpenAIModels || (OpenAIModels = {}));
|
|
173
|
+
/**
|
|
174
|
+
* Supported Models for Azure OpenAI
|
|
175
|
+
* Note: Azure uses deployment names, these are model identifiers
|
|
176
|
+
*/
|
|
177
|
+
export var AzureOpenAIModels;
|
|
178
|
+
(function (AzureOpenAIModels) {
|
|
179
|
+
// GPT-5.1 Series (Latest - December 2025)
|
|
180
|
+
AzureOpenAIModels["GPT_5_1"] = "gpt-5.1";
|
|
181
|
+
AzureOpenAIModels["GPT_5_1_CHAT"] = "gpt-5.1-chat";
|
|
182
|
+
AzureOpenAIModels["GPT_5_1_CODEX"] = "gpt-5.1-codex";
|
|
183
|
+
AzureOpenAIModels["GPT_5_1_CODEX_MINI"] = "gpt-5.1-codex-mini";
|
|
184
|
+
AzureOpenAIModels["GPT_5_1_CODEX_MAX"] = "gpt-5.1-codex-max";
|
|
185
|
+
// GPT-5.0 Series
|
|
186
|
+
AzureOpenAIModels["GPT_5"] = "gpt-5";
|
|
187
|
+
AzureOpenAIModels["GPT_5_MINI"] = "gpt-5-mini";
|
|
188
|
+
AzureOpenAIModels["GPT_5_NANO"] = "gpt-5-nano";
|
|
189
|
+
AzureOpenAIModels["GPT_5_CHAT"] = "gpt-5-chat";
|
|
190
|
+
AzureOpenAIModels["GPT_5_CODEX"] = "gpt-5-codex";
|
|
191
|
+
AzureOpenAIModels["GPT_5_PRO"] = "gpt-5-pro";
|
|
192
|
+
AzureOpenAIModels["GPT_5_TURBO"] = "gpt-5-turbo";
|
|
193
|
+
// O-Series Reasoning Models
|
|
194
|
+
AzureOpenAIModels["O4_MINI"] = "o4-mini";
|
|
195
|
+
AzureOpenAIModels["O3"] = "o3";
|
|
196
|
+
AzureOpenAIModels["O3_MINI"] = "o3-mini";
|
|
197
|
+
AzureOpenAIModels["O3_PRO"] = "o3-pro";
|
|
198
|
+
AzureOpenAIModels["O1"] = "o1";
|
|
199
|
+
AzureOpenAIModels["O1_MINI"] = "o1-mini";
|
|
200
|
+
AzureOpenAIModels["O1_PREVIEW"] = "o1-preview";
|
|
201
|
+
AzureOpenAIModels["CODEX_MINI"] = "codex-mini";
|
|
202
|
+
// GPT-4.1 Series
|
|
203
|
+
AzureOpenAIModels["GPT_4_1"] = "gpt-4.1";
|
|
204
|
+
AzureOpenAIModels["GPT_4_1_NANO"] = "gpt-4.1-nano";
|
|
205
|
+
AzureOpenAIModels["GPT_4_1_MINI"] = "gpt-4.1-mini";
|
|
206
|
+
// GPT-4o Series (Multimodal)
|
|
207
|
+
AzureOpenAIModels["GPT_4O"] = "gpt-4o";
|
|
208
|
+
AzureOpenAIModels["GPT_4O_MINI"] = "gpt-4o-mini";
|
|
209
|
+
// GPT-4 Turbo & GPT-4
|
|
210
|
+
AzureOpenAIModels["GPT_4_TURBO"] = "gpt-4-turbo";
|
|
211
|
+
AzureOpenAIModels["GPT_4"] = "gpt-4";
|
|
212
|
+
AzureOpenAIModels["GPT_4_32K"] = "gpt-4-32k";
|
|
213
|
+
// GPT-3.5 Turbo (Legacy)
|
|
214
|
+
AzureOpenAIModels["GPT_3_5_TURBO"] = "gpt-35-turbo";
|
|
215
|
+
AzureOpenAIModels["GPT_3_5_TURBO_INSTRUCT"] = "gpt-35-turbo-instruct";
|
|
216
|
+
})(AzureOpenAIModels || (AzureOpenAIModels = {}));
|
|
53
217
|
/**
|
|
54
218
|
* Supported Models for Google Vertex AI
|
|
55
219
|
*/
|
|
56
220
|
export var VertexModels;
|
|
57
221
|
(function (VertexModels) {
|
|
58
222
|
// Claude 4.5 Series (Latest - December 2025)
|
|
59
|
-
VertexModels["CLAUDE_4_5_SONNET"] = "claude-sonnet-4-5@20250929";
|
|
60
223
|
VertexModels["CLAUDE_4_5_OPUS"] = "claude-opus-4-5@20251124";
|
|
224
|
+
VertexModels["CLAUDE_4_5_SONNET"] = "claude-sonnet-4-5@20250929";
|
|
225
|
+
VertexModels["CLAUDE_4_5_HAIKU"] = "claude-haiku-4-5@20251001";
|
|
61
226
|
// Claude 4 Series (May 2025)
|
|
62
227
|
VertexModels["CLAUDE_4_0_SONNET"] = "claude-sonnet-4@20250514";
|
|
63
228
|
VertexModels["CLAUDE_4_0_OPUS"] = "claude-opus-4@20250514";
|
|
229
|
+
// Claude 3.7 Series (February 2025)
|
|
230
|
+
VertexModels["CLAUDE_3_7_SONNET"] = "claude-3-7-sonnet@20250219";
|
|
64
231
|
// Claude 3.5 Series (Still supported)
|
|
65
232
|
VertexModels["CLAUDE_3_5_SONNET"] = "claude-3-5-sonnet-20241022";
|
|
66
233
|
VertexModels["CLAUDE_3_5_HAIKU"] = "claude-3-5-haiku-20241022";
|
|
@@ -69,6 +236,8 @@ export var VertexModels;
|
|
|
69
236
|
VertexModels["CLAUDE_3_OPUS"] = "claude-3-opus-20240229";
|
|
70
237
|
VertexModels["CLAUDE_3_HAIKU"] = "claude-3-haiku-20240307";
|
|
71
238
|
// Gemini 3 Series (Preview)
|
|
239
|
+
/** Gemini 3 Pro - Base model with adaptive thinking */
|
|
240
|
+
VertexModels["GEMINI_3_PRO"] = "gemini-3-pro";
|
|
72
241
|
/** Gemini 3 Pro Preview - Versioned preview (November 2025) */
|
|
73
242
|
VertexModels["GEMINI_3_PRO_PREVIEW_11_2025"] = "gemini-3-pro-preview-11-2025";
|
|
74
243
|
/** Gemini 3 Pro Latest - Auto-updated alias (always points to latest preview) */
|
|
@@ -79,36 +248,41 @@ export var VertexModels;
|
|
|
79
248
|
VertexModels["GEMINI_2_5_PRO"] = "gemini-2.5-pro";
|
|
80
249
|
VertexModels["GEMINI_2_5_FLASH"] = "gemini-2.5-flash";
|
|
81
250
|
VertexModels["GEMINI_2_5_FLASH_LITE"] = "gemini-2.5-flash-lite";
|
|
251
|
+
VertexModels["GEMINI_2_5_FLASH_IMAGE"] = "gemini-2.5-flash-image";
|
|
82
252
|
// Gemini 2.0 Series
|
|
253
|
+
VertexModels["GEMINI_2_0_FLASH"] = "gemini-2.0-flash";
|
|
83
254
|
VertexModels["GEMINI_2_0_FLASH_001"] = "gemini-2.0-flash-001";
|
|
84
255
|
/** Gemini 2.0 Flash Lite - GA, production-ready, cost-optimized */
|
|
85
256
|
VertexModels["GEMINI_2_0_FLASH_LITE"] = "gemini-2.0-flash-lite";
|
|
86
257
|
// Gemini 1.5 Series (Legacy support)
|
|
87
|
-
VertexModels["GEMINI_1_5_PRO"] = "gemini-1.5-pro";
|
|
88
|
-
VertexModels["GEMINI_1_5_FLASH"] = "gemini-1.5-flash";
|
|
258
|
+
VertexModels["GEMINI_1_5_PRO"] = "gemini-1.5-pro-002";
|
|
259
|
+
VertexModels["GEMINI_1_5_FLASH"] = "gemini-1.5-flash-002";
|
|
89
260
|
})(VertexModels || (VertexModels = {}));
|
|
90
261
|
/**
|
|
91
262
|
* Supported Models for Google AI Studio
|
|
92
263
|
*/
|
|
93
264
|
export var GoogleAIModels;
|
|
94
265
|
(function (GoogleAIModels) {
|
|
95
|
-
// Gemini 3 Series
|
|
96
|
-
|
|
97
|
-
GoogleAIModels["
|
|
98
|
-
|
|
99
|
-
GoogleAIModels["GEMINI_3_PRO_LATEST"] = "gemini-3-pro-latest";
|
|
100
|
-
// Gemini 2.5 Series (Latest - 2025)
|
|
266
|
+
// Gemini 3 Series
|
|
267
|
+
GoogleAIModels["GEMINI_3_PRO_PREVIEW"] = "gemini-3-pro-preview";
|
|
268
|
+
GoogleAIModels["GEMINI_3_PRO_IMAGE_PREVIEW"] = "gemini-3-pro-image-preview";
|
|
269
|
+
// Gemini 2.5 Series
|
|
101
270
|
GoogleAIModels["GEMINI_2_5_PRO"] = "gemini-2.5-pro";
|
|
102
271
|
GoogleAIModels["GEMINI_2_5_FLASH"] = "gemini-2.5-flash";
|
|
103
272
|
GoogleAIModels["GEMINI_2_5_FLASH_LITE"] = "gemini-2.5-flash-lite";
|
|
273
|
+
GoogleAIModels["GEMINI_2_5_FLASH_IMAGE"] = "gemini-2.5-flash-image";
|
|
274
|
+
GoogleAIModels["GEMINI_2_5_FLASH_LIVE"] = "gemini-2.5-flash-native-audio-preview-09-2025";
|
|
104
275
|
// Gemini 2.0 Series
|
|
276
|
+
GoogleAIModels["GEMINI_2_0_FLASH"] = "gemini-2.0-flash";
|
|
105
277
|
GoogleAIModels["GEMINI_2_0_FLASH_001"] = "gemini-2.0-flash-001";
|
|
106
|
-
/** Gemini 2.0 Flash Lite - GA, production-ready, cost-optimized */
|
|
107
278
|
GoogleAIModels["GEMINI_2_0_FLASH_LITE"] = "gemini-2.0-flash-lite";
|
|
108
|
-
|
|
279
|
+
GoogleAIModels["GEMINI_2_0_FLASH_IMAGE"] = "gemini-2.0-flash-preview-image-generation";
|
|
280
|
+
// Gemini 1.5 Series (Legacy)
|
|
109
281
|
GoogleAIModels["GEMINI_1_5_PRO"] = "gemini-1.5-pro";
|
|
110
282
|
GoogleAIModels["GEMINI_1_5_FLASH"] = "gemini-1.5-flash";
|
|
111
|
-
|
|
283
|
+
// Embedding Models
|
|
284
|
+
GoogleAIModels["GEMINI_EMBEDDING"] = "gemini-embedding-001";
|
|
285
|
+
GoogleAIModels["TEXT_EMBEDDING_004"] = "text-embedding-004";
|
|
112
286
|
})(GoogleAIModels || (GoogleAIModels = {}));
|
|
113
287
|
/**
|
|
114
288
|
* Supported Models for Anthropic (Direct API)
|
|
@@ -116,17 +290,300 @@ export var GoogleAIModels;
|
|
|
116
290
|
export var AnthropicModels;
|
|
117
291
|
(function (AnthropicModels) {
|
|
118
292
|
// Claude 4.5 Series (Latest - September-November 2025)
|
|
293
|
+
AnthropicModels["CLAUDE_OPUS_4_5"] = "claude-opus-4-5-20251101";
|
|
119
294
|
AnthropicModels["CLAUDE_SONNET_4_5"] = "claude-sonnet-4-5-20250929";
|
|
120
|
-
AnthropicModels["CLAUDE_OPUS_4_5"] = "claude-opus-4-5-20251124";
|
|
121
295
|
AnthropicModels["CLAUDE_4_5_HAIKU"] = "claude-haiku-4-5-20251001";
|
|
122
|
-
// Claude
|
|
296
|
+
// Claude 4.1 Series (Legacy)
|
|
297
|
+
AnthropicModels["CLAUDE_OPUS_4_1"] = "claude-opus-4-1-20250805";
|
|
298
|
+
// Claude 4.0 Series (Legacy)
|
|
299
|
+
AnthropicModels["CLAUDE_OPUS_4_0"] = "claude-opus-4-20250514";
|
|
300
|
+
AnthropicModels["CLAUDE_SONNET_4_0"] = "claude-sonnet-4-20250514";
|
|
301
|
+
// Claude 3.7 Series (Legacy)
|
|
302
|
+
AnthropicModels["CLAUDE_SONNET_3_7"] = "claude-3-7-sonnet-20250219";
|
|
303
|
+
// Claude 3.5 Series (Legacy)
|
|
123
304
|
AnthropicModels["CLAUDE_3_5_SONNET"] = "claude-3-5-sonnet-20241022";
|
|
124
305
|
AnthropicModels["CLAUDE_3_5_HAIKU"] = "claude-3-5-haiku-20241022";
|
|
125
|
-
// Claude 3 Series (Legacy
|
|
306
|
+
// Claude 3 Series (Legacy - Deprecated)
|
|
126
307
|
AnthropicModels["CLAUDE_3_SONNET"] = "claude-3-sonnet-20240229";
|
|
127
308
|
AnthropicModels["CLAUDE_3_OPUS"] = "claude-3-opus-20240229";
|
|
128
309
|
AnthropicModels["CLAUDE_3_HAIKU"] = "claude-3-haiku-20240307";
|
|
129
310
|
})(AnthropicModels || (AnthropicModels = {}));
|
|
311
|
+
/**
|
|
312
|
+
* Supported Models for Mistral AI
|
|
313
|
+
*/
|
|
314
|
+
export var MistralModels;
|
|
315
|
+
(function (MistralModels) {
|
|
316
|
+
// Mistral Large (Latest)
|
|
317
|
+
MistralModels["MISTRAL_LARGE_LATEST"] = "mistral-large-latest";
|
|
318
|
+
MistralModels["MISTRAL_LARGE_2512"] = "mistral-large-2512";
|
|
319
|
+
// Mistral Medium
|
|
320
|
+
MistralModels["MISTRAL_MEDIUM_LATEST"] = "mistral-medium-latest";
|
|
321
|
+
MistralModels["MISTRAL_MEDIUM_2508"] = "mistral-medium-2508";
|
|
322
|
+
// Mistral Small
|
|
323
|
+
MistralModels["MISTRAL_SMALL_LATEST"] = "mistral-small-latest";
|
|
324
|
+
MistralModels["MISTRAL_SMALL_2506"] = "mistral-small-2506";
|
|
325
|
+
// Magistral (Reasoning)
|
|
326
|
+
MistralModels["MAGISTRAL_MEDIUM_LATEST"] = "magistral-medium-latest";
|
|
327
|
+
MistralModels["MAGISTRAL_SMALL_LATEST"] = "magistral-small-latest";
|
|
328
|
+
// Ministral (Edge Models)
|
|
329
|
+
MistralModels["MINISTRAL_14B_2512"] = "ministral-14b-2512";
|
|
330
|
+
MistralModels["MINISTRAL_8B_2512"] = "ministral-8b-2512";
|
|
331
|
+
MistralModels["MINISTRAL_3B_2512"] = "ministral-3b-2512";
|
|
332
|
+
// Codestral (Code Generation)
|
|
333
|
+
MistralModels["CODESTRAL_LATEST"] = "codestral-latest";
|
|
334
|
+
MistralModels["CODESTRAL_2508"] = "codestral-2508";
|
|
335
|
+
MistralModels["CODESTRAL_EMBED"] = "codestral-embed";
|
|
336
|
+
// Devstral (Software Development)
|
|
337
|
+
MistralModels["DEVSTRAL_MEDIUM_LATEST"] = "devstral-medium-latest";
|
|
338
|
+
MistralModels["DEVSTRAL_SMALL_LATEST"] = "devstral-small-latest";
|
|
339
|
+
// Pixtral (Multimodal/Vision)
|
|
340
|
+
MistralModels["PIXTRAL_LARGE"] = "pixtral-large";
|
|
341
|
+
MistralModels["PIXTRAL_12B"] = "pixtral-12b";
|
|
342
|
+
// Voxtral (Audio)
|
|
343
|
+
MistralModels["VOXTRAL_SMALL_LATEST"] = "voxtral-small-latest";
|
|
344
|
+
MistralModels["VOXTRAL_MINI_LATEST"] = "voxtral-mini-latest";
|
|
345
|
+
// Specialized Models
|
|
346
|
+
MistralModels["MISTRAL_NEMO"] = "mistral-nemo";
|
|
347
|
+
MistralModels["MISTRAL_EMBED"] = "mistral-embed";
|
|
348
|
+
MistralModels["MISTRAL_MODERATION_LATEST"] = "mistral-moderation-latest";
|
|
349
|
+
})(MistralModels || (MistralModels = {}));
|
|
350
|
+
/**
|
|
351
|
+
* Supported Models for Ollama (Local)
|
|
352
|
+
* All models can be run locally without requiring API keys or cloud services
|
|
353
|
+
*/
|
|
354
|
+
export var OllamaModels;
|
|
355
|
+
(function (OllamaModels) {
|
|
356
|
+
// Llama 4 Series - Multimodal with vision and tool capabilities
|
|
357
|
+
OllamaModels["LLAMA4_SCOUT"] = "llama4:scout";
|
|
358
|
+
OllamaModels["LLAMA4_MAVERICK"] = "llama4:maverick";
|
|
359
|
+
OllamaModels["LLAMA4_LATEST"] = "llama4:latest";
|
|
360
|
+
// Llama 3.3 Series - High-performance models
|
|
361
|
+
OllamaModels["LLAMA3_3_LATEST"] = "llama3.3:latest";
|
|
362
|
+
OllamaModels["LLAMA3_3_70B"] = "llama3.3:70b";
|
|
363
|
+
// Llama 3.2 Series - Optimized for edge and mobile deployment
|
|
364
|
+
OllamaModels["LLAMA3_2_LATEST"] = "llama3.2:latest";
|
|
365
|
+
OllamaModels["LLAMA3_2_3B"] = "llama3.2:3b";
|
|
366
|
+
OllamaModels["LLAMA3_2_1B"] = "llama3.2:1b";
|
|
367
|
+
// Llama 3.1 Series - Open models rivaling proprietary models
|
|
368
|
+
OllamaModels["LLAMA3_1_8B"] = "llama3.1:8b";
|
|
369
|
+
OllamaModels["LLAMA3_1_70B"] = "llama3.1:70b";
|
|
370
|
+
OllamaModels["LLAMA3_1_405B"] = "llama3.1:405b";
|
|
371
|
+
// Qwen 3 Series - Advanced reasoning and multilingual support
|
|
372
|
+
OllamaModels["QWEN3_4B"] = "qwen3:4b";
|
|
373
|
+
OllamaModels["QWEN3_8B"] = "qwen3:8b";
|
|
374
|
+
OllamaModels["QWEN3_14B"] = "qwen3:14b";
|
|
375
|
+
OllamaModels["QWEN3_32B"] = "qwen3:32b";
|
|
376
|
+
OllamaModels["QWEN3_72B"] = "qwen3:72b";
|
|
377
|
+
// Qwen 2.5 Series - Enhanced coding and mathematics
|
|
378
|
+
OllamaModels["QWEN2_5_3B"] = "qwen2.5:3b";
|
|
379
|
+
OllamaModels["QWEN2_5_7B"] = "qwen2.5:7b";
|
|
380
|
+
OllamaModels["QWEN2_5_14B"] = "qwen2.5:14b";
|
|
381
|
+
OllamaModels["QWEN2_5_32B"] = "qwen2.5:32b";
|
|
382
|
+
OllamaModels["QWEN2_5_72B"] = "qwen2.5:72b";
|
|
383
|
+
// Qwen Reasoning Model
|
|
384
|
+
OllamaModels["QWQ_32B"] = "qwq:32b";
|
|
385
|
+
OllamaModels["QWQ_LATEST"] = "qwq:latest";
|
|
386
|
+
// DeepSeek-R1 Series - State-of-the-art reasoning models
|
|
387
|
+
OllamaModels["DEEPSEEK_R1_1_5B"] = "deepseek-r1:1.5b";
|
|
388
|
+
OllamaModels["DEEPSEEK_R1_7B"] = "deepseek-r1:7b";
|
|
389
|
+
OllamaModels["DEEPSEEK_R1_8B"] = "deepseek-r1:8b";
|
|
390
|
+
OllamaModels["DEEPSEEK_R1_14B"] = "deepseek-r1:14b";
|
|
391
|
+
OllamaModels["DEEPSEEK_R1_32B"] = "deepseek-r1:32b";
|
|
392
|
+
OllamaModels["DEEPSEEK_R1_70B"] = "deepseek-r1:70b";
|
|
393
|
+
// DeepSeek-V3 Series - Mixture of Experts model
|
|
394
|
+
OllamaModels["DEEPSEEK_V3_671B"] = "deepseek-v3:671b";
|
|
395
|
+
OllamaModels["DEEPSEEK_V3_LATEST"] = "deepseek-v3:latest";
|
|
396
|
+
// Mistral AI Series - Efficient general-purpose models
|
|
397
|
+
OllamaModels["MISTRAL_LATEST"] = "mistral:latest";
|
|
398
|
+
OllamaModels["MISTRAL_7B"] = "mistral:7b";
|
|
399
|
+
OllamaModels["MISTRAL_SMALL_LATEST"] = "mistral-small:latest";
|
|
400
|
+
OllamaModels["MISTRAL_NEMO_LATEST"] = "mistral-nemo:latest";
|
|
401
|
+
OllamaModels["MISTRAL_LARGE_LATEST"] = "mistral-large:latest";
|
|
402
|
+
// Google Gemma Series - Efficient edge and cloud models
|
|
403
|
+
OllamaModels["GEMMA3_LATEST"] = "gemma3:latest";
|
|
404
|
+
OllamaModels["GEMMA2_2B"] = "gemma2:2b";
|
|
405
|
+
OllamaModels["GEMMA2_9B"] = "gemma2:9b";
|
|
406
|
+
OllamaModels["GEMMA2_27B"] = "gemma2:27b";
|
|
407
|
+
// Microsoft Phi Series - Compact, efficient models
|
|
408
|
+
OllamaModels["PHI4_LATEST"] = "phi4:latest";
|
|
409
|
+
OllamaModels["PHI4_14B"] = "phi4:14b";
|
|
410
|
+
OllamaModels["PHI3_MINI"] = "phi3:mini";
|
|
411
|
+
OllamaModels["PHI3_3_8B"] = "phi3:3.8b";
|
|
412
|
+
OllamaModels["PHI3_MEDIUM"] = "phi3:medium";
|
|
413
|
+
OllamaModels["PHI3_14B"] = "phi3:14b";
|
|
414
|
+
// Vision-Language Models
|
|
415
|
+
OllamaModels["LLAVA_7B"] = "llava:7b";
|
|
416
|
+
OllamaModels["LLAVA_13B"] = "llava:13b";
|
|
417
|
+
OllamaModels["LLAVA_34B"] = "llava:34b";
|
|
418
|
+
OllamaModels["LLAVA_LLAMA3_8B"] = "llava-llama3:8b";
|
|
419
|
+
// Code-Specialized Models
|
|
420
|
+
OllamaModels["CODELLAMA_7B"] = "codellama:7b";
|
|
421
|
+
OllamaModels["CODELLAMA_13B"] = "codellama:13b";
|
|
422
|
+
OllamaModels["CODELLAMA_34B"] = "codellama:34b";
|
|
423
|
+
OllamaModels["CODELLAMA_70B"] = "codellama:70b";
|
|
424
|
+
OllamaModels["QWEN2_5_CODER_7B"] = "qwen2.5-coder:7b";
|
|
425
|
+
OllamaModels["QWEN2_5_CODER_32B"] = "qwen2.5-coder:32b";
|
|
426
|
+
OllamaModels["STARCODER2_3B"] = "starcoder2:3b";
|
|
427
|
+
OllamaModels["STARCODER2_7B"] = "starcoder2:7b";
|
|
428
|
+
OllamaModels["STARCODER2_15B"] = "starcoder2:15b";
|
|
429
|
+
// Mixture of Experts Models
|
|
430
|
+
OllamaModels["MIXTRAL_8X7B"] = "mixtral:8x7b";
|
|
431
|
+
OllamaModels["MIXTRAL_8X22B"] = "mixtral:8x22b";
|
|
432
|
+
// Enterprise Models
|
|
433
|
+
OllamaModels["COMMAND_R_PLUS"] = "command-r-plus:104b";
|
|
434
|
+
})(OllamaModels || (OllamaModels = {}));
|
|
435
|
+
/**
|
|
436
|
+
* Common Models for LiteLLM Proxy
|
|
437
|
+
* LiteLLM supports 100+ models through unified proxy interface
|
|
438
|
+
* Models use provider-specific prefixes (e.g., "openai/", "anthropic/")
|
|
439
|
+
*/
|
|
440
|
+
export var LiteLLMModels;
|
|
441
|
+
(function (LiteLLMModels) {
|
|
442
|
+
// OpenAI via LiteLLM
|
|
443
|
+
LiteLLMModels["OPENAI_GPT_5"] = "openai/gpt-5";
|
|
444
|
+
LiteLLMModels["OPENAI_GPT_4O"] = "openai/gpt-4o";
|
|
445
|
+
LiteLLMModels["OPENAI_GPT_4O_MINI"] = "openai/gpt-4o-mini";
|
|
446
|
+
LiteLLMModels["OPENAI_GPT_4_TURBO"] = "openai/gpt-4-turbo";
|
|
447
|
+
LiteLLMModels["OPENAI_GPT_4"] = "openai/gpt-4";
|
|
448
|
+
LiteLLMModels["OPENAI_GPT_3_5_TURBO"] = "openai/gpt-3.5-turbo";
|
|
449
|
+
// Anthropic via LiteLLM
|
|
450
|
+
LiteLLMModels["ANTHROPIC_CLAUDE_SONNET_4_5"] = "anthropic/claude-sonnet-4-5-20250929";
|
|
451
|
+
LiteLLMModels["ANTHROPIC_CLAUDE_OPUS_4_1"] = "anthropic/claude-opus-4-1-20250805";
|
|
452
|
+
LiteLLMModels["ANTHROPIC_CLAUDE_3_5_SONNET"] = "anthropic/claude-3-5-sonnet-20240620";
|
|
453
|
+
LiteLLMModels["ANTHROPIC_CLAUDE_3_HAIKU"] = "anthropic/claude-3-haiku-20240307";
|
|
454
|
+
// Google Vertex AI via LiteLLM
|
|
455
|
+
LiteLLMModels["VERTEX_GEMINI_2_5_PRO"] = "vertex_ai/gemini-2.5-pro";
|
|
456
|
+
LiteLLMModels["VERTEX_GEMINI_1_5_PRO"] = "vertex_ai/gemini-1.5-pro";
|
|
457
|
+
LiteLLMModels["VERTEX_GEMINI_1_5_FLASH"] = "vertex_ai/gemini-1.5-flash";
|
|
458
|
+
// Google AI Studio (Gemini) via LiteLLM
|
|
459
|
+
LiteLLMModels["GEMINI_2_5_PRO"] = "gemini/gemini-2.5-pro";
|
|
460
|
+
LiteLLMModels["GEMINI_2_0_FLASH"] = "gemini/gemini-2.0-flash";
|
|
461
|
+
LiteLLMModels["GEMINI_1_5_PRO"] = "gemini/gemini-1.5-pro";
|
|
462
|
+
LiteLLMModels["GEMINI_1_5_FLASH"] = "gemini/gemini-1.5-flash";
|
|
463
|
+
// Groq via LiteLLM
|
|
464
|
+
LiteLLMModels["GROQ_LLAMA_3_1_70B_VERSATILE"] = "groq/llama-3.1-70b-versatile";
|
|
465
|
+
LiteLLMModels["GROQ_LLAMA_3_1_8B_INSTANT"] = "groq/llama-3.1-8b-instant";
|
|
466
|
+
LiteLLMModels["GROQ_LLAMA_3_2_11B_VISION"] = "groq/llama-3.2-11b-vision-preview";
|
|
467
|
+
LiteLLMModels["GROQ_MIXTRAL_8X7B"] = "groq/mixtral-8x7b-32768";
|
|
468
|
+
// Together AI via LiteLLM
|
|
469
|
+
LiteLLMModels["TOGETHER_LLAMA_2_70B_CHAT"] = "together_ai/togethercomputer/llama-2-70b-chat";
|
|
470
|
+
LiteLLMModels["TOGETHER_MIXTRAL_8X7B"] = "together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1";
|
|
471
|
+
LiteLLMModels["TOGETHER_CODELLAMA_34B"] = "together_ai/codellama/CodeLlama-34b-Instruct-hf";
|
|
472
|
+
// DeepInfra via LiteLLM
|
|
473
|
+
LiteLLMModels["DEEPINFRA_LLAMA_3_70B"] = "deepinfra/meta-llama/Meta-Llama-3-70B-Instruct";
|
|
474
|
+
LiteLLMModels["DEEPINFRA_LLAMA_2_70B"] = "deepinfra/meta-llama/Llama-2-70b-chat-hf";
|
|
475
|
+
LiteLLMModels["DEEPINFRA_MISTRAL_7B"] = "deepinfra/mistralai/Mistral-7B-Instruct-v0.1";
|
|
476
|
+
// Mistral AI via LiteLLM
|
|
477
|
+
LiteLLMModels["MISTRAL_LARGE"] = "mistral/mistral-large-latest";
|
|
478
|
+
LiteLLMModels["MISTRAL_SMALL"] = "mistral/mistral-small-latest";
|
|
479
|
+
LiteLLMModels["MISTRAL_MAGISTRAL_MEDIUM"] = "mistral/magistral-medium-2506";
|
|
480
|
+
// AWS Bedrock via LiteLLM
|
|
481
|
+
LiteLLMModels["BEDROCK_CLAUDE_3_5_SONNET"] = "bedrock/anthropic.claude-3-5-sonnet-20240620-v1:0";
|
|
482
|
+
LiteLLMModels["BEDROCK_CLAUDE_3_HAIKU"] = "bedrock/anthropic.claude-3-haiku-20240307-v1:0";
|
|
483
|
+
// Perplexity AI via LiteLLM
|
|
484
|
+
LiteLLMModels["PERPLEXITY_SONAR_PRO"] = "perplexity/sonar-pro";
|
|
485
|
+
LiteLLMModels["PERPLEXITY_SONAR_REASONING_PRO"] = "perplexity/sonar-reasoning-pro";
|
|
486
|
+
})(LiteLLMModels || (LiteLLMModels = {}));
|
|
487
|
+
/**
|
|
488
|
+
* Supported Models for Hugging Face Inference API
|
|
489
|
+
*/
|
|
490
|
+
export var HuggingFaceModels;
|
|
491
|
+
(function (HuggingFaceModels) {
|
|
492
|
+
// Meta Llama 3.3
|
|
493
|
+
HuggingFaceModels["LLAMA_3_3_70B_INSTRUCT"] = "meta-llama/Llama-3.3-70B-Instruct";
|
|
494
|
+
// Meta Llama 3.2
|
|
495
|
+
HuggingFaceModels["LLAMA_3_2_1B"] = "meta-llama/Llama-3.2-1B";
|
|
496
|
+
HuggingFaceModels["LLAMA_3_2_3B_INSTRUCT"] = "meta-llama/Llama-3.2-3B-Instruct";
|
|
497
|
+
// Meta Llama 3.1
|
|
498
|
+
HuggingFaceModels["LLAMA_3_1_8B"] = "meta-llama/Llama-3.1-8B";
|
|
499
|
+
HuggingFaceModels["LLAMA_3_1_70B_INSTRUCT"] = "meta-llama/Llama-3.1-70B-Instruct";
|
|
500
|
+
HuggingFaceModels["LLAMA_3_1_405B_INSTRUCT"] = "meta-llama/Llama-3.1-405B-Instruct";
|
|
501
|
+
// Meta Llama 3.0
|
|
502
|
+
HuggingFaceModels["LLAMA_3_8B_INSTRUCT"] = "meta-llama/Meta-Llama-3-8B-Instruct";
|
|
503
|
+
HuggingFaceModels["LLAMA_3_70B_INSTRUCT"] = "meta-llama/Meta-Llama-3-70B-Instruct";
|
|
504
|
+
// Mistral Large
|
|
505
|
+
HuggingFaceModels["MISTRAL_LARGE_3_675B"] = "mistralai/Mistral-Large-3-675B-Instruct-2512";
|
|
506
|
+
// Mistral Small
|
|
507
|
+
HuggingFaceModels["MISTRAL_SMALL_3_1_24B"] = "mistralai/Mistral-Small-3.1-24B-Instruct-2503";
|
|
508
|
+
HuggingFaceModels["MISTRAL_SMALL_24B"] = "mistralai/Mistral-Small-24B-Instruct-2501";
|
|
509
|
+
// Mistral
|
|
510
|
+
HuggingFaceModels["MISTRAL_7B_INSTRUCT"] = "mistralai/Mistral-7B-Instruct-v0.2";
|
|
511
|
+
HuggingFaceModels["MIXTRAL_8X7B_INSTRUCT"] = "mistralai/Mixtral-8x7B-Instruct-v0.1";
|
|
512
|
+
// Mistral Devstral
|
|
513
|
+
HuggingFaceModels["DEVSTRAL_2"] = "mistralai/Devstral-2";
|
|
514
|
+
// Qwen 2.5
|
|
515
|
+
HuggingFaceModels["QWEN_2_5_7B"] = "Qwen/Qwen2.5-7B";
|
|
516
|
+
HuggingFaceModels["QWEN_2_5_32B"] = "Qwen/Qwen2.5-32B";
|
|
517
|
+
HuggingFaceModels["QWEN_2_5_72B_INSTRUCT"] = "Qwen/Qwen2.5-72B-Instruct";
|
|
518
|
+
// Qwen 2.5 Coder
|
|
519
|
+
HuggingFaceModels["QWEN_2_5_CODER_7B"] = "Qwen/Qwen2.5-Coder-7B";
|
|
520
|
+
HuggingFaceModels["QWEN_2_5_CODER_32B_INSTRUCT"] = "Qwen/Qwen2.5-Coder-32B-Instruct";
|
|
521
|
+
// Qwen QwQ
|
|
522
|
+
HuggingFaceModels["QWQ_32B"] = "Qwen/QwQ-32B";
|
|
523
|
+
// Qwen 2.5 VL (Multimodal)
|
|
524
|
+
HuggingFaceModels["QWEN_2_5_VL_32B"] = "Qwen/Qwen2.5-VL-32B-Instruct";
|
|
525
|
+
// DeepSeek
|
|
526
|
+
HuggingFaceModels["DEEPSEEK_R1"] = "deepseek-ai/DeepSeek-R1";
|
|
527
|
+
HuggingFaceModels["DEEPSEEK_V3"] = "deepseek-ai/DeepSeek-V3";
|
|
528
|
+
HuggingFaceModels["DEEPSEEK_V3_1"] = "deepseek-ai/DeepSeek-V3.1";
|
|
529
|
+
HuggingFaceModels["DEEPSEEK_V3_2_EXP"] = "deepseek-ai/DeepSeek-V3.2-Exp";
|
|
530
|
+
// Microsoft Phi
|
|
531
|
+
HuggingFaceModels["PHI_4"] = "microsoft/phi-4";
|
|
532
|
+
HuggingFaceModels["PHI_4_REASONING"] = "microsoft/Phi-4-reasoning";
|
|
533
|
+
HuggingFaceModels["PHI_4_MINI_INSTRUCT"] = "microsoft/Phi-4-mini-instruct";
|
|
534
|
+
HuggingFaceModels["PHI_4_MINI_REASONING"] = "microsoft/Phi-4-mini-reasoning";
|
|
535
|
+
HuggingFaceModels["PHI_3_MINI_128K_INSTRUCT"] = "microsoft/Phi-3-mini-128k-instruct";
|
|
536
|
+
HuggingFaceModels["PHI_3_VISION_128K_INSTRUCT"] = "microsoft/Phi-3-vision-128k-instruct";
|
|
537
|
+
// Google Gemma 3
|
|
538
|
+
HuggingFaceModels["GEMMA_3_270M"] = "google/gemma-3-270m";
|
|
539
|
+
HuggingFaceModels["GEMMA_3_1B_IT"] = "google/gemma-3-1b-it";
|
|
540
|
+
HuggingFaceModels["GEMMA_3_4B_IT"] = "google/gemma-3-4b-it";
|
|
541
|
+
HuggingFaceModels["GEMMA_3_12B_IT"] = "google/gemma-3-12b-it";
|
|
542
|
+
HuggingFaceModels["GEMMA_3_27B_IT"] = "google/gemma-3-27b-it";
|
|
543
|
+
// Google Gemma 2
|
|
544
|
+
HuggingFaceModels["GEMMA_2_9B"] = "google/gemma-2-9b";
|
|
545
|
+
HuggingFaceModels["GEMMA_2_27B"] = "google/gemma-2-27b";
|
|
546
|
+
// Google Gemma 1
|
|
547
|
+
HuggingFaceModels["GEMMA_2B"] = "google/gemma-2b";
|
|
548
|
+
HuggingFaceModels["GEMMA_7B"] = "google/gemma-7b";
|
|
549
|
+
// Falcon
|
|
550
|
+
HuggingFaceModels["FALCON_40B_INSTRUCT"] = "tiiuae/falcon-40b-instruct";
|
|
551
|
+
HuggingFaceModels["FALCON_180B_CHAT"] = "tiiuae/falcon-180B-chat";
|
|
552
|
+
// Code Models
|
|
553
|
+
HuggingFaceModels["STARCODER2_15B"] = "bigcode/starcoder2-15b";
|
|
554
|
+
HuggingFaceModels["CODELLAMA_34B_INSTRUCT"] = "codellama/CodeLlama-34b-Instruct-hf";
|
|
555
|
+
// BLOOM
|
|
556
|
+
HuggingFaceModels["BLOOM_7B1"] = "bigscience/bloom-7b1";
|
|
557
|
+
HuggingFaceModels["BLOOM_1B3"] = "bigscience/bloom-1b3";
|
|
558
|
+
})(HuggingFaceModels || (HuggingFaceModels = {}));
|
|
559
|
+
/**
|
|
560
|
+
* Supported Models for AWS SageMaker JumpStart
|
|
561
|
+
* https://docs.aws.amazon.com/sagemaker/latest/dg/jumpstart-foundation-models-latest.html
|
|
562
|
+
*/
|
|
563
|
+
export var SageMakerModels;
|
|
564
|
+
(function (SageMakerModels) {
|
|
565
|
+
// Meta Llama 4 Series (Latest - 2025)
|
|
566
|
+
SageMakerModels["LLAMA_4_SCOUT_17B_16E"] = "meta-llama-4-scout-17b-16e-instruct";
|
|
567
|
+
SageMakerModels["LLAMA_4_MAVERICK_17B_128E"] = "meta-llama-4-maverick-17b-128e-instruct";
|
|
568
|
+
SageMakerModels["LLAMA_4_MAVERICK_17B_128E_FP8"] = "meta-llama-4-maverick-17b-128e-instruct-fp8";
|
|
569
|
+
// Meta Llama 3 Series
|
|
570
|
+
SageMakerModels["LLAMA_3_8B"] = "meta-llama-3-8b-instruct";
|
|
571
|
+
SageMakerModels["LLAMA_3_70B"] = "meta-llama-3-70b-instruct";
|
|
572
|
+
// Meta Code Llama Series
|
|
573
|
+
SageMakerModels["CODE_LLAMA_7B"] = "meta-code-llama-7b";
|
|
574
|
+
SageMakerModels["CODE_LLAMA_13B"] = "meta-code-llama-13b";
|
|
575
|
+
SageMakerModels["CODE_LLAMA_34B"] = "meta-code-llama-34b";
|
|
576
|
+
// Mistral AI Models
|
|
577
|
+
SageMakerModels["MISTRAL_SMALL_24B"] = "mistral-small-24b-instruct-2501";
|
|
578
|
+
SageMakerModels["MISTRAL_7B_INSTRUCT"] = "mistral-7b-instruct-v0.3";
|
|
579
|
+
SageMakerModels["MIXTRAL_8X7B"] = "mistral-mixtral-8x7b-instruct-v0.1";
|
|
580
|
+
SageMakerModels["MIXTRAL_8X22B"] = "mistral-mixtral-8x22b-instruct-v0.1";
|
|
581
|
+
// Falcon Models
|
|
582
|
+
SageMakerModels["FALCON_3_7B"] = "tii-falcon-3-7b-instruct";
|
|
583
|
+
SageMakerModels["FALCON_3_10B"] = "tii-falcon-3-10b-instruct";
|
|
584
|
+
SageMakerModels["FALCON_40B"] = "tii-falcon-40b-instruct";
|
|
585
|
+
SageMakerModels["FALCON_180B"] = "tii-falcon-180b";
|
|
586
|
+
})(SageMakerModels || (SageMakerModels = {}));
|
|
130
587
|
/**
|
|
131
588
|
* API Versions for various providers
|
|
132
589
|
*/
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { ProviderFactory } from "./providerFactory.js";
|
|
2
2
|
import { logger } from "../utils/logger.js";
|
|
3
|
-
import { AIProviderName, GoogleAIModels, OpenAIModels, } from "../constants/enums.js";
|
|
3
|
+
import { AIProviderName, GoogleAIModels, OpenAIModels, AnthropicModels, VertexModels, MistralModels, OllamaModels, LiteLLMModels, HuggingFaceModels, } from "../constants/enums.js";
|
|
4
4
|
/**
|
|
5
5
|
* Provider Registry - registers all providers with the factory
|
|
6
6
|
* This is where we migrate providers one by one to the new pattern
|
|
@@ -34,7 +34,7 @@ export class ProviderRegistry {
|
|
|
34
34
|
ProviderFactory.registerProvider(AIProviderName.ANTHROPIC, async (modelName, _providerName, sdk) => {
|
|
35
35
|
const { AnthropicProvider } = await import("../providers/anthropic.js");
|
|
36
36
|
return new AnthropicProvider(modelName, sdk);
|
|
37
|
-
},
|
|
37
|
+
}, AnthropicModels.CLAUDE_SONNET_4_0, ["claude", "anthropic"]);
|
|
38
38
|
// Register Amazon Bedrock provider
|
|
39
39
|
ProviderFactory.registerProvider(AIProviderName.BEDROCK, async (modelName, _providerName, sdk, region) => {
|
|
40
40
|
const { AmazonBedrockProvider } = await import("../providers/amazonBedrock.js");
|
|
@@ -54,27 +54,28 @@ export class ProviderRegistry {
|
|
|
54
54
|
ProviderFactory.registerProvider(AIProviderName.VERTEX, async (modelName, providerName, sdk, region) => {
|
|
55
55
|
const { GoogleVertexProvider } = await import("../providers/googleVertex.js");
|
|
56
56
|
return new GoogleVertexProvider(modelName, providerName, sdk, region);
|
|
57
|
-
},
|
|
57
|
+
}, VertexModels.CLAUDE_4_0_SONNET, ["vertex", "googleVertex"]);
|
|
58
58
|
// Register Hugging Face provider (Unified Router implementation)
|
|
59
59
|
ProviderFactory.registerProvider(AIProviderName.HUGGINGFACE, async (modelName) => {
|
|
60
60
|
const { HuggingFaceProvider } = await import("../providers/huggingFace.js");
|
|
61
61
|
return new HuggingFaceProvider(modelName);
|
|
62
|
-
}, process.env.HUGGINGFACE_MODEL ||
|
|
62
|
+
}, process.env.HUGGINGFACE_MODEL ||
|
|
63
|
+
HuggingFaceModels.QWEN_2_5_72B_INSTRUCT, ["huggingface", "hf"]);
|
|
63
64
|
// Register Mistral AI provider
|
|
64
65
|
ProviderFactory.registerProvider(AIProviderName.MISTRAL, async (modelName, _providerName, sdk) => {
|
|
65
66
|
const { MistralProvider } = await import("../providers/mistral.js");
|
|
66
67
|
return new MistralProvider(modelName, sdk);
|
|
67
|
-
},
|
|
68
|
+
}, MistralModels.MISTRAL_LARGE_LATEST, ["mistral"]);
|
|
68
69
|
// Register Ollama provider
|
|
69
70
|
ProviderFactory.registerProvider(AIProviderName.OLLAMA, async (modelName) => {
|
|
70
71
|
const { OllamaProvider } = await import("../providers/ollama.js");
|
|
71
72
|
return new OllamaProvider(modelName);
|
|
72
|
-
}, process.env.OLLAMA_MODEL ||
|
|
73
|
+
}, process.env.OLLAMA_MODEL || OllamaModels.LLAMA3_2_LATEST, ["ollama", "local"]);
|
|
73
74
|
// Register LiteLLM provider
|
|
74
75
|
ProviderFactory.registerProvider(AIProviderName.LITELLM, async (modelName, _providerName, sdk) => {
|
|
75
76
|
const { LiteLLMProvider } = await import("../providers/litellm.js");
|
|
76
77
|
return new LiteLLMProvider(modelName, sdk);
|
|
77
|
-
}, process.env.LITELLM_MODEL ||
|
|
78
|
+
}, process.env.LITELLM_MODEL || LiteLLMModels.OPENAI_GPT_4O_MINI, ["litellm"]);
|
|
78
79
|
// Register OpenAI Compatible provider
|
|
79
80
|
ProviderFactory.registerProvider(AIProviderName.OPENAI_COMPATIBLE, async (modelName, _providerName, sdk) => {
|
|
80
81
|
const { OpenAICompatibleProvider } = await import("../providers/openaiCompatible.js");
|