@everworker/oneringai 0.4.4 → 0.4.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +25 -12
- package/dist/{ImageModel-OWbA277F.d.ts → ImageModel-1uP-2vk7.d.ts} +8 -2
- package/dist/{ImageModel-Ds5_6sf7.d.cts → ImageModel-BDI37OED.d.cts} +8 -2
- package/dist/capabilities/agents/index.d.cts +1 -1
- package/dist/capabilities/agents/index.d.ts +1 -1
- package/dist/capabilities/images/index.cjs +149 -7
- package/dist/capabilities/images/index.cjs.map +1 -1
- package/dist/capabilities/images/index.d.cts +1 -1
- package/dist/capabilities/images/index.d.ts +1 -1
- package/dist/capabilities/images/index.js +149 -7
- package/dist/capabilities/images/index.js.map +1 -1
- package/dist/{index-DmYrHH7d.d.cts → index-DJ-qAK15.d.cts} +25 -5
- package/dist/{index-Dyl6pHfq.d.ts → index-oBtp-8Qn.d.ts} +25 -5
- package/dist/index.cjs +1348 -199
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +47 -13
- package/dist/index.d.ts +47 -13
- package/dist/index.js +1348 -199
- package/dist/index.js.map +1 -1
- package/dist/shared/index.cjs +785 -133
- package/dist/shared/index.cjs.map +1 -1
- package/dist/shared/index.d.cts +20 -2
- package/dist/shared/index.d.ts +20 -2
- package/dist/shared/index.js +785 -133
- package/dist/shared/index.js.map +1 -1
- package/package.json +1 -1
package/dist/shared/index.js
CHANGED
|
@@ -22,18 +22,30 @@ function isVendor(value) {
|
|
|
22
22
|
// src/domain/entities/Model.ts
|
|
23
23
|
var LLM_MODELS = {
|
|
24
24
|
[Vendor.OpenAI]: {
|
|
25
|
+
// GPT-5.3 Series
|
|
26
|
+
GPT_5_3_CODEX: "gpt-5.3-codex",
|
|
27
|
+
GPT_5_3_CHAT: "gpt-5.3-chat-latest",
|
|
25
28
|
// GPT-5.2 Series (Current Flagship)
|
|
26
29
|
GPT_5_2: "gpt-5.2",
|
|
27
30
|
GPT_5_2_PRO: "gpt-5.2-pro",
|
|
31
|
+
GPT_5_2_CODEX: "gpt-5.2-codex",
|
|
32
|
+
GPT_5_2_CHAT: "gpt-5.2-chat-latest",
|
|
33
|
+
// GPT-5.1 Series
|
|
34
|
+
GPT_5_1: "gpt-5.1",
|
|
35
|
+
GPT_5_1_CODEX: "gpt-5.1-codex",
|
|
36
|
+
GPT_5_1_CODEX_MAX: "gpt-5.1-codex-max",
|
|
37
|
+
GPT_5_1_CODEX_MINI: "gpt-5.1-codex-mini",
|
|
38
|
+
GPT_5_1_CHAT: "gpt-5.1-chat-latest",
|
|
28
39
|
// GPT-5 Series
|
|
29
40
|
GPT_5: "gpt-5",
|
|
30
41
|
GPT_5_MINI: "gpt-5-mini",
|
|
31
42
|
GPT_5_NANO: "gpt-5-nano",
|
|
43
|
+
GPT_5_CHAT: "gpt-5-chat-latest",
|
|
32
44
|
// GPT-4.1 Series
|
|
33
45
|
GPT_4_1: "gpt-4.1",
|
|
34
46
|
GPT_4_1_MINI: "gpt-4.1-mini",
|
|
35
47
|
GPT_4_1_NANO: "gpt-4.1-nano",
|
|
36
|
-
// GPT-4o Series (Legacy
|
|
48
|
+
// GPT-4o Series (Legacy)
|
|
37
49
|
GPT_4O: "gpt-4o",
|
|
38
50
|
GPT_4O_MINI: "gpt-4o-mini",
|
|
39
51
|
// Reasoning Models (o-series)
|
|
@@ -41,18 +53,26 @@ var LLM_MODELS = {
|
|
|
41
53
|
O1: "o1"
|
|
42
54
|
},
|
|
43
55
|
[Vendor.Anthropic]: {
|
|
44
|
-
// Claude 4.
|
|
56
|
+
// Claude 4.6 Series (Current)
|
|
57
|
+
CLAUDE_OPUS_4_6: "claude-opus-4-6",
|
|
58
|
+
CLAUDE_SONNET_4_6: "claude-sonnet-4-6",
|
|
59
|
+
// Claude 4.5 Series
|
|
45
60
|
CLAUDE_OPUS_4_5: "claude-opus-4-5-20251101",
|
|
46
61
|
CLAUDE_SONNET_4_5: "claude-sonnet-4-5-20250929",
|
|
47
62
|
CLAUDE_HAIKU_4_5: "claude-haiku-4-5-20251001",
|
|
48
63
|
// Claude 4.x Legacy
|
|
49
64
|
CLAUDE_OPUS_4_1: "claude-opus-4-1-20250805",
|
|
65
|
+
CLAUDE_OPUS_4: "claude-opus-4-20250514",
|
|
50
66
|
CLAUDE_SONNET_4: "claude-sonnet-4-20250514",
|
|
51
67
|
CLAUDE_SONNET_3_7: "claude-3-7-sonnet-20250219",
|
|
52
|
-
// Claude 3.x Legacy
|
|
68
|
+
// Claude 3.x Legacy (Deprecated)
|
|
53
69
|
CLAUDE_HAIKU_3: "claude-3-haiku-20240307"
|
|
54
70
|
},
|
|
55
71
|
[Vendor.Google]: {
|
|
72
|
+
// Gemini 3.1 Series (Preview)
|
|
73
|
+
GEMINI_3_1_PRO_PREVIEW: "gemini-3.1-pro-preview",
|
|
74
|
+
GEMINI_3_1_FLASH_LITE_PREVIEW: "gemini-3.1-flash-lite-preview",
|
|
75
|
+
GEMINI_3_1_FLASH_IMAGE_PREVIEW: "gemini-3.1-flash-image-preview",
|
|
56
76
|
// Gemini 3 Series (Preview)
|
|
57
77
|
GEMINI_3_FLASH_PREVIEW: "gemini-3-flash-preview",
|
|
58
78
|
GEMINI_3_PRO_PREVIEW: "gemini-3-pro-preview",
|
|
@@ -84,12 +104,88 @@ var MODEL_REGISTRY = {
|
|
|
84
104
|
// ============================================================================
|
|
85
105
|
// OpenAI Models (Verified from platform.openai.com)
|
|
86
106
|
// ============================================================================
|
|
107
|
+
// GPT-5.3 Series
|
|
108
|
+
"gpt-5.3-codex": {
|
|
109
|
+
name: "gpt-5.3-codex",
|
|
110
|
+
provider: Vendor.OpenAI,
|
|
111
|
+
description: "Latest codex model for coding and agentic tasks. Reasoning.effort: low, medium, high, xhigh",
|
|
112
|
+
isActive: true,
|
|
113
|
+
releaseDate: "2026-02-01",
|
|
114
|
+
knowledgeCutoff: "2025-08-31",
|
|
115
|
+
features: {
|
|
116
|
+
reasoning: true,
|
|
117
|
+
streaming: true,
|
|
118
|
+
structuredOutput: true,
|
|
119
|
+
functionCalling: true,
|
|
120
|
+
fineTuning: false,
|
|
121
|
+
predictedOutputs: false,
|
|
122
|
+
realtime: false,
|
|
123
|
+
vision: true,
|
|
124
|
+
audio: false,
|
|
125
|
+
video: false,
|
|
126
|
+
batchAPI: true,
|
|
127
|
+
promptCaching: true,
|
|
128
|
+
parameters: {
|
|
129
|
+
temperature: false,
|
|
130
|
+
topP: false,
|
|
131
|
+
frequencyPenalty: false,
|
|
132
|
+
presencePenalty: false
|
|
133
|
+
},
|
|
134
|
+
input: {
|
|
135
|
+
tokens: 4e5,
|
|
136
|
+
text: true,
|
|
137
|
+
image: true,
|
|
138
|
+
cpm: 1.75,
|
|
139
|
+
cpmCached: 0.175
|
|
140
|
+
},
|
|
141
|
+
output: {
|
|
142
|
+
tokens: 128e3,
|
|
143
|
+
text: true,
|
|
144
|
+
cpm: 14
|
|
145
|
+
}
|
|
146
|
+
}
|
|
147
|
+
},
|
|
148
|
+
"gpt-5.3-chat-latest": {
|
|
149
|
+
name: "gpt-5.3-chat-latest",
|
|
150
|
+
provider: Vendor.OpenAI,
|
|
151
|
+
description: "Latest GPT-5.3 chat model for general-purpose use",
|
|
152
|
+
isActive: true,
|
|
153
|
+
releaseDate: "2026-02-01",
|
|
154
|
+
knowledgeCutoff: "2025-08-31",
|
|
155
|
+
features: {
|
|
156
|
+
reasoning: false,
|
|
157
|
+
streaming: true,
|
|
158
|
+
structuredOutput: true,
|
|
159
|
+
functionCalling: true,
|
|
160
|
+
fineTuning: false,
|
|
161
|
+
predictedOutputs: false,
|
|
162
|
+
realtime: false,
|
|
163
|
+
vision: true,
|
|
164
|
+
audio: false,
|
|
165
|
+
video: false,
|
|
166
|
+
batchAPI: true,
|
|
167
|
+
promptCaching: true,
|
|
168
|
+
input: {
|
|
169
|
+
tokens: 128e3,
|
|
170
|
+
text: true,
|
|
171
|
+
image: true,
|
|
172
|
+
cpm: 1.75,
|
|
173
|
+
cpmCached: 0.175
|
|
174
|
+
},
|
|
175
|
+
output: {
|
|
176
|
+
tokens: 16e3,
|
|
177
|
+
text: true,
|
|
178
|
+
cpm: 14
|
|
179
|
+
}
|
|
180
|
+
}
|
|
181
|
+
},
|
|
87
182
|
// GPT-5.2 Series (Current Flagship)
|
|
88
183
|
"gpt-5.2": {
|
|
89
184
|
name: "gpt-5.2",
|
|
90
185
|
provider: Vendor.OpenAI,
|
|
91
186
|
description: "Flagship model for coding and agentic tasks. Reasoning.effort: none, low, medium, high, xhigh",
|
|
92
187
|
isActive: true,
|
|
188
|
+
preferred: true,
|
|
93
189
|
releaseDate: "2025-12-01",
|
|
94
190
|
knowledgeCutoff: "2025-08-31",
|
|
95
191
|
features: {
|
|
@@ -115,7 +211,8 @@ var MODEL_REGISTRY = {
|
|
|
115
211
|
tokens: 4e5,
|
|
116
212
|
text: true,
|
|
117
213
|
image: true,
|
|
118
|
-
cpm: 1.75
|
|
214
|
+
cpm: 1.75,
|
|
215
|
+
cpmCached: 0.175
|
|
119
216
|
},
|
|
120
217
|
output: {
|
|
121
218
|
tokens: 128e3,
|
|
@@ -134,7 +231,7 @@ var MODEL_REGISTRY = {
|
|
|
134
231
|
features: {
|
|
135
232
|
reasoning: true,
|
|
136
233
|
streaming: true,
|
|
137
|
-
structuredOutput:
|
|
234
|
+
structuredOutput: false,
|
|
138
235
|
functionCalling: true,
|
|
139
236
|
fineTuning: false,
|
|
140
237
|
predictedOutputs: false,
|
|
@@ -163,6 +260,276 @@ var MODEL_REGISTRY = {
|
|
|
163
260
|
}
|
|
164
261
|
}
|
|
165
262
|
},
|
|
263
|
+
"gpt-5.2-codex": {
|
|
264
|
+
name: "gpt-5.2-codex",
|
|
265
|
+
provider: Vendor.OpenAI,
|
|
266
|
+
description: "GPT-5.2 codex for coding and agentic tasks. Reasoning.effort: low, medium, high, xhigh",
|
|
267
|
+
isActive: true,
|
|
268
|
+
preferred: true,
|
|
269
|
+
releaseDate: "2025-12-01",
|
|
270
|
+
knowledgeCutoff: "2025-08-31",
|
|
271
|
+
features: {
|
|
272
|
+
reasoning: true,
|
|
273
|
+
streaming: true,
|
|
274
|
+
structuredOutput: true,
|
|
275
|
+
functionCalling: true,
|
|
276
|
+
fineTuning: false,
|
|
277
|
+
predictedOutputs: false,
|
|
278
|
+
realtime: false,
|
|
279
|
+
vision: true,
|
|
280
|
+
audio: false,
|
|
281
|
+
video: false,
|
|
282
|
+
batchAPI: true,
|
|
283
|
+
promptCaching: true,
|
|
284
|
+
parameters: {
|
|
285
|
+
temperature: false,
|
|
286
|
+
topP: false,
|
|
287
|
+
frequencyPenalty: false,
|
|
288
|
+
presencePenalty: false
|
|
289
|
+
},
|
|
290
|
+
input: {
|
|
291
|
+
tokens: 4e5,
|
|
292
|
+
text: true,
|
|
293
|
+
image: true,
|
|
294
|
+
cpm: 1.75,
|
|
295
|
+
cpmCached: 0.175
|
|
296
|
+
},
|
|
297
|
+
output: {
|
|
298
|
+
tokens: 128e3,
|
|
299
|
+
text: true,
|
|
300
|
+
cpm: 14
|
|
301
|
+
}
|
|
302
|
+
}
|
|
303
|
+
},
|
|
304
|
+
"gpt-5.2-chat-latest": {
|
|
305
|
+
name: "gpt-5.2-chat-latest",
|
|
306
|
+
provider: Vendor.OpenAI,
|
|
307
|
+
description: "GPT-5.2 chat model for general-purpose use",
|
|
308
|
+
isActive: true,
|
|
309
|
+
releaseDate: "2025-12-01",
|
|
310
|
+
knowledgeCutoff: "2025-08-31",
|
|
311
|
+
features: {
|
|
312
|
+
reasoning: false,
|
|
313
|
+
streaming: true,
|
|
314
|
+
structuredOutput: true,
|
|
315
|
+
functionCalling: true,
|
|
316
|
+
fineTuning: false,
|
|
317
|
+
predictedOutputs: false,
|
|
318
|
+
realtime: false,
|
|
319
|
+
vision: true,
|
|
320
|
+
audio: false,
|
|
321
|
+
video: false,
|
|
322
|
+
batchAPI: true,
|
|
323
|
+
promptCaching: true,
|
|
324
|
+
input: {
|
|
325
|
+
tokens: 128e3,
|
|
326
|
+
text: true,
|
|
327
|
+
image: true,
|
|
328
|
+
cpm: 1.75,
|
|
329
|
+
cpmCached: 0.175
|
|
330
|
+
},
|
|
331
|
+
output: {
|
|
332
|
+
tokens: 16e3,
|
|
333
|
+
text: true,
|
|
334
|
+
cpm: 14
|
|
335
|
+
}
|
|
336
|
+
}
|
|
337
|
+
},
|
|
338
|
+
// GPT-5.1 Series
|
|
339
|
+
"gpt-5.1": {
|
|
340
|
+
name: "gpt-5.1",
|
|
341
|
+
provider: Vendor.OpenAI,
|
|
342
|
+
description: "Intelligent reasoning model for coding and agentic tasks. Reasoning.effort: none, low, medium, high",
|
|
343
|
+
isActive: true,
|
|
344
|
+
releaseDate: "2025-10-01",
|
|
345
|
+
knowledgeCutoff: "2024-09-30",
|
|
346
|
+
features: {
|
|
347
|
+
reasoning: true,
|
|
348
|
+
streaming: true,
|
|
349
|
+
structuredOutput: true,
|
|
350
|
+
functionCalling: true,
|
|
351
|
+
fineTuning: false,
|
|
352
|
+
predictedOutputs: false,
|
|
353
|
+
realtime: false,
|
|
354
|
+
vision: true,
|
|
355
|
+
audio: false,
|
|
356
|
+
video: false,
|
|
357
|
+
batchAPI: true,
|
|
358
|
+
promptCaching: true,
|
|
359
|
+
parameters: {
|
|
360
|
+
temperature: false,
|
|
361
|
+
topP: false,
|
|
362
|
+
frequencyPenalty: false,
|
|
363
|
+
presencePenalty: false
|
|
364
|
+
},
|
|
365
|
+
input: {
|
|
366
|
+
tokens: 4e5,
|
|
367
|
+
text: true,
|
|
368
|
+
image: true,
|
|
369
|
+
cpm: 1.25,
|
|
370
|
+
cpmCached: 0.125
|
|
371
|
+
},
|
|
372
|
+
output: {
|
|
373
|
+
tokens: 128e3,
|
|
374
|
+
text: true,
|
|
375
|
+
cpm: 10
|
|
376
|
+
}
|
|
377
|
+
}
|
|
378
|
+
},
|
|
379
|
+
"gpt-5.1-codex": {
|
|
380
|
+
name: "gpt-5.1-codex",
|
|
381
|
+
provider: Vendor.OpenAI,
|
|
382
|
+
description: "GPT-5.1 codex for coding and agentic tasks with reasoning",
|
|
383
|
+
isActive: true,
|
|
384
|
+
releaseDate: "2025-10-01",
|
|
385
|
+
knowledgeCutoff: "2024-09-30",
|
|
386
|
+
features: {
|
|
387
|
+
reasoning: true,
|
|
388
|
+
streaming: true,
|
|
389
|
+
structuredOutput: true,
|
|
390
|
+
functionCalling: true,
|
|
391
|
+
fineTuning: false,
|
|
392
|
+
predictedOutputs: false,
|
|
393
|
+
realtime: false,
|
|
394
|
+
vision: true,
|
|
395
|
+
audio: false,
|
|
396
|
+
video: false,
|
|
397
|
+
batchAPI: true,
|
|
398
|
+
promptCaching: true,
|
|
399
|
+
parameters: {
|
|
400
|
+
temperature: false,
|
|
401
|
+
topP: false,
|
|
402
|
+
frequencyPenalty: false,
|
|
403
|
+
presencePenalty: false
|
|
404
|
+
},
|
|
405
|
+
input: {
|
|
406
|
+
tokens: 4e5,
|
|
407
|
+
text: true,
|
|
408
|
+
image: true,
|
|
409
|
+
cpm: 1.25,
|
|
410
|
+
cpmCached: 0.125
|
|
411
|
+
},
|
|
412
|
+
output: {
|
|
413
|
+
tokens: 128e3,
|
|
414
|
+
text: true,
|
|
415
|
+
cpm: 10
|
|
416
|
+
}
|
|
417
|
+
}
|
|
418
|
+
},
|
|
419
|
+
"gpt-5.1-codex-max": {
|
|
420
|
+
name: "gpt-5.1-codex-max",
|
|
421
|
+
provider: Vendor.OpenAI,
|
|
422
|
+
description: "GPT-5.1 codex max for maximum reasoning depth on coding tasks",
|
|
423
|
+
isActive: true,
|
|
424
|
+
releaseDate: "2025-10-01",
|
|
425
|
+
knowledgeCutoff: "2024-09-30",
|
|
426
|
+
features: {
|
|
427
|
+
reasoning: true,
|
|
428
|
+
streaming: true,
|
|
429
|
+
structuredOutput: true,
|
|
430
|
+
functionCalling: true,
|
|
431
|
+
fineTuning: false,
|
|
432
|
+
predictedOutputs: false,
|
|
433
|
+
realtime: false,
|
|
434
|
+
vision: true,
|
|
435
|
+
audio: false,
|
|
436
|
+
video: false,
|
|
437
|
+
batchAPI: true,
|
|
438
|
+
promptCaching: true,
|
|
439
|
+
parameters: {
|
|
440
|
+
temperature: false,
|
|
441
|
+
topP: false,
|
|
442
|
+
frequencyPenalty: false,
|
|
443
|
+
presencePenalty: false
|
|
444
|
+
},
|
|
445
|
+
input: {
|
|
446
|
+
tokens: 4e5,
|
|
447
|
+
text: true,
|
|
448
|
+
image: true,
|
|
449
|
+
cpm: 1.25,
|
|
450
|
+
cpmCached: 0.125
|
|
451
|
+
},
|
|
452
|
+
output: {
|
|
453
|
+
tokens: 128e3,
|
|
454
|
+
text: true,
|
|
455
|
+
cpm: 10
|
|
456
|
+
}
|
|
457
|
+
}
|
|
458
|
+
},
|
|
459
|
+
"gpt-5.1-codex-mini": {
|
|
460
|
+
name: "gpt-5.1-codex-mini",
|
|
461
|
+
provider: Vendor.OpenAI,
|
|
462
|
+
description: "GPT-5.1 codex mini for cost-efficient coding tasks",
|
|
463
|
+
isActive: true,
|
|
464
|
+
releaseDate: "2025-10-01",
|
|
465
|
+
knowledgeCutoff: "2024-09-30",
|
|
466
|
+
features: {
|
|
467
|
+
reasoning: true,
|
|
468
|
+
streaming: true,
|
|
469
|
+
structuredOutput: true,
|
|
470
|
+
functionCalling: true,
|
|
471
|
+
fineTuning: false,
|
|
472
|
+
predictedOutputs: false,
|
|
473
|
+
realtime: false,
|
|
474
|
+
vision: true,
|
|
475
|
+
audio: false,
|
|
476
|
+
video: false,
|
|
477
|
+
batchAPI: true,
|
|
478
|
+
promptCaching: true,
|
|
479
|
+
parameters: {
|
|
480
|
+
temperature: false,
|
|
481
|
+
topP: false,
|
|
482
|
+
frequencyPenalty: false,
|
|
483
|
+
presencePenalty: false
|
|
484
|
+
},
|
|
485
|
+
input: {
|
|
486
|
+
tokens: 4e5,
|
|
487
|
+
text: true,
|
|
488
|
+
image: true,
|
|
489
|
+
cpm: 0.25,
|
|
490
|
+
cpmCached: 0.025
|
|
491
|
+
},
|
|
492
|
+
output: {
|
|
493
|
+
tokens: 128e3,
|
|
494
|
+
text: true,
|
|
495
|
+
cpm: 2
|
|
496
|
+
}
|
|
497
|
+
}
|
|
498
|
+
},
|
|
499
|
+
"gpt-5.1-chat-latest": {
|
|
500
|
+
name: "gpt-5.1-chat-latest",
|
|
501
|
+
provider: Vendor.OpenAI,
|
|
502
|
+
description: "GPT-5.1 chat model for general-purpose use",
|
|
503
|
+
isActive: true,
|
|
504
|
+
releaseDate: "2025-10-01",
|
|
505
|
+
knowledgeCutoff: "2024-09-30",
|
|
506
|
+
features: {
|
|
507
|
+
reasoning: false,
|
|
508
|
+
streaming: true,
|
|
509
|
+
structuredOutput: true,
|
|
510
|
+
functionCalling: true,
|
|
511
|
+
fineTuning: false,
|
|
512
|
+
predictedOutputs: false,
|
|
513
|
+
realtime: false,
|
|
514
|
+
vision: true,
|
|
515
|
+
audio: false,
|
|
516
|
+
video: false,
|
|
517
|
+
batchAPI: true,
|
|
518
|
+
promptCaching: true,
|
|
519
|
+
input: {
|
|
520
|
+
tokens: 128e3,
|
|
521
|
+
text: true,
|
|
522
|
+
image: true,
|
|
523
|
+
cpm: 1.25,
|
|
524
|
+
cpmCached: 0.125
|
|
525
|
+
},
|
|
526
|
+
output: {
|
|
527
|
+
tokens: 16e3,
|
|
528
|
+
text: true,
|
|
529
|
+
cpm: 10
|
|
530
|
+
}
|
|
531
|
+
}
|
|
532
|
+
},
|
|
166
533
|
// GPT-5 Series
|
|
167
534
|
"gpt-5": {
|
|
168
535
|
name: "gpt-5",
|
|
@@ -194,7 +561,8 @@ var MODEL_REGISTRY = {
|
|
|
194
561
|
tokens: 4e5,
|
|
195
562
|
text: true,
|
|
196
563
|
image: true,
|
|
197
|
-
cpm: 1.25
|
|
564
|
+
cpm: 1.25,
|
|
565
|
+
cpmCached: 0.125
|
|
198
566
|
},
|
|
199
567
|
output: {
|
|
200
568
|
tokens: 128e3,
|
|
@@ -233,7 +601,8 @@ var MODEL_REGISTRY = {
|
|
|
233
601
|
tokens: 4e5,
|
|
234
602
|
text: true,
|
|
235
603
|
image: true,
|
|
236
|
-
cpm: 0.25
|
|
604
|
+
cpm: 0.25,
|
|
605
|
+
cpmCached: 0.025
|
|
237
606
|
},
|
|
238
607
|
output: {
|
|
239
608
|
tokens: 128e3,
|
|
@@ -272,7 +641,8 @@ var MODEL_REGISTRY = {
|
|
|
272
641
|
tokens: 4e5,
|
|
273
642
|
text: true,
|
|
274
643
|
image: true,
|
|
275
|
-
cpm: 0.05
|
|
644
|
+
cpm: 0.05,
|
|
645
|
+
cpmCached: 5e-3
|
|
276
646
|
},
|
|
277
647
|
output: {
|
|
278
648
|
tokens: 128e3,
|
|
@@ -281,6 +651,40 @@ var MODEL_REGISTRY = {
|
|
|
281
651
|
}
|
|
282
652
|
}
|
|
283
653
|
},
|
|
654
|
+
"gpt-5-chat-latest": {
|
|
655
|
+
name: "gpt-5-chat-latest",
|
|
656
|
+
provider: Vendor.OpenAI,
|
|
657
|
+
description: "GPT-5 chat model for general-purpose use",
|
|
658
|
+
isActive: true,
|
|
659
|
+
releaseDate: "2025-08-01",
|
|
660
|
+
knowledgeCutoff: "2024-09-30",
|
|
661
|
+
features: {
|
|
662
|
+
reasoning: false,
|
|
663
|
+
streaming: true,
|
|
664
|
+
structuredOutput: true,
|
|
665
|
+
functionCalling: true,
|
|
666
|
+
fineTuning: false,
|
|
667
|
+
predictedOutputs: false,
|
|
668
|
+
realtime: false,
|
|
669
|
+
vision: true,
|
|
670
|
+
audio: false,
|
|
671
|
+
video: false,
|
|
672
|
+
batchAPI: true,
|
|
673
|
+
promptCaching: true,
|
|
674
|
+
input: {
|
|
675
|
+
tokens: 128e3,
|
|
676
|
+
text: true,
|
|
677
|
+
image: true,
|
|
678
|
+
cpm: 1.25,
|
|
679
|
+
cpmCached: 0.125
|
|
680
|
+
},
|
|
681
|
+
output: {
|
|
682
|
+
tokens: 16e3,
|
|
683
|
+
text: true,
|
|
684
|
+
cpm: 10
|
|
685
|
+
}
|
|
686
|
+
}
|
|
687
|
+
},
|
|
284
688
|
// GPT-4.1 Series
|
|
285
689
|
"gpt-4.1": {
|
|
286
690
|
name: "gpt-4.1",
|
|
@@ -288,7 +692,7 @@ var MODEL_REGISTRY = {
|
|
|
288
692
|
description: "GPT-4.1 specialized for coding with 1M token context window",
|
|
289
693
|
isActive: true,
|
|
290
694
|
releaseDate: "2025-04-14",
|
|
291
|
-
knowledgeCutoff: "
|
|
695
|
+
knowledgeCutoff: "2024-06-01",
|
|
292
696
|
features: {
|
|
293
697
|
reasoning: false,
|
|
294
698
|
streaming: true,
|
|
@@ -306,7 +710,8 @@ var MODEL_REGISTRY = {
|
|
|
306
710
|
tokens: 1e6,
|
|
307
711
|
text: true,
|
|
308
712
|
image: true,
|
|
309
|
-
cpm: 2
|
|
713
|
+
cpm: 2,
|
|
714
|
+
cpmCached: 0.5
|
|
310
715
|
},
|
|
311
716
|
output: {
|
|
312
717
|
tokens: 32768,
|
|
@@ -321,7 +726,7 @@ var MODEL_REGISTRY = {
|
|
|
321
726
|
description: "Efficient GPT-4.1 model, beats GPT-4o in many benchmarks at 83% lower cost",
|
|
322
727
|
isActive: true,
|
|
323
728
|
releaseDate: "2025-04-14",
|
|
324
|
-
knowledgeCutoff: "
|
|
729
|
+
knowledgeCutoff: "2024-06-01",
|
|
325
730
|
features: {
|
|
326
731
|
reasoning: false,
|
|
327
732
|
streaming: true,
|
|
@@ -339,7 +744,8 @@ var MODEL_REGISTRY = {
|
|
|
339
744
|
tokens: 1e6,
|
|
340
745
|
text: true,
|
|
341
746
|
image: true,
|
|
342
|
-
cpm: 0.4
|
|
747
|
+
cpm: 0.4,
|
|
748
|
+
cpmCached: 0.1
|
|
343
749
|
},
|
|
344
750
|
output: {
|
|
345
751
|
tokens: 16384,
|
|
@@ -354,7 +760,7 @@ var MODEL_REGISTRY = {
|
|
|
354
760
|
description: "Fastest and cheapest model with 1M context. 80.1% MMLU, ideal for classification/autocompletion",
|
|
355
761
|
isActive: true,
|
|
356
762
|
releaseDate: "2025-04-14",
|
|
357
|
-
knowledgeCutoff: "
|
|
763
|
+
knowledgeCutoff: "2024-06-01",
|
|
358
764
|
features: {
|
|
359
765
|
reasoning: false,
|
|
360
766
|
streaming: true,
|
|
@@ -372,7 +778,8 @@ var MODEL_REGISTRY = {
|
|
|
372
778
|
tokens: 1e6,
|
|
373
779
|
text: true,
|
|
374
780
|
image: true,
|
|
375
|
-
cpm: 0.1
|
|
781
|
+
cpm: 0.1,
|
|
782
|
+
cpmCached: 0.025
|
|
376
783
|
},
|
|
377
784
|
output: {
|
|
378
785
|
tokens: 16384,
|
|
@@ -381,14 +788,14 @@ var MODEL_REGISTRY = {
|
|
|
381
788
|
}
|
|
382
789
|
}
|
|
383
790
|
},
|
|
384
|
-
// GPT-4o Series (Legacy
|
|
791
|
+
// GPT-4o Series (Legacy)
|
|
385
792
|
"gpt-4o": {
|
|
386
793
|
name: "gpt-4o",
|
|
387
794
|
provider: Vendor.OpenAI,
|
|
388
|
-
description: "Versatile omni model
|
|
795
|
+
description: "Versatile omni model. Legacy but still available",
|
|
389
796
|
isActive: true,
|
|
390
797
|
releaseDate: "2024-05-13",
|
|
391
|
-
knowledgeCutoff: "
|
|
798
|
+
knowledgeCutoff: "2023-10-01",
|
|
392
799
|
features: {
|
|
393
800
|
reasoning: false,
|
|
394
801
|
streaming: true,
|
|
@@ -398,7 +805,7 @@ var MODEL_REGISTRY = {
|
|
|
398
805
|
predictedOutputs: true,
|
|
399
806
|
realtime: true,
|
|
400
807
|
vision: true,
|
|
401
|
-
audio:
|
|
808
|
+
audio: false,
|
|
402
809
|
video: false,
|
|
403
810
|
batchAPI: true,
|
|
404
811
|
promptCaching: true,
|
|
@@ -406,13 +813,12 @@ var MODEL_REGISTRY = {
|
|
|
406
813
|
tokens: 128e3,
|
|
407
814
|
text: true,
|
|
408
815
|
image: true,
|
|
409
|
-
|
|
410
|
-
|
|
816
|
+
cpm: 2.5,
|
|
817
|
+
cpmCached: 1.25
|
|
411
818
|
},
|
|
412
819
|
output: {
|
|
413
820
|
tokens: 16384,
|
|
414
821
|
text: true,
|
|
415
|
-
audio: true,
|
|
416
822
|
cpm: 10
|
|
417
823
|
}
|
|
418
824
|
}
|
|
@@ -420,10 +826,10 @@ var MODEL_REGISTRY = {
|
|
|
420
826
|
"gpt-4o-mini": {
|
|
421
827
|
name: "gpt-4o-mini",
|
|
422
828
|
provider: Vendor.OpenAI,
|
|
423
|
-
description: "Fast, affordable omni model
|
|
829
|
+
description: "Fast, affordable omni model",
|
|
424
830
|
isActive: true,
|
|
425
831
|
releaseDate: "2024-07-18",
|
|
426
|
-
knowledgeCutoff: "
|
|
832
|
+
knowledgeCutoff: "2023-10-01",
|
|
427
833
|
features: {
|
|
428
834
|
reasoning: false,
|
|
429
835
|
streaming: true,
|
|
@@ -433,7 +839,7 @@ var MODEL_REGISTRY = {
|
|
|
433
839
|
predictedOutputs: false,
|
|
434
840
|
realtime: true,
|
|
435
841
|
vision: true,
|
|
436
|
-
audio:
|
|
842
|
+
audio: false,
|
|
437
843
|
video: false,
|
|
438
844
|
batchAPI: true,
|
|
439
845
|
promptCaching: true,
|
|
@@ -441,13 +847,12 @@ var MODEL_REGISTRY = {
|
|
|
441
847
|
tokens: 128e3,
|
|
442
848
|
text: true,
|
|
443
849
|
image: true,
|
|
444
|
-
|
|
445
|
-
|
|
850
|
+
cpm: 0.15,
|
|
851
|
+
cpmCached: 0.075
|
|
446
852
|
},
|
|
447
853
|
output: {
|
|
448
854
|
tokens: 16384,
|
|
449
855
|
text: true,
|
|
450
|
-
audio: true,
|
|
451
856
|
cpm: 0.6
|
|
452
857
|
}
|
|
453
858
|
}
|
|
@@ -459,7 +864,46 @@ var MODEL_REGISTRY = {
|
|
|
459
864
|
description: "Fast reasoning model tailored for coding, math, and science",
|
|
460
865
|
isActive: true,
|
|
461
866
|
releaseDate: "2025-01-31",
|
|
462
|
-
knowledgeCutoff: "
|
|
867
|
+
knowledgeCutoff: "2023-10-01",
|
|
868
|
+
features: {
|
|
869
|
+
reasoning: true,
|
|
870
|
+
streaming: true,
|
|
871
|
+
structuredOutput: true,
|
|
872
|
+
functionCalling: true,
|
|
873
|
+
fineTuning: false,
|
|
874
|
+
predictedOutputs: false,
|
|
875
|
+
realtime: false,
|
|
876
|
+
vision: false,
|
|
877
|
+
audio: false,
|
|
878
|
+
video: false,
|
|
879
|
+
batchAPI: true,
|
|
880
|
+
promptCaching: true,
|
|
881
|
+
parameters: {
|
|
882
|
+
temperature: false,
|
|
883
|
+
topP: false,
|
|
884
|
+
frequencyPenalty: false,
|
|
885
|
+
presencePenalty: false
|
|
886
|
+
},
|
|
887
|
+
input: {
|
|
888
|
+
tokens: 2e5,
|
|
889
|
+
text: true,
|
|
890
|
+
cpm: 1.1,
|
|
891
|
+
cpmCached: 0.55
|
|
892
|
+
},
|
|
893
|
+
output: {
|
|
894
|
+
tokens: 1e5,
|
|
895
|
+
text: true,
|
|
896
|
+
cpm: 4.4
|
|
897
|
+
}
|
|
898
|
+
}
|
|
899
|
+
},
|
|
900
|
+
"o1": {
|
|
901
|
+
name: "o1",
|
|
902
|
+
provider: Vendor.OpenAI,
|
|
903
|
+
description: "Advanced reasoning model for complex problems",
|
|
904
|
+
isActive: true,
|
|
905
|
+
releaseDate: "2024-12-17",
|
|
906
|
+
knowledgeCutoff: "2023-10-01",
|
|
463
907
|
features: {
|
|
464
908
|
reasoning: true,
|
|
465
909
|
streaming: true,
|
|
@@ -472,35 +916,78 @@ var MODEL_REGISTRY = {
|
|
|
472
916
|
audio: false,
|
|
473
917
|
video: false,
|
|
474
918
|
batchAPI: true,
|
|
475
|
-
promptCaching:
|
|
476
|
-
parameters: {
|
|
477
|
-
temperature: false,
|
|
478
|
-
topP: false,
|
|
479
|
-
frequencyPenalty: false,
|
|
480
|
-
presencePenalty: false
|
|
481
|
-
},
|
|
919
|
+
promptCaching: true,
|
|
920
|
+
parameters: {
|
|
921
|
+
temperature: false,
|
|
922
|
+
topP: false,
|
|
923
|
+
frequencyPenalty: false,
|
|
924
|
+
presencePenalty: false
|
|
925
|
+
},
|
|
926
|
+
input: {
|
|
927
|
+
tokens: 2e5,
|
|
928
|
+
text: true,
|
|
929
|
+
image: true,
|
|
930
|
+
cpm: 15,
|
|
931
|
+
cpmCached: 7.5
|
|
932
|
+
},
|
|
933
|
+
output: {
|
|
934
|
+
tokens: 1e5,
|
|
935
|
+
text: true,
|
|
936
|
+
cpm: 60
|
|
937
|
+
}
|
|
938
|
+
}
|
|
939
|
+
},
|
|
940
|
+
// ============================================================================
|
|
941
|
+
// Anthropic Models (Verified from platform.claude.com - March 2026)
|
|
942
|
+
// ============================================================================
|
|
943
|
+
// Claude 4.6 Series (Current)
|
|
944
|
+
"claude-opus-4-6": {
|
|
945
|
+
name: "claude-opus-4-6",
|
|
946
|
+
provider: Vendor.Anthropic,
|
|
947
|
+
description: "The most intelligent model for building agents and coding. 128K output, adaptive thinking",
|
|
948
|
+
isActive: true,
|
|
949
|
+
preferred: true,
|
|
950
|
+
releaseDate: "2026-02-01",
|
|
951
|
+
knowledgeCutoff: "2025-05-01",
|
|
952
|
+
features: {
|
|
953
|
+
reasoning: false,
|
|
954
|
+
streaming: true,
|
|
955
|
+
structuredOutput: true,
|
|
956
|
+
functionCalling: true,
|
|
957
|
+
fineTuning: false,
|
|
958
|
+
predictedOutputs: false,
|
|
959
|
+
realtime: false,
|
|
960
|
+
vision: true,
|
|
961
|
+
audio: false,
|
|
962
|
+
video: false,
|
|
963
|
+
extendedThinking: true,
|
|
964
|
+
batchAPI: true,
|
|
965
|
+
promptCaching: true,
|
|
482
966
|
input: {
|
|
483
967
|
tokens: 2e5,
|
|
968
|
+
// 1M with beta header
|
|
484
969
|
text: true,
|
|
485
970
|
image: true,
|
|
486
|
-
cpm:
|
|
971
|
+
cpm: 5,
|
|
972
|
+
cpmCached: 0.5
|
|
487
973
|
},
|
|
488
974
|
output: {
|
|
489
|
-
tokens:
|
|
975
|
+
tokens: 128e3,
|
|
490
976
|
text: true,
|
|
491
|
-
cpm:
|
|
977
|
+
cpm: 25
|
|
492
978
|
}
|
|
493
979
|
}
|
|
494
980
|
},
|
|
495
|
-
"
|
|
496
|
-
name: "
|
|
497
|
-
provider: Vendor.
|
|
498
|
-
description: "
|
|
981
|
+
"claude-sonnet-4-6": {
|
|
982
|
+
name: "claude-sonnet-4-6",
|
|
983
|
+
provider: Vendor.Anthropic,
|
|
984
|
+
description: "Best combination of speed and intelligence. Adaptive thinking, 1M context beta",
|
|
499
985
|
isActive: true,
|
|
500
|
-
|
|
501
|
-
|
|
986
|
+
preferred: true,
|
|
987
|
+
releaseDate: "2026-02-01",
|
|
988
|
+
knowledgeCutoff: "2025-08-01",
|
|
502
989
|
features: {
|
|
503
|
-
reasoning:
|
|
990
|
+
reasoning: false,
|
|
504
991
|
streaming: true,
|
|
505
992
|
structuredOutput: true,
|
|
506
993
|
functionCalling: true,
|
|
@@ -510,35 +997,29 @@ var MODEL_REGISTRY = {
|
|
|
510
997
|
vision: true,
|
|
511
998
|
audio: false,
|
|
512
999
|
video: false,
|
|
1000
|
+
extendedThinking: true,
|
|
513
1001
|
batchAPI: true,
|
|
514
|
-
promptCaching:
|
|
515
|
-
parameters: {
|
|
516
|
-
temperature: false,
|
|
517
|
-
topP: false,
|
|
518
|
-
frequencyPenalty: false,
|
|
519
|
-
presencePenalty: false
|
|
520
|
-
},
|
|
1002
|
+
promptCaching: true,
|
|
521
1003
|
input: {
|
|
522
1004
|
tokens: 2e5,
|
|
1005
|
+
// 1M with beta header
|
|
523
1006
|
text: true,
|
|
524
1007
|
image: true,
|
|
525
|
-
cpm:
|
|
1008
|
+
cpm: 3,
|
|
1009
|
+
cpmCached: 0.3
|
|
526
1010
|
},
|
|
527
1011
|
output: {
|
|
528
|
-
tokens:
|
|
1012
|
+
tokens: 64e3,
|
|
529
1013
|
text: true,
|
|
530
|
-
cpm:
|
|
1014
|
+
cpm: 15
|
|
531
1015
|
}
|
|
532
1016
|
}
|
|
533
1017
|
},
|
|
534
|
-
//
|
|
535
|
-
// Anthropic Models (Verified from platform.claude.com)
|
|
536
|
-
// ============================================================================
|
|
537
|
-
// Claude 4.5 Series (Current)
|
|
1018
|
+
// Claude 4.5 Series
|
|
538
1019
|
"claude-opus-4-5-20251101": {
|
|
539
1020
|
name: "claude-opus-4-5-20251101",
|
|
540
1021
|
provider: Vendor.Anthropic,
|
|
541
|
-
description: "Premium model combining maximum intelligence with practical performance",
|
|
1022
|
+
description: "Legacy Opus 4.5. Premium model combining maximum intelligence with practical performance",
|
|
542
1023
|
isActive: true,
|
|
543
1024
|
releaseDate: "2025-11-01",
|
|
544
1025
|
knowledgeCutoff: "2025-05-01",
|
|
@@ -573,7 +1054,7 @@ var MODEL_REGISTRY = {
|
|
|
573
1054
|
"claude-sonnet-4-5-20250929": {
|
|
574
1055
|
name: "claude-sonnet-4-5-20250929",
|
|
575
1056
|
provider: Vendor.Anthropic,
|
|
576
|
-
description: "Smart model for complex agents and coding
|
|
1057
|
+
description: "Legacy Sonnet 4.5. Smart model for complex agents and coding",
|
|
577
1058
|
isActive: true,
|
|
578
1059
|
releaseDate: "2025-09-29",
|
|
579
1060
|
knowledgeCutoff: "2025-01-01",
|
|
@@ -593,6 +1074,7 @@ var MODEL_REGISTRY = {
|
|
|
593
1074
|
promptCaching: true,
|
|
594
1075
|
input: {
|
|
595
1076
|
tokens: 2e5,
|
|
1077
|
+
// 1M with beta header
|
|
596
1078
|
text: true,
|
|
597
1079
|
image: true,
|
|
598
1080
|
cpm: 3,
|
|
@@ -676,10 +1158,45 @@ var MODEL_REGISTRY = {
|
|
|
676
1158
|
}
|
|
677
1159
|
}
|
|
678
1160
|
},
|
|
1161
|
+
"claude-opus-4-20250514": {
|
|
1162
|
+
name: "claude-opus-4-20250514",
|
|
1163
|
+
provider: Vendor.Anthropic,
|
|
1164
|
+
description: "Legacy Opus 4. Agentic tasks and reasoning",
|
|
1165
|
+
isActive: true,
|
|
1166
|
+
releaseDate: "2025-05-14",
|
|
1167
|
+
knowledgeCutoff: "2025-01-01",
|
|
1168
|
+
features: {
|
|
1169
|
+
reasoning: false,
|
|
1170
|
+
streaming: true,
|
|
1171
|
+
structuredOutput: true,
|
|
1172
|
+
functionCalling: true,
|
|
1173
|
+
fineTuning: false,
|
|
1174
|
+
predictedOutputs: false,
|
|
1175
|
+
realtime: false,
|
|
1176
|
+
vision: true,
|
|
1177
|
+
audio: false,
|
|
1178
|
+
video: false,
|
|
1179
|
+
extendedThinking: true,
|
|
1180
|
+
batchAPI: true,
|
|
1181
|
+
promptCaching: true,
|
|
1182
|
+
input: {
|
|
1183
|
+
tokens: 2e5,
|
|
1184
|
+
text: true,
|
|
1185
|
+
image: true,
|
|
1186
|
+
cpm: 15,
|
|
1187
|
+
cpmCached: 1.5
|
|
1188
|
+
},
|
|
1189
|
+
output: {
|
|
1190
|
+
tokens: 32e3,
|
|
1191
|
+
text: true,
|
|
1192
|
+
cpm: 75
|
|
1193
|
+
}
|
|
1194
|
+
}
|
|
1195
|
+
},
|
|
679
1196
|
"claude-sonnet-4-20250514": {
|
|
680
1197
|
name: "claude-sonnet-4-20250514",
|
|
681
1198
|
provider: Vendor.Anthropic,
|
|
682
|
-
description: "Legacy Sonnet 4.
|
|
1199
|
+
description: "Legacy Sonnet 4. Supports 1M context beta",
|
|
683
1200
|
isActive: true,
|
|
684
1201
|
releaseDate: "2025-05-14",
|
|
685
1202
|
knowledgeCutoff: "2025-01-01",
|
|
@@ -715,7 +1232,7 @@ var MODEL_REGISTRY = {
|
|
|
715
1232
|
"claude-3-7-sonnet-20250219": {
|
|
716
1233
|
name: "claude-3-7-sonnet-20250219",
|
|
717
1234
|
provider: Vendor.Anthropic,
|
|
718
|
-
description: "Claude 3.7 Sonnet with extended thinking
|
|
1235
|
+
description: "Deprecated. Claude 3.7 Sonnet with extended thinking",
|
|
719
1236
|
isActive: true,
|
|
720
1237
|
releaseDate: "2025-02-19",
|
|
721
1238
|
knowledgeCutoff: "2024-10-01",
|
|
@@ -742,17 +1259,16 @@ var MODEL_REGISTRY = {
|
|
|
742
1259
|
},
|
|
743
1260
|
output: {
|
|
744
1261
|
tokens: 64e3,
|
|
745
|
-
// 128K with beta header
|
|
746
1262
|
text: true,
|
|
747
1263
|
cpm: 15
|
|
748
1264
|
}
|
|
749
1265
|
}
|
|
750
1266
|
},
|
|
751
|
-
// Claude 3.x Legacy
|
|
1267
|
+
// Claude 3.x Legacy (Deprecated - retiring April 19, 2026)
|
|
752
1268
|
"claude-3-haiku-20240307": {
|
|
753
1269
|
name: "claude-3-haiku-20240307",
|
|
754
1270
|
provider: Vendor.Anthropic,
|
|
755
|
-
description: "
|
|
1271
|
+
description: "Deprecated. Retiring April 19, 2026. Migrate to Haiku 4.5",
|
|
756
1272
|
isActive: true,
|
|
757
1273
|
releaseDate: "2024-03-07",
|
|
758
1274
|
knowledgeCutoff: "2023-08-01",
|
|
@@ -785,16 +1301,124 @@ var MODEL_REGISTRY = {
|
|
|
785
1301
|
}
|
|
786
1302
|
},
|
|
787
1303
|
// ============================================================================
|
|
788
|
-
// Google Models (Verified from ai.google.dev)
|
|
1304
|
+
// Google Models (Verified from ai.google.dev - March 2026)
|
|
789
1305
|
// ============================================================================
|
|
1306
|
+
// Gemini 3.1 Series (Preview)
|
|
1307
|
+
"gemini-3.1-pro-preview": {
|
|
1308
|
+
name: "gemini-3.1-pro-preview",
|
|
1309
|
+
provider: Vendor.Google,
|
|
1310
|
+
description: "Advanced intelligence with powerful agentic and coding capabilities. Replaces gemini-3-pro-preview",
|
|
1311
|
+
isActive: true,
|
|
1312
|
+
preferred: true,
|
|
1313
|
+
releaseDate: "2026-02-01",
|
|
1314
|
+
knowledgeCutoff: "2025-01-01",
|
|
1315
|
+
features: {
|
|
1316
|
+
reasoning: true,
|
|
1317
|
+
streaming: true,
|
|
1318
|
+
structuredOutput: true,
|
|
1319
|
+
functionCalling: true,
|
|
1320
|
+
fineTuning: false,
|
|
1321
|
+
predictedOutputs: false,
|
|
1322
|
+
realtime: false,
|
|
1323
|
+
vision: true,
|
|
1324
|
+
audio: true,
|
|
1325
|
+
video: true,
|
|
1326
|
+
batchAPI: true,
|
|
1327
|
+
promptCaching: true,
|
|
1328
|
+
input: {
|
|
1329
|
+
tokens: 1048576,
|
|
1330
|
+
text: true,
|
|
1331
|
+
image: true,
|
|
1332
|
+
audio: true,
|
|
1333
|
+
video: true,
|
|
1334
|
+
cpm: 2,
|
|
1335
|
+
cpmCached: 0.2
|
|
1336
|
+
},
|
|
1337
|
+
output: {
|
|
1338
|
+
tokens: 65536,
|
|
1339
|
+
text: true,
|
|
1340
|
+
cpm: 12
|
|
1341
|
+
}
|
|
1342
|
+
}
|
|
1343
|
+
},
|
|
1344
|
+
"gemini-3.1-flash-lite-preview": {
|
|
1345
|
+
name: "gemini-3.1-flash-lite-preview",
|
|
1346
|
+
provider: Vendor.Google,
|
|
1347
|
+
description: "High performance, budget-friendly for high-volume agentic tasks and data extraction",
|
|
1348
|
+
isActive: true,
|
|
1349
|
+
releaseDate: "2026-03-01",
|
|
1350
|
+
knowledgeCutoff: "2025-01-01",
|
|
1351
|
+
features: {
|
|
1352
|
+
reasoning: true,
|
|
1353
|
+
streaming: true,
|
|
1354
|
+
structuredOutput: true,
|
|
1355
|
+
functionCalling: true,
|
|
1356
|
+
fineTuning: false,
|
|
1357
|
+
predictedOutputs: false,
|
|
1358
|
+
realtime: false,
|
|
1359
|
+
vision: true,
|
|
1360
|
+
audio: true,
|
|
1361
|
+
video: true,
|
|
1362
|
+
batchAPI: true,
|
|
1363
|
+
promptCaching: true,
|
|
1364
|
+
input: {
|
|
1365
|
+
tokens: 1048576,
|
|
1366
|
+
text: true,
|
|
1367
|
+
image: true,
|
|
1368
|
+
audio: true,
|
|
1369
|
+
video: true,
|
|
1370
|
+
cpm: 0.25
|
|
1371
|
+
},
|
|
1372
|
+
output: {
|
|
1373
|
+
tokens: 65536,
|
|
1374
|
+
text: true,
|
|
1375
|
+
cpm: 1.5
|
|
1376
|
+
}
|
|
1377
|
+
}
|
|
1378
|
+
},
|
|
1379
|
+
"gemini-3.1-flash-image-preview": {
|
|
1380
|
+
name: "gemini-3.1-flash-image-preview",
|
|
1381
|
+
provider: Vendor.Google,
|
|
1382
|
+
description: "High-efficiency image generation with up to 4K output, search grounding support",
|
|
1383
|
+
isActive: true,
|
|
1384
|
+
releaseDate: "2026-02-01",
|
|
1385
|
+
knowledgeCutoff: "2025-01-01",
|
|
1386
|
+
features: {
|
|
1387
|
+
reasoning: true,
|
|
1388
|
+
streaming: true,
|
|
1389
|
+
structuredOutput: false,
|
|
1390
|
+
functionCalling: false,
|
|
1391
|
+
fineTuning: false,
|
|
1392
|
+
predictedOutputs: false,
|
|
1393
|
+
realtime: false,
|
|
1394
|
+
vision: true,
|
|
1395
|
+
audio: false,
|
|
1396
|
+
video: false,
|
|
1397
|
+
batchAPI: true,
|
|
1398
|
+
promptCaching: false,
|
|
1399
|
+
input: {
|
|
1400
|
+
tokens: 131072,
|
|
1401
|
+
text: true,
|
|
1402
|
+
image: true,
|
|
1403
|
+
cpm: 0.25
|
|
1404
|
+
},
|
|
1405
|
+
output: {
|
|
1406
|
+
tokens: 32768,
|
|
1407
|
+
text: true,
|
|
1408
|
+
image: true,
|
|
1409
|
+
cpm: 1.5
|
|
1410
|
+
}
|
|
1411
|
+
}
|
|
1412
|
+
},
|
|
790
1413
|
// Gemini 3 Series (Preview)
|
|
791
1414
|
"gemini-3-flash-preview": {
|
|
792
1415
|
name: "gemini-3-flash-preview",
|
|
793
1416
|
provider: Vendor.Google,
|
|
794
|
-
description: "
|
|
1417
|
+
description: "Most powerful agentic and coding model with frontier-class reasoning",
|
|
795
1418
|
isActive: true,
|
|
796
|
-
|
|
797
|
-
|
|
1419
|
+
preferred: true,
|
|
1420
|
+
releaseDate: "2025-12-01",
|
|
1421
|
+
knowledgeCutoff: "2025-01-01",
|
|
798
1422
|
features: {
|
|
799
1423
|
reasoning: true,
|
|
800
1424
|
streaming: true,
|
|
@@ -809,27 +1433,28 @@ var MODEL_REGISTRY = {
|
|
|
809
1433
|
batchAPI: true,
|
|
810
1434
|
promptCaching: true,
|
|
811
1435
|
input: {
|
|
812
|
-
tokens:
|
|
1436
|
+
tokens: 1048576,
|
|
813
1437
|
text: true,
|
|
814
1438
|
image: true,
|
|
815
1439
|
audio: true,
|
|
816
1440
|
video: true,
|
|
817
|
-
cpm: 0.
|
|
1441
|
+
cpm: 0.5,
|
|
1442
|
+
cpmCached: 0.05
|
|
818
1443
|
},
|
|
819
1444
|
output: {
|
|
820
1445
|
tokens: 65536,
|
|
821
1446
|
text: true,
|
|
822
|
-
cpm:
|
|
1447
|
+
cpm: 3
|
|
823
1448
|
}
|
|
824
1449
|
}
|
|
825
1450
|
},
|
|
826
1451
|
"gemini-3-pro-preview": {
|
|
827
1452
|
name: "gemini-3-pro-preview",
|
|
828
1453
|
provider: Vendor.Google,
|
|
829
|
-
description: "
|
|
1454
|
+
description: "Deprecated. Shutting down March 9, 2026. Migrate to gemini-3.1-pro-preview",
|
|
830
1455
|
isActive: true,
|
|
831
1456
|
releaseDate: "2025-11-18",
|
|
832
|
-
knowledgeCutoff: "2025-
|
|
1457
|
+
knowledgeCutoff: "2025-01-01",
|
|
833
1458
|
features: {
|
|
834
1459
|
reasoning: true,
|
|
835
1460
|
streaming: true,
|
|
@@ -844,7 +1469,7 @@ var MODEL_REGISTRY = {
|
|
|
844
1469
|
batchAPI: true,
|
|
845
1470
|
promptCaching: true,
|
|
846
1471
|
input: {
|
|
847
|
-
tokens:
|
|
1472
|
+
tokens: 1048576,
|
|
848
1473
|
text: true,
|
|
849
1474
|
image: true,
|
|
850
1475
|
audio: true,
|
|
@@ -861,14 +1486,14 @@ var MODEL_REGISTRY = {
|
|
|
861
1486
|
"gemini-3-pro-image-preview": {
|
|
862
1487
|
name: "gemini-3-pro-image-preview",
|
|
863
1488
|
provider: Vendor.Google,
|
|
864
|
-
description: "
|
|
1489
|
+
description: "Professional-grade image generation and editing with reasoning",
|
|
865
1490
|
isActive: true,
|
|
866
1491
|
releaseDate: "2025-11-18",
|
|
867
|
-
knowledgeCutoff: "2025-
|
|
1492
|
+
knowledgeCutoff: "2025-01-01",
|
|
868
1493
|
features: {
|
|
869
1494
|
reasoning: true,
|
|
870
1495
|
streaming: true,
|
|
871
|
-
structuredOutput:
|
|
1496
|
+
structuredOutput: true,
|
|
872
1497
|
functionCalling: false,
|
|
873
1498
|
fineTuning: false,
|
|
874
1499
|
predictedOutputs: false,
|
|
@@ -877,15 +1502,15 @@ var MODEL_REGISTRY = {
|
|
|
877
1502
|
audio: false,
|
|
878
1503
|
video: false,
|
|
879
1504
|
batchAPI: true,
|
|
880
|
-
promptCaching:
|
|
1505
|
+
promptCaching: false,
|
|
881
1506
|
input: {
|
|
882
|
-
tokens:
|
|
1507
|
+
tokens: 65536,
|
|
883
1508
|
text: true,
|
|
884
1509
|
image: true,
|
|
885
1510
|
cpm: 1.25
|
|
886
1511
|
},
|
|
887
1512
|
output: {
|
|
888
|
-
tokens:
|
|
1513
|
+
tokens: 32768,
|
|
889
1514
|
text: true,
|
|
890
1515
|
image: true,
|
|
891
1516
|
cpm: 10
|
|
@@ -896,7 +1521,7 @@ var MODEL_REGISTRY = {
|
|
|
896
1521
|
"gemini-2.5-pro": {
|
|
897
1522
|
name: "gemini-2.5-pro",
|
|
898
1523
|
provider: Vendor.Google,
|
|
899
|
-
description: "
|
|
1524
|
+
description: "Most advanced model for complex tasks with deep reasoning and coding",
|
|
900
1525
|
isActive: true,
|
|
901
1526
|
releaseDate: "2025-03-01",
|
|
902
1527
|
knowledgeCutoff: "2025-01-01",
|
|
@@ -914,12 +1539,13 @@ var MODEL_REGISTRY = {
|
|
|
914
1539
|
batchAPI: true,
|
|
915
1540
|
promptCaching: true,
|
|
916
1541
|
input: {
|
|
917
|
-
tokens:
|
|
1542
|
+
tokens: 1048576,
|
|
918
1543
|
text: true,
|
|
919
1544
|
image: true,
|
|
920
1545
|
audio: true,
|
|
921
1546
|
video: true,
|
|
922
|
-
cpm: 1.25
|
|
1547
|
+
cpm: 1.25,
|
|
1548
|
+
cpmCached: 0.125
|
|
923
1549
|
},
|
|
924
1550
|
output: {
|
|
925
1551
|
tokens: 65536,
|
|
@@ -931,7 +1557,7 @@ var MODEL_REGISTRY = {
|
|
|
931
1557
|
"gemini-2.5-flash": {
|
|
932
1558
|
name: "gemini-2.5-flash",
|
|
933
1559
|
provider: Vendor.Google,
|
|
934
|
-
description: "
|
|
1560
|
+
description: "Best price-performance for low-latency, high-volume tasks with reasoning",
|
|
935
1561
|
isActive: true,
|
|
936
1562
|
releaseDate: "2025-06-17",
|
|
937
1563
|
knowledgeCutoff: "2025-01-01",
|
|
@@ -949,24 +1575,25 @@ var MODEL_REGISTRY = {
|
|
|
949
1575
|
batchAPI: true,
|
|
950
1576
|
promptCaching: true,
|
|
951
1577
|
input: {
|
|
952
|
-
tokens:
|
|
1578
|
+
tokens: 1048576,
|
|
953
1579
|
text: true,
|
|
954
1580
|
image: true,
|
|
955
1581
|
audio: true,
|
|
956
1582
|
video: true,
|
|
957
|
-
cpm: 0.
|
|
1583
|
+
cpm: 0.3,
|
|
1584
|
+
cpmCached: 0.03
|
|
958
1585
|
},
|
|
959
1586
|
output: {
|
|
960
1587
|
tokens: 65536,
|
|
961
1588
|
text: true,
|
|
962
|
-
cpm:
|
|
1589
|
+
cpm: 2.5
|
|
963
1590
|
}
|
|
964
1591
|
}
|
|
965
1592
|
},
|
|
966
1593
|
"gemini-2.5-flash-lite": {
|
|
967
1594
|
name: "gemini-2.5-flash-lite",
|
|
968
1595
|
provider: Vendor.Google,
|
|
969
|
-
description: "
|
|
1596
|
+
description: "Fastest and most budget-friendly multimodal model in the 2.5 family",
|
|
970
1597
|
isActive: true,
|
|
971
1598
|
releaseDate: "2025-06-17",
|
|
972
1599
|
knowledgeCutoff: "2025-01-01",
|
|
@@ -984,31 +1611,31 @@ var MODEL_REGISTRY = {
|
|
|
984
1611
|
batchAPI: true,
|
|
985
1612
|
promptCaching: true,
|
|
986
1613
|
input: {
|
|
987
|
-
tokens:
|
|
1614
|
+
tokens: 1048576,
|
|
988
1615
|
text: true,
|
|
989
1616
|
image: true,
|
|
990
1617
|
audio: true,
|
|
991
1618
|
video: true,
|
|
992
|
-
cpm: 0.
|
|
1619
|
+
cpm: 0.1
|
|
993
1620
|
},
|
|
994
1621
|
output: {
|
|
995
1622
|
tokens: 65536,
|
|
996
1623
|
text: true,
|
|
997
|
-
cpm: 0.
|
|
1624
|
+
cpm: 0.4
|
|
998
1625
|
}
|
|
999
1626
|
}
|
|
1000
1627
|
},
|
|
1001
1628
|
"gemini-2.5-flash-image": {
|
|
1002
1629
|
name: "gemini-2.5-flash-image",
|
|
1003
1630
|
provider: Vendor.Google,
|
|
1004
|
-
description: "
|
|
1631
|
+
description: "Fast native image generation and editing (Nano Banana)",
|
|
1005
1632
|
isActive: true,
|
|
1006
|
-
releaseDate: "2025-
|
|
1007
|
-
knowledgeCutoff: "2025-
|
|
1633
|
+
releaseDate: "2025-10-01",
|
|
1634
|
+
knowledgeCutoff: "2025-06-01",
|
|
1008
1635
|
features: {
|
|
1009
|
-
reasoning:
|
|
1636
|
+
reasoning: false,
|
|
1010
1637
|
streaming: true,
|
|
1011
|
-
structuredOutput:
|
|
1638
|
+
structuredOutput: true,
|
|
1012
1639
|
functionCalling: false,
|
|
1013
1640
|
fineTuning: false,
|
|
1014
1641
|
predictedOutputs: false,
|
|
@@ -1019,13 +1646,13 @@ var MODEL_REGISTRY = {
|
|
|
1019
1646
|
batchAPI: true,
|
|
1020
1647
|
promptCaching: true,
|
|
1021
1648
|
input: {
|
|
1022
|
-
tokens:
|
|
1649
|
+
tokens: 65536,
|
|
1023
1650
|
text: true,
|
|
1024
1651
|
image: true,
|
|
1025
1652
|
cpm: 0.15
|
|
1026
1653
|
},
|
|
1027
1654
|
output: {
|
|
1028
|
-
tokens:
|
|
1655
|
+
tokens: 32768,
|
|
1029
1656
|
text: true,
|
|
1030
1657
|
image: true,
|
|
1031
1658
|
cpm: 0.6
|
|
@@ -1033,7 +1660,7 @@ var MODEL_REGISTRY = {
|
|
|
1033
1660
|
}
|
|
1034
1661
|
},
|
|
1035
1662
|
// ============================================================================
|
|
1036
|
-
// xAI Grok Models (Verified from docs.x.ai -
|
|
1663
|
+
// xAI Grok Models (Verified from docs.x.ai - March 2026)
|
|
1037
1664
|
// ============================================================================
|
|
1038
1665
|
// Grok 4.1 Series (2M context, fast)
|
|
1039
1666
|
"grok-4-1-fast-reasoning": {
|
|
@@ -1054,13 +1681,14 @@ var MODEL_REGISTRY = {
|
|
|
1054
1681
|
vision: true,
|
|
1055
1682
|
audio: false,
|
|
1056
1683
|
video: false,
|
|
1057
|
-
batchAPI:
|
|
1058
|
-
promptCaching:
|
|
1684
|
+
batchAPI: true,
|
|
1685
|
+
promptCaching: true,
|
|
1059
1686
|
input: {
|
|
1060
1687
|
tokens: 2e6,
|
|
1061
1688
|
text: true,
|
|
1062
1689
|
image: true,
|
|
1063
|
-
cpm: 0.2
|
|
1690
|
+
cpm: 0.2,
|
|
1691
|
+
cpmCached: 0.05
|
|
1064
1692
|
},
|
|
1065
1693
|
output: {
|
|
1066
1694
|
tokens: 65536,
|
|
@@ -1087,13 +1715,14 @@ var MODEL_REGISTRY = {
|
|
|
1087
1715
|
vision: true,
|
|
1088
1716
|
audio: false,
|
|
1089
1717
|
video: false,
|
|
1090
|
-
batchAPI:
|
|
1091
|
-
promptCaching:
|
|
1718
|
+
batchAPI: true,
|
|
1719
|
+
promptCaching: true,
|
|
1092
1720
|
input: {
|
|
1093
1721
|
tokens: 2e6,
|
|
1094
1722
|
text: true,
|
|
1095
1723
|
image: true,
|
|
1096
|
-
cpm: 0.2
|
|
1724
|
+
cpm: 0.2,
|
|
1725
|
+
cpmCached: 0.05
|
|
1097
1726
|
},
|
|
1098
1727
|
output: {
|
|
1099
1728
|
tokens: 65536,
|
|
@@ -1121,12 +1750,13 @@ var MODEL_REGISTRY = {
|
|
|
1121
1750
|
vision: false,
|
|
1122
1751
|
audio: false,
|
|
1123
1752
|
video: false,
|
|
1124
|
-
batchAPI:
|
|
1125
|
-
promptCaching:
|
|
1753
|
+
batchAPI: true,
|
|
1754
|
+
promptCaching: true,
|
|
1126
1755
|
input: {
|
|
1127
1756
|
tokens: 256e3,
|
|
1128
1757
|
text: true,
|
|
1129
|
-
cpm: 0.2
|
|
1758
|
+
cpm: 0.2,
|
|
1759
|
+
cpmCached: 0.02
|
|
1130
1760
|
},
|
|
1131
1761
|
output: {
|
|
1132
1762
|
tokens: 32768,
|
|
@@ -1139,7 +1769,7 @@ var MODEL_REGISTRY = {
|
|
|
1139
1769
|
"grok-4-fast-reasoning": {
|
|
1140
1770
|
name: "grok-4-fast-reasoning",
|
|
1141
1771
|
provider: Vendor.Grok,
|
|
1142
|
-
description: "Fast Grok 4 with reasoning capabilities, 2M context window",
|
|
1772
|
+
description: "Fast Grok 4 with reasoning capabilities, 2M context window, vision support",
|
|
1143
1773
|
isActive: true,
|
|
1144
1774
|
releaseDate: "2025-09-01",
|
|
1145
1775
|
knowledgeCutoff: "2024-11-01",
|
|
@@ -1151,15 +1781,17 @@ var MODEL_REGISTRY = {
|
|
|
1151
1781
|
fineTuning: false,
|
|
1152
1782
|
predictedOutputs: false,
|
|
1153
1783
|
realtime: false,
|
|
1154
|
-
vision:
|
|
1784
|
+
vision: true,
|
|
1155
1785
|
audio: false,
|
|
1156
1786
|
video: false,
|
|
1157
|
-
batchAPI:
|
|
1158
|
-
promptCaching:
|
|
1787
|
+
batchAPI: true,
|
|
1788
|
+
promptCaching: true,
|
|
1159
1789
|
input: {
|
|
1160
1790
|
tokens: 2e6,
|
|
1161
1791
|
text: true,
|
|
1162
|
-
|
|
1792
|
+
image: true,
|
|
1793
|
+
cpm: 0.2,
|
|
1794
|
+
cpmCached: 0.05
|
|
1163
1795
|
},
|
|
1164
1796
|
output: {
|
|
1165
1797
|
tokens: 65536,
|
|
@@ -1186,13 +1818,14 @@ var MODEL_REGISTRY = {
|
|
|
1186
1818
|
vision: true,
|
|
1187
1819
|
audio: false,
|
|
1188
1820
|
video: false,
|
|
1189
|
-
batchAPI:
|
|
1190
|
-
promptCaching:
|
|
1821
|
+
batchAPI: true,
|
|
1822
|
+
promptCaching: true,
|
|
1191
1823
|
input: {
|
|
1192
1824
|
tokens: 2e6,
|
|
1193
1825
|
text: true,
|
|
1194
1826
|
image: true,
|
|
1195
|
-
cpm: 0.2
|
|
1827
|
+
cpm: 0.2,
|
|
1828
|
+
cpmCached: 0.05
|
|
1196
1829
|
},
|
|
1197
1830
|
output: {
|
|
1198
1831
|
tokens: 65536,
|
|
@@ -1204,12 +1837,12 @@ var MODEL_REGISTRY = {
|
|
|
1204
1837
|
"grok-4-0709": {
|
|
1205
1838
|
name: "grok-4-0709",
|
|
1206
1839
|
provider: Vendor.Grok,
|
|
1207
|
-
description: "Grok 4 flagship model (July 2025 release), 256K context, vision support",
|
|
1840
|
+
description: "Grok 4 flagship model (July 2025 release), 256K context, vision support, reasoning",
|
|
1208
1841
|
isActive: true,
|
|
1209
1842
|
releaseDate: "2025-07-09",
|
|
1210
1843
|
knowledgeCutoff: "2024-11-01",
|
|
1211
1844
|
features: {
|
|
1212
|
-
reasoning:
|
|
1845
|
+
reasoning: true,
|
|
1213
1846
|
streaming: true,
|
|
1214
1847
|
structuredOutput: true,
|
|
1215
1848
|
functionCalling: true,
|
|
@@ -1219,13 +1852,14 @@ var MODEL_REGISTRY = {
|
|
|
1219
1852
|
vision: true,
|
|
1220
1853
|
audio: false,
|
|
1221
1854
|
video: false,
|
|
1222
|
-
batchAPI:
|
|
1223
|
-
promptCaching:
|
|
1855
|
+
batchAPI: true,
|
|
1856
|
+
promptCaching: true,
|
|
1224
1857
|
input: {
|
|
1225
1858
|
tokens: 256e3,
|
|
1226
1859
|
text: true,
|
|
1227
1860
|
image: true,
|
|
1228
|
-
cpm: 3
|
|
1861
|
+
cpm: 3,
|
|
1862
|
+
cpmCached: 0.75
|
|
1229
1863
|
},
|
|
1230
1864
|
output: {
|
|
1231
1865
|
tokens: 32768,
|
|
@@ -1238,12 +1872,12 @@ var MODEL_REGISTRY = {
|
|
|
1238
1872
|
"grok-3-mini": {
|
|
1239
1873
|
name: "grok-3-mini",
|
|
1240
1874
|
provider: Vendor.Grok,
|
|
1241
|
-
description: "Lightweight, cost-efficient model
|
|
1875
|
+
description: "Lightweight, cost-efficient model with reasoning, 131K context",
|
|
1242
1876
|
isActive: true,
|
|
1243
1877
|
releaseDate: "2025-06-01",
|
|
1244
1878
|
knowledgeCutoff: "2024-11-01",
|
|
1245
1879
|
features: {
|
|
1246
|
-
reasoning:
|
|
1880
|
+
reasoning: true,
|
|
1247
1881
|
streaming: true,
|
|
1248
1882
|
structuredOutput: true,
|
|
1249
1883
|
functionCalling: true,
|
|
@@ -1253,12 +1887,13 @@ var MODEL_REGISTRY = {
|
|
|
1253
1887
|
vision: false,
|
|
1254
1888
|
audio: false,
|
|
1255
1889
|
video: false,
|
|
1256
|
-
batchAPI:
|
|
1257
|
-
promptCaching:
|
|
1890
|
+
batchAPI: true,
|
|
1891
|
+
promptCaching: true,
|
|
1258
1892
|
input: {
|
|
1259
1893
|
tokens: 131072,
|
|
1260
1894
|
text: true,
|
|
1261
|
-
cpm: 0.3
|
|
1895
|
+
cpm: 0.3,
|
|
1896
|
+
cpmCached: 0.07
|
|
1262
1897
|
},
|
|
1263
1898
|
output: {
|
|
1264
1899
|
tokens: 32768,
|
|
@@ -1285,12 +1920,13 @@ var MODEL_REGISTRY = {
|
|
|
1285
1920
|
vision: false,
|
|
1286
1921
|
audio: false,
|
|
1287
1922
|
video: false,
|
|
1288
|
-
batchAPI:
|
|
1289
|
-
promptCaching:
|
|
1923
|
+
batchAPI: true,
|
|
1924
|
+
promptCaching: true,
|
|
1290
1925
|
input: {
|
|
1291
1926
|
tokens: 131072,
|
|
1292
1927
|
text: true,
|
|
1293
|
-
cpm: 3
|
|
1928
|
+
cpm: 3,
|
|
1929
|
+
cpmCached: 0.75
|
|
1294
1930
|
},
|
|
1295
1931
|
output: {
|
|
1296
1932
|
tokens: 32768,
|
|
@@ -1299,11 +1935,11 @@ var MODEL_REGISTRY = {
|
|
|
1299
1935
|
}
|
|
1300
1936
|
}
|
|
1301
1937
|
},
|
|
1302
|
-
// Grok 2 Series (
|
|
1938
|
+
// Grok 2 Series (Legacy - not in current docs)
|
|
1303
1939
|
"grok-2-vision-1212": {
|
|
1304
1940
|
name: "grok-2-vision-1212",
|
|
1305
1941
|
provider: Vendor.Grok,
|
|
1306
|
-
description: "
|
|
1942
|
+
description: "Legacy vision model for image understanding, 32K context. Not in current xAI docs",
|
|
1307
1943
|
isActive: true,
|
|
1308
1944
|
releaseDate: "2024-12-12",
|
|
1309
1945
|
knowledgeCutoff: "2024-11-01",
|
|
@@ -1576,6 +2212,14 @@ var SERVICE_DEFINITIONS = [
|
|
|
1576
2212
|
baseURL: "https://aws.amazon.com",
|
|
1577
2213
|
docsURL: "https://docs.aws.amazon.com/"
|
|
1578
2214
|
},
|
|
2215
|
+
{
|
|
2216
|
+
id: "cloudflare",
|
|
2217
|
+
name: "Cloudflare",
|
|
2218
|
+
category: "cloud",
|
|
2219
|
+
urlPattern: /api\.cloudflare\.com/i,
|
|
2220
|
+
baseURL: "https://api.cloudflare.com/client/v4",
|
|
2221
|
+
docsURL: "https://developers.cloudflare.com/api/"
|
|
2222
|
+
},
|
|
1579
2223
|
// ============ Storage ============
|
|
1580
2224
|
{
|
|
1581
2225
|
id: "dropbox",
|
|
@@ -1619,6 +2263,14 @@ var SERVICE_DEFINITIONS = [
|
|
|
1619
2263
|
baseURL: "https://api.postmarkapp.com",
|
|
1620
2264
|
docsURL: "https://postmarkapp.com/developer"
|
|
1621
2265
|
},
|
|
2266
|
+
{
|
|
2267
|
+
id: "mailgun",
|
|
2268
|
+
name: "Mailgun",
|
|
2269
|
+
category: "email",
|
|
2270
|
+
urlPattern: /api\.mailgun\.net|api\.eu\.mailgun\.net/i,
|
|
2271
|
+
baseURL: "https://api.mailgun.net/v3",
|
|
2272
|
+
docsURL: "https://documentation.mailgun.com/docs/mailgun/api-reference/"
|
|
2273
|
+
},
|
|
1622
2274
|
// ============ Monitoring & Observability ============
|
|
1623
2275
|
{
|
|
1624
2276
|
id: "datadog",
|