@everworker/oneringai 0.4.4 → 0.4.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +25 -12
- package/dist/{ImageModel-OWbA277F.d.ts → ImageModel-1uP-2vk7.d.ts} +8 -2
- package/dist/{ImageModel-Ds5_6sf7.d.cts → ImageModel-BDI37OED.d.cts} +8 -2
- package/dist/capabilities/agents/index.d.cts +1 -1
- package/dist/capabilities/agents/index.d.ts +1 -1
- package/dist/capabilities/images/index.cjs +149 -7
- package/dist/capabilities/images/index.cjs.map +1 -1
- package/dist/capabilities/images/index.d.cts +1 -1
- package/dist/capabilities/images/index.d.ts +1 -1
- package/dist/capabilities/images/index.js +149 -7
- package/dist/capabilities/images/index.js.map +1 -1
- package/dist/{index-DmYrHH7d.d.cts → index-DJ-qAK15.d.cts} +25 -5
- package/dist/{index-Dyl6pHfq.d.ts → index-oBtp-8Qn.d.ts} +25 -5
- package/dist/index.cjs +1348 -199
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +47 -13
- package/dist/index.d.ts +47 -13
- package/dist/index.js +1348 -199
- package/dist/index.js.map +1 -1
- package/dist/shared/index.cjs +785 -133
- package/dist/shared/index.cjs.map +1 -1
- package/dist/shared/index.d.cts +20 -2
- package/dist/shared/index.d.ts +20 -2
- package/dist/shared/index.js +785 -133
- package/dist/shared/index.js.map +1 -1
- package/package.json +1 -1
package/dist/shared/index.cjs
CHANGED
|
@@ -24,18 +24,30 @@ function isVendor(value) {
|
|
|
24
24
|
// src/domain/entities/Model.ts
|
|
25
25
|
var LLM_MODELS = {
|
|
26
26
|
[Vendor.OpenAI]: {
|
|
27
|
+
// GPT-5.3 Series
|
|
28
|
+
GPT_5_3_CODEX: "gpt-5.3-codex",
|
|
29
|
+
GPT_5_3_CHAT: "gpt-5.3-chat-latest",
|
|
27
30
|
// GPT-5.2 Series (Current Flagship)
|
|
28
31
|
GPT_5_2: "gpt-5.2",
|
|
29
32
|
GPT_5_2_PRO: "gpt-5.2-pro",
|
|
33
|
+
GPT_5_2_CODEX: "gpt-5.2-codex",
|
|
34
|
+
GPT_5_2_CHAT: "gpt-5.2-chat-latest",
|
|
35
|
+
// GPT-5.1 Series
|
|
36
|
+
GPT_5_1: "gpt-5.1",
|
|
37
|
+
GPT_5_1_CODEX: "gpt-5.1-codex",
|
|
38
|
+
GPT_5_1_CODEX_MAX: "gpt-5.1-codex-max",
|
|
39
|
+
GPT_5_1_CODEX_MINI: "gpt-5.1-codex-mini",
|
|
40
|
+
GPT_5_1_CHAT: "gpt-5.1-chat-latest",
|
|
30
41
|
// GPT-5 Series
|
|
31
42
|
GPT_5: "gpt-5",
|
|
32
43
|
GPT_5_MINI: "gpt-5-mini",
|
|
33
44
|
GPT_5_NANO: "gpt-5-nano",
|
|
45
|
+
GPT_5_CHAT: "gpt-5-chat-latest",
|
|
34
46
|
// GPT-4.1 Series
|
|
35
47
|
GPT_4_1: "gpt-4.1",
|
|
36
48
|
GPT_4_1_MINI: "gpt-4.1-mini",
|
|
37
49
|
GPT_4_1_NANO: "gpt-4.1-nano",
|
|
38
|
-
// GPT-4o Series (Legacy
|
|
50
|
+
// GPT-4o Series (Legacy)
|
|
39
51
|
GPT_4O: "gpt-4o",
|
|
40
52
|
GPT_4O_MINI: "gpt-4o-mini",
|
|
41
53
|
// Reasoning Models (o-series)
|
|
@@ -43,18 +55,26 @@ var LLM_MODELS = {
|
|
|
43
55
|
O1: "o1"
|
|
44
56
|
},
|
|
45
57
|
[Vendor.Anthropic]: {
|
|
46
|
-
// Claude 4.
|
|
58
|
+
// Claude 4.6 Series (Current)
|
|
59
|
+
CLAUDE_OPUS_4_6: "claude-opus-4-6",
|
|
60
|
+
CLAUDE_SONNET_4_6: "claude-sonnet-4-6",
|
|
61
|
+
// Claude 4.5 Series
|
|
47
62
|
CLAUDE_OPUS_4_5: "claude-opus-4-5-20251101",
|
|
48
63
|
CLAUDE_SONNET_4_5: "claude-sonnet-4-5-20250929",
|
|
49
64
|
CLAUDE_HAIKU_4_5: "claude-haiku-4-5-20251001",
|
|
50
65
|
// Claude 4.x Legacy
|
|
51
66
|
CLAUDE_OPUS_4_1: "claude-opus-4-1-20250805",
|
|
67
|
+
CLAUDE_OPUS_4: "claude-opus-4-20250514",
|
|
52
68
|
CLAUDE_SONNET_4: "claude-sonnet-4-20250514",
|
|
53
69
|
CLAUDE_SONNET_3_7: "claude-3-7-sonnet-20250219",
|
|
54
|
-
// Claude 3.x Legacy
|
|
70
|
+
// Claude 3.x Legacy (Deprecated)
|
|
55
71
|
CLAUDE_HAIKU_3: "claude-3-haiku-20240307"
|
|
56
72
|
},
|
|
57
73
|
[Vendor.Google]: {
|
|
74
|
+
// Gemini 3.1 Series (Preview)
|
|
75
|
+
GEMINI_3_1_PRO_PREVIEW: "gemini-3.1-pro-preview",
|
|
76
|
+
GEMINI_3_1_FLASH_LITE_PREVIEW: "gemini-3.1-flash-lite-preview",
|
|
77
|
+
GEMINI_3_1_FLASH_IMAGE_PREVIEW: "gemini-3.1-flash-image-preview",
|
|
58
78
|
// Gemini 3 Series (Preview)
|
|
59
79
|
GEMINI_3_FLASH_PREVIEW: "gemini-3-flash-preview",
|
|
60
80
|
GEMINI_3_PRO_PREVIEW: "gemini-3-pro-preview",
|
|
@@ -86,12 +106,88 @@ var MODEL_REGISTRY = {
|
|
|
86
106
|
// ============================================================================
|
|
87
107
|
// OpenAI Models (Verified from platform.openai.com)
|
|
88
108
|
// ============================================================================
|
|
109
|
+
// GPT-5.3 Series
|
|
110
|
+
"gpt-5.3-codex": {
|
|
111
|
+
name: "gpt-5.3-codex",
|
|
112
|
+
provider: Vendor.OpenAI,
|
|
113
|
+
description: "Latest codex model for coding and agentic tasks. Reasoning.effort: low, medium, high, xhigh",
|
|
114
|
+
isActive: true,
|
|
115
|
+
releaseDate: "2026-02-01",
|
|
116
|
+
knowledgeCutoff: "2025-08-31",
|
|
117
|
+
features: {
|
|
118
|
+
reasoning: true,
|
|
119
|
+
streaming: true,
|
|
120
|
+
structuredOutput: true,
|
|
121
|
+
functionCalling: true,
|
|
122
|
+
fineTuning: false,
|
|
123
|
+
predictedOutputs: false,
|
|
124
|
+
realtime: false,
|
|
125
|
+
vision: true,
|
|
126
|
+
audio: false,
|
|
127
|
+
video: false,
|
|
128
|
+
batchAPI: true,
|
|
129
|
+
promptCaching: true,
|
|
130
|
+
parameters: {
|
|
131
|
+
temperature: false,
|
|
132
|
+
topP: false,
|
|
133
|
+
frequencyPenalty: false,
|
|
134
|
+
presencePenalty: false
|
|
135
|
+
},
|
|
136
|
+
input: {
|
|
137
|
+
tokens: 4e5,
|
|
138
|
+
text: true,
|
|
139
|
+
image: true,
|
|
140
|
+
cpm: 1.75,
|
|
141
|
+
cpmCached: 0.175
|
|
142
|
+
},
|
|
143
|
+
output: {
|
|
144
|
+
tokens: 128e3,
|
|
145
|
+
text: true,
|
|
146
|
+
cpm: 14
|
|
147
|
+
}
|
|
148
|
+
}
|
|
149
|
+
},
|
|
150
|
+
"gpt-5.3-chat-latest": {
|
|
151
|
+
name: "gpt-5.3-chat-latest",
|
|
152
|
+
provider: Vendor.OpenAI,
|
|
153
|
+
description: "Latest GPT-5.3 chat model for general-purpose use",
|
|
154
|
+
isActive: true,
|
|
155
|
+
releaseDate: "2026-02-01",
|
|
156
|
+
knowledgeCutoff: "2025-08-31",
|
|
157
|
+
features: {
|
|
158
|
+
reasoning: false,
|
|
159
|
+
streaming: true,
|
|
160
|
+
structuredOutput: true,
|
|
161
|
+
functionCalling: true,
|
|
162
|
+
fineTuning: false,
|
|
163
|
+
predictedOutputs: false,
|
|
164
|
+
realtime: false,
|
|
165
|
+
vision: true,
|
|
166
|
+
audio: false,
|
|
167
|
+
video: false,
|
|
168
|
+
batchAPI: true,
|
|
169
|
+
promptCaching: true,
|
|
170
|
+
input: {
|
|
171
|
+
tokens: 128e3,
|
|
172
|
+
text: true,
|
|
173
|
+
image: true,
|
|
174
|
+
cpm: 1.75,
|
|
175
|
+
cpmCached: 0.175
|
|
176
|
+
},
|
|
177
|
+
output: {
|
|
178
|
+
tokens: 16e3,
|
|
179
|
+
text: true,
|
|
180
|
+
cpm: 14
|
|
181
|
+
}
|
|
182
|
+
}
|
|
183
|
+
},
|
|
89
184
|
// GPT-5.2 Series (Current Flagship)
|
|
90
185
|
"gpt-5.2": {
|
|
91
186
|
name: "gpt-5.2",
|
|
92
187
|
provider: Vendor.OpenAI,
|
|
93
188
|
description: "Flagship model for coding and agentic tasks. Reasoning.effort: none, low, medium, high, xhigh",
|
|
94
189
|
isActive: true,
|
|
190
|
+
preferred: true,
|
|
95
191
|
releaseDate: "2025-12-01",
|
|
96
192
|
knowledgeCutoff: "2025-08-31",
|
|
97
193
|
features: {
|
|
@@ -117,7 +213,8 @@ var MODEL_REGISTRY = {
|
|
|
117
213
|
tokens: 4e5,
|
|
118
214
|
text: true,
|
|
119
215
|
image: true,
|
|
120
|
-
cpm: 1.75
|
|
216
|
+
cpm: 1.75,
|
|
217
|
+
cpmCached: 0.175
|
|
121
218
|
},
|
|
122
219
|
output: {
|
|
123
220
|
tokens: 128e3,
|
|
@@ -136,7 +233,7 @@ var MODEL_REGISTRY = {
|
|
|
136
233
|
features: {
|
|
137
234
|
reasoning: true,
|
|
138
235
|
streaming: true,
|
|
139
|
-
structuredOutput:
|
|
236
|
+
structuredOutput: false,
|
|
140
237
|
functionCalling: true,
|
|
141
238
|
fineTuning: false,
|
|
142
239
|
predictedOutputs: false,
|
|
@@ -165,6 +262,276 @@ var MODEL_REGISTRY = {
|
|
|
165
262
|
}
|
|
166
263
|
}
|
|
167
264
|
},
|
|
265
|
+
"gpt-5.2-codex": {
|
|
266
|
+
name: "gpt-5.2-codex",
|
|
267
|
+
provider: Vendor.OpenAI,
|
|
268
|
+
description: "GPT-5.2 codex for coding and agentic tasks. Reasoning.effort: low, medium, high, xhigh",
|
|
269
|
+
isActive: true,
|
|
270
|
+
preferred: true,
|
|
271
|
+
releaseDate: "2025-12-01",
|
|
272
|
+
knowledgeCutoff: "2025-08-31",
|
|
273
|
+
features: {
|
|
274
|
+
reasoning: true,
|
|
275
|
+
streaming: true,
|
|
276
|
+
structuredOutput: true,
|
|
277
|
+
functionCalling: true,
|
|
278
|
+
fineTuning: false,
|
|
279
|
+
predictedOutputs: false,
|
|
280
|
+
realtime: false,
|
|
281
|
+
vision: true,
|
|
282
|
+
audio: false,
|
|
283
|
+
video: false,
|
|
284
|
+
batchAPI: true,
|
|
285
|
+
promptCaching: true,
|
|
286
|
+
parameters: {
|
|
287
|
+
temperature: false,
|
|
288
|
+
topP: false,
|
|
289
|
+
frequencyPenalty: false,
|
|
290
|
+
presencePenalty: false
|
|
291
|
+
},
|
|
292
|
+
input: {
|
|
293
|
+
tokens: 4e5,
|
|
294
|
+
text: true,
|
|
295
|
+
image: true,
|
|
296
|
+
cpm: 1.75,
|
|
297
|
+
cpmCached: 0.175
|
|
298
|
+
},
|
|
299
|
+
output: {
|
|
300
|
+
tokens: 128e3,
|
|
301
|
+
text: true,
|
|
302
|
+
cpm: 14
|
|
303
|
+
}
|
|
304
|
+
}
|
|
305
|
+
},
|
|
306
|
+
"gpt-5.2-chat-latest": {
|
|
307
|
+
name: "gpt-5.2-chat-latest",
|
|
308
|
+
provider: Vendor.OpenAI,
|
|
309
|
+
description: "GPT-5.2 chat model for general-purpose use",
|
|
310
|
+
isActive: true,
|
|
311
|
+
releaseDate: "2025-12-01",
|
|
312
|
+
knowledgeCutoff: "2025-08-31",
|
|
313
|
+
features: {
|
|
314
|
+
reasoning: false,
|
|
315
|
+
streaming: true,
|
|
316
|
+
structuredOutput: true,
|
|
317
|
+
functionCalling: true,
|
|
318
|
+
fineTuning: false,
|
|
319
|
+
predictedOutputs: false,
|
|
320
|
+
realtime: false,
|
|
321
|
+
vision: true,
|
|
322
|
+
audio: false,
|
|
323
|
+
video: false,
|
|
324
|
+
batchAPI: true,
|
|
325
|
+
promptCaching: true,
|
|
326
|
+
input: {
|
|
327
|
+
tokens: 128e3,
|
|
328
|
+
text: true,
|
|
329
|
+
image: true,
|
|
330
|
+
cpm: 1.75,
|
|
331
|
+
cpmCached: 0.175
|
|
332
|
+
},
|
|
333
|
+
output: {
|
|
334
|
+
tokens: 16e3,
|
|
335
|
+
text: true,
|
|
336
|
+
cpm: 14
|
|
337
|
+
}
|
|
338
|
+
}
|
|
339
|
+
},
|
|
340
|
+
// GPT-5.1 Series
|
|
341
|
+
"gpt-5.1": {
|
|
342
|
+
name: "gpt-5.1",
|
|
343
|
+
provider: Vendor.OpenAI,
|
|
344
|
+
description: "Intelligent reasoning model for coding and agentic tasks. Reasoning.effort: none, low, medium, high",
|
|
345
|
+
isActive: true,
|
|
346
|
+
releaseDate: "2025-10-01",
|
|
347
|
+
knowledgeCutoff: "2024-09-30",
|
|
348
|
+
features: {
|
|
349
|
+
reasoning: true,
|
|
350
|
+
streaming: true,
|
|
351
|
+
structuredOutput: true,
|
|
352
|
+
functionCalling: true,
|
|
353
|
+
fineTuning: false,
|
|
354
|
+
predictedOutputs: false,
|
|
355
|
+
realtime: false,
|
|
356
|
+
vision: true,
|
|
357
|
+
audio: false,
|
|
358
|
+
video: false,
|
|
359
|
+
batchAPI: true,
|
|
360
|
+
promptCaching: true,
|
|
361
|
+
parameters: {
|
|
362
|
+
temperature: false,
|
|
363
|
+
topP: false,
|
|
364
|
+
frequencyPenalty: false,
|
|
365
|
+
presencePenalty: false
|
|
366
|
+
},
|
|
367
|
+
input: {
|
|
368
|
+
tokens: 4e5,
|
|
369
|
+
text: true,
|
|
370
|
+
image: true,
|
|
371
|
+
cpm: 1.25,
|
|
372
|
+
cpmCached: 0.125
|
|
373
|
+
},
|
|
374
|
+
output: {
|
|
375
|
+
tokens: 128e3,
|
|
376
|
+
text: true,
|
|
377
|
+
cpm: 10
|
|
378
|
+
}
|
|
379
|
+
}
|
|
380
|
+
},
|
|
381
|
+
"gpt-5.1-codex": {
|
|
382
|
+
name: "gpt-5.1-codex",
|
|
383
|
+
provider: Vendor.OpenAI,
|
|
384
|
+
description: "GPT-5.1 codex for coding and agentic tasks with reasoning",
|
|
385
|
+
isActive: true,
|
|
386
|
+
releaseDate: "2025-10-01",
|
|
387
|
+
knowledgeCutoff: "2024-09-30",
|
|
388
|
+
features: {
|
|
389
|
+
reasoning: true,
|
|
390
|
+
streaming: true,
|
|
391
|
+
structuredOutput: true,
|
|
392
|
+
functionCalling: true,
|
|
393
|
+
fineTuning: false,
|
|
394
|
+
predictedOutputs: false,
|
|
395
|
+
realtime: false,
|
|
396
|
+
vision: true,
|
|
397
|
+
audio: false,
|
|
398
|
+
video: false,
|
|
399
|
+
batchAPI: true,
|
|
400
|
+
promptCaching: true,
|
|
401
|
+
parameters: {
|
|
402
|
+
temperature: false,
|
|
403
|
+
topP: false,
|
|
404
|
+
frequencyPenalty: false,
|
|
405
|
+
presencePenalty: false
|
|
406
|
+
},
|
|
407
|
+
input: {
|
|
408
|
+
tokens: 4e5,
|
|
409
|
+
text: true,
|
|
410
|
+
image: true,
|
|
411
|
+
cpm: 1.25,
|
|
412
|
+
cpmCached: 0.125
|
|
413
|
+
},
|
|
414
|
+
output: {
|
|
415
|
+
tokens: 128e3,
|
|
416
|
+
text: true,
|
|
417
|
+
cpm: 10
|
|
418
|
+
}
|
|
419
|
+
}
|
|
420
|
+
},
|
|
421
|
+
"gpt-5.1-codex-max": {
|
|
422
|
+
name: "gpt-5.1-codex-max",
|
|
423
|
+
provider: Vendor.OpenAI,
|
|
424
|
+
description: "GPT-5.1 codex max for maximum reasoning depth on coding tasks",
|
|
425
|
+
isActive: true,
|
|
426
|
+
releaseDate: "2025-10-01",
|
|
427
|
+
knowledgeCutoff: "2024-09-30",
|
|
428
|
+
features: {
|
|
429
|
+
reasoning: true,
|
|
430
|
+
streaming: true,
|
|
431
|
+
structuredOutput: true,
|
|
432
|
+
functionCalling: true,
|
|
433
|
+
fineTuning: false,
|
|
434
|
+
predictedOutputs: false,
|
|
435
|
+
realtime: false,
|
|
436
|
+
vision: true,
|
|
437
|
+
audio: false,
|
|
438
|
+
video: false,
|
|
439
|
+
batchAPI: true,
|
|
440
|
+
promptCaching: true,
|
|
441
|
+
parameters: {
|
|
442
|
+
temperature: false,
|
|
443
|
+
topP: false,
|
|
444
|
+
frequencyPenalty: false,
|
|
445
|
+
presencePenalty: false
|
|
446
|
+
},
|
|
447
|
+
input: {
|
|
448
|
+
tokens: 4e5,
|
|
449
|
+
text: true,
|
|
450
|
+
image: true,
|
|
451
|
+
cpm: 1.25,
|
|
452
|
+
cpmCached: 0.125
|
|
453
|
+
},
|
|
454
|
+
output: {
|
|
455
|
+
tokens: 128e3,
|
|
456
|
+
text: true,
|
|
457
|
+
cpm: 10
|
|
458
|
+
}
|
|
459
|
+
}
|
|
460
|
+
},
|
|
461
|
+
"gpt-5.1-codex-mini": {
|
|
462
|
+
name: "gpt-5.1-codex-mini",
|
|
463
|
+
provider: Vendor.OpenAI,
|
|
464
|
+
description: "GPT-5.1 codex mini for cost-efficient coding tasks",
|
|
465
|
+
isActive: true,
|
|
466
|
+
releaseDate: "2025-10-01",
|
|
467
|
+
knowledgeCutoff: "2024-09-30",
|
|
468
|
+
features: {
|
|
469
|
+
reasoning: true,
|
|
470
|
+
streaming: true,
|
|
471
|
+
structuredOutput: true,
|
|
472
|
+
functionCalling: true,
|
|
473
|
+
fineTuning: false,
|
|
474
|
+
predictedOutputs: false,
|
|
475
|
+
realtime: false,
|
|
476
|
+
vision: true,
|
|
477
|
+
audio: false,
|
|
478
|
+
video: false,
|
|
479
|
+
batchAPI: true,
|
|
480
|
+
promptCaching: true,
|
|
481
|
+
parameters: {
|
|
482
|
+
temperature: false,
|
|
483
|
+
topP: false,
|
|
484
|
+
frequencyPenalty: false,
|
|
485
|
+
presencePenalty: false
|
|
486
|
+
},
|
|
487
|
+
input: {
|
|
488
|
+
tokens: 4e5,
|
|
489
|
+
text: true,
|
|
490
|
+
image: true,
|
|
491
|
+
cpm: 0.25,
|
|
492
|
+
cpmCached: 0.025
|
|
493
|
+
},
|
|
494
|
+
output: {
|
|
495
|
+
tokens: 128e3,
|
|
496
|
+
text: true,
|
|
497
|
+
cpm: 2
|
|
498
|
+
}
|
|
499
|
+
}
|
|
500
|
+
},
|
|
501
|
+
"gpt-5.1-chat-latest": {
|
|
502
|
+
name: "gpt-5.1-chat-latest",
|
|
503
|
+
provider: Vendor.OpenAI,
|
|
504
|
+
description: "GPT-5.1 chat model for general-purpose use",
|
|
505
|
+
isActive: true,
|
|
506
|
+
releaseDate: "2025-10-01",
|
|
507
|
+
knowledgeCutoff: "2024-09-30",
|
|
508
|
+
features: {
|
|
509
|
+
reasoning: false,
|
|
510
|
+
streaming: true,
|
|
511
|
+
structuredOutput: true,
|
|
512
|
+
functionCalling: true,
|
|
513
|
+
fineTuning: false,
|
|
514
|
+
predictedOutputs: false,
|
|
515
|
+
realtime: false,
|
|
516
|
+
vision: true,
|
|
517
|
+
audio: false,
|
|
518
|
+
video: false,
|
|
519
|
+
batchAPI: true,
|
|
520
|
+
promptCaching: true,
|
|
521
|
+
input: {
|
|
522
|
+
tokens: 128e3,
|
|
523
|
+
text: true,
|
|
524
|
+
image: true,
|
|
525
|
+
cpm: 1.25,
|
|
526
|
+
cpmCached: 0.125
|
|
527
|
+
},
|
|
528
|
+
output: {
|
|
529
|
+
tokens: 16e3,
|
|
530
|
+
text: true,
|
|
531
|
+
cpm: 10
|
|
532
|
+
}
|
|
533
|
+
}
|
|
534
|
+
},
|
|
168
535
|
// GPT-5 Series
|
|
169
536
|
"gpt-5": {
|
|
170
537
|
name: "gpt-5",
|
|
@@ -196,7 +563,8 @@ var MODEL_REGISTRY = {
|
|
|
196
563
|
tokens: 4e5,
|
|
197
564
|
text: true,
|
|
198
565
|
image: true,
|
|
199
|
-
cpm: 1.25
|
|
566
|
+
cpm: 1.25,
|
|
567
|
+
cpmCached: 0.125
|
|
200
568
|
},
|
|
201
569
|
output: {
|
|
202
570
|
tokens: 128e3,
|
|
@@ -235,7 +603,8 @@ var MODEL_REGISTRY = {
|
|
|
235
603
|
tokens: 4e5,
|
|
236
604
|
text: true,
|
|
237
605
|
image: true,
|
|
238
|
-
cpm: 0.25
|
|
606
|
+
cpm: 0.25,
|
|
607
|
+
cpmCached: 0.025
|
|
239
608
|
},
|
|
240
609
|
output: {
|
|
241
610
|
tokens: 128e3,
|
|
@@ -274,7 +643,8 @@ var MODEL_REGISTRY = {
|
|
|
274
643
|
tokens: 4e5,
|
|
275
644
|
text: true,
|
|
276
645
|
image: true,
|
|
277
|
-
cpm: 0.05
|
|
646
|
+
cpm: 0.05,
|
|
647
|
+
cpmCached: 5e-3
|
|
278
648
|
},
|
|
279
649
|
output: {
|
|
280
650
|
tokens: 128e3,
|
|
@@ -283,6 +653,40 @@ var MODEL_REGISTRY = {
|
|
|
283
653
|
}
|
|
284
654
|
}
|
|
285
655
|
},
|
|
656
|
+
"gpt-5-chat-latest": {
|
|
657
|
+
name: "gpt-5-chat-latest",
|
|
658
|
+
provider: Vendor.OpenAI,
|
|
659
|
+
description: "GPT-5 chat model for general-purpose use",
|
|
660
|
+
isActive: true,
|
|
661
|
+
releaseDate: "2025-08-01",
|
|
662
|
+
knowledgeCutoff: "2024-09-30",
|
|
663
|
+
features: {
|
|
664
|
+
reasoning: false,
|
|
665
|
+
streaming: true,
|
|
666
|
+
structuredOutput: true,
|
|
667
|
+
functionCalling: true,
|
|
668
|
+
fineTuning: false,
|
|
669
|
+
predictedOutputs: false,
|
|
670
|
+
realtime: false,
|
|
671
|
+
vision: true,
|
|
672
|
+
audio: false,
|
|
673
|
+
video: false,
|
|
674
|
+
batchAPI: true,
|
|
675
|
+
promptCaching: true,
|
|
676
|
+
input: {
|
|
677
|
+
tokens: 128e3,
|
|
678
|
+
text: true,
|
|
679
|
+
image: true,
|
|
680
|
+
cpm: 1.25,
|
|
681
|
+
cpmCached: 0.125
|
|
682
|
+
},
|
|
683
|
+
output: {
|
|
684
|
+
tokens: 16e3,
|
|
685
|
+
text: true,
|
|
686
|
+
cpm: 10
|
|
687
|
+
}
|
|
688
|
+
}
|
|
689
|
+
},
|
|
286
690
|
// GPT-4.1 Series
|
|
287
691
|
"gpt-4.1": {
|
|
288
692
|
name: "gpt-4.1",
|
|
@@ -290,7 +694,7 @@ var MODEL_REGISTRY = {
|
|
|
290
694
|
description: "GPT-4.1 specialized for coding with 1M token context window",
|
|
291
695
|
isActive: true,
|
|
292
696
|
releaseDate: "2025-04-14",
|
|
293
|
-
knowledgeCutoff: "
|
|
697
|
+
knowledgeCutoff: "2024-06-01",
|
|
294
698
|
features: {
|
|
295
699
|
reasoning: false,
|
|
296
700
|
streaming: true,
|
|
@@ -308,7 +712,8 @@ var MODEL_REGISTRY = {
|
|
|
308
712
|
tokens: 1e6,
|
|
309
713
|
text: true,
|
|
310
714
|
image: true,
|
|
311
|
-
cpm: 2
|
|
715
|
+
cpm: 2,
|
|
716
|
+
cpmCached: 0.5
|
|
312
717
|
},
|
|
313
718
|
output: {
|
|
314
719
|
tokens: 32768,
|
|
@@ -323,7 +728,7 @@ var MODEL_REGISTRY = {
|
|
|
323
728
|
description: "Efficient GPT-4.1 model, beats GPT-4o in many benchmarks at 83% lower cost",
|
|
324
729
|
isActive: true,
|
|
325
730
|
releaseDate: "2025-04-14",
|
|
326
|
-
knowledgeCutoff: "
|
|
731
|
+
knowledgeCutoff: "2024-06-01",
|
|
327
732
|
features: {
|
|
328
733
|
reasoning: false,
|
|
329
734
|
streaming: true,
|
|
@@ -341,7 +746,8 @@ var MODEL_REGISTRY = {
|
|
|
341
746
|
tokens: 1e6,
|
|
342
747
|
text: true,
|
|
343
748
|
image: true,
|
|
344
|
-
cpm: 0.4
|
|
749
|
+
cpm: 0.4,
|
|
750
|
+
cpmCached: 0.1
|
|
345
751
|
},
|
|
346
752
|
output: {
|
|
347
753
|
tokens: 16384,
|
|
@@ -356,7 +762,7 @@ var MODEL_REGISTRY = {
|
|
|
356
762
|
description: "Fastest and cheapest model with 1M context. 80.1% MMLU, ideal for classification/autocompletion",
|
|
357
763
|
isActive: true,
|
|
358
764
|
releaseDate: "2025-04-14",
|
|
359
|
-
knowledgeCutoff: "
|
|
765
|
+
knowledgeCutoff: "2024-06-01",
|
|
360
766
|
features: {
|
|
361
767
|
reasoning: false,
|
|
362
768
|
streaming: true,
|
|
@@ -374,7 +780,8 @@ var MODEL_REGISTRY = {
|
|
|
374
780
|
tokens: 1e6,
|
|
375
781
|
text: true,
|
|
376
782
|
image: true,
|
|
377
|
-
cpm: 0.1
|
|
783
|
+
cpm: 0.1,
|
|
784
|
+
cpmCached: 0.025
|
|
378
785
|
},
|
|
379
786
|
output: {
|
|
380
787
|
tokens: 16384,
|
|
@@ -383,14 +790,14 @@ var MODEL_REGISTRY = {
|
|
|
383
790
|
}
|
|
384
791
|
}
|
|
385
792
|
},
|
|
386
|
-
// GPT-4o Series (Legacy
|
|
793
|
+
// GPT-4o Series (Legacy)
|
|
387
794
|
"gpt-4o": {
|
|
388
795
|
name: "gpt-4o",
|
|
389
796
|
provider: Vendor.OpenAI,
|
|
390
|
-
description: "Versatile omni model
|
|
797
|
+
description: "Versatile omni model. Legacy but still available",
|
|
391
798
|
isActive: true,
|
|
392
799
|
releaseDate: "2024-05-13",
|
|
393
|
-
knowledgeCutoff: "
|
|
800
|
+
knowledgeCutoff: "2023-10-01",
|
|
394
801
|
features: {
|
|
395
802
|
reasoning: false,
|
|
396
803
|
streaming: true,
|
|
@@ -400,7 +807,7 @@ var MODEL_REGISTRY = {
|
|
|
400
807
|
predictedOutputs: true,
|
|
401
808
|
realtime: true,
|
|
402
809
|
vision: true,
|
|
403
|
-
audio:
|
|
810
|
+
audio: false,
|
|
404
811
|
video: false,
|
|
405
812
|
batchAPI: true,
|
|
406
813
|
promptCaching: true,
|
|
@@ -408,13 +815,12 @@ var MODEL_REGISTRY = {
|
|
|
408
815
|
tokens: 128e3,
|
|
409
816
|
text: true,
|
|
410
817
|
image: true,
|
|
411
|
-
|
|
412
|
-
|
|
818
|
+
cpm: 2.5,
|
|
819
|
+
cpmCached: 1.25
|
|
413
820
|
},
|
|
414
821
|
output: {
|
|
415
822
|
tokens: 16384,
|
|
416
823
|
text: true,
|
|
417
|
-
audio: true,
|
|
418
824
|
cpm: 10
|
|
419
825
|
}
|
|
420
826
|
}
|
|
@@ -422,10 +828,10 @@ var MODEL_REGISTRY = {
|
|
|
422
828
|
"gpt-4o-mini": {
|
|
423
829
|
name: "gpt-4o-mini",
|
|
424
830
|
provider: Vendor.OpenAI,
|
|
425
|
-
description: "Fast, affordable omni model
|
|
831
|
+
description: "Fast, affordable omni model",
|
|
426
832
|
isActive: true,
|
|
427
833
|
releaseDate: "2024-07-18",
|
|
428
|
-
knowledgeCutoff: "
|
|
834
|
+
knowledgeCutoff: "2023-10-01",
|
|
429
835
|
features: {
|
|
430
836
|
reasoning: false,
|
|
431
837
|
streaming: true,
|
|
@@ -435,7 +841,7 @@ var MODEL_REGISTRY = {
|
|
|
435
841
|
predictedOutputs: false,
|
|
436
842
|
realtime: true,
|
|
437
843
|
vision: true,
|
|
438
|
-
audio:
|
|
844
|
+
audio: false,
|
|
439
845
|
video: false,
|
|
440
846
|
batchAPI: true,
|
|
441
847
|
promptCaching: true,
|
|
@@ -443,13 +849,12 @@ var MODEL_REGISTRY = {
|
|
|
443
849
|
tokens: 128e3,
|
|
444
850
|
text: true,
|
|
445
851
|
image: true,
|
|
446
|
-
|
|
447
|
-
|
|
852
|
+
cpm: 0.15,
|
|
853
|
+
cpmCached: 0.075
|
|
448
854
|
},
|
|
449
855
|
output: {
|
|
450
856
|
tokens: 16384,
|
|
451
857
|
text: true,
|
|
452
|
-
audio: true,
|
|
453
858
|
cpm: 0.6
|
|
454
859
|
}
|
|
455
860
|
}
|
|
@@ -461,7 +866,46 @@ var MODEL_REGISTRY = {
|
|
|
461
866
|
description: "Fast reasoning model tailored for coding, math, and science",
|
|
462
867
|
isActive: true,
|
|
463
868
|
releaseDate: "2025-01-31",
|
|
464
|
-
knowledgeCutoff: "
|
|
869
|
+
knowledgeCutoff: "2023-10-01",
|
|
870
|
+
features: {
|
|
871
|
+
reasoning: true,
|
|
872
|
+
streaming: true,
|
|
873
|
+
structuredOutput: true,
|
|
874
|
+
functionCalling: true,
|
|
875
|
+
fineTuning: false,
|
|
876
|
+
predictedOutputs: false,
|
|
877
|
+
realtime: false,
|
|
878
|
+
vision: false,
|
|
879
|
+
audio: false,
|
|
880
|
+
video: false,
|
|
881
|
+
batchAPI: true,
|
|
882
|
+
promptCaching: true,
|
|
883
|
+
parameters: {
|
|
884
|
+
temperature: false,
|
|
885
|
+
topP: false,
|
|
886
|
+
frequencyPenalty: false,
|
|
887
|
+
presencePenalty: false
|
|
888
|
+
},
|
|
889
|
+
input: {
|
|
890
|
+
tokens: 2e5,
|
|
891
|
+
text: true,
|
|
892
|
+
cpm: 1.1,
|
|
893
|
+
cpmCached: 0.55
|
|
894
|
+
},
|
|
895
|
+
output: {
|
|
896
|
+
tokens: 1e5,
|
|
897
|
+
text: true,
|
|
898
|
+
cpm: 4.4
|
|
899
|
+
}
|
|
900
|
+
}
|
|
901
|
+
},
|
|
902
|
+
"o1": {
|
|
903
|
+
name: "o1",
|
|
904
|
+
provider: Vendor.OpenAI,
|
|
905
|
+
description: "Advanced reasoning model for complex problems",
|
|
906
|
+
isActive: true,
|
|
907
|
+
releaseDate: "2024-12-17",
|
|
908
|
+
knowledgeCutoff: "2023-10-01",
|
|
465
909
|
features: {
|
|
466
910
|
reasoning: true,
|
|
467
911
|
streaming: true,
|
|
@@ -474,35 +918,78 @@ var MODEL_REGISTRY = {
|
|
|
474
918
|
audio: false,
|
|
475
919
|
video: false,
|
|
476
920
|
batchAPI: true,
|
|
477
|
-
promptCaching:
|
|
478
|
-
parameters: {
|
|
479
|
-
temperature: false,
|
|
480
|
-
topP: false,
|
|
481
|
-
frequencyPenalty: false,
|
|
482
|
-
presencePenalty: false
|
|
483
|
-
},
|
|
921
|
+
promptCaching: true,
|
|
922
|
+
parameters: {
|
|
923
|
+
temperature: false,
|
|
924
|
+
topP: false,
|
|
925
|
+
frequencyPenalty: false,
|
|
926
|
+
presencePenalty: false
|
|
927
|
+
},
|
|
928
|
+
input: {
|
|
929
|
+
tokens: 2e5,
|
|
930
|
+
text: true,
|
|
931
|
+
image: true,
|
|
932
|
+
cpm: 15,
|
|
933
|
+
cpmCached: 7.5
|
|
934
|
+
},
|
|
935
|
+
output: {
|
|
936
|
+
tokens: 1e5,
|
|
937
|
+
text: true,
|
|
938
|
+
cpm: 60
|
|
939
|
+
}
|
|
940
|
+
}
|
|
941
|
+
},
|
|
942
|
+
// ============================================================================
|
|
943
|
+
// Anthropic Models (Verified from platform.claude.com - March 2026)
|
|
944
|
+
// ============================================================================
|
|
945
|
+
// Claude 4.6 Series (Current)
|
|
946
|
+
"claude-opus-4-6": {
|
|
947
|
+
name: "claude-opus-4-6",
|
|
948
|
+
provider: Vendor.Anthropic,
|
|
949
|
+
description: "The most intelligent model for building agents and coding. 128K output, adaptive thinking",
|
|
950
|
+
isActive: true,
|
|
951
|
+
preferred: true,
|
|
952
|
+
releaseDate: "2026-02-01",
|
|
953
|
+
knowledgeCutoff: "2025-05-01",
|
|
954
|
+
features: {
|
|
955
|
+
reasoning: false,
|
|
956
|
+
streaming: true,
|
|
957
|
+
structuredOutput: true,
|
|
958
|
+
functionCalling: true,
|
|
959
|
+
fineTuning: false,
|
|
960
|
+
predictedOutputs: false,
|
|
961
|
+
realtime: false,
|
|
962
|
+
vision: true,
|
|
963
|
+
audio: false,
|
|
964
|
+
video: false,
|
|
965
|
+
extendedThinking: true,
|
|
966
|
+
batchAPI: true,
|
|
967
|
+
promptCaching: true,
|
|
484
968
|
input: {
|
|
485
969
|
tokens: 2e5,
|
|
970
|
+
// 1M with beta header
|
|
486
971
|
text: true,
|
|
487
972
|
image: true,
|
|
488
|
-
cpm:
|
|
973
|
+
cpm: 5,
|
|
974
|
+
cpmCached: 0.5
|
|
489
975
|
},
|
|
490
976
|
output: {
|
|
491
|
-
tokens:
|
|
977
|
+
tokens: 128e3,
|
|
492
978
|
text: true,
|
|
493
|
-
cpm:
|
|
979
|
+
cpm: 25
|
|
494
980
|
}
|
|
495
981
|
}
|
|
496
982
|
},
|
|
497
|
-
"
|
|
498
|
-
name: "
|
|
499
|
-
provider: Vendor.
|
|
500
|
-
description: "
|
|
983
|
+
"claude-sonnet-4-6": {
|
|
984
|
+
name: "claude-sonnet-4-6",
|
|
985
|
+
provider: Vendor.Anthropic,
|
|
986
|
+
description: "Best combination of speed and intelligence. Adaptive thinking, 1M context beta",
|
|
501
987
|
isActive: true,
|
|
502
|
-
|
|
503
|
-
|
|
988
|
+
preferred: true,
|
|
989
|
+
releaseDate: "2026-02-01",
|
|
990
|
+
knowledgeCutoff: "2025-08-01",
|
|
504
991
|
features: {
|
|
505
|
-
reasoning:
|
|
992
|
+
reasoning: false,
|
|
506
993
|
streaming: true,
|
|
507
994
|
structuredOutput: true,
|
|
508
995
|
functionCalling: true,
|
|
@@ -512,35 +999,29 @@ var MODEL_REGISTRY = {
|
|
|
512
999
|
vision: true,
|
|
513
1000
|
audio: false,
|
|
514
1001
|
video: false,
|
|
1002
|
+
extendedThinking: true,
|
|
515
1003
|
batchAPI: true,
|
|
516
|
-
promptCaching:
|
|
517
|
-
parameters: {
|
|
518
|
-
temperature: false,
|
|
519
|
-
topP: false,
|
|
520
|
-
frequencyPenalty: false,
|
|
521
|
-
presencePenalty: false
|
|
522
|
-
},
|
|
1004
|
+
promptCaching: true,
|
|
523
1005
|
input: {
|
|
524
1006
|
tokens: 2e5,
|
|
1007
|
+
// 1M with beta header
|
|
525
1008
|
text: true,
|
|
526
1009
|
image: true,
|
|
527
|
-
cpm:
|
|
1010
|
+
cpm: 3,
|
|
1011
|
+
cpmCached: 0.3
|
|
528
1012
|
},
|
|
529
1013
|
output: {
|
|
530
|
-
tokens:
|
|
1014
|
+
tokens: 64e3,
|
|
531
1015
|
text: true,
|
|
532
|
-
cpm:
|
|
1016
|
+
cpm: 15
|
|
533
1017
|
}
|
|
534
1018
|
}
|
|
535
1019
|
},
|
|
536
|
-
//
|
|
537
|
-
// Anthropic Models (Verified from platform.claude.com)
|
|
538
|
-
// ============================================================================
|
|
539
|
-
// Claude 4.5 Series (Current)
|
|
1020
|
+
// Claude 4.5 Series
|
|
540
1021
|
"claude-opus-4-5-20251101": {
|
|
541
1022
|
name: "claude-opus-4-5-20251101",
|
|
542
1023
|
provider: Vendor.Anthropic,
|
|
543
|
-
description: "Premium model combining maximum intelligence with practical performance",
|
|
1024
|
+
description: "Legacy Opus 4.5. Premium model combining maximum intelligence with practical performance",
|
|
544
1025
|
isActive: true,
|
|
545
1026
|
releaseDate: "2025-11-01",
|
|
546
1027
|
knowledgeCutoff: "2025-05-01",
|
|
@@ -575,7 +1056,7 @@ var MODEL_REGISTRY = {
|
|
|
575
1056
|
"claude-sonnet-4-5-20250929": {
|
|
576
1057
|
name: "claude-sonnet-4-5-20250929",
|
|
577
1058
|
provider: Vendor.Anthropic,
|
|
578
|
-
description: "Smart model for complex agents and coding
|
|
1059
|
+
description: "Legacy Sonnet 4.5. Smart model for complex agents and coding",
|
|
579
1060
|
isActive: true,
|
|
580
1061
|
releaseDate: "2025-09-29",
|
|
581
1062
|
knowledgeCutoff: "2025-01-01",
|
|
@@ -595,6 +1076,7 @@ var MODEL_REGISTRY = {
|
|
|
595
1076
|
promptCaching: true,
|
|
596
1077
|
input: {
|
|
597
1078
|
tokens: 2e5,
|
|
1079
|
+
// 1M with beta header
|
|
598
1080
|
text: true,
|
|
599
1081
|
image: true,
|
|
600
1082
|
cpm: 3,
|
|
@@ -678,10 +1160,45 @@ var MODEL_REGISTRY = {
|
|
|
678
1160
|
}
|
|
679
1161
|
}
|
|
680
1162
|
},
|
|
1163
|
+
"claude-opus-4-20250514": {
|
|
1164
|
+
name: "claude-opus-4-20250514",
|
|
1165
|
+
provider: Vendor.Anthropic,
|
|
1166
|
+
description: "Legacy Opus 4. Agentic tasks and reasoning",
|
|
1167
|
+
isActive: true,
|
|
1168
|
+
releaseDate: "2025-05-14",
|
|
1169
|
+
knowledgeCutoff: "2025-01-01",
|
|
1170
|
+
features: {
|
|
1171
|
+
reasoning: false,
|
|
1172
|
+
streaming: true,
|
|
1173
|
+
structuredOutput: true,
|
|
1174
|
+
functionCalling: true,
|
|
1175
|
+
fineTuning: false,
|
|
1176
|
+
predictedOutputs: false,
|
|
1177
|
+
realtime: false,
|
|
1178
|
+
vision: true,
|
|
1179
|
+
audio: false,
|
|
1180
|
+
video: false,
|
|
1181
|
+
extendedThinking: true,
|
|
1182
|
+
batchAPI: true,
|
|
1183
|
+
promptCaching: true,
|
|
1184
|
+
input: {
|
|
1185
|
+
tokens: 2e5,
|
|
1186
|
+
text: true,
|
|
1187
|
+
image: true,
|
|
1188
|
+
cpm: 15,
|
|
1189
|
+
cpmCached: 1.5
|
|
1190
|
+
},
|
|
1191
|
+
output: {
|
|
1192
|
+
tokens: 32e3,
|
|
1193
|
+
text: true,
|
|
1194
|
+
cpm: 75
|
|
1195
|
+
}
|
|
1196
|
+
}
|
|
1197
|
+
},
|
|
681
1198
|
"claude-sonnet-4-20250514": {
|
|
682
1199
|
name: "claude-sonnet-4-20250514",
|
|
683
1200
|
provider: Vendor.Anthropic,
|
|
684
|
-
description: "Legacy Sonnet 4.
|
|
1201
|
+
description: "Legacy Sonnet 4. Supports 1M context beta",
|
|
685
1202
|
isActive: true,
|
|
686
1203
|
releaseDate: "2025-05-14",
|
|
687
1204
|
knowledgeCutoff: "2025-01-01",
|
|
@@ -717,7 +1234,7 @@ var MODEL_REGISTRY = {
|
|
|
717
1234
|
"claude-3-7-sonnet-20250219": {
|
|
718
1235
|
name: "claude-3-7-sonnet-20250219",
|
|
719
1236
|
provider: Vendor.Anthropic,
|
|
720
|
-
description: "Claude 3.7 Sonnet with extended thinking
|
|
1237
|
+
description: "Deprecated. Claude 3.7 Sonnet with extended thinking",
|
|
721
1238
|
isActive: true,
|
|
722
1239
|
releaseDate: "2025-02-19",
|
|
723
1240
|
knowledgeCutoff: "2024-10-01",
|
|
@@ -744,17 +1261,16 @@ var MODEL_REGISTRY = {
|
|
|
744
1261
|
},
|
|
745
1262
|
output: {
|
|
746
1263
|
tokens: 64e3,
|
|
747
|
-
// 128K with beta header
|
|
748
1264
|
text: true,
|
|
749
1265
|
cpm: 15
|
|
750
1266
|
}
|
|
751
1267
|
}
|
|
752
1268
|
},
|
|
753
|
-
// Claude 3.x Legacy
|
|
1269
|
+
// Claude 3.x Legacy (Deprecated - retiring April 19, 2026)
|
|
754
1270
|
"claude-3-haiku-20240307": {
|
|
755
1271
|
name: "claude-3-haiku-20240307",
|
|
756
1272
|
provider: Vendor.Anthropic,
|
|
757
|
-
description: "
|
|
1273
|
+
description: "Deprecated. Retiring April 19, 2026. Migrate to Haiku 4.5",
|
|
758
1274
|
isActive: true,
|
|
759
1275
|
releaseDate: "2024-03-07",
|
|
760
1276
|
knowledgeCutoff: "2023-08-01",
|
|
@@ -787,16 +1303,124 @@ var MODEL_REGISTRY = {
|
|
|
787
1303
|
}
|
|
788
1304
|
},
|
|
789
1305
|
// ============================================================================
|
|
790
|
-
// Google Models (Verified from ai.google.dev)
|
|
1306
|
+
// Google Models (Verified from ai.google.dev - March 2026)
|
|
791
1307
|
// ============================================================================
|
|
1308
|
+
// Gemini 3.1 Series (Preview)
|
|
1309
|
+
"gemini-3.1-pro-preview": {
|
|
1310
|
+
name: "gemini-3.1-pro-preview",
|
|
1311
|
+
provider: Vendor.Google,
|
|
1312
|
+
description: "Advanced intelligence with powerful agentic and coding capabilities. Replaces gemini-3-pro-preview",
|
|
1313
|
+
isActive: true,
|
|
1314
|
+
preferred: true,
|
|
1315
|
+
releaseDate: "2026-02-01",
|
|
1316
|
+
knowledgeCutoff: "2025-01-01",
|
|
1317
|
+
features: {
|
|
1318
|
+
reasoning: true,
|
|
1319
|
+
streaming: true,
|
|
1320
|
+
structuredOutput: true,
|
|
1321
|
+
functionCalling: true,
|
|
1322
|
+
fineTuning: false,
|
|
1323
|
+
predictedOutputs: false,
|
|
1324
|
+
realtime: false,
|
|
1325
|
+
vision: true,
|
|
1326
|
+
audio: true,
|
|
1327
|
+
video: true,
|
|
1328
|
+
batchAPI: true,
|
|
1329
|
+
promptCaching: true,
|
|
1330
|
+
input: {
|
|
1331
|
+
tokens: 1048576,
|
|
1332
|
+
text: true,
|
|
1333
|
+
image: true,
|
|
1334
|
+
audio: true,
|
|
1335
|
+
video: true,
|
|
1336
|
+
cpm: 2,
|
|
1337
|
+
cpmCached: 0.2
|
|
1338
|
+
},
|
|
1339
|
+
output: {
|
|
1340
|
+
tokens: 65536,
|
|
1341
|
+
text: true,
|
|
1342
|
+
cpm: 12
|
|
1343
|
+
}
|
|
1344
|
+
}
|
|
1345
|
+
},
|
|
1346
|
+
"gemini-3.1-flash-lite-preview": {
|
|
1347
|
+
name: "gemini-3.1-flash-lite-preview",
|
|
1348
|
+
provider: Vendor.Google,
|
|
1349
|
+
description: "High performance, budget-friendly for high-volume agentic tasks and data extraction",
|
|
1350
|
+
isActive: true,
|
|
1351
|
+
releaseDate: "2026-03-01",
|
|
1352
|
+
knowledgeCutoff: "2025-01-01",
|
|
1353
|
+
features: {
|
|
1354
|
+
reasoning: true,
|
|
1355
|
+
streaming: true,
|
|
1356
|
+
structuredOutput: true,
|
|
1357
|
+
functionCalling: true,
|
|
1358
|
+
fineTuning: false,
|
|
1359
|
+
predictedOutputs: false,
|
|
1360
|
+
realtime: false,
|
|
1361
|
+
vision: true,
|
|
1362
|
+
audio: true,
|
|
1363
|
+
video: true,
|
|
1364
|
+
batchAPI: true,
|
|
1365
|
+
promptCaching: true,
|
|
1366
|
+
input: {
|
|
1367
|
+
tokens: 1048576,
|
|
1368
|
+
text: true,
|
|
1369
|
+
image: true,
|
|
1370
|
+
audio: true,
|
|
1371
|
+
video: true,
|
|
1372
|
+
cpm: 0.25
|
|
1373
|
+
},
|
|
1374
|
+
output: {
|
|
1375
|
+
tokens: 65536,
|
|
1376
|
+
text: true,
|
|
1377
|
+
cpm: 1.5
|
|
1378
|
+
}
|
|
1379
|
+
}
|
|
1380
|
+
},
|
|
1381
|
+
"gemini-3.1-flash-image-preview": {
|
|
1382
|
+
name: "gemini-3.1-flash-image-preview",
|
|
1383
|
+
provider: Vendor.Google,
|
|
1384
|
+
description: "High-efficiency image generation with up to 4K output, search grounding support",
|
|
1385
|
+
isActive: true,
|
|
1386
|
+
releaseDate: "2026-02-01",
|
|
1387
|
+
knowledgeCutoff: "2025-01-01",
|
|
1388
|
+
features: {
|
|
1389
|
+
reasoning: true,
|
|
1390
|
+
streaming: true,
|
|
1391
|
+
structuredOutput: false,
|
|
1392
|
+
functionCalling: false,
|
|
1393
|
+
fineTuning: false,
|
|
1394
|
+
predictedOutputs: false,
|
|
1395
|
+
realtime: false,
|
|
1396
|
+
vision: true,
|
|
1397
|
+
audio: false,
|
|
1398
|
+
video: false,
|
|
1399
|
+
batchAPI: true,
|
|
1400
|
+
promptCaching: false,
|
|
1401
|
+
input: {
|
|
1402
|
+
tokens: 131072,
|
|
1403
|
+
text: true,
|
|
1404
|
+
image: true,
|
|
1405
|
+
cpm: 0.25
|
|
1406
|
+
},
|
|
1407
|
+
output: {
|
|
1408
|
+
tokens: 32768,
|
|
1409
|
+
text: true,
|
|
1410
|
+
image: true,
|
|
1411
|
+
cpm: 1.5
|
|
1412
|
+
}
|
|
1413
|
+
}
|
|
1414
|
+
},
|
|
792
1415
|
// Gemini 3 Series (Preview)
|
|
793
1416
|
"gemini-3-flash-preview": {
|
|
794
1417
|
name: "gemini-3-flash-preview",
|
|
795
1418
|
provider: Vendor.Google,
|
|
796
|
-
description: "
|
|
1419
|
+
description: "Most powerful agentic and coding model with frontier-class reasoning",
|
|
797
1420
|
isActive: true,
|
|
798
|
-
|
|
799
|
-
|
|
1421
|
+
preferred: true,
|
|
1422
|
+
releaseDate: "2025-12-01",
|
|
1423
|
+
knowledgeCutoff: "2025-01-01",
|
|
800
1424
|
features: {
|
|
801
1425
|
reasoning: true,
|
|
802
1426
|
streaming: true,
|
|
@@ -811,27 +1435,28 @@ var MODEL_REGISTRY = {
|
|
|
811
1435
|
batchAPI: true,
|
|
812
1436
|
promptCaching: true,
|
|
813
1437
|
input: {
|
|
814
|
-
tokens:
|
|
1438
|
+
tokens: 1048576,
|
|
815
1439
|
text: true,
|
|
816
1440
|
image: true,
|
|
817
1441
|
audio: true,
|
|
818
1442
|
video: true,
|
|
819
|
-
cpm: 0.
|
|
1443
|
+
cpm: 0.5,
|
|
1444
|
+
cpmCached: 0.05
|
|
820
1445
|
},
|
|
821
1446
|
output: {
|
|
822
1447
|
tokens: 65536,
|
|
823
1448
|
text: true,
|
|
824
|
-
cpm:
|
|
1449
|
+
cpm: 3
|
|
825
1450
|
}
|
|
826
1451
|
}
|
|
827
1452
|
},
|
|
828
1453
|
"gemini-3-pro-preview": {
|
|
829
1454
|
name: "gemini-3-pro-preview",
|
|
830
1455
|
provider: Vendor.Google,
|
|
831
|
-
description: "
|
|
1456
|
+
description: "Deprecated. Shutting down March 9, 2026. Migrate to gemini-3.1-pro-preview",
|
|
832
1457
|
isActive: true,
|
|
833
1458
|
releaseDate: "2025-11-18",
|
|
834
|
-
knowledgeCutoff: "2025-
|
|
1459
|
+
knowledgeCutoff: "2025-01-01",
|
|
835
1460
|
features: {
|
|
836
1461
|
reasoning: true,
|
|
837
1462
|
streaming: true,
|
|
@@ -846,7 +1471,7 @@ var MODEL_REGISTRY = {
|
|
|
846
1471
|
batchAPI: true,
|
|
847
1472
|
promptCaching: true,
|
|
848
1473
|
input: {
|
|
849
|
-
tokens:
|
|
1474
|
+
tokens: 1048576,
|
|
850
1475
|
text: true,
|
|
851
1476
|
image: true,
|
|
852
1477
|
audio: true,
|
|
@@ -863,14 +1488,14 @@ var MODEL_REGISTRY = {
|
|
|
863
1488
|
"gemini-3-pro-image-preview": {
|
|
864
1489
|
name: "gemini-3-pro-image-preview",
|
|
865
1490
|
provider: Vendor.Google,
|
|
866
|
-
description: "
|
|
1491
|
+
description: "Professional-grade image generation and editing with reasoning",
|
|
867
1492
|
isActive: true,
|
|
868
1493
|
releaseDate: "2025-11-18",
|
|
869
|
-
knowledgeCutoff: "2025-
|
|
1494
|
+
knowledgeCutoff: "2025-01-01",
|
|
870
1495
|
features: {
|
|
871
1496
|
reasoning: true,
|
|
872
1497
|
streaming: true,
|
|
873
|
-
structuredOutput:
|
|
1498
|
+
structuredOutput: true,
|
|
874
1499
|
functionCalling: false,
|
|
875
1500
|
fineTuning: false,
|
|
876
1501
|
predictedOutputs: false,
|
|
@@ -879,15 +1504,15 @@ var MODEL_REGISTRY = {
|
|
|
879
1504
|
audio: false,
|
|
880
1505
|
video: false,
|
|
881
1506
|
batchAPI: true,
|
|
882
|
-
promptCaching:
|
|
1507
|
+
promptCaching: false,
|
|
883
1508
|
input: {
|
|
884
|
-
tokens:
|
|
1509
|
+
tokens: 65536,
|
|
885
1510
|
text: true,
|
|
886
1511
|
image: true,
|
|
887
1512
|
cpm: 1.25
|
|
888
1513
|
},
|
|
889
1514
|
output: {
|
|
890
|
-
tokens:
|
|
1515
|
+
tokens: 32768,
|
|
891
1516
|
text: true,
|
|
892
1517
|
image: true,
|
|
893
1518
|
cpm: 10
|
|
@@ -898,7 +1523,7 @@ var MODEL_REGISTRY = {
|
|
|
898
1523
|
"gemini-2.5-pro": {
|
|
899
1524
|
name: "gemini-2.5-pro",
|
|
900
1525
|
provider: Vendor.Google,
|
|
901
|
-
description: "
|
|
1526
|
+
description: "Most advanced model for complex tasks with deep reasoning and coding",
|
|
902
1527
|
isActive: true,
|
|
903
1528
|
releaseDate: "2025-03-01",
|
|
904
1529
|
knowledgeCutoff: "2025-01-01",
|
|
@@ -916,12 +1541,13 @@ var MODEL_REGISTRY = {
|
|
|
916
1541
|
batchAPI: true,
|
|
917
1542
|
promptCaching: true,
|
|
918
1543
|
input: {
|
|
919
|
-
tokens:
|
|
1544
|
+
tokens: 1048576,
|
|
920
1545
|
text: true,
|
|
921
1546
|
image: true,
|
|
922
1547
|
audio: true,
|
|
923
1548
|
video: true,
|
|
924
|
-
cpm: 1.25
|
|
1549
|
+
cpm: 1.25,
|
|
1550
|
+
cpmCached: 0.125
|
|
925
1551
|
},
|
|
926
1552
|
output: {
|
|
927
1553
|
tokens: 65536,
|
|
@@ -933,7 +1559,7 @@ var MODEL_REGISTRY = {
|
|
|
933
1559
|
"gemini-2.5-flash": {
|
|
934
1560
|
name: "gemini-2.5-flash",
|
|
935
1561
|
provider: Vendor.Google,
|
|
936
|
-
description: "
|
|
1562
|
+
description: "Best price-performance for low-latency, high-volume tasks with reasoning",
|
|
937
1563
|
isActive: true,
|
|
938
1564
|
releaseDate: "2025-06-17",
|
|
939
1565
|
knowledgeCutoff: "2025-01-01",
|
|
@@ -951,24 +1577,25 @@ var MODEL_REGISTRY = {
|
|
|
951
1577
|
batchAPI: true,
|
|
952
1578
|
promptCaching: true,
|
|
953
1579
|
input: {
|
|
954
|
-
tokens:
|
|
1580
|
+
tokens: 1048576,
|
|
955
1581
|
text: true,
|
|
956
1582
|
image: true,
|
|
957
1583
|
audio: true,
|
|
958
1584
|
video: true,
|
|
959
|
-
cpm: 0.
|
|
1585
|
+
cpm: 0.3,
|
|
1586
|
+
cpmCached: 0.03
|
|
960
1587
|
},
|
|
961
1588
|
output: {
|
|
962
1589
|
tokens: 65536,
|
|
963
1590
|
text: true,
|
|
964
|
-
cpm:
|
|
1591
|
+
cpm: 2.5
|
|
965
1592
|
}
|
|
966
1593
|
}
|
|
967
1594
|
},
|
|
968
1595
|
"gemini-2.5-flash-lite": {
|
|
969
1596
|
name: "gemini-2.5-flash-lite",
|
|
970
1597
|
provider: Vendor.Google,
|
|
971
|
-
description: "
|
|
1598
|
+
description: "Fastest and most budget-friendly multimodal model in the 2.5 family",
|
|
972
1599
|
isActive: true,
|
|
973
1600
|
releaseDate: "2025-06-17",
|
|
974
1601
|
knowledgeCutoff: "2025-01-01",
|
|
@@ -986,31 +1613,31 @@ var MODEL_REGISTRY = {
|
|
|
986
1613
|
batchAPI: true,
|
|
987
1614
|
promptCaching: true,
|
|
988
1615
|
input: {
|
|
989
|
-
tokens:
|
|
1616
|
+
tokens: 1048576,
|
|
990
1617
|
text: true,
|
|
991
1618
|
image: true,
|
|
992
1619
|
audio: true,
|
|
993
1620
|
video: true,
|
|
994
|
-
cpm: 0.
|
|
1621
|
+
cpm: 0.1
|
|
995
1622
|
},
|
|
996
1623
|
output: {
|
|
997
1624
|
tokens: 65536,
|
|
998
1625
|
text: true,
|
|
999
|
-
cpm: 0.
|
|
1626
|
+
cpm: 0.4
|
|
1000
1627
|
}
|
|
1001
1628
|
}
|
|
1002
1629
|
},
|
|
1003
1630
|
"gemini-2.5-flash-image": {
|
|
1004
1631
|
name: "gemini-2.5-flash-image",
|
|
1005
1632
|
provider: Vendor.Google,
|
|
1006
|
-
description: "
|
|
1633
|
+
description: "Fast native image generation and editing (Nano Banana)",
|
|
1007
1634
|
isActive: true,
|
|
1008
|
-
releaseDate: "2025-
|
|
1009
|
-
knowledgeCutoff: "2025-
|
|
1635
|
+
releaseDate: "2025-10-01",
|
|
1636
|
+
knowledgeCutoff: "2025-06-01",
|
|
1010
1637
|
features: {
|
|
1011
|
-
reasoning:
|
|
1638
|
+
reasoning: false,
|
|
1012
1639
|
streaming: true,
|
|
1013
|
-
structuredOutput:
|
|
1640
|
+
structuredOutput: true,
|
|
1014
1641
|
functionCalling: false,
|
|
1015
1642
|
fineTuning: false,
|
|
1016
1643
|
predictedOutputs: false,
|
|
@@ -1021,13 +1648,13 @@ var MODEL_REGISTRY = {
|
|
|
1021
1648
|
batchAPI: true,
|
|
1022
1649
|
promptCaching: true,
|
|
1023
1650
|
input: {
|
|
1024
|
-
tokens:
|
|
1651
|
+
tokens: 65536,
|
|
1025
1652
|
text: true,
|
|
1026
1653
|
image: true,
|
|
1027
1654
|
cpm: 0.15
|
|
1028
1655
|
},
|
|
1029
1656
|
output: {
|
|
1030
|
-
tokens:
|
|
1657
|
+
tokens: 32768,
|
|
1031
1658
|
text: true,
|
|
1032
1659
|
image: true,
|
|
1033
1660
|
cpm: 0.6
|
|
@@ -1035,7 +1662,7 @@ var MODEL_REGISTRY = {
|
|
|
1035
1662
|
}
|
|
1036
1663
|
},
|
|
1037
1664
|
// ============================================================================
|
|
1038
|
-
// xAI Grok Models (Verified from docs.x.ai -
|
|
1665
|
+
// xAI Grok Models (Verified from docs.x.ai - March 2026)
|
|
1039
1666
|
// ============================================================================
|
|
1040
1667
|
// Grok 4.1 Series (2M context, fast)
|
|
1041
1668
|
"grok-4-1-fast-reasoning": {
|
|
@@ -1056,13 +1683,14 @@ var MODEL_REGISTRY = {
|
|
|
1056
1683
|
vision: true,
|
|
1057
1684
|
audio: false,
|
|
1058
1685
|
video: false,
|
|
1059
|
-
batchAPI:
|
|
1060
|
-
promptCaching:
|
|
1686
|
+
batchAPI: true,
|
|
1687
|
+
promptCaching: true,
|
|
1061
1688
|
input: {
|
|
1062
1689
|
tokens: 2e6,
|
|
1063
1690
|
text: true,
|
|
1064
1691
|
image: true,
|
|
1065
|
-
cpm: 0.2
|
|
1692
|
+
cpm: 0.2,
|
|
1693
|
+
cpmCached: 0.05
|
|
1066
1694
|
},
|
|
1067
1695
|
output: {
|
|
1068
1696
|
tokens: 65536,
|
|
@@ -1089,13 +1717,14 @@ var MODEL_REGISTRY = {
|
|
|
1089
1717
|
vision: true,
|
|
1090
1718
|
audio: false,
|
|
1091
1719
|
video: false,
|
|
1092
|
-
batchAPI:
|
|
1093
|
-
promptCaching:
|
|
1720
|
+
batchAPI: true,
|
|
1721
|
+
promptCaching: true,
|
|
1094
1722
|
input: {
|
|
1095
1723
|
tokens: 2e6,
|
|
1096
1724
|
text: true,
|
|
1097
1725
|
image: true,
|
|
1098
|
-
cpm: 0.2
|
|
1726
|
+
cpm: 0.2,
|
|
1727
|
+
cpmCached: 0.05
|
|
1099
1728
|
},
|
|
1100
1729
|
output: {
|
|
1101
1730
|
tokens: 65536,
|
|
@@ -1123,12 +1752,13 @@ var MODEL_REGISTRY = {
|
|
|
1123
1752
|
vision: false,
|
|
1124
1753
|
audio: false,
|
|
1125
1754
|
video: false,
|
|
1126
|
-
batchAPI:
|
|
1127
|
-
promptCaching:
|
|
1755
|
+
batchAPI: true,
|
|
1756
|
+
promptCaching: true,
|
|
1128
1757
|
input: {
|
|
1129
1758
|
tokens: 256e3,
|
|
1130
1759
|
text: true,
|
|
1131
|
-
cpm: 0.2
|
|
1760
|
+
cpm: 0.2,
|
|
1761
|
+
cpmCached: 0.02
|
|
1132
1762
|
},
|
|
1133
1763
|
output: {
|
|
1134
1764
|
tokens: 32768,
|
|
@@ -1141,7 +1771,7 @@ var MODEL_REGISTRY = {
|
|
|
1141
1771
|
"grok-4-fast-reasoning": {
|
|
1142
1772
|
name: "grok-4-fast-reasoning",
|
|
1143
1773
|
provider: Vendor.Grok,
|
|
1144
|
-
description: "Fast Grok 4 with reasoning capabilities, 2M context window",
|
|
1774
|
+
description: "Fast Grok 4 with reasoning capabilities, 2M context window, vision support",
|
|
1145
1775
|
isActive: true,
|
|
1146
1776
|
releaseDate: "2025-09-01",
|
|
1147
1777
|
knowledgeCutoff: "2024-11-01",
|
|
@@ -1153,15 +1783,17 @@ var MODEL_REGISTRY = {
|
|
|
1153
1783
|
fineTuning: false,
|
|
1154
1784
|
predictedOutputs: false,
|
|
1155
1785
|
realtime: false,
|
|
1156
|
-
vision:
|
|
1786
|
+
vision: true,
|
|
1157
1787
|
audio: false,
|
|
1158
1788
|
video: false,
|
|
1159
|
-
batchAPI:
|
|
1160
|
-
promptCaching:
|
|
1789
|
+
batchAPI: true,
|
|
1790
|
+
promptCaching: true,
|
|
1161
1791
|
input: {
|
|
1162
1792
|
tokens: 2e6,
|
|
1163
1793
|
text: true,
|
|
1164
|
-
|
|
1794
|
+
image: true,
|
|
1795
|
+
cpm: 0.2,
|
|
1796
|
+
cpmCached: 0.05
|
|
1165
1797
|
},
|
|
1166
1798
|
output: {
|
|
1167
1799
|
tokens: 65536,
|
|
@@ -1188,13 +1820,14 @@ var MODEL_REGISTRY = {
|
|
|
1188
1820
|
vision: true,
|
|
1189
1821
|
audio: false,
|
|
1190
1822
|
video: false,
|
|
1191
|
-
batchAPI:
|
|
1192
|
-
promptCaching:
|
|
1823
|
+
batchAPI: true,
|
|
1824
|
+
promptCaching: true,
|
|
1193
1825
|
input: {
|
|
1194
1826
|
tokens: 2e6,
|
|
1195
1827
|
text: true,
|
|
1196
1828
|
image: true,
|
|
1197
|
-
cpm: 0.2
|
|
1829
|
+
cpm: 0.2,
|
|
1830
|
+
cpmCached: 0.05
|
|
1198
1831
|
},
|
|
1199
1832
|
output: {
|
|
1200
1833
|
tokens: 65536,
|
|
@@ -1206,12 +1839,12 @@ var MODEL_REGISTRY = {
|
|
|
1206
1839
|
"grok-4-0709": {
|
|
1207
1840
|
name: "grok-4-0709",
|
|
1208
1841
|
provider: Vendor.Grok,
|
|
1209
|
-
description: "Grok 4 flagship model (July 2025 release), 256K context, vision support",
|
|
1842
|
+
description: "Grok 4 flagship model (July 2025 release), 256K context, vision support, reasoning",
|
|
1210
1843
|
isActive: true,
|
|
1211
1844
|
releaseDate: "2025-07-09",
|
|
1212
1845
|
knowledgeCutoff: "2024-11-01",
|
|
1213
1846
|
features: {
|
|
1214
|
-
reasoning:
|
|
1847
|
+
reasoning: true,
|
|
1215
1848
|
streaming: true,
|
|
1216
1849
|
structuredOutput: true,
|
|
1217
1850
|
functionCalling: true,
|
|
@@ -1221,13 +1854,14 @@ var MODEL_REGISTRY = {
|
|
|
1221
1854
|
vision: true,
|
|
1222
1855
|
audio: false,
|
|
1223
1856
|
video: false,
|
|
1224
|
-
batchAPI:
|
|
1225
|
-
promptCaching:
|
|
1857
|
+
batchAPI: true,
|
|
1858
|
+
promptCaching: true,
|
|
1226
1859
|
input: {
|
|
1227
1860
|
tokens: 256e3,
|
|
1228
1861
|
text: true,
|
|
1229
1862
|
image: true,
|
|
1230
|
-
cpm: 3
|
|
1863
|
+
cpm: 3,
|
|
1864
|
+
cpmCached: 0.75
|
|
1231
1865
|
},
|
|
1232
1866
|
output: {
|
|
1233
1867
|
tokens: 32768,
|
|
@@ -1240,12 +1874,12 @@ var MODEL_REGISTRY = {
|
|
|
1240
1874
|
"grok-3-mini": {
|
|
1241
1875
|
name: "grok-3-mini",
|
|
1242
1876
|
provider: Vendor.Grok,
|
|
1243
|
-
description: "Lightweight, cost-efficient model
|
|
1877
|
+
description: "Lightweight, cost-efficient model with reasoning, 131K context",
|
|
1244
1878
|
isActive: true,
|
|
1245
1879
|
releaseDate: "2025-06-01",
|
|
1246
1880
|
knowledgeCutoff: "2024-11-01",
|
|
1247
1881
|
features: {
|
|
1248
|
-
reasoning:
|
|
1882
|
+
reasoning: true,
|
|
1249
1883
|
streaming: true,
|
|
1250
1884
|
structuredOutput: true,
|
|
1251
1885
|
functionCalling: true,
|
|
@@ -1255,12 +1889,13 @@ var MODEL_REGISTRY = {
|
|
|
1255
1889
|
vision: false,
|
|
1256
1890
|
audio: false,
|
|
1257
1891
|
video: false,
|
|
1258
|
-
batchAPI:
|
|
1259
|
-
promptCaching:
|
|
1892
|
+
batchAPI: true,
|
|
1893
|
+
promptCaching: true,
|
|
1260
1894
|
input: {
|
|
1261
1895
|
tokens: 131072,
|
|
1262
1896
|
text: true,
|
|
1263
|
-
cpm: 0.3
|
|
1897
|
+
cpm: 0.3,
|
|
1898
|
+
cpmCached: 0.07
|
|
1264
1899
|
},
|
|
1265
1900
|
output: {
|
|
1266
1901
|
tokens: 32768,
|
|
@@ -1287,12 +1922,13 @@ var MODEL_REGISTRY = {
|
|
|
1287
1922
|
vision: false,
|
|
1288
1923
|
audio: false,
|
|
1289
1924
|
video: false,
|
|
1290
|
-
batchAPI:
|
|
1291
|
-
promptCaching:
|
|
1925
|
+
batchAPI: true,
|
|
1926
|
+
promptCaching: true,
|
|
1292
1927
|
input: {
|
|
1293
1928
|
tokens: 131072,
|
|
1294
1929
|
text: true,
|
|
1295
|
-
cpm: 3
|
|
1930
|
+
cpm: 3,
|
|
1931
|
+
cpmCached: 0.75
|
|
1296
1932
|
},
|
|
1297
1933
|
output: {
|
|
1298
1934
|
tokens: 32768,
|
|
@@ -1301,11 +1937,11 @@ var MODEL_REGISTRY = {
|
|
|
1301
1937
|
}
|
|
1302
1938
|
}
|
|
1303
1939
|
},
|
|
1304
|
-
// Grok 2 Series (
|
|
1940
|
+
// Grok 2 Series (Legacy - not in current docs)
|
|
1305
1941
|
"grok-2-vision-1212": {
|
|
1306
1942
|
name: "grok-2-vision-1212",
|
|
1307
1943
|
provider: Vendor.Grok,
|
|
1308
|
-
description: "
|
|
1944
|
+
description: "Legacy vision model for image understanding, 32K context. Not in current xAI docs",
|
|
1309
1945
|
isActive: true,
|
|
1310
1946
|
releaseDate: "2024-12-12",
|
|
1311
1947
|
knowledgeCutoff: "2024-11-01",
|
|
@@ -1578,6 +2214,14 @@ var SERVICE_DEFINITIONS = [
|
|
|
1578
2214
|
baseURL: "https://aws.amazon.com",
|
|
1579
2215
|
docsURL: "https://docs.aws.amazon.com/"
|
|
1580
2216
|
},
|
|
2217
|
+
{
|
|
2218
|
+
id: "cloudflare",
|
|
2219
|
+
name: "Cloudflare",
|
|
2220
|
+
category: "cloud",
|
|
2221
|
+
urlPattern: /api\.cloudflare\.com/i,
|
|
2222
|
+
baseURL: "https://api.cloudflare.com/client/v4",
|
|
2223
|
+
docsURL: "https://developers.cloudflare.com/api/"
|
|
2224
|
+
},
|
|
1581
2225
|
// ============ Storage ============
|
|
1582
2226
|
{
|
|
1583
2227
|
id: "dropbox",
|
|
@@ -1621,6 +2265,14 @@ var SERVICE_DEFINITIONS = [
|
|
|
1621
2265
|
baseURL: "https://api.postmarkapp.com",
|
|
1622
2266
|
docsURL: "https://postmarkapp.com/developer"
|
|
1623
2267
|
},
|
|
2268
|
+
{
|
|
2269
|
+
id: "mailgun",
|
|
2270
|
+
name: "Mailgun",
|
|
2271
|
+
category: "email",
|
|
2272
|
+
urlPattern: /api\.mailgun\.net|api\.eu\.mailgun\.net/i,
|
|
2273
|
+
baseURL: "https://api.mailgun.net/v3",
|
|
2274
|
+
docsURL: "https://documentation.mailgun.com/docs/mailgun/api-reference/"
|
|
2275
|
+
},
|
|
1624
2276
|
// ============ Monitoring & Observability ============
|
|
1625
2277
|
{
|
|
1626
2278
|
id: "datadog",
|