@aeye/models 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (74) hide show
  1. package/ReplicateScrape.md +54 -0
  2. package/dist/scripts/codegen.d.ts +21 -0
  3. package/dist/scripts/codegen.d.ts.map +1 -0
  4. package/dist/scripts/codegen.js +102 -0
  5. package/dist/scripts/codegen.js.map +1 -0
  6. package/dist/scripts/scrape.d.ts +19 -0
  7. package/dist/scripts/scrape.d.ts.map +1 -0
  8. package/dist/scripts/scrape.js +146 -0
  9. package/dist/scripts/scrape.js.map +1 -0
  10. package/dist/scripts/scrapers/__tests__/aws.test.d.ts +8 -0
  11. package/dist/scripts/scrapers/__tests__/aws.test.d.ts.map +1 -0
  12. package/dist/scripts/scrapers/__tests__/aws.test.js +73 -0
  13. package/dist/scripts/scrapers/__tests__/aws.test.js.map +1 -0
  14. package/dist/scripts/scrapers/aws.d.ts +12 -0
  15. package/dist/scripts/scrapers/aws.d.ts.map +1 -0
  16. package/dist/scripts/scrapers/aws.js +314 -0
  17. package/dist/scripts/scrapers/aws.js.map +1 -0
  18. package/dist/scripts/scrapers/openai.d.ts +12 -0
  19. package/dist/scripts/scrapers/openai.d.ts.map +1 -0
  20. package/dist/scripts/scrapers/openai.js +490 -0
  21. package/dist/scripts/scrapers/openai.js.map +1 -0
  22. package/dist/scripts/scrapers/openrouter.d.ts +13 -0
  23. package/dist/scripts/scrapers/openrouter.d.ts.map +1 -0
  24. package/dist/scripts/scrapers/openrouter.js +156 -0
  25. package/dist/scripts/scrapers/openrouter.js.map +1 -0
  26. package/dist/scripts/scrapers/replicate.d.ts +12 -0
  27. package/dist/scripts/scrapers/replicate.d.ts.map +1 -0
  28. package/dist/scripts/scrapers/replicate.js +305 -0
  29. package/dist/scripts/scrapers/replicate.js.map +1 -0
  30. package/dist/src/index.d.ts +11 -0
  31. package/dist/src/index.d.ts.map +1 -0
  32. package/dist/src/index.js +11 -0
  33. package/dist/src/index.js.map +1 -0
  34. package/dist/src/models/aws.d.ts +11 -0
  35. package/dist/src/models/aws.d.ts.map +1 -0
  36. package/dist/src/models/aws.js +2632 -0
  37. package/dist/src/models/aws.js.map +1 -0
  38. package/dist/src/models/index.d.ts +15 -0
  39. package/dist/src/models/index.d.ts.map +1 -0
  40. package/dist/src/models/index.js +18 -0
  41. package/dist/src/models/index.js.map +1 -0
  42. package/dist/src/models/openai.d.ts +11 -0
  43. package/dist/src/models/openai.d.ts.map +1 -0
  44. package/dist/src/models/openai.js +2207 -0
  45. package/dist/src/models/openai.js.map +1 -0
  46. package/dist/src/models/openrouter.d.ts +11 -0
  47. package/dist/src/models/openrouter.d.ts.map +1 -0
  48. package/dist/src/models/openrouter.js +9786 -0
  49. package/dist/src/models/openrouter.js.map +1 -0
  50. package/dist/src/models/replicate.d.ts +11 -0
  51. package/dist/src/models/replicate.d.ts.map +1 -0
  52. package/dist/src/models/replicate.js +4106 -0
  53. package/dist/src/models/replicate.js.map +1 -0
  54. package/dist/src/transformers/index.d.ts +23 -0
  55. package/dist/src/transformers/index.d.ts.map +1 -0
  56. package/dist/src/transformers/index.js +24 -0
  57. package/dist/src/transformers/index.js.map +1 -0
  58. package/package.json +50 -0
  59. package/scripts/codegen.ts +117 -0
  60. package/scripts/scrape.ts +182 -0
  61. package/scripts/scrapers/__tests__/aws.test.ts +86 -0
  62. package/scripts/scrapers/aws.ts +370 -0
  63. package/scripts/scrapers/openai.ts +619 -0
  64. package/scripts/scrapers/openrouter.ts +214 -0
  65. package/scripts/scrapers/replicate.ts +448 -0
  66. package/scripts/tsconfig.json +24 -0
  67. package/src/index.ts +11 -0
  68. package/src/models/aws.ts +2634 -0
  69. package/src/models/index.ts +21 -0
  70. package/src/models/openai.ts +2209 -0
  71. package/src/models/openrouter.ts +9788 -0
  72. package/src/models/replicate.ts +4108 -0
  73. package/src/transformers/index.ts +26 -0
  74. package/tsconfig.json +14 -0
@@ -0,0 +1,2209 @@
1
+ /**
2
+ * Generated model data
3
+ *
4
+ * This file is auto-generated by the scraper scripts.
5
+ * Do not edit manually - your changes will be overwritten.
6
+ *
7
+ * To regenerate, run: npm run scrape
8
+ */
9
+
10
+ import type { ModelInfo } from '@aeye/ai';
11
+
12
+ export const openaiModels: ModelInfo[] = [{
13
+ id: 'gpt-4-0613',
14
+ provider: 'openai',
15
+ name: 'gpt-4-0613',
16
+ contextWindow: 8192,
17
+ maxOutputTokens: 8192,
18
+ tier: 'legacy',
19
+ capabilities: new Set(['chat', 'streaming']),
20
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
21
+ pricing: {
22
+ text: {
23
+ input: 30,
24
+ output: 60
25
+ }
26
+ },
27
+ metadata: {
28
+ knowledgeCutoff: "dec 01, 2023",
29
+ performance: "average",
30
+ speed: "medium"
31
+ }
32
+ }, {
33
+ id: 'gpt-4',
34
+ provider: 'openai',
35
+ name: 'gpt-4',
36
+ contextWindow: 8192,
37
+ maxOutputTokens: 8192,
38
+ tier: 'legacy',
39
+ capabilities: new Set(['chat', 'streaming']),
40
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
41
+ pricing: {
42
+ text: {
43
+ input: 30,
44
+ output: 60
45
+ }
46
+ },
47
+ metadata: {
48
+ knowledgeCutoff: "dec 01, 2023",
49
+ performance: "average",
50
+ speed: "medium"
51
+ }
52
+ }, {
53
+ id: 'gpt-3.5-turbo',
54
+ provider: 'openai',
55
+ name: 'gpt-3.5-turbo',
56
+ contextWindow: 16385,
57
+ maxOutputTokens: 4096,
58
+ tier: 'legacy',
59
+ capabilities: new Set(['chat']),
60
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
61
+ pricing: {
62
+ text: {
63
+ input: 0.5,
64
+ output: 1.5
65
+ }
66
+ },
67
+ metadata: {
68
+ knowledgeCutoff: "sep 01, 2021",
69
+ performance: "low",
70
+ speed: "slow"
71
+ }
72
+ }, {
73
+ id: 'gpt-realtime-mini',
74
+ provider: 'openai',
75
+ name: 'gpt-realtime-mini',
76
+ contextWindow: 32000,
77
+ maxOutputTokens: 4096,
78
+ tier: 'flagship',
79
+ capabilities: new Set(['vision', 'hearing', 'audio', 'chat', 'tools']),
80
+ pricing: {
81
+ text: {
82
+ input: 0.6,
83
+ output: 2.4,
84
+ cached: 0.06
85
+ },
86
+ audio: {
87
+ input: 10,
88
+ output: 20
89
+ },
90
+ image: {
91
+ input: 0.8
92
+ }
93
+ },
94
+ metadata: {
95
+ knowledgeCutoff: "oct 01, 2023",
96
+ intelligence: "higher",
97
+ speed: "very fast"
98
+ }
99
+ }, {
100
+ id: 'gpt-realtime-mini-2025-10-06',
101
+ provider: 'openai',
102
+ name: 'gpt-realtime-mini-2025-10-06',
103
+ contextWindow: 32000,
104
+ maxOutputTokens: 4096,
105
+ tier: 'flagship',
106
+ capabilities: new Set(['vision', 'hearing', 'audio', 'chat', 'tools']),
107
+ pricing: {
108
+ text: {
109
+ input: 0.6,
110
+ output: 2.4,
111
+ cached: 0.06
112
+ },
113
+ audio: {
114
+ input: 10,
115
+ output: 20
116
+ },
117
+ image: {
118
+ input: 0.8
119
+ }
120
+ },
121
+ metadata: {
122
+ knowledgeCutoff: "oct 01, 2023",
123
+ intelligence: "higher",
124
+ speed: "very fast"
125
+ }
126
+ }, {
127
+ id: 'sora-2',
128
+ provider: 'openai',
129
+ name: 'sora-2',
130
+ contextWindow: 0,
131
+ tier: 'flagship',
132
+ capabilities: new Set(['chat', 'vision', 'audio']),
133
+ pricing: {},
134
+ metadata: {
135
+ intelligence: "higher",
136
+ speed: "slow"
137
+ }
138
+ }, {
139
+ id: 'sora-2-pro',
140
+ provider: 'openai',
141
+ name: 'sora-2-pro',
142
+ contextWindow: 0,
143
+ tier: 'flagship',
144
+ capabilities: new Set(['chat', 'vision', 'audio']),
145
+ pricing: {},
146
+ metadata: {
147
+ intelligence: "highest",
148
+ speed: "slow"
149
+ }
150
+ }, {
151
+ id: 'davinci-002',
152
+ provider: 'openai',
153
+ name: 'davinci-002',
154
+ contextWindow: 0,
155
+ maxOutputTokens: 16384,
156
+ tier: 'legacy',
157
+ capabilities: new Set(['chat']),
158
+ pricing: {
159
+ text: {
160
+ input: 2,
161
+ output: 2
162
+ }
163
+ },
164
+ metadata: {
165
+ knowledgeCutoff: "sep 01, 2021",
166
+ performance: "low",
167
+ speed: "medium"
168
+ }
169
+ }, {
170
+ id: 'babbage-002',
171
+ provider: 'openai',
172
+ name: 'babbage-002',
173
+ contextWindow: 0,
174
+ maxOutputTokens: 16384,
175
+ tier: 'legacy',
176
+ capabilities: new Set(['chat']),
177
+ pricing: {
178
+ text: {
179
+ input: 0.4,
180
+ output: 0.4
181
+ }
182
+ },
183
+ metadata: {
184
+ knowledgeCutoff: "sep 01, 2021",
185
+ performance: "low",
186
+ speed: "medium"
187
+ }
188
+ }, {
189
+ id: 'gpt-3.5-turbo-instruct',
190
+ provider: 'openai',
191
+ name: 'gpt-3.5-turbo-instruct',
192
+ contextWindow: 4096,
193
+ maxOutputTokens: 4096,
194
+ tier: 'legacy',
195
+ capabilities: new Set(['chat']),
196
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
197
+ pricing: {
198
+ text: {
199
+ input: 1.5,
200
+ output: 2
201
+ }
202
+ },
203
+ metadata: {
204
+ knowledgeCutoff: "sep 01, 2021",
205
+ performance: "low",
206
+ speed: "slow"
207
+ }
208
+ }, {
209
+ id: 'gpt-3.5-turbo-instruct-0914',
210
+ provider: 'openai',
211
+ name: 'gpt-3.5-turbo-instruct-0914',
212
+ contextWindow: 4096,
213
+ maxOutputTokens: 4096,
214
+ tier: 'legacy',
215
+ capabilities: new Set(['chat']),
216
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
217
+ pricing: {
218
+ text: {
219
+ input: 1.5,
220
+ output: 2
221
+ }
222
+ },
223
+ metadata: {
224
+ knowledgeCutoff: "sep 01, 2021",
225
+ performance: "low",
226
+ speed: "slow"
227
+ }
228
+ }, {
229
+ id: 'dall-e-3',
230
+ provider: 'openai',
231
+ name: 'dall-e-3',
232
+ contextWindow: 0,
233
+ tier: 'legacy',
234
+ capabilities: new Set(['image', 'chat']),
235
+ supportedParameters: new Set([]),
236
+ pricing: {
237
+ image: {
238
+ output: [
239
+ {
240
+ quality: "standard",
241
+ sizes: [
242
+ {
243
+ width: 1024,
244
+ height: 1024,
245
+ cost: 0.04
246
+ },
247
+ {
248
+ width: 1024,
249
+ height: 1536,
250
+ cost: 0.08
251
+ },
252
+ {
253
+ width: 1536,
254
+ height: 1024,
255
+ cost: 0.08
256
+ }
257
+ ]
258
+ },
259
+ {
260
+ quality: "hd",
261
+ sizes: [
262
+ {
263
+ width: 1024,
264
+ height: 1024,
265
+ cost: 0.08
266
+ },
267
+ {
268
+ width: 1024,
269
+ height: 1536,
270
+ cost: 0.12
271
+ },
272
+ {
273
+ width: 1536,
274
+ height: 1024,
275
+ cost: 0.12
276
+ }
277
+ ]
278
+ }
279
+ ]
280
+ }
281
+ },
282
+ metadata: {
283
+ intelligence: "high",
284
+ speed: "slow"
285
+ }
286
+ }, {
287
+ id: 'dall-e-2',
288
+ provider: 'openai',
289
+ name: 'dall-e-2',
290
+ contextWindow: 0,
291
+ tier: 'legacy',
292
+ capabilities: new Set(['image', 'chat']),
293
+ supportedParameters: new Set(['imageStyle', 'imageMultiple']),
294
+ pricing: {
295
+ image: {
296
+ output: [
297
+ {
298
+ quality: "standard",
299
+ sizes: [
300
+ {
301
+ width: 1024,
302
+ height: 1024,
303
+ cost: 0.016
304
+ },
305
+ {
306
+ width: 1024,
307
+ height: 1536,
308
+ cost: 0.018
309
+ },
310
+ {
311
+ width: 1536,
312
+ height: 1024,
313
+ cost: 0.02
314
+ }
315
+ ]
316
+ }
317
+ ]
318
+ }
319
+ },
320
+ metadata: {
321
+ intelligence: "low",
322
+ speed: "slow"
323
+ }
324
+ }, {
325
+ id: 'gpt-3.5-turbo-1106',
326
+ provider: 'openai',
327
+ name: 'gpt-3.5-turbo-1106',
328
+ contextWindow: 16385,
329
+ maxOutputTokens: 4096,
330
+ tier: 'legacy',
331
+ capabilities: new Set(['chat']),
332
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
333
+ pricing: {
334
+ text: {
335
+ input: 1,
336
+ output: 2
337
+ }
338
+ },
339
+ metadata: {
340
+ knowledgeCutoff: "sep 01, 2021",
341
+ performance: "low",
342
+ speed: "slow"
343
+ }
344
+ }, {
345
+ id: 'tts-1-hd',
346
+ provider: 'openai',
347
+ name: 'tts-1-hd',
348
+ contextWindow: 0,
349
+ tier: 'legacy',
350
+ capabilities: new Set(['audio', 'chat']),
351
+ supportedParameters: new Set([]),
352
+ pricing: {
353
+ audio: {
354
+ output: 30
355
+ }
356
+ },
357
+ metadata: {
358
+ intelligence: "high",
359
+ speed: "medium"
360
+ }
361
+ }, {
362
+ id: 'tts-1-1106',
363
+ provider: 'openai',
364
+ name: 'tts-1-1106',
365
+ contextWindow: 0,
366
+ tier: 'efficient',
367
+ capabilities: new Set(['audio', 'chat']),
368
+ supportedParameters: new Set([]),
369
+ pricing: {
370
+ audio: {
371
+ output: 15
372
+ }
373
+ },
374
+ metadata: {
375
+ intelligence: "average",
376
+ speed: "fast"
377
+ }
378
+ }, {
379
+ id: 'tts-1-hd-1106',
380
+ provider: 'openai',
381
+ name: 'tts-1-hd-1106',
382
+ contextWindow: 0,
383
+ tier: 'legacy',
384
+ capabilities: new Set(['audio', 'chat']),
385
+ supportedParameters: new Set([]),
386
+ pricing: {
387
+ audio: {
388
+ output: 30
389
+ }
390
+ },
391
+ metadata: {
392
+ intelligence: "high",
393
+ speed: "medium"
394
+ }
395
+ }, {
396
+ id: 'text-embedding-3-small',
397
+ provider: 'openai',
398
+ name: 'text-embedding-3-small',
399
+ contextWindow: 0,
400
+ tier: 'legacy',
401
+ capabilities: new Set(['embedding', 'chat']),
402
+ supportedParameters: new Set(['embeddingDimensions']),
403
+ pricing: {
404
+ embeddings: {
405
+ cost: 0.02
406
+ }
407
+ },
408
+ metadata: {
409
+ intelligence: "average",
410
+ speed: "medium"
411
+ }
412
+ }, {
413
+ id: 'text-embedding-3-large',
414
+ provider: 'openai',
415
+ name: 'text-embedding-3-large',
416
+ contextWindow: 0,
417
+ tier: 'legacy',
418
+ capabilities: new Set(['embedding', 'chat']),
419
+ supportedParameters: new Set(['embeddingDimensions']),
420
+ pricing: {
421
+ embeddings: {
422
+ cost: 0.13
423
+ }
424
+ },
425
+ metadata: {
426
+ intelligence: "high",
427
+ speed: "slow"
428
+ }
429
+ }, {
430
+ id: 'gpt-4-turbo-preview',
431
+ provider: 'openai',
432
+ name: 'gpt-4-turbo-preview',
433
+ contextWindow: 128000,
434
+ maxOutputTokens: 4096,
435
+ tier: 'legacy',
436
+ capabilities: new Set(['chat']),
437
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
438
+ pricing: {
439
+ text: {
440
+ input: 10,
441
+ output: 30
442
+ }
443
+ },
444
+ metadata: {
445
+ knowledgeCutoff: "dec 01, 2023",
446
+ performance: "average",
447
+ speed: "medium"
448
+ }
449
+ }, {
450
+ id: 'gpt-3.5-turbo-0125',
451
+ provider: 'openai',
452
+ name: 'gpt-3.5-turbo-0125',
453
+ contextWindow: 16385,
454
+ maxOutputTokens: 4096,
455
+ tier: 'legacy',
456
+ capabilities: new Set(['chat']),
457
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
458
+ pricing: {
459
+ text: {
460
+ input: 0.5,
461
+ output: 1.5
462
+ }
463
+ },
464
+ metadata: {
465
+ knowledgeCutoff: "sep 01, 2021",
466
+ performance: "low",
467
+ speed: "slow"
468
+ }
469
+ }, {
470
+ id: 'gpt-4-turbo',
471
+ provider: 'openai',
472
+ name: 'gpt-4-turbo',
473
+ contextWindow: 128000,
474
+ maxOutputTokens: 4096,
475
+ tier: 'legacy',
476
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools']),
477
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
478
+ pricing: {
479
+ text: {
480
+ input: 10,
481
+ output: 30
482
+ }
483
+ },
484
+ metadata: {
485
+ knowledgeCutoff: "dec 01, 2023",
486
+ performance: "average",
487
+ speed: "medium"
488
+ }
489
+ }, {
490
+ id: 'gpt-4-turbo-2024-04-09',
491
+ provider: 'openai',
492
+ name: 'gpt-4-turbo-2024-04-09',
493
+ contextWindow: 128000,
494
+ maxOutputTokens: 4096,
495
+ tier: 'legacy',
496
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools']),
497
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
498
+ pricing: {
499
+ text: {
500
+ input: 10,
501
+ output: 30
502
+ }
503
+ },
504
+ metadata: {
505
+ knowledgeCutoff: "dec 01, 2023",
506
+ performance: "average",
507
+ speed: "medium"
508
+ }
509
+ }, {
510
+ id: 'gpt-4o',
511
+ provider: 'openai',
512
+ name: 'gpt-4o',
513
+ contextWindow: 128000,
514
+ maxOutputTokens: 16384,
515
+ tier: 'legacy',
516
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
517
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
518
+ pricing: {
519
+ text: {
520
+ input: 2.5,
521
+ output: 10,
522
+ cached: 1.25
523
+ }
524
+ },
525
+ metadata: {
526
+ knowledgeCutoff: "oct 01, 2023",
527
+ performance: "high",
528
+ speed: "medium"
529
+ }
530
+ }, {
531
+ id: 'gpt-4o-2024-05-13',
532
+ provider: 'openai',
533
+ name: 'gpt-4o-2024-05-13',
534
+ contextWindow: 128000,
535
+ maxOutputTokens: 4096,
536
+ tier: 'legacy',
537
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools']),
538
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
539
+ pricing: {
540
+ text: {
541
+ input: 5,
542
+ output: 15
543
+ }
544
+ },
545
+ metadata: {
546
+ knowledgeCutoff: "oct 01, 2023",
547
+ performance: "high",
548
+ speed: "medium"
549
+ }
550
+ }, {
551
+ id: 'gpt-4o-mini-2024-07-18',
552
+ provider: 'openai',
553
+ name: 'gpt-4o-mini-2024-07-18',
554
+ contextWindow: 128000,
555
+ maxOutputTokens: 16384,
556
+ tier: 'efficient',
557
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
558
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
559
+ pricing: {
560
+ text: {
561
+ input: 0.15,
562
+ output: 0.6,
563
+ cached: 0.075
564
+ }
565
+ },
566
+ metadata: {
567
+ knowledgeCutoff: "oct 01, 2023",
568
+ performance: "average",
569
+ speed: "fast"
570
+ }
571
+ }, {
572
+ id: 'gpt-4o-mini',
573
+ provider: 'openai',
574
+ name: 'gpt-4o-mini',
575
+ contextWindow: 128000,
576
+ maxOutputTokens: 16384,
577
+ tier: 'efficient',
578
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
579
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
580
+ pricing: {
581
+ text: {
582
+ input: 0.15,
583
+ output: 0.6,
584
+ cached: 0.075
585
+ }
586
+ },
587
+ metadata: {
588
+ knowledgeCutoff: "oct 01, 2023",
589
+ performance: "average",
590
+ speed: "fast"
591
+ }
592
+ }, {
593
+ id: 'gpt-4o-2024-08-06',
594
+ provider: 'openai',
595
+ name: 'gpt-4o-2024-08-06',
596
+ contextWindow: 128000,
597
+ maxOutputTokens: 16384,
598
+ tier: 'legacy',
599
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
600
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
601
+ pricing: {
602
+ text: {
603
+ input: 2.5,
604
+ output: 10,
605
+ cached: 1.25
606
+ }
607
+ },
608
+ metadata: {
609
+ knowledgeCutoff: "oct 01, 2023",
610
+ performance: "high",
611
+ speed: "medium"
612
+ }
613
+ }, {
614
+ id: 'chatgpt-4o-latest',
615
+ provider: 'openai',
616
+ name: 'chatgpt-4o-latest',
617
+ contextWindow: 128000,
618
+ maxOutputTokens: 16384,
619
+ tier: 'legacy',
620
+ capabilities: new Set(['chat', 'vision', 'streaming']),
621
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
622
+ pricing: {
623
+ text: {
624
+ input: 5,
625
+ output: 15
626
+ }
627
+ },
628
+ metadata: {
629
+ knowledgeCutoff: "oct 01, 2023",
630
+ performance: "high",
631
+ speed: "medium"
632
+ }
633
+ }, {
634
+ id: 'o1-mini-2024-09-12',
635
+ provider: 'openai',
636
+ name: 'o1-mini-2024-09-12',
637
+ contextWindow: 128000,
638
+ maxOutputTokens: 65536,
639
+ tier: 'legacy',
640
+ capabilities: new Set(['chat', 'streaming', 'reasoning']),
641
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'stop']),
642
+ pricing: {
643
+ text: {
644
+ input: 1.1,
645
+ output: 4.4,
646
+ cached: 0.55
647
+ }
648
+ },
649
+ metadata: {
650
+ knowledgeCutoff: "oct 01, 2023",
651
+ reasoning: "high",
652
+ speed: "slow"
653
+ }
654
+ }, {
655
+ id: 'o1-mini',
656
+ provider: 'openai',
657
+ name: 'o1-mini',
658
+ contextWindow: 128000,
659
+ maxOutputTokens: 65536,
660
+ tier: 'legacy',
661
+ capabilities: new Set(['chat', 'streaming', 'reasoning']),
662
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'stop']),
663
+ pricing: {
664
+ text: {
665
+ input: 1.1,
666
+ output: 4.4,
667
+ cached: 0.55
668
+ }
669
+ },
670
+ metadata: {
671
+ knowledgeCutoff: "oct 01, 2023",
672
+ reasoning: "high",
673
+ speed: "slow"
674
+ }
675
+ }, {
676
+ id: 'gpt-4o-realtime-preview-2024-10-01',
677
+ provider: 'openai',
678
+ name: 'gpt-4o-realtime-preview-2024-10-01',
679
+ contextWindow: 16000,
680
+ maxOutputTokens: 4096,
681
+ tier: 'efficient',
682
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
683
+ pricing: {
684
+ text: {
685
+ input: 5,
686
+ output: 20,
687
+ cached: 2.5
688
+ },
689
+ audio: {
690
+ input: 100,
691
+ output: 200
692
+ }
693
+ },
694
+ metadata: {
695
+ knowledgeCutoff: "oct 01, 2023",
696
+ intelligence: "average",
697
+ speed: "fast"
698
+ }
699
+ }, {
700
+ id: 'gpt-4o-audio-preview-2024-10-01',
701
+ provider: 'openai',
702
+ name: 'gpt-4o-audio-preview-2024-10-01',
703
+ contextWindow: 128000,
704
+ maxOutputTokens: 16384,
705
+ tier: 'legacy',
706
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
707
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
708
+ pricing: {
709
+ text: {
710
+ input: 2.5,
711
+ output: 10
712
+ },
713
+ audio: {
714
+ input: 100,
715
+ output: 200
716
+ }
717
+ },
718
+ metadata: {
719
+ knowledgeCutoff: "oct 01, 2023",
720
+ performance: "high",
721
+ speed: "medium"
722
+ }
723
+ }, {
724
+ id: 'gpt-4o-audio-preview',
725
+ provider: 'openai',
726
+ name: 'gpt-4o-audio-preview',
727
+ contextWindow: 128000,
728
+ maxOutputTokens: 16384,
729
+ tier: 'legacy',
730
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
731
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
732
+ pricing: {
733
+ text: {
734
+ input: 2.5,
735
+ output: 10
736
+ },
737
+ audio: {
738
+ input: 40,
739
+ output: 80
740
+ }
741
+ },
742
+ metadata: {
743
+ knowledgeCutoff: "oct 01, 2023",
744
+ performance: "high",
745
+ speed: "medium"
746
+ }
747
+ }, {
748
+ id: 'gpt-4o-realtime-preview',
749
+ provider: 'openai',
750
+ name: 'gpt-4o-realtime-preview',
751
+ contextWindow: 32000,
752
+ maxOutputTokens: 4096,
753
+ tier: 'efficient',
754
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
755
+ pricing: {
756
+ text: {
757
+ input: 5,
758
+ output: 20,
759
+ cached: 2.5
760
+ },
761
+ audio: {
762
+ input: 40,
763
+ output: 80
764
+ }
765
+ },
766
+ metadata: {
767
+ knowledgeCutoff: "oct 01, 2023",
768
+ intelligence: "high",
769
+ speed: "fast"
770
+ }
771
+ }, {
772
+ id: 'omni-moderation-latest',
773
+ provider: 'openai',
774
+ name: 'omni-moderation-latest',
775
+ contextWindow: 0,
776
+ tier: 'legacy',
777
+ capabilities: new Set(['chat', 'vision']),
778
+ pricing: {},
779
+ metadata: {
780
+ intelligence: "high",
781
+ speed: "medium"
782
+ }
783
+ }, {
784
+ id: 'gpt-4o-realtime-preview-2024-12-17',
785
+ provider: 'openai',
786
+ name: 'gpt-4o-realtime-preview-2024-12-17',
787
+ contextWindow: 16000,
788
+ maxOutputTokens: 4096,
789
+ tier: 'efficient',
790
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
791
+ pricing: {
792
+ text: {
793
+ input: 5,
794
+ output: 20,
795
+ cached: 2.5
796
+ },
797
+ audio: {
798
+ input: 40,
799
+ output: 80
800
+ }
801
+ },
802
+ metadata: {
803
+ knowledgeCutoff: "oct 01, 2023",
804
+ intelligence: "high",
805
+ speed: "fast"
806
+ }
807
+ }, {
808
+ id: 'gpt-4o-audio-preview-2024-12-17',
809
+ provider: 'openai',
810
+ name: 'gpt-4o-audio-preview-2024-12-17',
811
+ contextWindow: 128000,
812
+ maxOutputTokens: 16384,
813
+ tier: 'legacy',
814
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
815
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
816
+ pricing: {
817
+ text: {
818
+ input: 2.5,
819
+ output: 10
820
+ },
821
+ audio: {
822
+ input: 40,
823
+ output: 80
824
+ }
825
+ },
826
+ metadata: {
827
+ knowledgeCutoff: "oct 01, 2023",
828
+ performance: "high",
829
+ speed: "medium"
830
+ }
831
+ }, {
832
+ id: 'gpt-4o-mini-realtime-preview-2024-12-17',
833
+ provider: 'openai',
834
+ name: 'gpt-4o-mini-realtime-preview-2024-12-17',
835
+ contextWindow: 16000,
836
+ maxOutputTokens: 4096,
837
+ tier: 'efficient',
838
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
839
+ pricing: {
840
+ text: {
841
+ input: 0.6,
842
+ output: 2.4,
843
+ cached: 0.3
844
+ },
845
+ audio: {
846
+ input: 10,
847
+ output: 20
848
+ }
849
+ },
850
+ metadata: {
851
+ knowledgeCutoff: "oct 01, 2023",
852
+ intelligence: "average",
853
+ speed: "very fast"
854
+ }
855
+ }, {
856
+ id: 'gpt-4o-mini-audio-preview-2024-12-17',
857
+ provider: 'openai',
858
+ name: 'gpt-4o-mini-audio-preview-2024-12-17',
859
+ contextWindow: 128000,
860
+ maxOutputTokens: 16384,
861
+ tier: 'efficient',
862
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
863
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
864
+ pricing: {
865
+ text: {
866
+ input: 0.15,
867
+ output: 0.6
868
+ },
869
+ audio: {
870
+ input: 10,
871
+ output: 20
872
+ }
873
+ },
874
+ metadata: {
875
+ knowledgeCutoff: "oct 01, 2023",
876
+ performance: "average",
877
+ speed: "fast"
878
+ }
879
+ }, {
880
+ id: 'o1-2024-12-17',
881
+ provider: 'openai',
882
+ name: 'o1-2024-12-17',
883
+ contextWindow: 200000,
884
+ maxOutputTokens: 100000,
885
+ tier: 'flagship',
886
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
887
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
888
+ pricing: {
889
+ text: {
890
+ input: 15,
891
+ output: 60,
892
+ cached: 7.5
893
+ }
894
+ },
895
+ metadata: {
896
+ knowledgeCutoff: "oct 01, 2023",
897
+ reasoning: "higher",
898
+ speed: "slow"
899
+ }
900
+ }, {
901
+ id: 'o1',
902
+ provider: 'openai',
903
+ name: 'o1',
904
+ contextWindow: 200000,
905
+ maxOutputTokens: 100000,
906
+ tier: 'flagship',
907
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
908
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
909
+ pricing: {
910
+ text: {
911
+ input: 15,
912
+ output: 60,
913
+ cached: 7.5
914
+ }
915
+ },
916
+ metadata: {
917
+ knowledgeCutoff: "oct 01, 2023",
918
+ reasoning: "higher",
919
+ speed: "slow"
920
+ }
921
+ }, {
922
+ id: 'gpt-4o-mini-realtime-preview',
923
+ provider: 'openai',
924
+ name: 'gpt-4o-mini-realtime-preview',
925
+ contextWindow: 16000,
926
+ maxOutputTokens: 4096,
927
+ tier: 'efficient',
928
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
929
+ pricing: {
930
+ text: {
931
+ input: 0.6,
932
+ output: 2.4,
933
+ cached: 0.3
934
+ },
935
+ audio: {
936
+ input: 10,
937
+ output: 20
938
+ }
939
+ },
940
+ metadata: {
941
+ knowledgeCutoff: "oct 01, 2023",
942
+ intelligence: "average",
943
+ speed: "very fast"
944
+ }
945
+ }, {
946
+ id: 'gpt-4o-mini-audio-preview',
947
+ provider: 'openai',
948
+ name: 'gpt-4o-mini-audio-preview',
949
+ contextWindow: 128000,
950
+ maxOutputTokens: 16384,
951
+ tier: 'efficient',
952
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
953
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
954
+ pricing: {
955
+ text: {
956
+ input: 0.15,
957
+ output: 0.6
958
+ },
959
+ audio: {
960
+ input: 10,
961
+ output: 20
962
+ }
963
+ },
964
+ metadata: {
965
+ knowledgeCutoff: "oct 01, 2023",
966
+ performance: "average",
967
+ speed: "fast"
968
+ }
969
+ }, {
970
+ id: 'o3-mini',
971
+ provider: 'openai',
972
+ name: 'o3-mini',
973
+ contextWindow: 200000,
974
+ maxOutputTokens: 100000,
975
+ tier: 'flagship',
976
+ capabilities: new Set(['chat', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
977
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
978
+ pricing: {
979
+ text: {
980
+ input: 1.1,
981
+ output: 4.4,
982
+ cached: 0.55
983
+ }
984
+ },
985
+ metadata: {
986
+ knowledgeCutoff: "oct 01, 2023",
987
+ reasoning: "higher",
988
+ speed: "medium"
989
+ }
990
+ }, {
991
+ id: 'o3-mini-2025-01-31',
992
+ provider: 'openai',
993
+ name: 'o3-mini-2025-01-31',
994
+ contextWindow: 200000,
995
+ maxOutputTokens: 100000,
996
+ tier: 'flagship',
997
+ capabilities: new Set(['chat', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
998
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
999
+ pricing: {
1000
+ text: {
1001
+ input: 1.1,
1002
+ output: 4.4,
1003
+ cached: 0.55
1004
+ }
1005
+ },
1006
+ metadata: {
1007
+ knowledgeCutoff: "oct 01, 2023",
1008
+ reasoning: "higher",
1009
+ speed: "medium"
1010
+ }
1011
+ }, {
1012
+ id: 'gpt-4o-2024-11-20',
1013
+ provider: 'openai',
1014
+ name: 'gpt-4o-2024-11-20',
1015
+ contextWindow: 128000,
1016
+ maxOutputTokens: 16384,
1017
+ tier: 'legacy',
1018
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1019
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1020
+ pricing: {
1021
+ text: {
1022
+ input: 2.5,
1023
+ output: 10,
1024
+ cached: 1.25
1025
+ }
1026
+ },
1027
+ metadata: {
1028
+ knowledgeCutoff: "oct 01, 2023",
1029
+ performance: "high",
1030
+ speed: "medium"
1031
+ }
1032
+ }, {
1033
+ id: 'gpt-4o-search-preview-2025-03-11',
1034
+ provider: 'openai',
1035
+ name: 'gpt-4o-search-preview-2025-03-11',
1036
+ contextWindow: 128000,
1037
+ maxOutputTokens: 16384,
1038
+ tier: 'legacy',
1039
+ capabilities: new Set(['chat', 'streaming', 'structured', 'json']),
1040
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'responseFormat', 'structuredOutput', 'stop']),
1041
+ pricing: {
1042
+ text: {
1043
+ input: 2.5,
1044
+ output: 10
1045
+ }
1046
+ },
1047
+ metadata: {
1048
+ knowledgeCutoff: "oct 01, 2023",
1049
+ intelligence: "high",
1050
+ speed: "medium"
1051
+ }
1052
+ }, {
1053
+ id: 'gpt-4o-search-preview',
1054
+ provider: 'openai',
1055
+ name: 'gpt-4o-search-preview',
1056
+ contextWindow: 128000,
1057
+ maxOutputTokens: 16384,
1058
+ tier: 'legacy',
1059
+ capabilities: new Set(['chat', 'streaming', 'structured', 'json']),
1060
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'responseFormat', 'structuredOutput', 'stop']),
1061
+ pricing: {
1062
+ text: {
1063
+ input: 2.5,
1064
+ output: 10
1065
+ }
1066
+ },
1067
+ metadata: {
1068
+ knowledgeCutoff: "oct 01, 2023",
1069
+ intelligence: "high",
1070
+ speed: "medium"
1071
+ }
1072
+ }, {
1073
+ id: 'gpt-4o-mini-search-preview-2025-03-11',
1074
+ provider: 'openai',
1075
+ name: 'gpt-4o-mini-search-preview-2025-03-11',
1076
+ contextWindow: 128000,
1077
+ maxOutputTokens: 16384,
1078
+ tier: 'efficient',
1079
+ capabilities: new Set(['chat', 'streaming', 'structured', 'json']),
1080
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'responseFormat', 'structuredOutput', 'stop']),
1081
+ pricing: {
1082
+ text: {
1083
+ input: 0.15,
1084
+ output: 0.6
1085
+ }
1086
+ },
1087
+ metadata: {
1088
+ knowledgeCutoff: "oct 01, 2023",
1089
+ intelligence: "average",
1090
+ speed: "fast"
1091
+ }
1092
+ }, {
1093
+ id: 'gpt-4o-mini-search-preview',
1094
+ provider: 'openai',
1095
+ name: 'gpt-4o-mini-search-preview',
1096
+ contextWindow: 128000,
1097
+ maxOutputTokens: 16384,
1098
+ tier: 'efficient',
1099
+ capabilities: new Set(['chat', 'streaming', 'structured', 'json']),
1100
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'responseFormat', 'structuredOutput', 'stop']),
1101
+ pricing: {
1102
+ text: {
1103
+ input: 0.15,
1104
+ output: 0.6
1105
+ }
1106
+ },
1107
+ metadata: {
1108
+ knowledgeCutoff: "oct 01, 2023",
1109
+ intelligence: "average",
1110
+ speed: "fast"
1111
+ }
1112
+ }, {
1113
+ id: 'gpt-4o-transcribe',
1114
+ provider: 'openai',
1115
+ name: 'gpt-4o-transcribe',
1116
+ contextWindow: 16000,
1117
+ maxOutputTokens: 2000,
1118
+ tier: 'flagship',
1119
+ capabilities: new Set(['hearing', 'chat']),
1120
+ supportedParameters: new Set(['transcribePrompt', 'transcribeStream']),
1121
+ pricing: {
1122
+ text: {
1123
+ input: 2.5,
1124
+ output: 10
1125
+ },
1126
+ audio: {
1127
+ input: 6
1128
+ }
1129
+ },
1130
+ metadata: {
1131
+ knowledgeCutoff: "jun 01, 2024",
1132
+ intelligence: "higher",
1133
+ speed: "medium"
1134
+ }
1135
+ }, {
1136
+ id: 'gpt-4o-mini-transcribe',
1137
+ provider: 'openai',
1138
+ name: 'gpt-4o-mini-transcribe',
1139
+ contextWindow: 16000,
1140
+ maxOutputTokens: 2000,
1141
+ tier: 'efficient',
1142
+ capabilities: new Set(['hearing', 'chat']),
1143
+ supportedParameters: new Set(['transcribePrompt', 'transcribeStream']),
1144
+ pricing: {
1145
+ text: {
1146
+ input: 1.25,
1147
+ output: 5
1148
+ },
1149
+ audio: {
1150
+ input: 3
1151
+ }
1152
+ },
1153
+ metadata: {
1154
+ knowledgeCutoff: "jun 01, 2024",
1155
+ intelligence: "high",
1156
+ speed: "fast"
1157
+ }
1158
+ }, {
1159
+ id: 'o1-pro-2025-03-19',
1160
+ provider: 'openai',
1161
+ name: 'o1-pro-2025-03-19',
1162
+ contextWindow: 200000,
1163
+ maxOutputTokens: 100000,
1164
+ tier: 'flagship',
1165
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
1166
+ pricing: {
1167
+ text: {
1168
+ input: 150,
1169
+ output: 600
1170
+ }
1171
+ },
1172
+ metadata: {
1173
+ knowledgeCutoff: "oct 01, 2023",
1174
+ reasoning: "higher",
1175
+ speed: "slow"
1176
+ }
1177
+ }, {
1178
+ id: 'o1-pro',
1179
+ provider: 'openai',
1180
+ name: 'o1-pro',
1181
+ contextWindow: 200000,
1182
+ maxOutputTokens: 100000,
1183
+ tier: 'flagship',
1184
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
1185
+ pricing: {
1186
+ text: {
1187
+ input: 150,
1188
+ output: 600
1189
+ }
1190
+ },
1191
+ metadata: {
1192
+ knowledgeCutoff: "oct 01, 2023",
1193
+ reasoning: "higher",
1194
+ speed: "slow"
1195
+ }
1196
+ }, {
1197
+ id: 'gpt-4o-mini-tts',
1198
+ provider: 'openai',
1199
+ name: 'gpt-4o-mini-tts',
1200
+ contextWindow: 0,
1201
+ tier: 'flagship',
1202
+ capabilities: new Set(['audio', 'chat']),
1203
+ supportedParameters: new Set(['speechInstructions']),
1204
+ pricing: {
1205
+ text: {
1206
+ input: 0.6
1207
+ },
1208
+ audio: {
1209
+ output: 12
1210
+ }
1211
+ },
1212
+ metadata: {
1213
+ intelligence: "higher",
1214
+ speed: "fast"
1215
+ }
1216
+ }, {
1217
+ id: 'o3-2025-04-16',
1218
+ provider: 'openai',
1219
+ name: 'o3-2025-04-16',
1220
+ contextWindow: 200000,
1221
+ maxOutputTokens: 100000,
1222
+ tier: 'flagship',
1223
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1224
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
1225
+ pricing: {
1226
+ text: {
1227
+ input: 1,
1228
+ output: 4,
1229
+ cached: 0.25
1230
+ }
1231
+ },
1232
+ metadata: {
1233
+ knowledgeCutoff: "jun 01, 2024",
1234
+ reasoning: "highest",
1235
+ speed: "slow"
1236
+ }
1237
+ }, {
1238
+ id: 'o4-mini-2025-04-16',
1239
+ provider: 'openai',
1240
+ name: 'o4-mini-2025-04-16',
1241
+ contextWindow: 200000,
1242
+ maxOutputTokens: 100000,
1243
+ tier: 'flagship',
1244
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1245
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
1246
+ pricing: {
1247
+ text: {
1248
+ input: 1.1,
1249
+ output: 4.4,
1250
+ cached: 0.275
1251
+ }
1252
+ },
1253
+ metadata: {
1254
+ knowledgeCutoff: "jun 01, 2024",
1255
+ reasoning: "higher",
1256
+ speed: "medium"
1257
+ }
1258
+ }, {
1259
+ id: 'o3',
1260
+ provider: 'openai',
1261
+ name: 'o3',
1262
+ contextWindow: 200000,
1263
+ maxOutputTokens: 100000,
1264
+ tier: 'flagship',
1265
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1266
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
1267
+ pricing: {
1268
+ text: {
1269
+ input: 1,
1270
+ output: 4,
1271
+ cached: 0.25
1272
+ }
1273
+ },
1274
+ metadata: {
1275
+ knowledgeCutoff: "jun 01, 2024",
1276
+ reasoning: "highest",
1277
+ speed: "slow"
1278
+ }
1279
+ }, {
1280
+ id: 'o4-mini',
1281
+ provider: 'openai',
1282
+ name: 'o4-mini',
1283
+ contextWindow: 200000,
1284
+ maxOutputTokens: 100000,
1285
+ tier: 'flagship',
1286
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1287
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
1288
+ pricing: {
1289
+ text: {
1290
+ input: 1.1,
1291
+ output: 4.4,
1292
+ cached: 0.275
1293
+ }
1294
+ },
1295
+ metadata: {
1296
+ knowledgeCutoff: "jun 01, 2024",
1297
+ reasoning: "higher",
1298
+ speed: "medium"
1299
+ }
1300
+ }, {
1301
+ id: 'gpt-4.1-2025-04-14',
1302
+ provider: 'openai',
1303
+ name: 'gpt-4.1-2025-04-14',
1304
+ contextWindow: 1047576,
1305
+ maxOutputTokens: 32768,
1306
+ tier: 'flagship',
1307
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1308
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1309
+ pricing: {
1310
+ text: {
1311
+ input: 2,
1312
+ output: 8,
1313
+ cached: 0.5
1314
+ }
1315
+ },
1316
+ metadata: {
1317
+ knowledgeCutoff: "jun 01, 2024",
1318
+ performance: "higher",
1319
+ speed: "medium"
1320
+ }
1321
+ }, {
1322
+ id: 'gpt-4.1',
1323
+ provider: 'openai',
1324
+ name: 'gpt-4.1',
1325
+ contextWindow: 1047576,
1326
+ maxOutputTokens: 32768,
1327
+ tier: 'flagship',
1328
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1329
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1330
+ pricing: {
1331
+ text: {
1332
+ input: 2,
1333
+ output: 8,
1334
+ cached: 0.5
1335
+ }
1336
+ },
1337
+ metadata: {
1338
+ knowledgeCutoff: "jun 01, 2024",
1339
+ performance: "higher",
1340
+ speed: "medium"
1341
+ }
1342
+ }, {
1343
+ id: 'gpt-4.1-mini-2025-04-14',
1344
+ provider: 'openai',
1345
+ name: 'gpt-4.1-mini-2025-04-14',
1346
+ contextWindow: 1047576,
1347
+ maxOutputTokens: 32768,
1348
+ tier: 'efficient',
1349
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1350
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1351
+ pricing: {
1352
+ text: {
1353
+ input: 0.4,
1354
+ output: 1.6,
1355
+ cached: 0.1
1356
+ }
1357
+ },
1358
+ metadata: {
1359
+ knowledgeCutoff: "jun 01, 2024",
1360
+ performance: "high",
1361
+ speed: "fast"
1362
+ }
1363
+ }, {
1364
+ id: 'gpt-4.1-mini',
1365
+ provider: 'openai',
1366
+ name: 'gpt-4.1-mini',
1367
+ contextWindow: 1047576,
1368
+ maxOutputTokens: 32768,
1369
+ tier: 'efficient',
1370
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1371
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1372
+ pricing: {
1373
+ text: {
1374
+ input: 0.4,
1375
+ output: 1.6,
1376
+ cached: 0.1
1377
+ }
1378
+ },
1379
+ metadata: {
1380
+ knowledgeCutoff: "jun 01, 2024",
1381
+ performance: "high",
1382
+ speed: "fast"
1383
+ }
1384
+ }, {
1385
+ id: 'gpt-4.1-nano-2025-04-14',
1386
+ provider: 'openai',
1387
+ name: 'gpt-4.1-nano-2025-04-14',
1388
+ contextWindow: 1047576,
1389
+ maxOutputTokens: 32768,
1390
+ tier: 'efficient',
1391
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1392
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1393
+ pricing: {
1394
+ text: {
1395
+ input: 0.1,
1396
+ output: 0.4,
1397
+ cached: 0.025
1398
+ }
1399
+ },
1400
+ metadata: {
1401
+ knowledgeCutoff: "jun 01, 2024",
1402
+ performance: "average",
1403
+ speed: "very fast"
1404
+ }
1405
+ }, {
1406
+ id: 'gpt-4.1-nano',
1407
+ provider: 'openai',
1408
+ name: 'gpt-4.1-nano',
1409
+ contextWindow: 1047576,
1410
+ maxOutputTokens: 32768,
1411
+ tier: 'efficient',
1412
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1413
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1414
+ pricing: {
1415
+ text: {
1416
+ input: 0.1,
1417
+ output: 0.4,
1418
+ cached: 0.025
1419
+ }
1420
+ },
1421
+ metadata: {
1422
+ knowledgeCutoff: "jun 01, 2024",
1423
+ performance: "average",
1424
+ speed: "very fast"
1425
+ }
1426
+ }, {
1427
+ id: 'gpt-image-1',
1428
+ provider: 'openai',
1429
+ name: 'gpt-image-1',
1430
+ contextWindow: 0,
1431
+ tier: 'flagship',
1432
+ capabilities: new Set(['vision', 'image', 'chat']),
1433
+ supportedParameters: new Set(['imageStyle', 'imageMultiple', 'imageBackground', 'imageStream', 'imageFormat']),
1434
+ pricing: {
1435
+ text: {
1436
+ input: 5,
1437
+ cached: 1.25
1438
+ },
1439
+ image: {
1440
+ input: 10,
1441
+ output: [
1442
+ {
1443
+ quality: "low",
1444
+ sizes: [
1445
+ {
1446
+ width: 1024,
1447
+ height: 1024,
1448
+ cost: 0.011
1449
+ },
1450
+ {
1451
+ width: 1024,
1452
+ height: 1536,
1453
+ cost: 0.016
1454
+ },
1455
+ {
1456
+ width: 1536,
1457
+ height: 1024,
1458
+ cost: 0.016
1459
+ }
1460
+ ]
1461
+ },
1462
+ {
1463
+ quality: "medium",
1464
+ sizes: [
1465
+ {
1466
+ width: 1024,
1467
+ height: 1024,
1468
+ cost: 0.042
1469
+ },
1470
+ {
1471
+ width: 1024,
1472
+ height: 1536,
1473
+ cost: 0.063
1474
+ },
1475
+ {
1476
+ width: 1536,
1477
+ height: 1024,
1478
+ cost: 0.063
1479
+ }
1480
+ ]
1481
+ },
1482
+ {
1483
+ quality: "high",
1484
+ sizes: [
1485
+ {
1486
+ width: 1024,
1487
+ height: 1024,
1488
+ cost: 0.167
1489
+ },
1490
+ {
1491
+ width: 1024,
1492
+ height: 1536,
1493
+ cost: 0.25
1494
+ },
1495
+ {
1496
+ width: 1536,
1497
+ height: 1024,
1498
+ cost: 0.25
1499
+ }
1500
+ ]
1501
+ }
1502
+ ]
1503
+ }
1504
+ },
1505
+ metadata: {
1506
+ intelligence: "higher",
1507
+ speed: "slow"
1508
+ }
1509
+ }, {
1510
+ id: 'codex-mini-latest',
1511
+ provider: 'openai',
1512
+ name: 'codex-mini-latest',
1513
+ contextWindow: 200000,
1514
+ maxOutputTokens: 100000,
1515
+ tier: 'flagship',
1516
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1517
+ pricing: {
1518
+ text: {
1519
+ input: 1.5,
1520
+ output: 6,
1521
+ cached: 0.375
1522
+ }
1523
+ },
1524
+ metadata: {
1525
+ knowledgeCutoff: "jun 01, 2024",
1526
+ intelligence: "higher",
1527
+ speed: "medium"
1528
+ }
1529
+ }, {
1530
+ id: 'o3-pro',
1531
+ provider: 'openai',
1532
+ name: 'o3-pro',
1533
+ contextWindow: 200000,
1534
+ maxOutputTokens: 100000,
1535
+ tier: 'flagship',
1536
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
1537
+ pricing: {
1538
+ text: {
1539
+ input: 20,
1540
+ output: 80
1541
+ }
1542
+ },
1543
+ metadata: {
1544
+ knowledgeCutoff: "jun 01, 2024",
1545
+ reasoning: "highest",
1546
+ speed: "slow"
1547
+ }
1548
+ }, {
1549
+ id: 'gpt-4o-realtime-preview-2025-06-03',
1550
+ provider: 'openai',
1551
+ name: 'gpt-4o-realtime-preview-2025-06-03',
1552
+ contextWindow: 32000,
1553
+ maxOutputTokens: 4096,
1554
+ tier: 'efficient',
1555
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
1556
+ pricing: {
1557
+ text: {
1558
+ input: 5,
1559
+ output: 20,
1560
+ cached: 2.5
1561
+ },
1562
+ audio: {
1563
+ input: 40,
1564
+ output: 80
1565
+ }
1566
+ },
1567
+ metadata: {
1568
+ knowledgeCutoff: "oct 01, 2023",
1569
+ intelligence: "high",
1570
+ speed: "fast"
1571
+ }
1572
+ }, {
1573
+ id: 'gpt-4o-audio-preview-2025-06-03',
1574
+ provider: 'openai',
1575
+ name: 'gpt-4o-audio-preview-2025-06-03',
1576
+ contextWindow: 128000,
1577
+ maxOutputTokens: 16384,
1578
+ tier: 'legacy',
1579
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
1580
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
1581
+ pricing: {
1582
+ text: {
1583
+ input: 2.5,
1584
+ output: 10
1585
+ },
1586
+ audio: {
1587
+ input: 40,
1588
+ output: 80
1589
+ }
1590
+ },
1591
+ metadata: {
1592
+ knowledgeCutoff: "oct 01, 2023",
1593
+ performance: "high",
1594
+ speed: "medium"
1595
+ }
1596
+ }, {
1597
+ id: 'o3-pro-2025-06-10',
1598
+ provider: 'openai',
1599
+ name: 'o3-pro-2025-06-10',
1600
+ contextWindow: 200000,
1601
+ maxOutputTokens: 100000,
1602
+ tier: 'flagship',
1603
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
1604
+ pricing: {
1605
+ text: {
1606
+ input: 20,
1607
+ output: 80
1608
+ }
1609
+ },
1610
+ metadata: {
1611
+ knowledgeCutoff: "jun 01, 2024",
1612
+ reasoning: "highest",
1613
+ speed: "slow"
1614
+ }
1615
+ }, {
1616
+ id: 'o4-mini-deep-research',
1617
+ provider: 'openai',
1618
+ name: 'o4-mini-deep-research',
1619
+ contextWindow: 200000,
1620
+ maxOutputTokens: 100000,
1621
+ tier: 'flagship',
1622
+ capabilities: new Set(['chat', 'vision', 'streaming', 'reasoning']),
1623
+ pricing: {
1624
+ text: {
1625
+ input: 2,
1626
+ output: 8,
1627
+ cached: 0.5
1628
+ }
1629
+ },
1630
+ metadata: {
1631
+ knowledgeCutoff: "jun 01, 2024",
1632
+ reasoning: "higher",
1633
+ speed: "medium"
1634
+ }
1635
+ }, {
1636
+ id: 'o3-deep-research',
1637
+ provider: 'openai',
1638
+ name: 'o3-deep-research',
1639
+ contextWindow: 200000,
1640
+ maxOutputTokens: 100000,
1641
+ tier: 'flagship',
1642
+ capabilities: new Set(['chat', 'vision', 'streaming', 'reasoning']),
1643
+ pricing: {
1644
+ text: {
1645
+ input: 10,
1646
+ output: 40,
1647
+ cached: 2.5
1648
+ }
1649
+ },
1650
+ metadata: {
1651
+ knowledgeCutoff: "jun 01, 2024",
1652
+ reasoning: "highest",
1653
+ speed: "slow"
1654
+ }
1655
+ }, {
1656
+ id: 'gpt-4o-transcribe-diarize',
1657
+ provider: 'openai',
1658
+ name: 'gpt-4o-transcribe-diarize',
1659
+ contextWindow: 16000,
1660
+ maxOutputTokens: 2000,
1661
+ tier: 'flagship',
1662
+ capabilities: new Set(['hearing', 'chat']),
1663
+ supportedParameters: new Set(['transcribeStream']),
1664
+ pricing: {
1665
+ text: {
1666
+ input: 2.5,
1667
+ output: 10
1668
+ },
1669
+ audio: {
1670
+ input: 6
1671
+ }
1672
+ },
1673
+ metadata: {
1674
+ knowledgeCutoff: "jun 01, 2024",
1675
+ intelligence: "higher",
1676
+ speed: "medium"
1677
+ }
1678
+ }, {
1679
+ id: 'o3-deep-research-2025-06-26',
1680
+ provider: 'openai',
1681
+ name: 'o3-deep-research-2025-06-26',
1682
+ contextWindow: 200000,
1683
+ maxOutputTokens: 100000,
1684
+ tier: 'flagship',
1685
+ capabilities: new Set(['chat', 'vision', 'streaming', 'reasoning']),
1686
+ pricing: {
1687
+ text: {
1688
+ input: 10,
1689
+ output: 40,
1690
+ cached: 2.5
1691
+ }
1692
+ },
1693
+ metadata: {
1694
+ knowledgeCutoff: "jun 01, 2024",
1695
+ reasoning: "highest",
1696
+ speed: "slow"
1697
+ }
1698
+ }, {
1699
+ id: 'o4-mini-deep-research-2025-06-26',
1700
+ provider: 'openai',
1701
+ name: 'o4-mini-deep-research-2025-06-26',
1702
+ contextWindow: 200000,
1703
+ maxOutputTokens: 100000,
1704
+ tier: 'flagship',
1705
+ capabilities: new Set(['chat', 'vision', 'streaming', 'reasoning']),
1706
+ pricing: {
1707
+ text: {
1708
+ input: 2,
1709
+ output: 8,
1710
+ cached: 0.5
1711
+ }
1712
+ },
1713
+ metadata: {
1714
+ knowledgeCutoff: "jun 01, 2024",
1715
+ reasoning: "higher",
1716
+ speed: "medium"
1717
+ }
1718
+ }, {
1719
+ id: 'gpt-5-chat-latest',
1720
+ provider: 'openai',
1721
+ name: 'gpt-5-chat-latest',
1722
+ contextWindow: 128000,
1723
+ maxOutputTokens: 16384,
1724
+ tier: 'legacy',
1725
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1726
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1727
+ pricing: {
1728
+ text: {
1729
+ input: 1.25,
1730
+ output: 10,
1731
+ cached: 0.125
1732
+ }
1733
+ },
1734
+ metadata: {
1735
+ knowledgeCutoff: "sep 30, 2024",
1736
+ performance: "high",
1737
+ speed: "medium"
1738
+ }
1739
+ }, {
1740
+ id: 'gpt-5-2025-08-07',
1741
+ provider: 'openai',
1742
+ name: 'gpt-5-2025-08-07',
1743
+ contextWindow: 400000,
1744
+ maxOutputTokens: 128000,
1745
+ tier: 'flagship',
1746
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1747
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1748
+ pricing: {
1749
+ text: {
1750
+ input: 1.25,
1751
+ output: 10,
1752
+ cached: 0.125
1753
+ }
1754
+ },
1755
+ metadata: {
1756
+ knowledgeCutoff: "sep 30, 2024",
1757
+ reasoning: "higher",
1758
+ speed: "medium"
1759
+ }
1760
+ }, {
1761
+ id: 'gpt-5',
1762
+ provider: 'openai',
1763
+ name: 'gpt-5',
1764
+ contextWindow: 400000,
1765
+ maxOutputTokens: 128000,
1766
+ tier: 'flagship',
1767
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1768
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1769
+ pricing: {
1770
+ text: {
1771
+ input: 1.25,
1772
+ output: 10,
1773
+ cached: 0.125
1774
+ }
1775
+ },
1776
+ metadata: {
1777
+ knowledgeCutoff: "sep 30, 2024",
1778
+ reasoning: "higher",
1779
+ speed: "medium"
1780
+ }
1781
+ }, {
1782
+ id: 'gpt-5-mini-2025-08-07',
1783
+ provider: 'openai',
1784
+ name: 'gpt-5-mini-2025-08-07',
1785
+ contextWindow: 400000,
1786
+ maxOutputTokens: 128000,
1787
+ tier: 'efficient',
1788
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1789
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1790
+ pricing: {
1791
+ text: {
1792
+ input: 0.25,
1793
+ output: 2,
1794
+ cached: 0.025
1795
+ }
1796
+ },
1797
+ metadata: {
1798
+ knowledgeCutoff: "may 31, 2024",
1799
+ reasoning: "high",
1800
+ speed: "fast"
1801
+ }
1802
+ }, {
1803
+ id: 'gpt-5-mini',
1804
+ provider: 'openai',
1805
+ name: 'gpt-5-mini',
1806
+ contextWindow: 400000,
1807
+ maxOutputTokens: 128000,
1808
+ tier: 'efficient',
1809
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1810
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1811
+ pricing: {
1812
+ text: {
1813
+ input: 0.25,
1814
+ output: 2,
1815
+ cached: 0.025
1816
+ }
1817
+ },
1818
+ metadata: {
1819
+ knowledgeCutoff: "may 31, 2024",
1820
+ reasoning: "high",
1821
+ speed: "fast"
1822
+ }
1823
+ }, {
1824
+ id: 'gpt-5-nano-2025-08-07',
1825
+ provider: 'openai',
1826
+ name: 'gpt-5-nano-2025-08-07',
1827
+ contextWindow: 400000,
1828
+ maxOutputTokens: 128000,
1829
+ tier: 'efficient',
1830
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1831
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1832
+ pricing: {
1833
+ text: {
1834
+ input: 0.05,
1835
+ output: 0.4,
1836
+ cached: 0.005
1837
+ }
1838
+ },
1839
+ metadata: {
1840
+ knowledgeCutoff: "may 31, 2024",
1841
+ reasoning: "average",
1842
+ speed: "very fast"
1843
+ }
1844
+ }, {
1845
+ id: 'gpt-5-nano',
1846
+ provider: 'openai',
1847
+ name: 'gpt-5-nano',
1848
+ contextWindow: 400000,
1849
+ maxOutputTokens: 128000,
1850
+ tier: 'efficient',
1851
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1852
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1853
+ pricing: {
1854
+ text: {
1855
+ input: 0.05,
1856
+ output: 0.4,
1857
+ cached: 0.005
1858
+ }
1859
+ },
1860
+ metadata: {
1861
+ knowledgeCutoff: "may 31, 2024",
1862
+ reasoning: "average",
1863
+ speed: "very fast"
1864
+ }
1865
+ }, {
1866
+ id: 'gpt-audio-2025-08-28',
1867
+ provider: 'openai',
1868
+ name: 'gpt-audio-2025-08-28',
1869
+ contextWindow: 128000,
1870
+ maxOutputTokens: 16384,
1871
+ tier: 'flagship',
1872
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
1873
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
1874
+ pricing: {
1875
+ text: {
1876
+ input: 2.5,
1877
+ output: 10
1878
+ },
1879
+ audio: {
1880
+ input: 32,
1881
+ output: 64
1882
+ }
1883
+ },
1884
+ metadata: {
1885
+ knowledgeCutoff: "oct 01, 2023",
1886
+ performance: "higher",
1887
+ speed: "medium"
1888
+ }
1889
+ }, {
1890
+ id: 'gpt-realtime',
1891
+ provider: 'openai',
1892
+ name: 'gpt-realtime',
1893
+ contextWindow: 32000,
1894
+ maxOutputTokens: 4096,
1895
+ tier: 'flagship',
1896
+ capabilities: new Set(['vision', 'hearing', 'audio', 'chat', 'tools']),
1897
+ pricing: {
1898
+ text: {
1899
+ input: 4,
1900
+ output: 16,
1901
+ cached: 0.5
1902
+ },
1903
+ audio: {
1904
+ input: 32,
1905
+ output: 64
1906
+ },
1907
+ image: {
1908
+ input: 5
1909
+ }
1910
+ },
1911
+ metadata: {
1912
+ knowledgeCutoff: "oct 01, 2023",
1913
+ intelligence: "highest",
1914
+ speed: "fast"
1915
+ }
1916
+ }, {
1917
+ id: 'gpt-realtime-2025-08-28',
1918
+ provider: 'openai',
1919
+ name: 'gpt-realtime-2025-08-28',
1920
+ contextWindow: 32000,
1921
+ maxOutputTokens: 4096,
1922
+ tier: 'flagship',
1923
+ capabilities: new Set(['vision', 'hearing', 'audio', 'chat', 'tools']),
1924
+ pricing: {
1925
+ text: {
1926
+ input: 4,
1927
+ output: 16,
1928
+ cached: 0.5
1929
+ },
1930
+ audio: {
1931
+ input: 32,
1932
+ output: 64
1933
+ },
1934
+ image: {
1935
+ input: 5
1936
+ }
1937
+ },
1938
+ metadata: {
1939
+ knowledgeCutoff: "oct 01, 2023",
1940
+ intelligence: "highest",
1941
+ speed: "fast"
1942
+ }
1943
+ }, {
1944
+ id: 'gpt-audio',
1945
+ provider: 'openai',
1946
+ name: 'gpt-audio',
1947
+ contextWindow: 128000,
1948
+ maxOutputTokens: 16384,
1949
+ tier: 'flagship',
1950
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
1951
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
1952
+ pricing: {
1953
+ text: {
1954
+ input: 2.5,
1955
+ output: 10
1956
+ },
1957
+ audio: {
1958
+ input: 32,
1959
+ output: 64
1960
+ }
1961
+ },
1962
+ metadata: {
1963
+ knowledgeCutoff: "oct 01, 2023",
1964
+ performance: "higher",
1965
+ speed: "medium"
1966
+ }
1967
+ }, {
1968
+ id: 'gpt-5-codex',
1969
+ provider: 'openai',
1970
+ name: 'gpt-5-codex',
1971
+ contextWindow: 400000,
1972
+ maxOutputTokens: 128000,
1973
+ tier: 'flagship',
1974
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1975
+ pricing: {
1976
+ text: {
1977
+ input: 1.25,
1978
+ output: 10,
1979
+ cached: 0.125
1980
+ }
1981
+ },
1982
+ metadata: {
1983
+ knowledgeCutoff: "sep 30, 2024",
1984
+ reasoning: "higher",
1985
+ speed: "medium"
1986
+ }
1987
+ }, {
1988
+ id: 'gpt-image-1-mini',
1989
+ provider: 'openai',
1990
+ name: 'gpt-image-1-mini',
1991
+ contextWindow: 0,
1992
+ tier: 'flagship',
1993
+ capabilities: new Set(['vision', 'image', 'chat']),
1994
+ supportedParameters: new Set(['imageStyle', 'imageMultiple', 'imageBackground', 'imageStream', 'imageFormat']),
1995
+ pricing: {
1996
+ text: {
1997
+ input: 2,
1998
+ cached: 0.2
1999
+ },
2000
+ image: {
2001
+ input: 2.5,
2002
+ output: [
2003
+ {
2004
+ quality: "low",
2005
+ sizes: [
2006
+ {
2007
+ width: 1024,
2008
+ height: 1024,
2009
+ cost: 0.005
2010
+ },
2011
+ {
2012
+ width: 1024,
2013
+ height: 1536,
2014
+ cost: 0.006
2015
+ },
2016
+ {
2017
+ width: 1536,
2018
+ height: 1024,
2019
+ cost: 0.006
2020
+ }
2021
+ ]
2022
+ },
2023
+ {
2024
+ quality: "medium",
2025
+ sizes: [
2026
+ {
2027
+ width: 1024,
2028
+ height: 1024,
2029
+ cost: 0.011
2030
+ },
2031
+ {
2032
+ width: 1024,
2033
+ height: 1536,
2034
+ cost: 0.015
2035
+ },
2036
+ {
2037
+ width: 1536,
2038
+ height: 1024,
2039
+ cost: 0.015
2040
+ }
2041
+ ]
2042
+ },
2043
+ {
2044
+ quality: "high",
2045
+ sizes: [
2046
+ {
2047
+ width: 1024,
2048
+ height: 1024,
2049
+ cost: 0.036
2050
+ },
2051
+ {
2052
+ width: 1024,
2053
+ height: 1536,
2054
+ cost: 0.052
2055
+ },
2056
+ {
2057
+ width: 1536,
2058
+ height: 1024,
2059
+ cost: 0.052
2060
+ }
2061
+ ]
2062
+ }
2063
+ ]
2064
+ }
2065
+ },
2066
+ metadata: {
2067
+ intelligence: "higher",
2068
+ speed: "slow"
2069
+ }
2070
+ }, {
2071
+ id: 'gpt-5-pro-2025-10-06',
2072
+ provider: 'openai',
2073
+ name: 'gpt-5-pro-2025-10-06',
2074
+ contextWindow: 400000,
2075
+ maxOutputTokens: 272000,
2076
+ tier: 'flagship',
2077
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
2078
+ supportedParameters: new Set(['imageStyle', 'imageMultiple']),
2079
+ pricing: {
2080
+ text: {
2081
+ input: 15,
2082
+ output: 120
2083
+ }
2084
+ },
2085
+ metadata: {
2086
+ knowledgeCutoff: "sep 30, 2024",
2087
+ reasoning: "highest",
2088
+ speed: "slow"
2089
+ }
2090
+ }, {
2091
+ id: 'gpt-5-pro',
2092
+ provider: 'openai',
2093
+ name: 'gpt-5-pro',
2094
+ contextWindow: 400000,
2095
+ maxOutputTokens: 272000,
2096
+ tier: 'flagship',
2097
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
2098
+ supportedParameters: new Set(['imageStyle', 'imageMultiple']),
2099
+ pricing: {
2100
+ text: {
2101
+ input: 15,
2102
+ output: 120
2103
+ }
2104
+ },
2105
+ metadata: {
2106
+ knowledgeCutoff: "sep 30, 2024",
2107
+ reasoning: "highest",
2108
+ speed: "slow"
2109
+ }
2110
+ }, {
2111
+ id: 'gpt-audio-mini',
2112
+ provider: 'openai',
2113
+ name: 'gpt-audio-mini',
2114
+ contextWindow: 128000,
2115
+ maxOutputTokens: 16384,
2116
+ tier: 'flagship',
2117
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
2118
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
2119
+ pricing: {
2120
+ text: {
2121
+ input: 0.6,
2122
+ output: 2.4
2123
+ },
2124
+ audio: {
2125
+ input: 10,
2126
+ output: 20
2127
+ }
2128
+ },
2129
+ metadata: {
2130
+ knowledgeCutoff: "oct 01, 2023",
2131
+ performance: "higher",
2132
+ speed: "very fast"
2133
+ }
2134
+ }, {
2135
+ id: 'gpt-audio-mini-2025-10-06',
2136
+ provider: 'openai',
2137
+ name: 'gpt-audio-mini-2025-10-06',
2138
+ contextWindow: 128000,
2139
+ maxOutputTokens: 16384,
2140
+ tier: 'flagship',
2141
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
2142
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
2143
+ pricing: {
2144
+ text: {
2145
+ input: 0.6,
2146
+ output: 2.4
2147
+ },
2148
+ audio: {
2149
+ input: 10,
2150
+ output: 20
2151
+ }
2152
+ },
2153
+ metadata: {
2154
+ knowledgeCutoff: "oct 01, 2023",
2155
+ performance: "higher",
2156
+ speed: "very fast"
2157
+ }
2158
+ }, {
2159
+ id: 'tts-1',
2160
+ provider: 'openai',
2161
+ name: 'tts-1',
2162
+ contextWindow: 0,
2163
+ tier: 'efficient',
2164
+ capabilities: new Set(['audio', 'chat']),
2165
+ supportedParameters: new Set([]),
2166
+ pricing: {
2167
+ audio: {
2168
+ output: 15
2169
+ }
2170
+ },
2171
+ metadata: {
2172
+ intelligence: "average",
2173
+ speed: "fast"
2174
+ }
2175
+ }, {
2176
+ id: 'whisper-1',
2177
+ provider: 'openai',
2178
+ name: 'whisper-1',
2179
+ contextWindow: 0,
2180
+ tier: 'legacy',
2181
+ capabilities: new Set(['hearing']),
2182
+ supportedParameters: new Set(['transcribePrompt']),
2183
+ pricing: {
2184
+ audio: {
2185
+ input: 0.006
2186
+ }
2187
+ },
2188
+ metadata: {
2189
+ intelligence: "average",
2190
+ speed: "medium"
2191
+ }
2192
+ }, {
2193
+ id: 'text-embedding-ada-002',
2194
+ provider: 'openai',
2195
+ name: 'text-embedding-ada-002',
2196
+ contextWindow: 0,
2197
+ tier: 'legacy',
2198
+ capabilities: new Set(['embedding', 'chat']),
2199
+ supportedParameters: new Set([]),
2200
+ pricing: {
2201
+ embeddings: {
2202
+ cost: 0.1
2203
+ }
2204
+ },
2205
+ metadata: {
2206
+ intelligence: "low",
2207
+ speed: "slow"
2208
+ }
2209
+ }];