@aeye/models 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (74) hide show
  1. package/ReplicateScrape.md +54 -0
  2. package/dist/scripts/codegen.d.ts +21 -0
  3. package/dist/scripts/codegen.d.ts.map +1 -0
  4. package/dist/scripts/codegen.js +102 -0
  5. package/dist/scripts/codegen.js.map +1 -0
  6. package/dist/scripts/scrape.d.ts +19 -0
  7. package/dist/scripts/scrape.d.ts.map +1 -0
  8. package/dist/scripts/scrape.js +146 -0
  9. package/dist/scripts/scrape.js.map +1 -0
  10. package/dist/scripts/scrapers/__tests__/aws.test.d.ts +8 -0
  11. package/dist/scripts/scrapers/__tests__/aws.test.d.ts.map +1 -0
  12. package/dist/scripts/scrapers/__tests__/aws.test.js +73 -0
  13. package/dist/scripts/scrapers/__tests__/aws.test.js.map +1 -0
  14. package/dist/scripts/scrapers/aws.d.ts +12 -0
  15. package/dist/scripts/scrapers/aws.d.ts.map +1 -0
  16. package/dist/scripts/scrapers/aws.js +314 -0
  17. package/dist/scripts/scrapers/aws.js.map +1 -0
  18. package/dist/scripts/scrapers/openai.d.ts +12 -0
  19. package/dist/scripts/scrapers/openai.d.ts.map +1 -0
  20. package/dist/scripts/scrapers/openai.js +490 -0
  21. package/dist/scripts/scrapers/openai.js.map +1 -0
  22. package/dist/scripts/scrapers/openrouter.d.ts +13 -0
  23. package/dist/scripts/scrapers/openrouter.d.ts.map +1 -0
  24. package/dist/scripts/scrapers/openrouter.js +156 -0
  25. package/dist/scripts/scrapers/openrouter.js.map +1 -0
  26. package/dist/scripts/scrapers/replicate.d.ts +12 -0
  27. package/dist/scripts/scrapers/replicate.d.ts.map +1 -0
  28. package/dist/scripts/scrapers/replicate.js +305 -0
  29. package/dist/scripts/scrapers/replicate.js.map +1 -0
  30. package/dist/src/index.d.ts +11 -0
  31. package/dist/src/index.d.ts.map +1 -0
  32. package/dist/src/index.js +11 -0
  33. package/dist/src/index.js.map +1 -0
  34. package/dist/src/models/aws.d.ts +11 -0
  35. package/dist/src/models/aws.d.ts.map +1 -0
  36. package/dist/src/models/aws.js +2632 -0
  37. package/dist/src/models/aws.js.map +1 -0
  38. package/dist/src/models/index.d.ts +15 -0
  39. package/dist/src/models/index.d.ts.map +1 -0
  40. package/dist/src/models/index.js +18 -0
  41. package/dist/src/models/index.js.map +1 -0
  42. package/dist/src/models/openai.d.ts +11 -0
  43. package/dist/src/models/openai.d.ts.map +1 -0
  44. package/dist/src/models/openai.js +2207 -0
  45. package/dist/src/models/openai.js.map +1 -0
  46. package/dist/src/models/openrouter.d.ts +11 -0
  47. package/dist/src/models/openrouter.d.ts.map +1 -0
  48. package/dist/src/models/openrouter.js +9786 -0
  49. package/dist/src/models/openrouter.js.map +1 -0
  50. package/dist/src/models/replicate.d.ts +11 -0
  51. package/dist/src/models/replicate.d.ts.map +1 -0
  52. package/dist/src/models/replicate.js +4106 -0
  53. package/dist/src/models/replicate.js.map +1 -0
  54. package/dist/src/transformers/index.d.ts +23 -0
  55. package/dist/src/transformers/index.d.ts.map +1 -0
  56. package/dist/src/transformers/index.js +24 -0
  57. package/dist/src/transformers/index.js.map +1 -0
  58. package/package.json +50 -0
  59. package/scripts/codegen.ts +117 -0
  60. package/scripts/scrape.ts +182 -0
  61. package/scripts/scrapers/__tests__/aws.test.ts +86 -0
  62. package/scripts/scrapers/aws.ts +370 -0
  63. package/scripts/scrapers/openai.ts +619 -0
  64. package/scripts/scrapers/openrouter.ts +214 -0
  65. package/scripts/scrapers/replicate.ts +448 -0
  66. package/scripts/tsconfig.json +24 -0
  67. package/src/index.ts +11 -0
  68. package/src/models/aws.ts +2634 -0
  69. package/src/models/index.ts +21 -0
  70. package/src/models/openai.ts +2209 -0
  71. package/src/models/openrouter.ts +9788 -0
  72. package/src/models/replicate.ts +4108 -0
  73. package/src/transformers/index.ts +26 -0
  74. package/tsconfig.json +14 -0
@@ -0,0 +1,2207 @@
1
+ /**
2
+ * Generated model data
3
+ *
4
+ * This file is auto-generated by the scraper scripts.
5
+ * Do not edit manually - your changes will be overwritten.
6
+ *
7
+ * To regenerate, run: npm run scrape
8
+ */
9
+ export const openaiModels = [{
10
+ id: 'gpt-4-0613',
11
+ provider: 'openai',
12
+ name: 'gpt-4-0613',
13
+ contextWindow: 8192,
14
+ maxOutputTokens: 8192,
15
+ tier: 'legacy',
16
+ capabilities: new Set(['chat', 'streaming']),
17
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
18
+ pricing: {
19
+ text: {
20
+ input: 30,
21
+ output: 60
22
+ }
23
+ },
24
+ metadata: {
25
+ knowledgeCutoff: "dec 01, 2023",
26
+ performance: "average",
27
+ speed: "medium"
28
+ }
29
+ }, {
30
+ id: 'gpt-4',
31
+ provider: 'openai',
32
+ name: 'gpt-4',
33
+ contextWindow: 8192,
34
+ maxOutputTokens: 8192,
35
+ tier: 'legacy',
36
+ capabilities: new Set(['chat', 'streaming']),
37
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
38
+ pricing: {
39
+ text: {
40
+ input: 30,
41
+ output: 60
42
+ }
43
+ },
44
+ metadata: {
45
+ knowledgeCutoff: "dec 01, 2023",
46
+ performance: "average",
47
+ speed: "medium"
48
+ }
49
+ }, {
50
+ id: 'gpt-3.5-turbo',
51
+ provider: 'openai',
52
+ name: 'gpt-3.5-turbo',
53
+ contextWindow: 16385,
54
+ maxOutputTokens: 4096,
55
+ tier: 'legacy',
56
+ capabilities: new Set(['chat']),
57
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
58
+ pricing: {
59
+ text: {
60
+ input: 0.5,
61
+ output: 1.5
62
+ }
63
+ },
64
+ metadata: {
65
+ knowledgeCutoff: "sep 01, 2021",
66
+ performance: "low",
67
+ speed: "slow"
68
+ }
69
+ }, {
70
+ id: 'gpt-realtime-mini',
71
+ provider: 'openai',
72
+ name: 'gpt-realtime-mini',
73
+ contextWindow: 32000,
74
+ maxOutputTokens: 4096,
75
+ tier: 'flagship',
76
+ capabilities: new Set(['vision', 'hearing', 'audio', 'chat', 'tools']),
77
+ pricing: {
78
+ text: {
79
+ input: 0.6,
80
+ output: 2.4,
81
+ cached: 0.06
82
+ },
83
+ audio: {
84
+ input: 10,
85
+ output: 20
86
+ },
87
+ image: {
88
+ input: 0.8
89
+ }
90
+ },
91
+ metadata: {
92
+ knowledgeCutoff: "oct 01, 2023",
93
+ intelligence: "higher",
94
+ speed: "very fast"
95
+ }
96
+ }, {
97
+ id: 'gpt-realtime-mini-2025-10-06',
98
+ provider: 'openai',
99
+ name: 'gpt-realtime-mini-2025-10-06',
100
+ contextWindow: 32000,
101
+ maxOutputTokens: 4096,
102
+ tier: 'flagship',
103
+ capabilities: new Set(['vision', 'hearing', 'audio', 'chat', 'tools']),
104
+ pricing: {
105
+ text: {
106
+ input: 0.6,
107
+ output: 2.4,
108
+ cached: 0.06
109
+ },
110
+ audio: {
111
+ input: 10,
112
+ output: 20
113
+ },
114
+ image: {
115
+ input: 0.8
116
+ }
117
+ },
118
+ metadata: {
119
+ knowledgeCutoff: "oct 01, 2023",
120
+ intelligence: "higher",
121
+ speed: "very fast"
122
+ }
123
+ }, {
124
+ id: 'sora-2',
125
+ provider: 'openai',
126
+ name: 'sora-2',
127
+ contextWindow: 0,
128
+ tier: 'flagship',
129
+ capabilities: new Set(['chat', 'vision', 'audio']),
130
+ pricing: {},
131
+ metadata: {
132
+ intelligence: "higher",
133
+ speed: "slow"
134
+ }
135
+ }, {
136
+ id: 'sora-2-pro',
137
+ provider: 'openai',
138
+ name: 'sora-2-pro',
139
+ contextWindow: 0,
140
+ tier: 'flagship',
141
+ capabilities: new Set(['chat', 'vision', 'audio']),
142
+ pricing: {},
143
+ metadata: {
144
+ intelligence: "highest",
145
+ speed: "slow"
146
+ }
147
+ }, {
148
+ id: 'davinci-002',
149
+ provider: 'openai',
150
+ name: 'davinci-002',
151
+ contextWindow: 0,
152
+ maxOutputTokens: 16384,
153
+ tier: 'legacy',
154
+ capabilities: new Set(['chat']),
155
+ pricing: {
156
+ text: {
157
+ input: 2,
158
+ output: 2
159
+ }
160
+ },
161
+ metadata: {
162
+ knowledgeCutoff: "sep 01, 2021",
163
+ performance: "low",
164
+ speed: "medium"
165
+ }
166
+ }, {
167
+ id: 'babbage-002',
168
+ provider: 'openai',
169
+ name: 'babbage-002',
170
+ contextWindow: 0,
171
+ maxOutputTokens: 16384,
172
+ tier: 'legacy',
173
+ capabilities: new Set(['chat']),
174
+ pricing: {
175
+ text: {
176
+ input: 0.4,
177
+ output: 0.4
178
+ }
179
+ },
180
+ metadata: {
181
+ knowledgeCutoff: "sep 01, 2021",
182
+ performance: "low",
183
+ speed: "medium"
184
+ }
185
+ }, {
186
+ id: 'gpt-3.5-turbo-instruct',
187
+ provider: 'openai',
188
+ name: 'gpt-3.5-turbo-instruct',
189
+ contextWindow: 4096,
190
+ maxOutputTokens: 4096,
191
+ tier: 'legacy',
192
+ capabilities: new Set(['chat']),
193
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
194
+ pricing: {
195
+ text: {
196
+ input: 1.5,
197
+ output: 2
198
+ }
199
+ },
200
+ metadata: {
201
+ knowledgeCutoff: "sep 01, 2021",
202
+ performance: "low",
203
+ speed: "slow"
204
+ }
205
+ }, {
206
+ id: 'gpt-3.5-turbo-instruct-0914',
207
+ provider: 'openai',
208
+ name: 'gpt-3.5-turbo-instruct-0914',
209
+ contextWindow: 4096,
210
+ maxOutputTokens: 4096,
211
+ tier: 'legacy',
212
+ capabilities: new Set(['chat']),
213
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
214
+ pricing: {
215
+ text: {
216
+ input: 1.5,
217
+ output: 2
218
+ }
219
+ },
220
+ metadata: {
221
+ knowledgeCutoff: "sep 01, 2021",
222
+ performance: "low",
223
+ speed: "slow"
224
+ }
225
+ }, {
226
+ id: 'dall-e-3',
227
+ provider: 'openai',
228
+ name: 'dall-e-3',
229
+ contextWindow: 0,
230
+ tier: 'legacy',
231
+ capabilities: new Set(['image', 'chat']),
232
+ supportedParameters: new Set([]),
233
+ pricing: {
234
+ image: {
235
+ output: [
236
+ {
237
+ quality: "standard",
238
+ sizes: [
239
+ {
240
+ width: 1024,
241
+ height: 1024,
242
+ cost: 0.04
243
+ },
244
+ {
245
+ width: 1024,
246
+ height: 1536,
247
+ cost: 0.08
248
+ },
249
+ {
250
+ width: 1536,
251
+ height: 1024,
252
+ cost: 0.08
253
+ }
254
+ ]
255
+ },
256
+ {
257
+ quality: "hd",
258
+ sizes: [
259
+ {
260
+ width: 1024,
261
+ height: 1024,
262
+ cost: 0.08
263
+ },
264
+ {
265
+ width: 1024,
266
+ height: 1536,
267
+ cost: 0.12
268
+ },
269
+ {
270
+ width: 1536,
271
+ height: 1024,
272
+ cost: 0.12
273
+ }
274
+ ]
275
+ }
276
+ ]
277
+ }
278
+ },
279
+ metadata: {
280
+ intelligence: "high",
281
+ speed: "slow"
282
+ }
283
+ }, {
284
+ id: 'dall-e-2',
285
+ provider: 'openai',
286
+ name: 'dall-e-2',
287
+ contextWindow: 0,
288
+ tier: 'legacy',
289
+ capabilities: new Set(['image', 'chat']),
290
+ supportedParameters: new Set(['imageStyle', 'imageMultiple']),
291
+ pricing: {
292
+ image: {
293
+ output: [
294
+ {
295
+ quality: "standard",
296
+ sizes: [
297
+ {
298
+ width: 1024,
299
+ height: 1024,
300
+ cost: 0.016
301
+ },
302
+ {
303
+ width: 1024,
304
+ height: 1536,
305
+ cost: 0.018
306
+ },
307
+ {
308
+ width: 1536,
309
+ height: 1024,
310
+ cost: 0.02
311
+ }
312
+ ]
313
+ }
314
+ ]
315
+ }
316
+ },
317
+ metadata: {
318
+ intelligence: "low",
319
+ speed: "slow"
320
+ }
321
+ }, {
322
+ id: 'gpt-3.5-turbo-1106',
323
+ provider: 'openai',
324
+ name: 'gpt-3.5-turbo-1106',
325
+ contextWindow: 16385,
326
+ maxOutputTokens: 4096,
327
+ tier: 'legacy',
328
+ capabilities: new Set(['chat']),
329
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
330
+ pricing: {
331
+ text: {
332
+ input: 1,
333
+ output: 2
334
+ }
335
+ },
336
+ metadata: {
337
+ knowledgeCutoff: "sep 01, 2021",
338
+ performance: "low",
339
+ speed: "slow"
340
+ }
341
+ }, {
342
+ id: 'tts-1-hd',
343
+ provider: 'openai',
344
+ name: 'tts-1-hd',
345
+ contextWindow: 0,
346
+ tier: 'legacy',
347
+ capabilities: new Set(['audio', 'chat']),
348
+ supportedParameters: new Set([]),
349
+ pricing: {
350
+ audio: {
351
+ output: 30
352
+ }
353
+ },
354
+ metadata: {
355
+ intelligence: "high",
356
+ speed: "medium"
357
+ }
358
+ }, {
359
+ id: 'tts-1-1106',
360
+ provider: 'openai',
361
+ name: 'tts-1-1106',
362
+ contextWindow: 0,
363
+ tier: 'efficient',
364
+ capabilities: new Set(['audio', 'chat']),
365
+ supportedParameters: new Set([]),
366
+ pricing: {
367
+ audio: {
368
+ output: 15
369
+ }
370
+ },
371
+ metadata: {
372
+ intelligence: "average",
373
+ speed: "fast"
374
+ }
375
+ }, {
376
+ id: 'tts-1-hd-1106',
377
+ provider: 'openai',
378
+ name: 'tts-1-hd-1106',
379
+ contextWindow: 0,
380
+ tier: 'legacy',
381
+ capabilities: new Set(['audio', 'chat']),
382
+ supportedParameters: new Set([]),
383
+ pricing: {
384
+ audio: {
385
+ output: 30
386
+ }
387
+ },
388
+ metadata: {
389
+ intelligence: "high",
390
+ speed: "medium"
391
+ }
392
+ }, {
393
+ id: 'text-embedding-3-small',
394
+ provider: 'openai',
395
+ name: 'text-embedding-3-small',
396
+ contextWindow: 0,
397
+ tier: 'legacy',
398
+ capabilities: new Set(['embedding', 'chat']),
399
+ supportedParameters: new Set(['embeddingDimensions']),
400
+ pricing: {
401
+ embeddings: {
402
+ cost: 0.02
403
+ }
404
+ },
405
+ metadata: {
406
+ intelligence: "average",
407
+ speed: "medium"
408
+ }
409
+ }, {
410
+ id: 'text-embedding-3-large',
411
+ provider: 'openai',
412
+ name: 'text-embedding-3-large',
413
+ contextWindow: 0,
414
+ tier: 'legacy',
415
+ capabilities: new Set(['embedding', 'chat']),
416
+ supportedParameters: new Set(['embeddingDimensions']),
417
+ pricing: {
418
+ embeddings: {
419
+ cost: 0.13
420
+ }
421
+ },
422
+ metadata: {
423
+ intelligence: "high",
424
+ speed: "slow"
425
+ }
426
+ }, {
427
+ id: 'gpt-4-turbo-preview',
428
+ provider: 'openai',
429
+ name: 'gpt-4-turbo-preview',
430
+ contextWindow: 128000,
431
+ maxOutputTokens: 4096,
432
+ tier: 'legacy',
433
+ capabilities: new Set(['chat']),
434
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
435
+ pricing: {
436
+ text: {
437
+ input: 10,
438
+ output: 30
439
+ }
440
+ },
441
+ metadata: {
442
+ knowledgeCutoff: "dec 01, 2023",
443
+ performance: "average",
444
+ speed: "medium"
445
+ }
446
+ }, {
447
+ id: 'gpt-3.5-turbo-0125',
448
+ provider: 'openai',
449
+ name: 'gpt-3.5-turbo-0125',
450
+ contextWindow: 16385,
451
+ maxOutputTokens: 4096,
452
+ tier: 'legacy',
453
+ capabilities: new Set(['chat']),
454
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
455
+ pricing: {
456
+ text: {
457
+ input: 0.5,
458
+ output: 1.5
459
+ }
460
+ },
461
+ metadata: {
462
+ knowledgeCutoff: "sep 01, 2021",
463
+ performance: "low",
464
+ speed: "slow"
465
+ }
466
+ }, {
467
+ id: 'gpt-4-turbo',
468
+ provider: 'openai',
469
+ name: 'gpt-4-turbo',
470
+ contextWindow: 128000,
471
+ maxOutputTokens: 4096,
472
+ tier: 'legacy',
473
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools']),
474
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
475
+ pricing: {
476
+ text: {
477
+ input: 10,
478
+ output: 30
479
+ }
480
+ },
481
+ metadata: {
482
+ knowledgeCutoff: "dec 01, 2023",
483
+ performance: "average",
484
+ speed: "medium"
485
+ }
486
+ }, {
487
+ id: 'gpt-4-turbo-2024-04-09',
488
+ provider: 'openai',
489
+ name: 'gpt-4-turbo-2024-04-09',
490
+ contextWindow: 128000,
491
+ maxOutputTokens: 4096,
492
+ tier: 'legacy',
493
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools']),
494
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
495
+ pricing: {
496
+ text: {
497
+ input: 10,
498
+ output: 30
499
+ }
500
+ },
501
+ metadata: {
502
+ knowledgeCutoff: "dec 01, 2023",
503
+ performance: "average",
504
+ speed: "medium"
505
+ }
506
+ }, {
507
+ id: 'gpt-4o',
508
+ provider: 'openai',
509
+ name: 'gpt-4o',
510
+ contextWindow: 128000,
511
+ maxOutputTokens: 16384,
512
+ tier: 'legacy',
513
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
514
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
515
+ pricing: {
516
+ text: {
517
+ input: 2.5,
518
+ output: 10,
519
+ cached: 1.25
520
+ }
521
+ },
522
+ metadata: {
523
+ knowledgeCutoff: "oct 01, 2023",
524
+ performance: "high",
525
+ speed: "medium"
526
+ }
527
+ }, {
528
+ id: 'gpt-4o-2024-05-13',
529
+ provider: 'openai',
530
+ name: 'gpt-4o-2024-05-13',
531
+ contextWindow: 128000,
532
+ maxOutputTokens: 4096,
533
+ tier: 'legacy',
534
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools']),
535
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
536
+ pricing: {
537
+ text: {
538
+ input: 5,
539
+ output: 15
540
+ }
541
+ },
542
+ metadata: {
543
+ knowledgeCutoff: "oct 01, 2023",
544
+ performance: "high",
545
+ speed: "medium"
546
+ }
547
+ }, {
548
+ id: 'gpt-4o-mini-2024-07-18',
549
+ provider: 'openai',
550
+ name: 'gpt-4o-mini-2024-07-18',
551
+ contextWindow: 128000,
552
+ maxOutputTokens: 16384,
553
+ tier: 'efficient',
554
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
555
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
556
+ pricing: {
557
+ text: {
558
+ input: 0.15,
559
+ output: 0.6,
560
+ cached: 0.075
561
+ }
562
+ },
563
+ metadata: {
564
+ knowledgeCutoff: "oct 01, 2023",
565
+ performance: "average",
566
+ speed: "fast"
567
+ }
568
+ }, {
569
+ id: 'gpt-4o-mini',
570
+ provider: 'openai',
571
+ name: 'gpt-4o-mini',
572
+ contextWindow: 128000,
573
+ maxOutputTokens: 16384,
574
+ tier: 'efficient',
575
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
576
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
577
+ pricing: {
578
+ text: {
579
+ input: 0.15,
580
+ output: 0.6,
581
+ cached: 0.075
582
+ }
583
+ },
584
+ metadata: {
585
+ knowledgeCutoff: "oct 01, 2023",
586
+ performance: "average",
587
+ speed: "fast"
588
+ }
589
+ }, {
590
+ id: 'gpt-4o-2024-08-06',
591
+ provider: 'openai',
592
+ name: 'gpt-4o-2024-08-06',
593
+ contextWindow: 128000,
594
+ maxOutputTokens: 16384,
595
+ tier: 'legacy',
596
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
597
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
598
+ pricing: {
599
+ text: {
600
+ input: 2.5,
601
+ output: 10,
602
+ cached: 1.25
603
+ }
604
+ },
605
+ metadata: {
606
+ knowledgeCutoff: "oct 01, 2023",
607
+ performance: "high",
608
+ speed: "medium"
609
+ }
610
+ }, {
611
+ id: 'chatgpt-4o-latest',
612
+ provider: 'openai',
613
+ name: 'chatgpt-4o-latest',
614
+ contextWindow: 128000,
615
+ maxOutputTokens: 16384,
616
+ tier: 'legacy',
617
+ capabilities: new Set(['chat', 'vision', 'streaming']),
618
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'stop']),
619
+ pricing: {
620
+ text: {
621
+ input: 5,
622
+ output: 15
623
+ }
624
+ },
625
+ metadata: {
626
+ knowledgeCutoff: "oct 01, 2023",
627
+ performance: "high",
628
+ speed: "medium"
629
+ }
630
+ }, {
631
+ id: 'o1-mini-2024-09-12',
632
+ provider: 'openai',
633
+ name: 'o1-mini-2024-09-12',
634
+ contextWindow: 128000,
635
+ maxOutputTokens: 65536,
636
+ tier: 'legacy',
637
+ capabilities: new Set(['chat', 'streaming', 'reasoning']),
638
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'stop']),
639
+ pricing: {
640
+ text: {
641
+ input: 1.1,
642
+ output: 4.4,
643
+ cached: 0.55
644
+ }
645
+ },
646
+ metadata: {
647
+ knowledgeCutoff: "oct 01, 2023",
648
+ reasoning: "high",
649
+ speed: "slow"
650
+ }
651
+ }, {
652
+ id: 'o1-mini',
653
+ provider: 'openai',
654
+ name: 'o1-mini',
655
+ contextWindow: 128000,
656
+ maxOutputTokens: 65536,
657
+ tier: 'legacy',
658
+ capabilities: new Set(['chat', 'streaming', 'reasoning']),
659
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'stop']),
660
+ pricing: {
661
+ text: {
662
+ input: 1.1,
663
+ output: 4.4,
664
+ cached: 0.55
665
+ }
666
+ },
667
+ metadata: {
668
+ knowledgeCutoff: "oct 01, 2023",
669
+ reasoning: "high",
670
+ speed: "slow"
671
+ }
672
+ }, {
673
+ id: 'gpt-4o-realtime-preview-2024-10-01',
674
+ provider: 'openai',
675
+ name: 'gpt-4o-realtime-preview-2024-10-01',
676
+ contextWindow: 16000,
677
+ maxOutputTokens: 4096,
678
+ tier: 'efficient',
679
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
680
+ pricing: {
681
+ text: {
682
+ input: 5,
683
+ output: 20,
684
+ cached: 2.5
685
+ },
686
+ audio: {
687
+ input: 100,
688
+ output: 200
689
+ }
690
+ },
691
+ metadata: {
692
+ knowledgeCutoff: "oct 01, 2023",
693
+ intelligence: "average",
694
+ speed: "fast"
695
+ }
696
+ }, {
697
+ id: 'gpt-4o-audio-preview-2024-10-01',
698
+ provider: 'openai',
699
+ name: 'gpt-4o-audio-preview-2024-10-01',
700
+ contextWindow: 128000,
701
+ maxOutputTokens: 16384,
702
+ tier: 'legacy',
703
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
704
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
705
+ pricing: {
706
+ text: {
707
+ input: 2.5,
708
+ output: 10
709
+ },
710
+ audio: {
711
+ input: 100,
712
+ output: 200
713
+ }
714
+ },
715
+ metadata: {
716
+ knowledgeCutoff: "oct 01, 2023",
717
+ performance: "high",
718
+ speed: "medium"
719
+ }
720
+ }, {
721
+ id: 'gpt-4o-audio-preview',
722
+ provider: 'openai',
723
+ name: 'gpt-4o-audio-preview',
724
+ contextWindow: 128000,
725
+ maxOutputTokens: 16384,
726
+ tier: 'legacy',
727
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
728
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
729
+ pricing: {
730
+ text: {
731
+ input: 2.5,
732
+ output: 10
733
+ },
734
+ audio: {
735
+ input: 40,
736
+ output: 80
737
+ }
738
+ },
739
+ metadata: {
740
+ knowledgeCutoff: "oct 01, 2023",
741
+ performance: "high",
742
+ speed: "medium"
743
+ }
744
+ }, {
745
+ id: 'gpt-4o-realtime-preview',
746
+ provider: 'openai',
747
+ name: 'gpt-4o-realtime-preview',
748
+ contextWindow: 32000,
749
+ maxOutputTokens: 4096,
750
+ tier: 'efficient',
751
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
752
+ pricing: {
753
+ text: {
754
+ input: 5,
755
+ output: 20,
756
+ cached: 2.5
757
+ },
758
+ audio: {
759
+ input: 40,
760
+ output: 80
761
+ }
762
+ },
763
+ metadata: {
764
+ knowledgeCutoff: "oct 01, 2023",
765
+ intelligence: "high",
766
+ speed: "fast"
767
+ }
768
+ }, {
769
+ id: 'omni-moderation-latest',
770
+ provider: 'openai',
771
+ name: 'omni-moderation-latest',
772
+ contextWindow: 0,
773
+ tier: 'legacy',
774
+ capabilities: new Set(['chat', 'vision']),
775
+ pricing: {},
776
+ metadata: {
777
+ intelligence: "high",
778
+ speed: "medium"
779
+ }
780
+ }, {
781
+ id: 'gpt-4o-realtime-preview-2024-12-17',
782
+ provider: 'openai',
783
+ name: 'gpt-4o-realtime-preview-2024-12-17',
784
+ contextWindow: 16000,
785
+ maxOutputTokens: 4096,
786
+ tier: 'efficient',
787
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
788
+ pricing: {
789
+ text: {
790
+ input: 5,
791
+ output: 20,
792
+ cached: 2.5
793
+ },
794
+ audio: {
795
+ input: 40,
796
+ output: 80
797
+ }
798
+ },
799
+ metadata: {
800
+ knowledgeCutoff: "oct 01, 2023",
801
+ intelligence: "high",
802
+ speed: "fast"
803
+ }
804
+ }, {
805
+ id: 'gpt-4o-audio-preview-2024-12-17',
806
+ provider: 'openai',
807
+ name: 'gpt-4o-audio-preview-2024-12-17',
808
+ contextWindow: 128000,
809
+ maxOutputTokens: 16384,
810
+ tier: 'legacy',
811
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
812
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
813
+ pricing: {
814
+ text: {
815
+ input: 2.5,
816
+ output: 10
817
+ },
818
+ audio: {
819
+ input: 40,
820
+ output: 80
821
+ }
822
+ },
823
+ metadata: {
824
+ knowledgeCutoff: "oct 01, 2023",
825
+ performance: "high",
826
+ speed: "medium"
827
+ }
828
+ }, {
829
+ id: 'gpt-4o-mini-realtime-preview-2024-12-17',
830
+ provider: 'openai',
831
+ name: 'gpt-4o-mini-realtime-preview-2024-12-17',
832
+ contextWindow: 16000,
833
+ maxOutputTokens: 4096,
834
+ tier: 'efficient',
835
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
836
+ pricing: {
837
+ text: {
838
+ input: 0.6,
839
+ output: 2.4,
840
+ cached: 0.3
841
+ },
842
+ audio: {
843
+ input: 10,
844
+ output: 20
845
+ }
846
+ },
847
+ metadata: {
848
+ knowledgeCutoff: "oct 01, 2023",
849
+ intelligence: "average",
850
+ speed: "very fast"
851
+ }
852
+ }, {
853
+ id: 'gpt-4o-mini-audio-preview-2024-12-17',
854
+ provider: 'openai',
855
+ name: 'gpt-4o-mini-audio-preview-2024-12-17',
856
+ contextWindow: 128000,
857
+ maxOutputTokens: 16384,
858
+ tier: 'efficient',
859
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
860
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
861
+ pricing: {
862
+ text: {
863
+ input: 0.15,
864
+ output: 0.6
865
+ },
866
+ audio: {
867
+ input: 10,
868
+ output: 20
869
+ }
870
+ },
871
+ metadata: {
872
+ knowledgeCutoff: "oct 01, 2023",
873
+ performance: "average",
874
+ speed: "fast"
875
+ }
876
+ }, {
877
+ id: 'o1-2024-12-17',
878
+ provider: 'openai',
879
+ name: 'o1-2024-12-17',
880
+ contextWindow: 200000,
881
+ maxOutputTokens: 100000,
882
+ tier: 'flagship',
883
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
884
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
885
+ pricing: {
886
+ text: {
887
+ input: 15,
888
+ output: 60,
889
+ cached: 7.5
890
+ }
891
+ },
892
+ metadata: {
893
+ knowledgeCutoff: "oct 01, 2023",
894
+ reasoning: "higher",
895
+ speed: "slow"
896
+ }
897
+ }, {
898
+ id: 'o1',
899
+ provider: 'openai',
900
+ name: 'o1',
901
+ contextWindow: 200000,
902
+ maxOutputTokens: 100000,
903
+ tier: 'flagship',
904
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
905
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
906
+ pricing: {
907
+ text: {
908
+ input: 15,
909
+ output: 60,
910
+ cached: 7.5
911
+ }
912
+ },
913
+ metadata: {
914
+ knowledgeCutoff: "oct 01, 2023",
915
+ reasoning: "higher",
916
+ speed: "slow"
917
+ }
918
+ }, {
919
+ id: 'gpt-4o-mini-realtime-preview',
920
+ provider: 'openai',
921
+ name: 'gpt-4o-mini-realtime-preview',
922
+ contextWindow: 16000,
923
+ maxOutputTokens: 4096,
924
+ tier: 'efficient',
925
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
926
+ pricing: {
927
+ text: {
928
+ input: 0.6,
929
+ output: 2.4,
930
+ cached: 0.3
931
+ },
932
+ audio: {
933
+ input: 10,
934
+ output: 20
935
+ }
936
+ },
937
+ metadata: {
938
+ knowledgeCutoff: "oct 01, 2023",
939
+ intelligence: "average",
940
+ speed: "very fast"
941
+ }
942
+ }, {
943
+ id: 'gpt-4o-mini-audio-preview',
944
+ provider: 'openai',
945
+ name: 'gpt-4o-mini-audio-preview',
946
+ contextWindow: 128000,
947
+ maxOutputTokens: 16384,
948
+ tier: 'efficient',
949
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
950
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
951
+ pricing: {
952
+ text: {
953
+ input: 0.15,
954
+ output: 0.6
955
+ },
956
+ audio: {
957
+ input: 10,
958
+ output: 20
959
+ }
960
+ },
961
+ metadata: {
962
+ knowledgeCutoff: "oct 01, 2023",
963
+ performance: "average",
964
+ speed: "fast"
965
+ }
966
+ }, {
967
+ id: 'o3-mini',
968
+ provider: 'openai',
969
+ name: 'o3-mini',
970
+ contextWindow: 200000,
971
+ maxOutputTokens: 100000,
972
+ tier: 'flagship',
973
+ capabilities: new Set(['chat', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
974
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
975
+ pricing: {
976
+ text: {
977
+ input: 1.1,
978
+ output: 4.4,
979
+ cached: 0.55
980
+ }
981
+ },
982
+ metadata: {
983
+ knowledgeCutoff: "oct 01, 2023",
984
+ reasoning: "higher",
985
+ speed: "medium"
986
+ }
987
+ }, {
988
+ id: 'o3-mini-2025-01-31',
989
+ provider: 'openai',
990
+ name: 'o3-mini-2025-01-31',
991
+ contextWindow: 200000,
992
+ maxOutputTokens: 100000,
993
+ tier: 'flagship',
994
+ capabilities: new Set(['chat', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
995
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
996
+ pricing: {
997
+ text: {
998
+ input: 1.1,
999
+ output: 4.4,
1000
+ cached: 0.55
1001
+ }
1002
+ },
1003
+ metadata: {
1004
+ knowledgeCutoff: "oct 01, 2023",
1005
+ reasoning: "higher",
1006
+ speed: "medium"
1007
+ }
1008
+ }, {
1009
+ id: 'gpt-4o-2024-11-20',
1010
+ provider: 'openai',
1011
+ name: 'gpt-4o-2024-11-20',
1012
+ contextWindow: 128000,
1013
+ maxOutputTokens: 16384,
1014
+ tier: 'legacy',
1015
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1016
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1017
+ pricing: {
1018
+ text: {
1019
+ input: 2.5,
1020
+ output: 10,
1021
+ cached: 1.25
1022
+ }
1023
+ },
1024
+ metadata: {
1025
+ knowledgeCutoff: "oct 01, 2023",
1026
+ performance: "high",
1027
+ speed: "medium"
1028
+ }
1029
+ }, {
1030
+ id: 'gpt-4o-search-preview-2025-03-11',
1031
+ provider: 'openai',
1032
+ name: 'gpt-4o-search-preview-2025-03-11',
1033
+ contextWindow: 128000,
1034
+ maxOutputTokens: 16384,
1035
+ tier: 'legacy',
1036
+ capabilities: new Set(['chat', 'streaming', 'structured', 'json']),
1037
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'responseFormat', 'structuredOutput', 'stop']),
1038
+ pricing: {
1039
+ text: {
1040
+ input: 2.5,
1041
+ output: 10
1042
+ }
1043
+ },
1044
+ metadata: {
1045
+ knowledgeCutoff: "oct 01, 2023",
1046
+ intelligence: "high",
1047
+ speed: "medium"
1048
+ }
1049
+ }, {
1050
+ id: 'gpt-4o-search-preview',
1051
+ provider: 'openai',
1052
+ name: 'gpt-4o-search-preview',
1053
+ contextWindow: 128000,
1054
+ maxOutputTokens: 16384,
1055
+ tier: 'legacy',
1056
+ capabilities: new Set(['chat', 'streaming', 'structured', 'json']),
1057
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'responseFormat', 'structuredOutput', 'stop']),
1058
+ pricing: {
1059
+ text: {
1060
+ input: 2.5,
1061
+ output: 10
1062
+ }
1063
+ },
1064
+ metadata: {
1065
+ knowledgeCutoff: "oct 01, 2023",
1066
+ intelligence: "high",
1067
+ speed: "medium"
1068
+ }
1069
+ }, {
1070
+ id: 'gpt-4o-mini-search-preview-2025-03-11',
1071
+ provider: 'openai',
1072
+ name: 'gpt-4o-mini-search-preview-2025-03-11',
1073
+ contextWindow: 128000,
1074
+ maxOutputTokens: 16384,
1075
+ tier: 'efficient',
1076
+ capabilities: new Set(['chat', 'streaming', 'structured', 'json']),
1077
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'responseFormat', 'structuredOutput', 'stop']),
1078
+ pricing: {
1079
+ text: {
1080
+ input: 0.15,
1081
+ output: 0.6
1082
+ }
1083
+ },
1084
+ metadata: {
1085
+ knowledgeCutoff: "oct 01, 2023",
1086
+ intelligence: "average",
1087
+ speed: "fast"
1088
+ }
1089
+ }, {
1090
+ id: 'gpt-4o-mini-search-preview',
1091
+ provider: 'openai',
1092
+ name: 'gpt-4o-mini-search-preview',
1093
+ contextWindow: 128000,
1094
+ maxOutputTokens: 16384,
1095
+ tier: 'efficient',
1096
+ capabilities: new Set(['chat', 'streaming', 'structured', 'json']),
1097
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'responseFormat', 'structuredOutput', 'stop']),
1098
+ pricing: {
1099
+ text: {
1100
+ input: 0.15,
1101
+ output: 0.6
1102
+ }
1103
+ },
1104
+ metadata: {
1105
+ knowledgeCutoff: "oct 01, 2023",
1106
+ intelligence: "average",
1107
+ speed: "fast"
1108
+ }
1109
+ }, {
1110
+ id: 'gpt-4o-transcribe',
1111
+ provider: 'openai',
1112
+ name: 'gpt-4o-transcribe',
1113
+ contextWindow: 16000,
1114
+ maxOutputTokens: 2000,
1115
+ tier: 'flagship',
1116
+ capabilities: new Set(['hearing', 'chat']),
1117
+ supportedParameters: new Set(['transcribePrompt', 'transcribeStream']),
1118
+ pricing: {
1119
+ text: {
1120
+ input: 2.5,
1121
+ output: 10
1122
+ },
1123
+ audio: {
1124
+ input: 6
1125
+ }
1126
+ },
1127
+ metadata: {
1128
+ knowledgeCutoff: "jun 01, 2024",
1129
+ intelligence: "higher",
1130
+ speed: "medium"
1131
+ }
1132
+ }, {
1133
+ id: 'gpt-4o-mini-transcribe',
1134
+ provider: 'openai',
1135
+ name: 'gpt-4o-mini-transcribe',
1136
+ contextWindow: 16000,
1137
+ maxOutputTokens: 2000,
1138
+ tier: 'efficient',
1139
+ capabilities: new Set(['hearing', 'chat']),
1140
+ supportedParameters: new Set(['transcribePrompt', 'transcribeStream']),
1141
+ pricing: {
1142
+ text: {
1143
+ input: 1.25,
1144
+ output: 5
1145
+ },
1146
+ audio: {
1147
+ input: 3
1148
+ }
1149
+ },
1150
+ metadata: {
1151
+ knowledgeCutoff: "jun 01, 2024",
1152
+ intelligence: "high",
1153
+ speed: "fast"
1154
+ }
1155
+ }, {
1156
+ id: 'o1-pro-2025-03-19',
1157
+ provider: 'openai',
1158
+ name: 'o1-pro-2025-03-19',
1159
+ contextWindow: 200000,
1160
+ maxOutputTokens: 100000,
1161
+ tier: 'flagship',
1162
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
1163
+ pricing: {
1164
+ text: {
1165
+ input: 150,
1166
+ output: 600
1167
+ }
1168
+ },
1169
+ metadata: {
1170
+ knowledgeCutoff: "oct 01, 2023",
1171
+ reasoning: "higher",
1172
+ speed: "slow"
1173
+ }
1174
+ }, {
1175
+ id: 'o1-pro',
1176
+ provider: 'openai',
1177
+ name: 'o1-pro',
1178
+ contextWindow: 200000,
1179
+ maxOutputTokens: 100000,
1180
+ tier: 'flagship',
1181
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
1182
+ pricing: {
1183
+ text: {
1184
+ input: 150,
1185
+ output: 600
1186
+ }
1187
+ },
1188
+ metadata: {
1189
+ knowledgeCutoff: "oct 01, 2023",
1190
+ reasoning: "higher",
1191
+ speed: "slow"
1192
+ }
1193
+ }, {
1194
+ id: 'gpt-4o-mini-tts',
1195
+ provider: 'openai',
1196
+ name: 'gpt-4o-mini-tts',
1197
+ contextWindow: 0,
1198
+ tier: 'flagship',
1199
+ capabilities: new Set(['audio', 'chat']),
1200
+ supportedParameters: new Set(['speechInstructions']),
1201
+ pricing: {
1202
+ text: {
1203
+ input: 0.6
1204
+ },
1205
+ audio: {
1206
+ output: 12
1207
+ }
1208
+ },
1209
+ metadata: {
1210
+ intelligence: "higher",
1211
+ speed: "fast"
1212
+ }
1213
+ }, {
1214
+ id: 'o3-2025-04-16',
1215
+ provider: 'openai',
1216
+ name: 'o3-2025-04-16',
1217
+ contextWindow: 200000,
1218
+ maxOutputTokens: 100000,
1219
+ tier: 'flagship',
1220
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1221
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
1222
+ pricing: {
1223
+ text: {
1224
+ input: 1,
1225
+ output: 4,
1226
+ cached: 0.25
1227
+ }
1228
+ },
1229
+ metadata: {
1230
+ knowledgeCutoff: "jun 01, 2024",
1231
+ reasoning: "highest",
1232
+ speed: "slow"
1233
+ }
1234
+ }, {
1235
+ id: 'o4-mini-2025-04-16',
1236
+ provider: 'openai',
1237
+ name: 'o4-mini-2025-04-16',
1238
+ contextWindow: 200000,
1239
+ maxOutputTokens: 100000,
1240
+ tier: 'flagship',
1241
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1242
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
1243
+ pricing: {
1244
+ text: {
1245
+ input: 1.1,
1246
+ output: 4.4,
1247
+ cached: 0.275
1248
+ }
1249
+ },
1250
+ metadata: {
1251
+ knowledgeCutoff: "jun 01, 2024",
1252
+ reasoning: "higher",
1253
+ speed: "medium"
1254
+ }
1255
+ }, {
1256
+ id: 'o3',
1257
+ provider: 'openai',
1258
+ name: 'o3',
1259
+ contextWindow: 200000,
1260
+ maxOutputTokens: 100000,
1261
+ tier: 'flagship',
1262
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1263
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
1264
+ pricing: {
1265
+ text: {
1266
+ input: 1,
1267
+ output: 4,
1268
+ cached: 0.25
1269
+ }
1270
+ },
1271
+ metadata: {
1272
+ knowledgeCutoff: "jun 01, 2024",
1273
+ reasoning: "highest",
1274
+ speed: "slow"
1275
+ }
1276
+ }, {
1277
+ id: 'o4-mini',
1278
+ provider: 'openai',
1279
+ name: 'o4-mini',
1280
+ contextWindow: 200000,
1281
+ maxOutputTokens: 100000,
1282
+ tier: 'flagship',
1283
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1284
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput']),
1285
+ pricing: {
1286
+ text: {
1287
+ input: 1.1,
1288
+ output: 4.4,
1289
+ cached: 0.275
1290
+ }
1291
+ },
1292
+ metadata: {
1293
+ knowledgeCutoff: "jun 01, 2024",
1294
+ reasoning: "higher",
1295
+ speed: "medium"
1296
+ }
1297
+ }, {
1298
+ id: 'gpt-4.1-2025-04-14',
1299
+ provider: 'openai',
1300
+ name: 'gpt-4.1-2025-04-14',
1301
+ contextWindow: 1047576,
1302
+ maxOutputTokens: 32768,
1303
+ tier: 'flagship',
1304
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1305
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1306
+ pricing: {
1307
+ text: {
1308
+ input: 2,
1309
+ output: 8,
1310
+ cached: 0.5
1311
+ }
1312
+ },
1313
+ metadata: {
1314
+ knowledgeCutoff: "jun 01, 2024",
1315
+ performance: "higher",
1316
+ speed: "medium"
1317
+ }
1318
+ }, {
1319
+ id: 'gpt-4.1',
1320
+ provider: 'openai',
1321
+ name: 'gpt-4.1',
1322
+ contextWindow: 1047576,
1323
+ maxOutputTokens: 32768,
1324
+ tier: 'flagship',
1325
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1326
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1327
+ pricing: {
1328
+ text: {
1329
+ input: 2,
1330
+ output: 8,
1331
+ cached: 0.5
1332
+ }
1333
+ },
1334
+ metadata: {
1335
+ knowledgeCutoff: "jun 01, 2024",
1336
+ performance: "higher",
1337
+ speed: "medium"
1338
+ }
1339
+ }, {
1340
+ id: 'gpt-4.1-mini-2025-04-14',
1341
+ provider: 'openai',
1342
+ name: 'gpt-4.1-mini-2025-04-14',
1343
+ contextWindow: 1047576,
1344
+ maxOutputTokens: 32768,
1345
+ tier: 'efficient',
1346
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1347
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1348
+ pricing: {
1349
+ text: {
1350
+ input: 0.4,
1351
+ output: 1.6,
1352
+ cached: 0.1
1353
+ }
1354
+ },
1355
+ metadata: {
1356
+ knowledgeCutoff: "jun 01, 2024",
1357
+ performance: "high",
1358
+ speed: "fast"
1359
+ }
1360
+ }, {
1361
+ id: 'gpt-4.1-mini',
1362
+ provider: 'openai',
1363
+ name: 'gpt-4.1-mini',
1364
+ contextWindow: 1047576,
1365
+ maxOutputTokens: 32768,
1366
+ tier: 'efficient',
1367
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1368
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1369
+ pricing: {
1370
+ text: {
1371
+ input: 0.4,
1372
+ output: 1.6,
1373
+ cached: 0.1
1374
+ }
1375
+ },
1376
+ metadata: {
1377
+ knowledgeCutoff: "jun 01, 2024",
1378
+ performance: "high",
1379
+ speed: "fast"
1380
+ }
1381
+ }, {
1382
+ id: 'gpt-4.1-nano-2025-04-14',
1383
+ provider: 'openai',
1384
+ name: 'gpt-4.1-nano-2025-04-14',
1385
+ contextWindow: 1047576,
1386
+ maxOutputTokens: 32768,
1387
+ tier: 'efficient',
1388
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1389
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1390
+ pricing: {
1391
+ text: {
1392
+ input: 0.1,
1393
+ output: 0.4,
1394
+ cached: 0.025
1395
+ }
1396
+ },
1397
+ metadata: {
1398
+ knowledgeCutoff: "jun 01, 2024",
1399
+ performance: "average",
1400
+ speed: "very fast"
1401
+ }
1402
+ }, {
1403
+ id: 'gpt-4.1-nano',
1404
+ provider: 'openai',
1405
+ name: 'gpt-4.1-nano',
1406
+ contextWindow: 1047576,
1407
+ maxOutputTokens: 32768,
1408
+ tier: 'efficient',
1409
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1410
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1411
+ pricing: {
1412
+ text: {
1413
+ input: 0.1,
1414
+ output: 0.4,
1415
+ cached: 0.025
1416
+ }
1417
+ },
1418
+ metadata: {
1419
+ knowledgeCutoff: "jun 01, 2024",
1420
+ performance: "average",
1421
+ speed: "very fast"
1422
+ }
1423
+ }, {
1424
+ id: 'gpt-image-1',
1425
+ provider: 'openai',
1426
+ name: 'gpt-image-1',
1427
+ contextWindow: 0,
1428
+ tier: 'flagship',
1429
+ capabilities: new Set(['vision', 'image', 'chat']),
1430
+ supportedParameters: new Set(['imageStyle', 'imageMultiple', 'imageBackground', 'imageStream', 'imageFormat']),
1431
+ pricing: {
1432
+ text: {
1433
+ input: 5,
1434
+ cached: 1.25
1435
+ },
1436
+ image: {
1437
+ input: 10,
1438
+ output: [
1439
+ {
1440
+ quality: "low",
1441
+ sizes: [
1442
+ {
1443
+ width: 1024,
1444
+ height: 1024,
1445
+ cost: 0.011
1446
+ },
1447
+ {
1448
+ width: 1024,
1449
+ height: 1536,
1450
+ cost: 0.016
1451
+ },
1452
+ {
1453
+ width: 1536,
1454
+ height: 1024,
1455
+ cost: 0.016
1456
+ }
1457
+ ]
1458
+ },
1459
+ {
1460
+ quality: "medium",
1461
+ sizes: [
1462
+ {
1463
+ width: 1024,
1464
+ height: 1024,
1465
+ cost: 0.042
1466
+ },
1467
+ {
1468
+ width: 1024,
1469
+ height: 1536,
1470
+ cost: 0.063
1471
+ },
1472
+ {
1473
+ width: 1536,
1474
+ height: 1024,
1475
+ cost: 0.063
1476
+ }
1477
+ ]
1478
+ },
1479
+ {
1480
+ quality: "high",
1481
+ sizes: [
1482
+ {
1483
+ width: 1024,
1484
+ height: 1024,
1485
+ cost: 0.167
1486
+ },
1487
+ {
1488
+ width: 1024,
1489
+ height: 1536,
1490
+ cost: 0.25
1491
+ },
1492
+ {
1493
+ width: 1536,
1494
+ height: 1024,
1495
+ cost: 0.25
1496
+ }
1497
+ ]
1498
+ }
1499
+ ]
1500
+ }
1501
+ },
1502
+ metadata: {
1503
+ intelligence: "higher",
1504
+ speed: "slow"
1505
+ }
1506
+ }, {
1507
+ id: 'codex-mini-latest',
1508
+ provider: 'openai',
1509
+ name: 'codex-mini-latest',
1510
+ contextWindow: 200000,
1511
+ maxOutputTokens: 100000,
1512
+ tier: 'flagship',
1513
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1514
+ pricing: {
1515
+ text: {
1516
+ input: 1.5,
1517
+ output: 6,
1518
+ cached: 0.375
1519
+ }
1520
+ },
1521
+ metadata: {
1522
+ knowledgeCutoff: "jun 01, 2024",
1523
+ intelligence: "higher",
1524
+ speed: "medium"
1525
+ }
1526
+ }, {
1527
+ id: 'o3-pro',
1528
+ provider: 'openai',
1529
+ name: 'o3-pro',
1530
+ contextWindow: 200000,
1531
+ maxOutputTokens: 100000,
1532
+ tier: 'flagship',
1533
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
1534
+ pricing: {
1535
+ text: {
1536
+ input: 20,
1537
+ output: 80
1538
+ }
1539
+ },
1540
+ metadata: {
1541
+ knowledgeCutoff: "jun 01, 2024",
1542
+ reasoning: "highest",
1543
+ speed: "slow"
1544
+ }
1545
+ }, {
1546
+ id: 'gpt-4o-realtime-preview-2025-06-03',
1547
+ provider: 'openai',
1548
+ name: 'gpt-4o-realtime-preview-2025-06-03',
1549
+ contextWindow: 32000,
1550
+ maxOutputTokens: 4096,
1551
+ tier: 'efficient',
1552
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
1553
+ pricing: {
1554
+ text: {
1555
+ input: 5,
1556
+ output: 20,
1557
+ cached: 2.5
1558
+ },
1559
+ audio: {
1560
+ input: 40,
1561
+ output: 80
1562
+ }
1563
+ },
1564
+ metadata: {
1565
+ knowledgeCutoff: "oct 01, 2023",
1566
+ intelligence: "high",
1567
+ speed: "fast"
1568
+ }
1569
+ }, {
1570
+ id: 'gpt-4o-audio-preview-2025-06-03',
1571
+ provider: 'openai',
1572
+ name: 'gpt-4o-audio-preview-2025-06-03',
1573
+ contextWindow: 128000,
1574
+ maxOutputTokens: 16384,
1575
+ tier: 'legacy',
1576
+ capabilities: new Set(['hearing', 'audio', 'chat', 'streaming', 'tools']),
1577
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
1578
+ pricing: {
1579
+ text: {
1580
+ input: 2.5,
1581
+ output: 10
1582
+ },
1583
+ audio: {
1584
+ input: 40,
1585
+ output: 80
1586
+ }
1587
+ },
1588
+ metadata: {
1589
+ knowledgeCutoff: "oct 01, 2023",
1590
+ performance: "high",
1591
+ speed: "medium"
1592
+ }
1593
+ }, {
1594
+ id: 'o3-pro-2025-06-10',
1595
+ provider: 'openai',
1596
+ name: 'o3-pro-2025-06-10',
1597
+ contextWindow: 200000,
1598
+ maxOutputTokens: 100000,
1599
+ tier: 'flagship',
1600
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
1601
+ pricing: {
1602
+ text: {
1603
+ input: 20,
1604
+ output: 80
1605
+ }
1606
+ },
1607
+ metadata: {
1608
+ knowledgeCutoff: "jun 01, 2024",
1609
+ reasoning: "highest",
1610
+ speed: "slow"
1611
+ }
1612
+ }, {
1613
+ id: 'o4-mini-deep-research',
1614
+ provider: 'openai',
1615
+ name: 'o4-mini-deep-research',
1616
+ contextWindow: 200000,
1617
+ maxOutputTokens: 100000,
1618
+ tier: 'flagship',
1619
+ capabilities: new Set(['chat', 'vision', 'streaming', 'reasoning']),
1620
+ pricing: {
1621
+ text: {
1622
+ input: 2,
1623
+ output: 8,
1624
+ cached: 0.5
1625
+ }
1626
+ },
1627
+ metadata: {
1628
+ knowledgeCutoff: "jun 01, 2024",
1629
+ reasoning: "higher",
1630
+ speed: "medium"
1631
+ }
1632
+ }, {
1633
+ id: 'o3-deep-research',
1634
+ provider: 'openai',
1635
+ name: 'o3-deep-research',
1636
+ contextWindow: 200000,
1637
+ maxOutputTokens: 100000,
1638
+ tier: 'flagship',
1639
+ capabilities: new Set(['chat', 'vision', 'streaming', 'reasoning']),
1640
+ pricing: {
1641
+ text: {
1642
+ input: 10,
1643
+ output: 40,
1644
+ cached: 2.5
1645
+ }
1646
+ },
1647
+ metadata: {
1648
+ knowledgeCutoff: "jun 01, 2024",
1649
+ reasoning: "highest",
1650
+ speed: "slow"
1651
+ }
1652
+ }, {
1653
+ id: 'gpt-4o-transcribe-diarize',
1654
+ provider: 'openai',
1655
+ name: 'gpt-4o-transcribe-diarize',
1656
+ contextWindow: 16000,
1657
+ maxOutputTokens: 2000,
1658
+ tier: 'flagship',
1659
+ capabilities: new Set(['hearing', 'chat']),
1660
+ supportedParameters: new Set(['transcribeStream']),
1661
+ pricing: {
1662
+ text: {
1663
+ input: 2.5,
1664
+ output: 10
1665
+ },
1666
+ audio: {
1667
+ input: 6
1668
+ }
1669
+ },
1670
+ metadata: {
1671
+ knowledgeCutoff: "jun 01, 2024",
1672
+ intelligence: "higher",
1673
+ speed: "medium"
1674
+ }
1675
+ }, {
1676
+ id: 'o3-deep-research-2025-06-26',
1677
+ provider: 'openai',
1678
+ name: 'o3-deep-research-2025-06-26',
1679
+ contextWindow: 200000,
1680
+ maxOutputTokens: 100000,
1681
+ tier: 'flagship',
1682
+ capabilities: new Set(['chat', 'vision', 'streaming', 'reasoning']),
1683
+ pricing: {
1684
+ text: {
1685
+ input: 10,
1686
+ output: 40,
1687
+ cached: 2.5
1688
+ }
1689
+ },
1690
+ metadata: {
1691
+ knowledgeCutoff: "jun 01, 2024",
1692
+ reasoning: "highest",
1693
+ speed: "slow"
1694
+ }
1695
+ }, {
1696
+ id: 'o4-mini-deep-research-2025-06-26',
1697
+ provider: 'openai',
1698
+ name: 'o4-mini-deep-research-2025-06-26',
1699
+ contextWindow: 200000,
1700
+ maxOutputTokens: 100000,
1701
+ tier: 'flagship',
1702
+ capabilities: new Set(['chat', 'vision', 'streaming', 'reasoning']),
1703
+ pricing: {
1704
+ text: {
1705
+ input: 2,
1706
+ output: 8,
1707
+ cached: 0.5
1708
+ }
1709
+ },
1710
+ metadata: {
1711
+ knowledgeCutoff: "jun 01, 2024",
1712
+ reasoning: "higher",
1713
+ speed: "medium"
1714
+ }
1715
+ }, {
1716
+ id: 'gpt-5-chat-latest',
1717
+ provider: 'openai',
1718
+ name: 'gpt-5-chat-latest',
1719
+ contextWindow: 128000,
1720
+ maxOutputTokens: 16384,
1721
+ tier: 'legacy',
1722
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json']),
1723
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1724
+ pricing: {
1725
+ text: {
1726
+ input: 1.25,
1727
+ output: 10,
1728
+ cached: 0.125
1729
+ }
1730
+ },
1731
+ metadata: {
1732
+ knowledgeCutoff: "sep 30, 2024",
1733
+ performance: "high",
1734
+ speed: "medium"
1735
+ }
1736
+ }, {
1737
+ id: 'gpt-5-2025-08-07',
1738
+ provider: 'openai',
1739
+ name: 'gpt-5-2025-08-07',
1740
+ contextWindow: 400000,
1741
+ maxOutputTokens: 128000,
1742
+ tier: 'flagship',
1743
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1744
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1745
+ pricing: {
1746
+ text: {
1747
+ input: 1.25,
1748
+ output: 10,
1749
+ cached: 0.125
1750
+ }
1751
+ },
1752
+ metadata: {
1753
+ knowledgeCutoff: "sep 30, 2024",
1754
+ reasoning: "higher",
1755
+ speed: "medium"
1756
+ }
1757
+ }, {
1758
+ id: 'gpt-5',
1759
+ provider: 'openai',
1760
+ name: 'gpt-5',
1761
+ contextWindow: 400000,
1762
+ maxOutputTokens: 128000,
1763
+ tier: 'flagship',
1764
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1765
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1766
+ pricing: {
1767
+ text: {
1768
+ input: 1.25,
1769
+ output: 10,
1770
+ cached: 0.125
1771
+ }
1772
+ },
1773
+ metadata: {
1774
+ knowledgeCutoff: "sep 30, 2024",
1775
+ reasoning: "higher",
1776
+ speed: "medium"
1777
+ }
1778
+ }, {
1779
+ id: 'gpt-5-mini-2025-08-07',
1780
+ provider: 'openai',
1781
+ name: 'gpt-5-mini-2025-08-07',
1782
+ contextWindow: 400000,
1783
+ maxOutputTokens: 128000,
1784
+ tier: 'efficient',
1785
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1786
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1787
+ pricing: {
1788
+ text: {
1789
+ input: 0.25,
1790
+ output: 2,
1791
+ cached: 0.025
1792
+ }
1793
+ },
1794
+ metadata: {
1795
+ knowledgeCutoff: "may 31, 2024",
1796
+ reasoning: "high",
1797
+ speed: "fast"
1798
+ }
1799
+ }, {
1800
+ id: 'gpt-5-mini',
1801
+ provider: 'openai',
1802
+ name: 'gpt-5-mini',
1803
+ contextWindow: 400000,
1804
+ maxOutputTokens: 128000,
1805
+ tier: 'efficient',
1806
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1807
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop']),
1808
+ pricing: {
1809
+ text: {
1810
+ input: 0.25,
1811
+ output: 2,
1812
+ cached: 0.025
1813
+ }
1814
+ },
1815
+ metadata: {
1816
+ knowledgeCutoff: "may 31, 2024",
1817
+ reasoning: "high",
1818
+ speed: "fast"
1819
+ }
1820
+ }, {
1821
+ id: 'gpt-5-nano-2025-08-07',
1822
+ provider: 'openai',
1823
+ name: 'gpt-5-nano-2025-08-07',
1824
+ contextWindow: 400000,
1825
+ maxOutputTokens: 128000,
1826
+ tier: 'efficient',
1827
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1828
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1829
+ pricing: {
1830
+ text: {
1831
+ input: 0.05,
1832
+ output: 0.4,
1833
+ cached: 0.005
1834
+ }
1835
+ },
1836
+ metadata: {
1837
+ knowledgeCutoff: "may 31, 2024",
1838
+ reasoning: "average",
1839
+ speed: "very fast"
1840
+ }
1841
+ }, {
1842
+ id: 'gpt-5-nano',
1843
+ provider: 'openai',
1844
+ name: 'gpt-5-nano',
1845
+ contextWindow: 400000,
1846
+ maxOutputTokens: 128000,
1847
+ tier: 'efficient',
1848
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1849
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'reason', 'tools', 'toolChoice', 'responseFormat', 'structuredOutput', 'stop', 'imageStyle', 'imageMultiple']),
1850
+ pricing: {
1851
+ text: {
1852
+ input: 0.05,
1853
+ output: 0.4,
1854
+ cached: 0.005
1855
+ }
1856
+ },
1857
+ metadata: {
1858
+ knowledgeCutoff: "may 31, 2024",
1859
+ reasoning: "average",
1860
+ speed: "very fast"
1861
+ }
1862
+ }, {
1863
+ id: 'gpt-audio-2025-08-28',
1864
+ provider: 'openai',
1865
+ name: 'gpt-audio-2025-08-28',
1866
+ contextWindow: 128000,
1867
+ maxOutputTokens: 16384,
1868
+ tier: 'flagship',
1869
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
1870
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
1871
+ pricing: {
1872
+ text: {
1873
+ input: 2.5,
1874
+ output: 10
1875
+ },
1876
+ audio: {
1877
+ input: 32,
1878
+ output: 64
1879
+ }
1880
+ },
1881
+ metadata: {
1882
+ knowledgeCutoff: "oct 01, 2023",
1883
+ performance: "higher",
1884
+ speed: "medium"
1885
+ }
1886
+ }, {
1887
+ id: 'gpt-realtime',
1888
+ provider: 'openai',
1889
+ name: 'gpt-realtime',
1890
+ contextWindow: 32000,
1891
+ maxOutputTokens: 4096,
1892
+ tier: 'flagship',
1893
+ capabilities: new Set(['vision', 'hearing', 'audio', 'chat', 'tools']),
1894
+ pricing: {
1895
+ text: {
1896
+ input: 4,
1897
+ output: 16,
1898
+ cached: 0.5
1899
+ },
1900
+ audio: {
1901
+ input: 32,
1902
+ output: 64
1903
+ },
1904
+ image: {
1905
+ input: 5
1906
+ }
1907
+ },
1908
+ metadata: {
1909
+ knowledgeCutoff: "oct 01, 2023",
1910
+ intelligence: "highest",
1911
+ speed: "fast"
1912
+ }
1913
+ }, {
1914
+ id: 'gpt-realtime-2025-08-28',
1915
+ provider: 'openai',
1916
+ name: 'gpt-realtime-2025-08-28',
1917
+ contextWindow: 32000,
1918
+ maxOutputTokens: 4096,
1919
+ tier: 'flagship',
1920
+ capabilities: new Set(['vision', 'hearing', 'audio', 'chat', 'tools']),
1921
+ pricing: {
1922
+ text: {
1923
+ input: 4,
1924
+ output: 16,
1925
+ cached: 0.5
1926
+ },
1927
+ audio: {
1928
+ input: 32,
1929
+ output: 64
1930
+ },
1931
+ image: {
1932
+ input: 5
1933
+ }
1934
+ },
1935
+ metadata: {
1936
+ knowledgeCutoff: "oct 01, 2023",
1937
+ intelligence: "highest",
1938
+ speed: "fast"
1939
+ }
1940
+ }, {
1941
+ id: 'gpt-audio',
1942
+ provider: 'openai',
1943
+ name: 'gpt-audio',
1944
+ contextWindow: 128000,
1945
+ maxOutputTokens: 16384,
1946
+ tier: 'flagship',
1947
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
1948
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
1949
+ pricing: {
1950
+ text: {
1951
+ input: 2.5,
1952
+ output: 10
1953
+ },
1954
+ audio: {
1955
+ input: 32,
1956
+ output: 64
1957
+ }
1958
+ },
1959
+ metadata: {
1960
+ knowledgeCutoff: "oct 01, 2023",
1961
+ performance: "higher",
1962
+ speed: "medium"
1963
+ }
1964
+ }, {
1965
+ id: 'gpt-5-codex',
1966
+ provider: 'openai',
1967
+ name: 'gpt-5-codex',
1968
+ contextWindow: 400000,
1969
+ maxOutputTokens: 128000,
1970
+ tier: 'flagship',
1971
+ capabilities: new Set(['chat', 'vision', 'streaming', 'tools', 'structured', 'json', 'reasoning']),
1972
+ pricing: {
1973
+ text: {
1974
+ input: 1.25,
1975
+ output: 10,
1976
+ cached: 0.125
1977
+ }
1978
+ },
1979
+ metadata: {
1980
+ knowledgeCutoff: "sep 30, 2024",
1981
+ reasoning: "higher",
1982
+ speed: "medium"
1983
+ }
1984
+ }, {
1985
+ id: 'gpt-image-1-mini',
1986
+ provider: 'openai',
1987
+ name: 'gpt-image-1-mini',
1988
+ contextWindow: 0,
1989
+ tier: 'flagship',
1990
+ capabilities: new Set(['vision', 'image', 'chat']),
1991
+ supportedParameters: new Set(['imageStyle', 'imageMultiple', 'imageBackground', 'imageStream', 'imageFormat']),
1992
+ pricing: {
1993
+ text: {
1994
+ input: 2,
1995
+ cached: 0.2
1996
+ },
1997
+ image: {
1998
+ input: 2.5,
1999
+ output: [
2000
+ {
2001
+ quality: "low",
2002
+ sizes: [
2003
+ {
2004
+ width: 1024,
2005
+ height: 1024,
2006
+ cost: 0.005
2007
+ },
2008
+ {
2009
+ width: 1024,
2010
+ height: 1536,
2011
+ cost: 0.006
2012
+ },
2013
+ {
2014
+ width: 1536,
2015
+ height: 1024,
2016
+ cost: 0.006
2017
+ }
2018
+ ]
2019
+ },
2020
+ {
2021
+ quality: "medium",
2022
+ sizes: [
2023
+ {
2024
+ width: 1024,
2025
+ height: 1024,
2026
+ cost: 0.011
2027
+ },
2028
+ {
2029
+ width: 1024,
2030
+ height: 1536,
2031
+ cost: 0.015
2032
+ },
2033
+ {
2034
+ width: 1536,
2035
+ height: 1024,
2036
+ cost: 0.015
2037
+ }
2038
+ ]
2039
+ },
2040
+ {
2041
+ quality: "high",
2042
+ sizes: [
2043
+ {
2044
+ width: 1024,
2045
+ height: 1024,
2046
+ cost: 0.036
2047
+ },
2048
+ {
2049
+ width: 1024,
2050
+ height: 1536,
2051
+ cost: 0.052
2052
+ },
2053
+ {
2054
+ width: 1536,
2055
+ height: 1024,
2056
+ cost: 0.052
2057
+ }
2058
+ ]
2059
+ }
2060
+ ]
2061
+ }
2062
+ },
2063
+ metadata: {
2064
+ intelligence: "higher",
2065
+ speed: "slow"
2066
+ }
2067
+ }, {
2068
+ id: 'gpt-5-pro-2025-10-06',
2069
+ provider: 'openai',
2070
+ name: 'gpt-5-pro-2025-10-06',
2071
+ contextWindow: 400000,
2072
+ maxOutputTokens: 272000,
2073
+ tier: 'flagship',
2074
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
2075
+ supportedParameters: new Set(['imageStyle', 'imageMultiple']),
2076
+ pricing: {
2077
+ text: {
2078
+ input: 15,
2079
+ output: 120
2080
+ }
2081
+ },
2082
+ metadata: {
2083
+ knowledgeCutoff: "sep 30, 2024",
2084
+ reasoning: "highest",
2085
+ speed: "slow"
2086
+ }
2087
+ }, {
2088
+ id: 'gpt-5-pro',
2089
+ provider: 'openai',
2090
+ name: 'gpt-5-pro',
2091
+ contextWindow: 400000,
2092
+ maxOutputTokens: 272000,
2093
+ tier: 'flagship',
2094
+ capabilities: new Set(['chat', 'vision', 'tools', 'structured', 'json', 'reasoning']),
2095
+ supportedParameters: new Set(['imageStyle', 'imageMultiple']),
2096
+ pricing: {
2097
+ text: {
2098
+ input: 15,
2099
+ output: 120
2100
+ }
2101
+ },
2102
+ metadata: {
2103
+ knowledgeCutoff: "sep 30, 2024",
2104
+ reasoning: "highest",
2105
+ speed: "slow"
2106
+ }
2107
+ }, {
2108
+ id: 'gpt-audio-mini',
2109
+ provider: 'openai',
2110
+ name: 'gpt-audio-mini',
2111
+ contextWindow: 128000,
2112
+ maxOutputTokens: 16384,
2113
+ tier: 'flagship',
2114
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
2115
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
2116
+ pricing: {
2117
+ text: {
2118
+ input: 0.6,
2119
+ output: 2.4
2120
+ },
2121
+ audio: {
2122
+ input: 10,
2123
+ output: 20
2124
+ }
2125
+ },
2126
+ metadata: {
2127
+ knowledgeCutoff: "oct 01, 2023",
2128
+ performance: "higher",
2129
+ speed: "very fast"
2130
+ }
2131
+ }, {
2132
+ id: 'gpt-audio-mini-2025-10-06',
2133
+ provider: 'openai',
2134
+ name: 'gpt-audio-mini-2025-10-06',
2135
+ contextWindow: 128000,
2136
+ maxOutputTokens: 16384,
2137
+ tier: 'flagship',
2138
+ capabilities: new Set(['hearing', 'audio', 'chat', 'tools']),
2139
+ supportedParameters: new Set(['maxTokens', 'temperature', 'topP', 'frequencyPenalty', 'presencePenalty', 'logitBias', 'logProbabilities', 'tools', 'toolChoice', 'stop']),
2140
+ pricing: {
2141
+ text: {
2142
+ input: 0.6,
2143
+ output: 2.4
2144
+ },
2145
+ audio: {
2146
+ input: 10,
2147
+ output: 20
2148
+ }
2149
+ },
2150
+ metadata: {
2151
+ knowledgeCutoff: "oct 01, 2023",
2152
+ performance: "higher",
2153
+ speed: "very fast"
2154
+ }
2155
+ }, {
2156
+ id: 'tts-1',
2157
+ provider: 'openai',
2158
+ name: 'tts-1',
2159
+ contextWindow: 0,
2160
+ tier: 'efficient',
2161
+ capabilities: new Set(['audio', 'chat']),
2162
+ supportedParameters: new Set([]),
2163
+ pricing: {
2164
+ audio: {
2165
+ output: 15
2166
+ }
2167
+ },
2168
+ metadata: {
2169
+ intelligence: "average",
2170
+ speed: "fast"
2171
+ }
2172
+ }, {
2173
+ id: 'whisper-1',
2174
+ provider: 'openai',
2175
+ name: 'whisper-1',
2176
+ contextWindow: 0,
2177
+ tier: 'legacy',
2178
+ capabilities: new Set(['hearing']),
2179
+ supportedParameters: new Set(['transcribePrompt']),
2180
+ pricing: {
2181
+ audio: {
2182
+ input: 0.006
2183
+ }
2184
+ },
2185
+ metadata: {
2186
+ intelligence: "average",
2187
+ speed: "medium"
2188
+ }
2189
+ }, {
2190
+ id: 'text-embedding-ada-002',
2191
+ provider: 'openai',
2192
+ name: 'text-embedding-ada-002',
2193
+ contextWindow: 0,
2194
+ tier: 'legacy',
2195
+ capabilities: new Set(['embedding', 'chat']),
2196
+ supportedParameters: new Set([]),
2197
+ pricing: {
2198
+ embeddings: {
2199
+ cost: 0.1
2200
+ }
2201
+ },
2202
+ metadata: {
2203
+ intelligence: "low",
2204
+ speed: "slow"
2205
+ }
2206
+ }];
2207
+ //# sourceMappingURL=openai.js.map