@juspay/neurolink 8.8.0 → 8.10.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -4,7 +4,7 @@
4
4
  * Part of Phase 4.1 - Models Command System
5
5
  */
6
6
  import { DEFAULT_MODEL_ALIASES } from "../types/providers.js";
7
- import { AIProviderName, OpenAIModels, GoogleAIModels, AnthropicModels, } from "../constants/enums.js";
7
+ import { AIProviderName, OpenAIModels, AzureOpenAIModels, GoogleAIModels, AnthropicModels, BedrockModels, MistralModels, OllamaModels, } from "../constants/enums.js";
8
8
  /**
9
9
  * Comprehensive model registry
10
10
  */
@@ -98,12 +98,1610 @@ export const MODEL_REGISTRY = {
98
98
  releaseDate: "2024-07-18",
99
99
  category: "general",
100
100
  },
101
+ // OpenAI GPT-5 Series
102
+ [OpenAIModels.GPT_5]: {
103
+ id: OpenAIModels.GPT_5,
104
+ name: "GPT-5",
105
+ provider: AIProviderName.OPENAI,
106
+ description: "OpenAI's most advanced model with breakthrough reasoning and multimodal capabilities",
107
+ capabilities: {
108
+ vision: true,
109
+ functionCalling: true,
110
+ codeGeneration: true,
111
+ reasoning: true,
112
+ multimodal: true,
113
+ streaming: true,
114
+ jsonMode: true,
115
+ },
116
+ pricing: {
117
+ inputCostPer1K: 0.01,
118
+ outputCostPer1K: 0.03,
119
+ currency: "USD",
120
+ },
121
+ performance: {
122
+ speed: "medium",
123
+ quality: "high",
124
+ accuracy: "high",
125
+ },
126
+ limits: {
127
+ maxContextTokens: 256000,
128
+ maxOutputTokens: 32768,
129
+ maxRequestsPerMinute: 200,
130
+ },
131
+ useCases: {
132
+ coding: 10,
133
+ creative: 10,
134
+ analysis: 10,
135
+ conversation: 10,
136
+ reasoning: 10,
137
+ translation: 9,
138
+ summarization: 9,
139
+ },
140
+ aliases: ["gpt5", "gpt-5-flagship", "openai-latest"],
141
+ deprecated: false,
142
+ isLocal: false,
143
+ releaseDate: "2025-08-07",
144
+ category: "reasoning",
145
+ },
146
+ [OpenAIModels.GPT_5_MINI]: {
147
+ id: OpenAIModels.GPT_5_MINI,
148
+ name: "GPT-5 Mini",
149
+ provider: AIProviderName.OPENAI,
150
+ description: "Fast and efficient GPT-5 variant for everyday tasks",
151
+ capabilities: {
152
+ vision: true,
153
+ functionCalling: true,
154
+ codeGeneration: true,
155
+ reasoning: true,
156
+ multimodal: true,
157
+ streaming: true,
158
+ jsonMode: true,
159
+ },
160
+ pricing: {
161
+ inputCostPer1K: 0.002,
162
+ outputCostPer1K: 0.006,
163
+ currency: "USD",
164
+ },
165
+ performance: {
166
+ speed: "fast",
167
+ quality: "high",
168
+ accuracy: "high",
169
+ },
170
+ limits: {
171
+ maxContextTokens: 128000,
172
+ maxOutputTokens: 16384,
173
+ maxRequestsPerMinute: 500,
174
+ },
175
+ useCases: {
176
+ coding: 8,
177
+ creative: 8,
178
+ analysis: 8,
179
+ conversation: 9,
180
+ reasoning: 8,
181
+ translation: 8,
182
+ summarization: 9,
183
+ },
184
+ aliases: ["gpt5-mini", "gpt-5-fast"],
185
+ deprecated: false,
186
+ isLocal: false,
187
+ releaseDate: "2025-08-07",
188
+ category: "general",
189
+ },
190
+ // OpenAI O-Series Reasoning Models
191
+ [OpenAIModels.O3]: {
192
+ id: OpenAIModels.O3,
193
+ name: "O3",
194
+ provider: AIProviderName.OPENAI,
195
+ description: "Advanced reasoning model with extended thinking capabilities for complex tasks",
196
+ capabilities: {
197
+ vision: true,
198
+ functionCalling: true,
199
+ codeGeneration: true,
200
+ reasoning: true,
201
+ multimodal: true,
202
+ streaming: true,
203
+ jsonMode: true,
204
+ },
205
+ pricing: {
206
+ inputCostPer1K: 0.015,
207
+ outputCostPer1K: 0.06,
208
+ currency: "USD",
209
+ },
210
+ performance: {
211
+ speed: "slow",
212
+ quality: "high",
213
+ accuracy: "high",
214
+ },
215
+ limits: {
216
+ maxContextTokens: 200000,
217
+ maxOutputTokens: 100000,
218
+ maxRequestsPerMinute: 100,
219
+ },
220
+ useCases: {
221
+ coding: 10,
222
+ creative: 8,
223
+ analysis: 10,
224
+ conversation: 7,
225
+ reasoning: 10,
226
+ translation: 7,
227
+ summarization: 8,
228
+ },
229
+ aliases: ["o3-reasoning", "o3-thinking"],
230
+ deprecated: false,
231
+ isLocal: false,
232
+ releaseDate: "2025-01-31",
233
+ category: "reasoning",
234
+ },
235
+ [OpenAIModels.O3_MINI]: {
236
+ id: OpenAIModels.O3_MINI,
237
+ name: "O3 Mini",
238
+ provider: AIProviderName.OPENAI,
239
+ description: "Cost-effective reasoning model with strong logical capabilities",
240
+ capabilities: {
241
+ vision: false,
242
+ functionCalling: true,
243
+ codeGeneration: true,
244
+ reasoning: true,
245
+ multimodal: false,
246
+ streaming: true,
247
+ jsonMode: true,
248
+ },
249
+ pricing: {
250
+ inputCostPer1K: 0.003,
251
+ outputCostPer1K: 0.012,
252
+ currency: "USD",
253
+ },
254
+ performance: {
255
+ speed: "medium",
256
+ quality: "high",
257
+ accuracy: "high",
258
+ },
259
+ limits: {
260
+ maxContextTokens: 200000,
261
+ maxOutputTokens: 65536,
262
+ maxRequestsPerMinute: 200,
263
+ },
264
+ useCases: {
265
+ coding: 9,
266
+ creative: 6,
267
+ analysis: 9,
268
+ conversation: 7,
269
+ reasoning: 9,
270
+ translation: 6,
271
+ summarization: 7,
272
+ },
273
+ aliases: ["o3-mini-reasoning"],
274
+ deprecated: false,
275
+ isLocal: false,
276
+ releaseDate: "2025-01-31",
277
+ category: "reasoning",
278
+ },
279
+ [OpenAIModels.GPT_5_NANO]: {
280
+ id: OpenAIModels.GPT_5_NANO,
281
+ name: "GPT-5 Nano",
282
+ provider: AIProviderName.OPENAI,
283
+ description: "Fastest and most cost-effective GPT-5 variant for simple tasks",
284
+ capabilities: {
285
+ vision: true,
286
+ functionCalling: true,
287
+ codeGeneration: true,
288
+ reasoning: true,
289
+ multimodal: true,
290
+ streaming: true,
291
+ jsonMode: true,
292
+ },
293
+ pricing: {
294
+ inputCostPer1K: 0.00005,
295
+ outputCostPer1K: 0.0004,
296
+ currency: "USD",
297
+ },
298
+ performance: {
299
+ speed: "fast",
300
+ quality: "medium",
301
+ accuracy: "medium",
302
+ },
303
+ limits: {
304
+ maxContextTokens: 272000,
305
+ maxOutputTokens: 128000,
306
+ maxRequestsPerMinute: 2000,
307
+ },
308
+ useCases: {
309
+ coding: 6,
310
+ creative: 6,
311
+ analysis: 6,
312
+ conversation: 8,
313
+ reasoning: 6,
314
+ translation: 7,
315
+ summarization: 8,
316
+ },
317
+ aliases: ["gpt5-nano", "gpt-5-cheapest"],
318
+ deprecated: false,
319
+ isLocal: false,
320
+ releaseDate: "2025-08-07",
321
+ category: "general",
322
+ },
323
+ // OpenAI GPT-5.2 Series (Released December 11, 2025) - Latest flagship models
324
+ [OpenAIModels.GPT_5_2]: {
325
+ id: OpenAIModels.GPT_5_2,
326
+ name: "GPT-5.2 Thinking",
327
+ provider: AIProviderName.OPENAI,
328
+ description: "OpenAI's latest flagship model with deep reasoning capabilities, 100% on AIME 2025, 80% SWE-bench Verified",
329
+ capabilities: {
330
+ vision: true,
331
+ functionCalling: true,
332
+ codeGeneration: true,
333
+ reasoning: true,
334
+ multimodal: true,
335
+ streaming: true,
336
+ jsonMode: true,
337
+ },
338
+ pricing: {
339
+ inputCostPer1K: 0.00175,
340
+ outputCostPer1K: 0.014,
341
+ currency: "USD",
342
+ },
343
+ performance: {
344
+ speed: "medium",
345
+ quality: "high",
346
+ accuracy: "high",
347
+ },
348
+ limits: {
349
+ maxContextTokens: 256000,
350
+ maxOutputTokens: 64000,
351
+ maxRequestsPerMinute: 150,
352
+ },
353
+ useCases: {
354
+ coding: 10,
355
+ creative: 10,
356
+ analysis: 10,
357
+ conversation: 9,
358
+ reasoning: 10,
359
+ translation: 9,
360
+ summarization: 9,
361
+ },
362
+ aliases: ["gpt52", "gpt-5.2-thinking", "openai-latest-reasoning"],
363
+ deprecated: false,
364
+ isLocal: false,
365
+ releaseDate: "2025-12-11",
366
+ category: "reasoning",
367
+ },
368
+ [OpenAIModels.GPT_5_2_CHAT_LATEST]: {
369
+ id: OpenAIModels.GPT_5_2_CHAT_LATEST,
370
+ name: "GPT-5.2 Instant",
371
+ provider: AIProviderName.OPENAI,
372
+ description: "Fast everyday model for quick tasks with excellent performance across all domains",
373
+ capabilities: {
374
+ vision: true,
375
+ functionCalling: true,
376
+ codeGeneration: true,
377
+ reasoning: true,
378
+ multimodal: true,
379
+ streaming: true,
380
+ jsonMode: true,
381
+ },
382
+ pricing: {
383
+ inputCostPer1K: 0.00175,
384
+ outputCostPer1K: 0.014,
385
+ currency: "USD",
386
+ },
387
+ performance: {
388
+ speed: "fast",
389
+ quality: "high",
390
+ accuracy: "high",
391
+ },
392
+ limits: {
393
+ maxContextTokens: 256000,
394
+ maxOutputTokens: 32000,
395
+ maxRequestsPerMinute: 300,
396
+ },
397
+ useCases: {
398
+ coding: 9,
399
+ creative: 9,
400
+ analysis: 9,
401
+ conversation: 10,
402
+ reasoning: 9,
403
+ translation: 9,
404
+ summarization: 9,
405
+ },
406
+ aliases: ["gpt52-chat", "gpt-5.2-instant", "gpt52-fast"],
407
+ deprecated: false,
408
+ isLocal: false,
409
+ releaseDate: "2025-12-11",
410
+ category: "general",
411
+ },
412
+ [OpenAIModels.GPT_5_2_PRO]: {
413
+ id: OpenAIModels.GPT_5_2_PRO,
414
+ name: "GPT-5.2 Pro",
415
+ provider: AIProviderName.OPENAI,
416
+ description: "Highest quality model for science, math, and complex problem-solving with 92.4% GPQA Diamond performance",
417
+ capabilities: {
418
+ vision: true,
419
+ functionCalling: true,
420
+ codeGeneration: true,
421
+ reasoning: true,
422
+ multimodal: true,
423
+ streaming: true,
424
+ jsonMode: true,
425
+ },
426
+ pricing: {
427
+ inputCostPer1K: 0.021,
428
+ outputCostPer1K: 0.168,
429
+ currency: "USD",
430
+ },
431
+ performance: {
432
+ speed: "slow",
433
+ quality: "high",
434
+ accuracy: "high",
435
+ },
436
+ limits: {
437
+ maxContextTokens: 256000,
438
+ maxOutputTokens: 128000,
439
+ maxRequestsPerMinute: 50,
440
+ },
441
+ useCases: {
442
+ coding: 10,
443
+ creative: 9,
444
+ analysis: 10,
445
+ conversation: 8,
446
+ reasoning: 10,
447
+ translation: 9,
448
+ summarization: 9,
449
+ },
450
+ aliases: ["gpt52-pro", "gpt-5.2-professional", "openai-science"],
451
+ deprecated: false,
452
+ isLocal: false,
453
+ releaseDate: "2025-12-11",
454
+ category: "reasoning",
455
+ },
456
+ // OpenAI GPT-4.1 Series (1M context window)
457
+ [OpenAIModels.GPT_4_1]: {
458
+ id: OpenAIModels.GPT_4_1,
459
+ name: "GPT-4.1",
460
+ provider: AIProviderName.OPENAI,
461
+ description: "Advanced coding model with 1 million token context window",
462
+ capabilities: {
463
+ vision: true,
464
+ functionCalling: true,
465
+ codeGeneration: true,
466
+ reasoning: true,
467
+ multimodal: true,
468
+ streaming: true,
469
+ jsonMode: true,
470
+ },
471
+ pricing: {
472
+ inputCostPer1K: 0.002,
473
+ outputCostPer1K: 0.008,
474
+ currency: "USD",
475
+ },
476
+ performance: {
477
+ speed: "medium",
478
+ quality: "high",
479
+ accuracy: "high",
480
+ },
481
+ limits: {
482
+ maxContextTokens: 1000000,
483
+ maxOutputTokens: 128000,
484
+ maxRequestsPerMinute: 200,
485
+ },
486
+ useCases: {
487
+ coding: 10,
488
+ creative: 8,
489
+ analysis: 9,
490
+ conversation: 8,
491
+ reasoning: 9,
492
+ translation: 8,
493
+ summarization: 9,
494
+ },
495
+ aliases: ["gpt-4.1", "gpt41", "million-context"],
496
+ deprecated: false,
497
+ isLocal: false,
498
+ releaseDate: "2025-04-14",
499
+ category: "coding",
500
+ },
501
+ [OpenAIModels.GPT_4_1_MINI]: {
502
+ id: OpenAIModels.GPT_4_1_MINI,
503
+ name: "GPT-4.1 Mini",
504
+ provider: AIProviderName.OPENAI,
505
+ description: "Fast GPT-4.1 variant with 1M context for efficient coding",
506
+ capabilities: {
507
+ vision: true,
508
+ functionCalling: true,
509
+ codeGeneration: true,
510
+ reasoning: true,
511
+ multimodal: true,
512
+ streaming: true,
513
+ jsonMode: true,
514
+ },
515
+ pricing: {
516
+ inputCostPer1K: 0.0004,
517
+ outputCostPer1K: 0.0016,
518
+ currency: "USD",
519
+ },
520
+ performance: {
521
+ speed: "fast",
522
+ quality: "high",
523
+ accuracy: "high",
524
+ },
525
+ limits: {
526
+ maxContextTokens: 1000000,
527
+ maxOutputTokens: 128000,
528
+ maxRequestsPerMinute: 500,
529
+ },
530
+ useCases: {
531
+ coding: 9,
532
+ creative: 7,
533
+ analysis: 8,
534
+ conversation: 8,
535
+ reasoning: 8,
536
+ translation: 8,
537
+ summarization: 9,
538
+ },
539
+ aliases: ["gpt-4.1-mini", "gpt41-mini"],
540
+ deprecated: false,
541
+ isLocal: false,
542
+ releaseDate: "2025-04-14",
543
+ category: "coding",
544
+ },
545
+ [OpenAIModels.GPT_4_1_NANO]: {
546
+ id: OpenAIModels.GPT_4_1_NANO,
547
+ name: "GPT-4.1 Nano",
548
+ provider: AIProviderName.OPENAI,
549
+ description: "Most cost-effective GPT-4.1 variant with 1M context",
550
+ capabilities: {
551
+ vision: true,
552
+ functionCalling: true,
553
+ codeGeneration: true,
554
+ reasoning: true,
555
+ multimodal: true,
556
+ streaming: true,
557
+ jsonMode: true,
558
+ },
559
+ pricing: {
560
+ inputCostPer1K: 0.0001,
561
+ outputCostPer1K: 0.0004,
562
+ currency: "USD",
563
+ },
564
+ performance: {
565
+ speed: "fast",
566
+ quality: "medium",
567
+ accuracy: "medium",
568
+ },
569
+ limits: {
570
+ maxContextTokens: 1000000,
571
+ maxOutputTokens: 128000,
572
+ maxRequestsPerMinute: 1000,
573
+ },
574
+ useCases: {
575
+ coding: 7,
576
+ creative: 6,
577
+ analysis: 7,
578
+ conversation: 7,
579
+ reasoning: 7,
580
+ translation: 7,
581
+ summarization: 8,
582
+ },
583
+ aliases: ["gpt-4.1-nano", "gpt41-nano"],
584
+ deprecated: false,
585
+ isLocal: false,
586
+ releaseDate: "2025-04-14",
587
+ category: "coding",
588
+ },
589
+ // OpenAI O-Series Additional Models
590
+ [OpenAIModels.O3_PRO]: {
591
+ id: OpenAIModels.O3_PRO,
592
+ name: "O3 Pro",
593
+ provider: AIProviderName.OPENAI,
594
+ description: "Most powerful reasoning model for complex scientific and coding tasks",
595
+ capabilities: {
596
+ vision: true,
597
+ functionCalling: true,
598
+ codeGeneration: true,
599
+ reasoning: true,
600
+ multimodal: true,
601
+ streaming: true,
602
+ jsonMode: true,
603
+ },
604
+ pricing: {
605
+ inputCostPer1K: 0.03,
606
+ outputCostPer1K: 0.12,
607
+ currency: "USD",
608
+ },
609
+ performance: {
610
+ speed: "slow",
611
+ quality: "high",
612
+ accuracy: "high",
613
+ },
614
+ limits: {
615
+ maxContextTokens: 200000,
616
+ maxOutputTokens: 100000,
617
+ maxRequestsPerMinute: 50,
618
+ },
619
+ useCases: {
620
+ coding: 10,
621
+ creative: 7,
622
+ analysis: 10,
623
+ conversation: 6,
624
+ reasoning: 10,
625
+ translation: 6,
626
+ summarization: 7,
627
+ },
628
+ aliases: ["o3-pro", "o3-professional"],
629
+ deprecated: false,
630
+ isLocal: false,
631
+ releaseDate: "2025-04-16",
632
+ category: "reasoning",
633
+ },
634
+ [OpenAIModels.O4_MINI]: {
635
+ id: OpenAIModels.O4_MINI,
636
+ name: "O4 Mini",
637
+ provider: AIProviderName.OPENAI,
638
+ description: "Fast reasoning model optimized for math, coding, and visual tasks",
639
+ capabilities: {
640
+ vision: true,
641
+ functionCalling: true,
642
+ codeGeneration: true,
643
+ reasoning: true,
644
+ multimodal: true,
645
+ streaming: true,
646
+ jsonMode: true,
647
+ },
648
+ pricing: {
649
+ inputCostPer1K: 0.003,
650
+ outputCostPer1K: 0.012,
651
+ currency: "USD",
652
+ },
653
+ performance: {
654
+ speed: "medium",
655
+ quality: "high",
656
+ accuracy: "high",
657
+ },
658
+ limits: {
659
+ maxContextTokens: 200000,
660
+ maxOutputTokens: 100000,
661
+ maxRequestsPerMinute: 200,
662
+ },
663
+ useCases: {
664
+ coding: 9,
665
+ creative: 6,
666
+ analysis: 9,
667
+ conversation: 7,
668
+ reasoning: 10,
669
+ translation: 6,
670
+ summarization: 7,
671
+ },
672
+ aliases: ["o4-mini", "o4-fast"],
673
+ deprecated: false,
674
+ isLocal: false,
675
+ releaseDate: "2025-04-16",
676
+ category: "reasoning",
677
+ },
678
+ [OpenAIModels.O1]: {
679
+ id: OpenAIModels.O1,
680
+ name: "O1",
681
+ provider: AIProviderName.OPENAI,
682
+ description: "Premium reasoning model with highest capability for mission-critical tasks",
683
+ capabilities: {
684
+ vision: true,
685
+ functionCalling: true,
686
+ codeGeneration: true,
687
+ reasoning: true,
688
+ multimodal: true,
689
+ streaming: true,
690
+ jsonMode: true,
691
+ },
692
+ pricing: {
693
+ inputCostPer1K: 0.15,
694
+ outputCostPer1K: 0.6,
695
+ currency: "USD",
696
+ },
697
+ performance: {
698
+ speed: "slow",
699
+ quality: "high",
700
+ accuracy: "high",
701
+ },
702
+ limits: {
703
+ maxContextTokens: 128000,
704
+ maxOutputTokens: 32768,
705
+ maxRequestsPerMinute: 50,
706
+ },
707
+ useCases: {
708
+ coding: 10,
709
+ creative: 7,
710
+ analysis: 10,
711
+ conversation: 6,
712
+ reasoning: 10,
713
+ translation: 6,
714
+ summarization: 7,
715
+ },
716
+ aliases: ["o1-full", "o1-premium"],
717
+ deprecated: false,
718
+ isLocal: false,
719
+ releaseDate: "2024-09-12",
720
+ category: "reasoning",
721
+ },
722
+ [OpenAIModels.O1_PREVIEW]: {
723
+ id: OpenAIModels.O1_PREVIEW,
724
+ name: "O1 Preview",
725
+ provider: AIProviderName.OPENAI,
726
+ description: "Preview version of O1 reasoning model",
727
+ capabilities: {
728
+ vision: false,
729
+ functionCalling: true,
730
+ codeGeneration: true,
731
+ reasoning: true,
732
+ multimodal: false,
733
+ streaming: true,
734
+ jsonMode: true,
735
+ },
736
+ pricing: {
737
+ inputCostPer1K: 0.015,
738
+ outputCostPer1K: 0.06,
739
+ currency: "USD",
740
+ },
741
+ performance: {
742
+ speed: "slow",
743
+ quality: "high",
744
+ accuracy: "high",
745
+ },
746
+ limits: {
747
+ maxContextTokens: 128000,
748
+ maxOutputTokens: 32768,
749
+ maxRequestsPerMinute: 100,
750
+ },
751
+ useCases: {
752
+ coding: 9,
753
+ creative: 6,
754
+ analysis: 9,
755
+ conversation: 6,
756
+ reasoning: 9,
757
+ translation: 5,
758
+ summarization: 6,
759
+ },
760
+ aliases: ["o1-preview"],
761
+ deprecated: false,
762
+ isLocal: false,
763
+ releaseDate: "2024-09-12",
764
+ category: "reasoning",
765
+ },
766
+ [OpenAIModels.O1_MINI]: {
767
+ id: OpenAIModels.O1_MINI,
768
+ name: "O1 Mini",
769
+ provider: AIProviderName.OPENAI,
770
+ description: "Cost-effective O1 variant with strong reasoning capabilities",
771
+ capabilities: {
772
+ vision: false,
773
+ functionCalling: true,
774
+ codeGeneration: true,
775
+ reasoning: true,
776
+ multimodal: false,
777
+ streaming: true,
778
+ jsonMode: true,
779
+ },
780
+ pricing: {
781
+ inputCostPer1K: 0.003,
782
+ outputCostPer1K: 0.012,
783
+ currency: "USD",
784
+ },
785
+ performance: {
786
+ speed: "medium",
787
+ quality: "high",
788
+ accuracy: "high",
789
+ },
790
+ limits: {
791
+ maxContextTokens: 128000,
792
+ maxOutputTokens: 65536,
793
+ maxRequestsPerMinute: 200,
794
+ },
795
+ useCases: {
796
+ coding: 8,
797
+ creative: 5,
798
+ analysis: 8,
799
+ conversation: 6,
800
+ reasoning: 8,
801
+ translation: 5,
802
+ summarization: 6,
803
+ },
804
+ aliases: ["o1-mini", "o1-budget"],
805
+ deprecated: false,
806
+ isLocal: false,
807
+ releaseDate: "2024-09-12",
808
+ category: "reasoning",
809
+ },
810
+ // OpenAI Legacy Models
811
+ [OpenAIModels.GPT_4]: {
812
+ id: OpenAIModels.GPT_4,
813
+ name: "GPT-4",
814
+ provider: AIProviderName.OPENAI,
815
+ description: "Previous generation flagship model (legacy)",
816
+ capabilities: {
817
+ vision: false,
818
+ functionCalling: true,
819
+ codeGeneration: true,
820
+ reasoning: true,
821
+ multimodal: false,
822
+ streaming: true,
823
+ jsonMode: true,
824
+ },
825
+ pricing: {
826
+ inputCostPer1K: 0.03,
827
+ outputCostPer1K: 0.06,
828
+ currency: "USD",
829
+ },
830
+ performance: {
831
+ speed: "slow",
832
+ quality: "high",
833
+ accuracy: "high",
834
+ },
835
+ limits: {
836
+ maxContextTokens: 8192,
837
+ maxOutputTokens: 4096,
838
+ maxRequestsPerMinute: 200,
839
+ },
840
+ useCases: {
841
+ coding: 8,
842
+ creative: 8,
843
+ analysis: 8,
844
+ conversation: 8,
845
+ reasoning: 8,
846
+ translation: 8,
847
+ summarization: 8,
848
+ },
849
+ aliases: ["gpt4", "gpt-4-base"],
850
+ deprecated: true,
851
+ isLocal: false,
852
+ releaseDate: "2023-03-14",
853
+ category: "general",
854
+ },
855
+ [OpenAIModels.GPT_4_TURBO]: {
856
+ id: OpenAIModels.GPT_4_TURBO,
857
+ name: "GPT-4 Turbo",
858
+ provider: AIProviderName.OPENAI,
859
+ description: "Faster GPT-4 variant with extended context (legacy)",
860
+ capabilities: {
861
+ vision: true,
862
+ functionCalling: true,
863
+ codeGeneration: true,
864
+ reasoning: true,
865
+ multimodal: true,
866
+ streaming: true,
867
+ jsonMode: true,
868
+ },
869
+ pricing: {
870
+ inputCostPer1K: 0.01,
871
+ outputCostPer1K: 0.03,
872
+ currency: "USD",
873
+ },
874
+ performance: {
875
+ speed: "medium",
876
+ quality: "high",
877
+ accuracy: "high",
878
+ },
879
+ limits: {
880
+ maxContextTokens: 128000,
881
+ maxOutputTokens: 4096,
882
+ maxRequestsPerMinute: 500,
883
+ },
884
+ useCases: {
885
+ coding: 8,
886
+ creative: 8,
887
+ analysis: 9,
888
+ conversation: 8,
889
+ reasoning: 8,
890
+ translation: 8,
891
+ summarization: 8,
892
+ },
893
+ aliases: ["gpt4-turbo", "gpt-4-turbo-preview"],
894
+ deprecated: true,
895
+ isLocal: false,
896
+ releaseDate: "2024-04-09",
897
+ category: "general",
898
+ },
899
+ [OpenAIModels.GPT_3_5_TURBO]: {
900
+ id: OpenAIModels.GPT_3_5_TURBO,
901
+ name: "GPT-3.5 Turbo",
902
+ provider: AIProviderName.OPENAI,
903
+ description: "Fast and cost-effective model for simpler tasks (legacy)",
904
+ capabilities: {
905
+ vision: false,
906
+ functionCalling: true,
907
+ codeGeneration: true,
908
+ reasoning: false,
909
+ multimodal: false,
910
+ streaming: true,
911
+ jsonMode: true,
912
+ },
913
+ pricing: {
914
+ inputCostPer1K: 0.0005,
915
+ outputCostPer1K: 0.0015,
916
+ currency: "USD",
917
+ },
918
+ performance: {
919
+ speed: "fast",
920
+ quality: "medium",
921
+ accuracy: "medium",
922
+ },
923
+ limits: {
924
+ maxContextTokens: 16385,
925
+ maxOutputTokens: 4096,
926
+ maxRequestsPerMinute: 3500,
927
+ },
928
+ useCases: {
929
+ coding: 6,
930
+ creative: 6,
931
+ analysis: 6,
932
+ conversation: 7,
933
+ reasoning: 5,
934
+ translation: 7,
935
+ summarization: 7,
936
+ },
937
+ aliases: ["gpt35", "gpt-3.5", "chatgpt"],
938
+ deprecated: true,
939
+ isLocal: false,
940
+ releaseDate: "2023-03-01",
941
+ category: "general",
942
+ },
101
943
  // Google AI Studio Models
102
944
  [GoogleAIModels.GEMINI_2_5_PRO]: {
103
945
  id: GoogleAIModels.GEMINI_2_5_PRO,
104
946
  name: "Gemini 2.5 Pro",
105
947
  provider: AIProviderName.GOOGLE_AI,
106
- description: "Google's most capable multimodal model with large context window",
948
+ description: "Google's most capable multimodal model with large context window",
949
+ capabilities: {
950
+ vision: true,
951
+ functionCalling: true,
952
+ codeGeneration: true,
953
+ reasoning: true,
954
+ multimodal: true,
955
+ streaming: true,
956
+ jsonMode: true,
957
+ },
958
+ pricing: {
959
+ inputCostPer1K: 0.00125,
960
+ outputCostPer1K: 0.005,
961
+ currency: "USD",
962
+ },
963
+ performance: {
964
+ speed: "medium",
965
+ quality: "high",
966
+ accuracy: "high",
967
+ },
968
+ limits: {
969
+ maxContextTokens: 2097152, // 2M tokens
970
+ maxOutputTokens: 8192,
971
+ maxRequestsPerMinute: 360,
972
+ },
973
+ useCases: {
974
+ coding: 9,
975
+ creative: 8,
976
+ analysis: 10,
977
+ conversation: 8,
978
+ reasoning: 9,
979
+ translation: 9,
980
+ summarization: 9,
981
+ },
982
+ aliases: ["gemini-pro", "google-flagship", "best-analysis"],
983
+ deprecated: false,
984
+ isLocal: false, // Cloud-based model
985
+ releaseDate: "2024-12-11",
986
+ category: "reasoning",
987
+ },
988
+ [GoogleAIModels.GEMINI_2_5_FLASH]: {
989
+ id: GoogleAIModels.GEMINI_2_5_FLASH,
990
+ name: "Gemini 2.5 Flash",
991
+ provider: AIProviderName.GOOGLE_AI,
992
+ description: "Fast and efficient multimodal model with large context",
993
+ capabilities: {
994
+ vision: true,
995
+ functionCalling: true,
996
+ codeGeneration: true,
997
+ reasoning: true,
998
+ multimodal: true,
999
+ streaming: true,
1000
+ jsonMode: true,
1001
+ },
1002
+ pricing: {
1003
+ inputCostPer1K: 0.000075,
1004
+ outputCostPer1K: 0.0003,
1005
+ currency: "USD",
1006
+ },
1007
+ performance: {
1008
+ speed: "fast",
1009
+ quality: "high",
1010
+ accuracy: "high",
1011
+ },
1012
+ limits: {
1013
+ maxContextTokens: 1048576, // 1M tokens
1014
+ maxOutputTokens: 8192,
1015
+ maxRequestsPerMinute: 1000,
1016
+ },
1017
+ useCases: {
1018
+ coding: 8,
1019
+ creative: 7,
1020
+ analysis: 9,
1021
+ conversation: 8,
1022
+ reasoning: 8,
1023
+ translation: 8,
1024
+ summarization: 9,
1025
+ },
1026
+ aliases: ["gemini-flash", "google-fast", "best-value"],
1027
+ deprecated: false,
1028
+ isLocal: false, // Cloud-based model
1029
+ releaseDate: "2024-12-11",
1030
+ category: "general",
1031
+ },
1032
+ // Anthropic Models
1033
+ [AnthropicModels.CLAUDE_OPUS_4_5]: {
1034
+ id: AnthropicModels.CLAUDE_OPUS_4_5,
1035
+ name: "Claude Opus 4.5",
1036
+ provider: AIProviderName.ANTHROPIC,
1037
+ description: "Anthropic's most capable model with exceptional reasoning, coding, and multimodal capabilities",
1038
+ capabilities: {
1039
+ vision: true,
1040
+ functionCalling: true,
1041
+ codeGeneration: true,
1042
+ reasoning: true,
1043
+ multimodal: true,
1044
+ streaming: true,
1045
+ jsonMode: false,
1046
+ },
1047
+ pricing: {
1048
+ inputCostPer1K: 0.015,
1049
+ outputCostPer1K: 0.075,
1050
+ currency: "USD",
1051
+ },
1052
+ performance: {
1053
+ speed: "medium",
1054
+ quality: "high",
1055
+ accuracy: "high",
1056
+ },
1057
+ limits: {
1058
+ maxContextTokens: 200000,
1059
+ maxOutputTokens: 64000,
1060
+ maxRequestsPerMinute: 50,
1061
+ },
1062
+ useCases: {
1063
+ coding: 10,
1064
+ creative: 10,
1065
+ analysis: 10,
1066
+ conversation: 9,
1067
+ reasoning: 10,
1068
+ translation: 9,
1069
+ summarization: 9,
1070
+ },
1071
+ aliases: [
1072
+ "claude-4.5-opus",
1073
+ "claude-opus-latest",
1074
+ "opus-4.5",
1075
+ "anthropic-flagship",
1076
+ ],
1077
+ deprecated: false,
1078
+ isLocal: false,
1079
+ releaseDate: "2025-11-24",
1080
+ category: "reasoning",
1081
+ },
1082
+ [AnthropicModels.CLAUDE_SONNET_4_5]: {
1083
+ id: AnthropicModels.CLAUDE_SONNET_4_5,
1084
+ name: "Claude Sonnet 4.5",
1085
+ provider: AIProviderName.ANTHROPIC,
1086
+ description: "Balanced Claude model with excellent performance across all tasks including vision and reasoning",
1087
+ capabilities: {
1088
+ vision: true,
1089
+ functionCalling: true,
1090
+ codeGeneration: true,
1091
+ reasoning: true,
1092
+ multimodal: true,
1093
+ streaming: true,
1094
+ jsonMode: false,
1095
+ },
1096
+ pricing: {
1097
+ inputCostPer1K: 0.003,
1098
+ outputCostPer1K: 0.015,
1099
+ currency: "USD",
1100
+ },
1101
+ performance: {
1102
+ speed: "medium",
1103
+ quality: "high",
1104
+ accuracy: "high",
1105
+ },
1106
+ limits: {
1107
+ maxContextTokens: 200000,
1108
+ maxOutputTokens: 64000,
1109
+ maxRequestsPerMinute: 100,
1110
+ },
1111
+ useCases: {
1112
+ coding: 10,
1113
+ creative: 9,
1114
+ analysis: 9,
1115
+ conversation: 9,
1116
+ reasoning: 10,
1117
+ translation: 8,
1118
+ summarization: 8,
1119
+ },
1120
+ aliases: ["claude-4.5-sonnet", "claude-sonnet-latest", "sonnet-4.5"],
1121
+ deprecated: false,
1122
+ isLocal: false,
1123
+ releaseDate: "2025-09-29",
1124
+ category: "coding",
1125
+ },
1126
+ [AnthropicModels.CLAUDE_4_5_HAIKU]: {
1127
+ id: AnthropicModels.CLAUDE_4_5_HAIKU,
1128
+ name: "Claude 4.5 Haiku",
1129
+ provider: AIProviderName.ANTHROPIC,
1130
+ description: "Latest fast and efficient Claude model with vision support",
1131
+ capabilities: {
1132
+ vision: true,
1133
+ functionCalling: true,
1134
+ codeGeneration: true,
1135
+ reasoning: true,
1136
+ multimodal: true,
1137
+ streaming: true,
1138
+ jsonMode: false,
1139
+ },
1140
+ pricing: {
1141
+ inputCostPer1K: 0.001,
1142
+ outputCostPer1K: 0.005,
1143
+ currency: "USD",
1144
+ },
1145
+ performance: {
1146
+ speed: "fast",
1147
+ quality: "high",
1148
+ accuracy: "high",
1149
+ },
1150
+ limits: {
1151
+ maxContextTokens: 200000,
1152
+ maxOutputTokens: 64000,
1153
+ maxRequestsPerMinute: 100,
1154
+ },
1155
+ useCases: {
1156
+ coding: 8,
1157
+ creative: 8,
1158
+ analysis: 8,
1159
+ conversation: 9,
1160
+ reasoning: 8,
1161
+ translation: 8,
1162
+ summarization: 9,
1163
+ },
1164
+ aliases: ["claude-4.5-haiku", "claude-haiku-latest", "haiku-4.5"],
1165
+ deprecated: false,
1166
+ isLocal: false,
1167
+ releaseDate: "2025-10-15",
1168
+ category: "general",
1169
+ },
1170
+ [AnthropicModels.CLAUDE_3_5_SONNET]: {
1171
+ id: AnthropicModels.CLAUDE_3_5_SONNET,
1172
+ name: "Claude 3.5 Sonnet",
1173
+ provider: AIProviderName.ANTHROPIC,
1174
+ description: "Anthropic's most capable model with excellent reasoning and coding",
1175
+ capabilities: {
1176
+ vision: true,
1177
+ functionCalling: true,
1178
+ codeGeneration: true,
1179
+ reasoning: true,
1180
+ multimodal: true,
1181
+ streaming: true,
1182
+ jsonMode: false,
1183
+ },
1184
+ pricing: {
1185
+ inputCostPer1K: 0.003,
1186
+ outputCostPer1K: 0.015,
1187
+ currency: "USD",
1188
+ },
1189
+ performance: {
1190
+ speed: "medium",
1191
+ quality: "high",
1192
+ accuracy: "high",
1193
+ },
1194
+ limits: {
1195
+ maxContextTokens: 200000,
1196
+ maxOutputTokens: 8192,
1197
+ maxRequestsPerMinute: 50,
1198
+ },
1199
+ useCases: {
1200
+ coding: 10,
1201
+ creative: 9,
1202
+ analysis: 9,
1203
+ conversation: 9,
1204
+ reasoning: 10,
1205
+ translation: 8,
1206
+ summarization: 8,
1207
+ },
1208
+ aliases: [
1209
+ "claude-3.5-sonnet",
1210
+ "claude-sonnet",
1211
+ "best-coding",
1212
+ "claude-latest",
1213
+ ],
1214
+ deprecated: false,
1215
+ isLocal: false, // Cloud-based model
1216
+ releaseDate: "2024-10-22",
1217
+ category: "coding",
1218
+ },
1219
+ [AnthropicModels.CLAUDE_3_5_HAIKU]: {
1220
+ id: AnthropicModels.CLAUDE_3_5_HAIKU,
1221
+ name: "Claude 3.5 Haiku",
1222
+ provider: AIProviderName.ANTHROPIC,
1223
+ description: "Fast and efficient Claude model for quick tasks",
1224
+ capabilities: {
1225
+ vision: false,
1226
+ functionCalling: true,
1227
+ codeGeneration: true,
1228
+ reasoning: true,
1229
+ multimodal: false,
1230
+ streaming: true,
1231
+ jsonMode: false,
1232
+ },
1233
+ pricing: {
1234
+ inputCostPer1K: 0.001,
1235
+ outputCostPer1K: 0.005,
1236
+ currency: "USD",
1237
+ },
1238
+ performance: {
1239
+ speed: "fast",
1240
+ quality: "high",
1241
+ accuracy: "high",
1242
+ },
1243
+ limits: {
1244
+ maxContextTokens: 200000,
1245
+ maxOutputTokens: 8192,
1246
+ maxRequestsPerMinute: 100,
1247
+ },
1248
+ useCases: {
1249
+ coding: 8,
1250
+ creative: 7,
1251
+ analysis: 8,
1252
+ conversation: 8,
1253
+ reasoning: 8,
1254
+ translation: 8,
1255
+ summarization: 9,
1256
+ },
1257
+ aliases: ["claude-3.5-haiku", "claude-haiku", "claude-fast"],
1258
+ deprecated: false,
1259
+ isLocal: false, // Cloud-based model
1260
+ releaseDate: "2024-10-22",
1261
+ category: "general",
1262
+ },
1263
+ // Mistral Models
1264
+ [MistralModels.MISTRAL_LARGE_LATEST]: {
1265
+ id: MistralModels.MISTRAL_LARGE_LATEST,
1266
+ name: "Mistral Large",
1267
+ provider: AIProviderName.MISTRAL,
1268
+ description: "Mistral's flagship model with excellent reasoning and multilingual capabilities",
1269
+ capabilities: {
1270
+ vision: false,
1271
+ functionCalling: true,
1272
+ codeGeneration: true,
1273
+ reasoning: true,
1274
+ multimodal: false,
1275
+ streaming: true,
1276
+ jsonMode: true,
1277
+ },
1278
+ pricing: {
1279
+ inputCostPer1K: 0.002,
1280
+ outputCostPer1K: 0.006,
1281
+ currency: "USD",
1282
+ },
1283
+ performance: {
1284
+ speed: "medium",
1285
+ quality: "high",
1286
+ accuracy: "high",
1287
+ },
1288
+ limits: {
1289
+ maxContextTokens: 131072,
1290
+ maxOutputTokens: 8192,
1291
+ maxRequestsPerMinute: 100,
1292
+ },
1293
+ useCases: {
1294
+ coding: 9,
1295
+ creative: 8,
1296
+ analysis: 9,
1297
+ conversation: 8,
1298
+ reasoning: 9,
1299
+ translation: 9,
1300
+ summarization: 8,
1301
+ },
1302
+ aliases: ["mistral-large", "mistral-flagship"],
1303
+ deprecated: false,
1304
+ isLocal: false,
1305
+ releaseDate: "2025-12-01",
1306
+ category: "reasoning",
1307
+ },
1308
+ [MistralModels.MISTRAL_SMALL_LATEST]: {
1309
+ id: MistralModels.MISTRAL_SMALL_LATEST,
1310
+ name: "Mistral Small",
1311
+ provider: AIProviderName.MISTRAL,
1312
+ description: "Efficient model for simple tasks and cost-sensitive applications",
1313
+ capabilities: {
1314
+ vision: false,
1315
+ functionCalling: true,
1316
+ codeGeneration: true,
1317
+ reasoning: true,
1318
+ multimodal: false,
1319
+ streaming: true,
1320
+ jsonMode: true,
1321
+ },
1322
+ pricing: {
1323
+ inputCostPer1K: 0.001,
1324
+ outputCostPer1K: 0.003,
1325
+ currency: "USD",
1326
+ },
1327
+ performance: {
1328
+ speed: "fast",
1329
+ quality: "medium",
1330
+ accuracy: "medium",
1331
+ },
1332
+ limits: {
1333
+ maxContextTokens: 32768,
1334
+ maxOutputTokens: 8192,
1335
+ maxRequestsPerMinute: 200,
1336
+ },
1337
+ useCases: {
1338
+ coding: 6,
1339
+ creative: 6,
1340
+ analysis: 7,
1341
+ conversation: 7,
1342
+ reasoning: 6,
1343
+ translation: 7,
1344
+ summarization: 7,
1345
+ },
1346
+ aliases: ["mistral-small", "mistral-cheap"],
1347
+ deprecated: false,
1348
+ isLocal: false,
1349
+ releaseDate: "2024-02-26",
1350
+ category: "general",
1351
+ },
1352
+ [MistralModels.CODESTRAL_LATEST]: {
1353
+ id: MistralModels.CODESTRAL_LATEST,
1354
+ name: "Codestral",
1355
+ provider: AIProviderName.MISTRAL,
1356
+ description: "Specialized code generation model trained on 80+ programming languages",
1357
+ capabilities: {
1358
+ vision: false,
1359
+ functionCalling: true,
1360
+ codeGeneration: true,
1361
+ reasoning: true,
1362
+ multimodal: false,
1363
+ streaming: true,
1364
+ jsonMode: true,
1365
+ },
1366
+ pricing: {
1367
+ inputCostPer1K: 0.001,
1368
+ outputCostPer1K: 0.003,
1369
+ currency: "USD",
1370
+ },
1371
+ performance: {
1372
+ speed: "fast",
1373
+ quality: "high",
1374
+ accuracy: "high",
1375
+ },
1376
+ limits: {
1377
+ maxContextTokens: 32768,
1378
+ maxOutputTokens: 8192,
1379
+ maxRequestsPerMinute: 200,
1380
+ },
1381
+ useCases: {
1382
+ coding: 10,
1383
+ creative: 5,
1384
+ analysis: 7,
1385
+ conversation: 5,
1386
+ reasoning: 8,
1387
+ translation: 5,
1388
+ summarization: 6,
1389
+ },
1390
+ aliases: ["codestral", "mistral-code"],
1391
+ deprecated: false,
1392
+ isLocal: false,
1393
+ releaseDate: "2024-05-29",
1394
+ category: "coding",
1395
+ },
1396
+ [MistralModels.PIXTRAL_LARGE]: {
1397
+ id: MistralModels.PIXTRAL_LARGE,
1398
+ name: "Pixtral Large",
1399
+ provider: AIProviderName.MISTRAL,
1400
+ description: "Multimodal vision-language model for image understanding",
1401
+ capabilities: {
1402
+ vision: true,
1403
+ functionCalling: true,
1404
+ codeGeneration: true,
1405
+ reasoning: true,
1406
+ multimodal: true,
1407
+ streaming: true,
1408
+ jsonMode: true,
1409
+ },
1410
+ pricing: {
1411
+ inputCostPer1K: 0.002,
1412
+ outputCostPer1K: 0.006,
1413
+ currency: "USD",
1414
+ },
1415
+ performance: {
1416
+ speed: "medium",
1417
+ quality: "high",
1418
+ accuracy: "high",
1419
+ },
1420
+ limits: {
1421
+ maxContextTokens: 131072,
1422
+ maxOutputTokens: 8192,
1423
+ maxRequestsPerMinute: 100,
1424
+ },
1425
+ useCases: {
1426
+ coding: 8,
1427
+ creative: 8,
1428
+ analysis: 9,
1429
+ conversation: 7,
1430
+ reasoning: 8,
1431
+ translation: 7,
1432
+ summarization: 8,
1433
+ },
1434
+ aliases: ["pixtral", "mistral-vision"],
1435
+ deprecated: false,
1436
+ isLocal: false,
1437
+ releaseDate: "2024-09-01",
1438
+ category: "vision",
1439
+ },
1440
+ // Ollama Models (local)
1441
+ [OllamaModels.LLAMA4_LATEST]: {
1442
+ id: OllamaModels.LLAMA4_LATEST,
1443
+ name: "Llama 4",
1444
+ provider: AIProviderName.OLLAMA,
1445
+ description: "Latest Llama 4 with multimodal vision and tool capabilities, runs locally",
1446
+ capabilities: {
1447
+ vision: true,
1448
+ functionCalling: true,
1449
+ codeGeneration: true,
1450
+ reasoning: true,
1451
+ multimodal: true,
1452
+ streaming: true,
1453
+ jsonMode: true,
1454
+ },
1455
+ pricing: {
1456
+ inputCostPer1K: 0,
1457
+ outputCostPer1K: 0,
1458
+ currency: "USD",
1459
+ },
1460
+ performance: {
1461
+ speed: "medium",
1462
+ quality: "high",
1463
+ accuracy: "high",
1464
+ },
1465
+ limits: {
1466
+ maxContextTokens: 131072,
1467
+ maxOutputTokens: 8192,
1468
+ },
1469
+ useCases: {
1470
+ coding: 9,
1471
+ creative: 8,
1472
+ analysis: 9,
1473
+ conversation: 8,
1474
+ reasoning: 9,
1475
+ translation: 8,
1476
+ summarization: 8,
1477
+ },
1478
+ aliases: ["llama4", "llama4-local"],
1479
+ deprecated: false,
1480
+ isLocal: true,
1481
+ releaseDate: "2025-04-01",
1482
+ category: "reasoning",
1483
+ },
1484
+ [OllamaModels.LLAMA3_3_LATEST]: {
1485
+ id: OllamaModels.LLAMA3_3_LATEST,
1486
+ name: "Llama 3.3",
1487
+ provider: AIProviderName.OLLAMA,
1488
+ description: "High-performance Llama 3.3 for local inference",
1489
+ capabilities: {
1490
+ vision: false,
1491
+ functionCalling: true,
1492
+ codeGeneration: true,
1493
+ reasoning: true,
1494
+ multimodal: false,
1495
+ streaming: true,
1496
+ jsonMode: true,
1497
+ },
1498
+ pricing: {
1499
+ inputCostPer1K: 0,
1500
+ outputCostPer1K: 0,
1501
+ currency: "USD",
1502
+ },
1503
+ performance: {
1504
+ speed: "medium",
1505
+ quality: "high",
1506
+ accuracy: "high",
1507
+ },
1508
+ limits: {
1509
+ maxContextTokens: 131072,
1510
+ maxOutputTokens: 8192,
1511
+ },
1512
+ useCases: {
1513
+ coding: 8,
1514
+ creative: 8,
1515
+ analysis: 8,
1516
+ conversation: 8,
1517
+ reasoning: 8,
1518
+ translation: 8,
1519
+ summarization: 8,
1520
+ },
1521
+ aliases: ["llama3.3", "llama3.3-local"],
1522
+ deprecated: false,
1523
+ isLocal: true,
1524
+ releaseDate: "2024-12-01",
1525
+ category: "general",
1526
+ },
1527
+ [OllamaModels.LLAMA3_2_LATEST]: {
1528
+ id: OllamaModels.LLAMA3_2_LATEST,
1529
+ name: "Llama 3.2 Latest",
1530
+ provider: AIProviderName.OLLAMA,
1531
+ description: "Local Llama model for private, offline AI generation",
1532
+ capabilities: {
1533
+ vision: false,
1534
+ functionCalling: false,
1535
+ codeGeneration: true,
1536
+ reasoning: true,
1537
+ multimodal: false,
1538
+ streaming: true,
1539
+ jsonMode: false,
1540
+ },
1541
+ pricing: {
1542
+ inputCostPer1K: 0,
1543
+ outputCostPer1K: 0,
1544
+ currency: "USD",
1545
+ },
1546
+ performance: {
1547
+ speed: "fast",
1548
+ quality: "medium",
1549
+ accuracy: "medium",
1550
+ },
1551
+ limits: {
1552
+ maxContextTokens: 131072,
1553
+ maxOutputTokens: 8192,
1554
+ },
1555
+ useCases: {
1556
+ coding: 6,
1557
+ creative: 7,
1558
+ analysis: 6,
1559
+ conversation: 7,
1560
+ reasoning: 6,
1561
+ translation: 6,
1562
+ summarization: 6,
1563
+ },
1564
+ aliases: ["llama3.2", "llama", "local", "offline"],
1565
+ deprecated: false,
1566
+ isLocal: true,
1567
+ releaseDate: "2024-09-25",
1568
+ category: "general",
1569
+ },
1570
+ [OllamaModels.DEEPSEEK_R1_70B]: {
1571
+ id: OllamaModels.DEEPSEEK_R1_70B,
1572
+ name: "DeepSeek-R1 70B",
1573
+ provider: AIProviderName.OLLAMA,
1574
+ description: "State-of-the-art reasoning model rivaling OpenAI O1, runs locally",
1575
+ capabilities: {
1576
+ vision: false,
1577
+ functionCalling: false,
1578
+ codeGeneration: true,
1579
+ reasoning: true,
1580
+ multimodal: false,
1581
+ streaming: true,
1582
+ jsonMode: false,
1583
+ },
1584
+ pricing: {
1585
+ inputCostPer1K: 0,
1586
+ outputCostPer1K: 0,
1587
+ currency: "USD",
1588
+ },
1589
+ performance: {
1590
+ speed: "slow",
1591
+ quality: "high",
1592
+ accuracy: "high",
1593
+ },
1594
+ limits: {
1595
+ maxContextTokens: 65536,
1596
+ maxOutputTokens: 8192,
1597
+ },
1598
+ useCases: {
1599
+ coding: 10,
1600
+ creative: 7,
1601
+ analysis: 10,
1602
+ conversation: 6,
1603
+ reasoning: 10,
1604
+ translation: 7,
1605
+ summarization: 7,
1606
+ },
1607
+ aliases: ["deepseek-r1", "deepseek-reasoning", "local-reasoning"],
1608
+ deprecated: false,
1609
+ isLocal: true,
1610
+ releaseDate: "2025-01-20",
1611
+ category: "reasoning",
1612
+ },
1613
+ [OllamaModels.QWEN3_72B]: {
1614
+ id: OllamaModels.QWEN3_72B,
1615
+ name: "Qwen 3 72B",
1616
+ provider: AIProviderName.OLLAMA,
1617
+ description: "Advanced reasoning and multilingual model from Alibaba",
1618
+ capabilities: {
1619
+ vision: false,
1620
+ functionCalling: true,
1621
+ codeGeneration: true,
1622
+ reasoning: true,
1623
+ multimodal: false,
1624
+ streaming: true,
1625
+ jsonMode: true,
1626
+ },
1627
+ pricing: {
1628
+ inputCostPer1K: 0,
1629
+ outputCostPer1K: 0,
1630
+ currency: "USD",
1631
+ },
1632
+ performance: {
1633
+ speed: "slow",
1634
+ quality: "high",
1635
+ accuracy: "high",
1636
+ },
1637
+ limits: {
1638
+ maxContextTokens: 131072,
1639
+ maxOutputTokens: 8192,
1640
+ },
1641
+ useCases: {
1642
+ coding: 9,
1643
+ creative: 8,
1644
+ analysis: 9,
1645
+ conversation: 8,
1646
+ reasoning: 9,
1647
+ translation: 9,
1648
+ summarization: 8,
1649
+ },
1650
+ aliases: ["qwen3", "qwen3-72b-local"],
1651
+ deprecated: false,
1652
+ isLocal: true,
1653
+ releaseDate: "2025-04-01",
1654
+ category: "reasoning",
1655
+ },
1656
+ [OllamaModels.MISTRAL_LARGE_LATEST]: {
1657
+ id: OllamaModels.MISTRAL_LARGE_LATEST,
1658
+ name: "Mistral Large (Local)",
1659
+ provider: AIProviderName.OLLAMA,
1660
+ description: "Mistral Large model for local inference",
1661
+ capabilities: {
1662
+ vision: false,
1663
+ functionCalling: true,
1664
+ codeGeneration: true,
1665
+ reasoning: true,
1666
+ multimodal: false,
1667
+ streaming: true,
1668
+ jsonMode: true,
1669
+ },
1670
+ pricing: {
1671
+ inputCostPer1K: 0,
1672
+ outputCostPer1K: 0,
1673
+ currency: "USD",
1674
+ },
1675
+ performance: {
1676
+ speed: "slow",
1677
+ quality: "high",
1678
+ accuracy: "high",
1679
+ },
1680
+ limits: {
1681
+ maxContextTokens: 131072,
1682
+ maxOutputTokens: 8192,
1683
+ },
1684
+ useCases: {
1685
+ coding: 8,
1686
+ creative: 8,
1687
+ analysis: 8,
1688
+ conversation: 8,
1689
+ reasoning: 8,
1690
+ translation: 9,
1691
+ summarization: 8,
1692
+ },
1693
+ aliases: ["mistral-large-local"],
1694
+ deprecated: false,
1695
+ isLocal: true,
1696
+ releaseDate: "2024-02-26",
1697
+ category: "general",
1698
+ },
1699
+ // Bedrock Models
1700
+ [BedrockModels.NOVA_PREMIER]: {
1701
+ id: BedrockModels.NOVA_PREMIER,
1702
+ name: "Amazon Nova Premier",
1703
+ provider: AIProviderName.BEDROCK,
1704
+ description: "Amazon's most capable foundation model with advanced multimodal capabilities",
107
1705
  capabilities: {
108
1706
  vision: true,
109
1707
  functionCalling: true,
@@ -114,8 +1712,8 @@ export const MODEL_REGISTRY = {
114
1712
  jsonMode: true,
115
1713
  },
116
1714
  pricing: {
117
- inputCostPer1K: 0.00125,
118
- outputCostPer1K: 0.005,
1715
+ inputCostPer1K: 0.0025,
1716
+ outputCostPer1K: 0.0125,
119
1717
  currency: "USD",
120
1718
  },
121
1719
  performance: {
@@ -124,30 +1722,30 @@ export const MODEL_REGISTRY = {
124
1722
  accuracy: "high",
125
1723
  },
126
1724
  limits: {
127
- maxContextTokens: 2097152, // 2M tokens
128
- maxOutputTokens: 8192,
129
- maxRequestsPerMinute: 360,
1725
+ maxContextTokens: 300000,
1726
+ maxOutputTokens: 5000,
1727
+ maxRequestsPerMinute: 100,
130
1728
  },
131
1729
  useCases: {
132
1730
  coding: 9,
133
- creative: 8,
1731
+ creative: 9,
134
1732
  analysis: 10,
135
1733
  conversation: 8,
136
1734
  reasoning: 9,
137
- translation: 9,
1735
+ translation: 8,
138
1736
  summarization: 9,
139
1737
  },
140
- aliases: ["gemini-pro", "google-flagship", "best-analysis"],
1738
+ aliases: ["nova-premier", "aws-flagship"],
141
1739
  deprecated: false,
142
- isLocal: false, // Cloud-based model
143
- releaseDate: "2024-12-11",
1740
+ isLocal: false,
1741
+ releaseDate: "2025-01-01",
144
1742
  category: "reasoning",
145
1743
  },
146
- [GoogleAIModels.GEMINI_2_5_FLASH]: {
147
- id: GoogleAIModels.GEMINI_2_5_FLASH,
148
- name: "Gemini 2.5 Flash",
149
- provider: AIProviderName.GOOGLE_AI,
150
- description: "Fast and efficient multimodal model with large context",
1744
+ [BedrockModels.NOVA_PRO]: {
1745
+ id: BedrockModels.NOVA_PRO,
1746
+ name: "Amazon Nova Pro",
1747
+ provider: AIProviderName.BEDROCK,
1748
+ description: "Highly capable multimodal model balancing accuracy and speed",
151
1749
  capabilities: {
152
1750
  vision: true,
153
1751
  functionCalling: true,
@@ -158,8 +1756,140 @@ export const MODEL_REGISTRY = {
158
1756
  jsonMode: true,
159
1757
  },
160
1758
  pricing: {
161
- inputCostPer1K: 0.000075,
162
- outputCostPer1K: 0.0003,
1759
+ inputCostPer1K: 0.0008,
1760
+ outputCostPer1K: 0.0032,
1761
+ currency: "USD",
1762
+ },
1763
+ performance: {
1764
+ speed: "fast",
1765
+ quality: "high",
1766
+ accuracy: "high",
1767
+ },
1768
+ limits: {
1769
+ maxContextTokens: 300000,
1770
+ maxOutputTokens: 5000,
1771
+ maxRequestsPerMinute: 200,
1772
+ },
1773
+ useCases: {
1774
+ coding: 8,
1775
+ creative: 8,
1776
+ analysis: 9,
1777
+ conversation: 8,
1778
+ reasoning: 8,
1779
+ translation: 8,
1780
+ summarization: 9,
1781
+ },
1782
+ aliases: ["nova-pro", "aws-balanced"],
1783
+ deprecated: false,
1784
+ isLocal: false,
1785
+ releaseDate: "2024-12-03",
1786
+ category: "general",
1787
+ },
1788
+ [BedrockModels.NOVA_LITE]: {
1789
+ id: BedrockModels.NOVA_LITE,
1790
+ name: "Amazon Nova Lite",
1791
+ provider: AIProviderName.BEDROCK,
1792
+ description: "Fast and cost-effective multimodal model optimized for everyday tasks",
1793
+ capabilities: {
1794
+ vision: true,
1795
+ functionCalling: true,
1796
+ codeGeneration: true,
1797
+ reasoning: true,
1798
+ multimodal: true,
1799
+ streaming: true,
1800
+ jsonMode: true,
1801
+ },
1802
+ pricing: {
1803
+ inputCostPer1K: 0.00006,
1804
+ outputCostPer1K: 0.00024,
1805
+ currency: "USD",
1806
+ },
1807
+ performance: {
1808
+ speed: "fast",
1809
+ quality: "high",
1810
+ accuracy: "high",
1811
+ },
1812
+ limits: {
1813
+ maxContextTokens: 300000,
1814
+ maxOutputTokens: 5000,
1815
+ maxRequestsPerMinute: 500,
1816
+ },
1817
+ useCases: {
1818
+ coding: 7,
1819
+ creative: 7,
1820
+ analysis: 8,
1821
+ conversation: 8,
1822
+ reasoning: 7,
1823
+ translation: 8,
1824
+ summarization: 9,
1825
+ },
1826
+ aliases: ["nova-lite", "aws-lite", "aws-cheap"],
1827
+ deprecated: false,
1828
+ isLocal: false,
1829
+ releaseDate: "2024-12-03",
1830
+ category: "general",
1831
+ },
1832
+ [BedrockModels.CLAUDE_4_5_OPUS]: {
1833
+ id: BedrockModels.CLAUDE_4_5_OPUS,
1834
+ name: "Claude 4.5 Opus (Bedrock)",
1835
+ provider: AIProviderName.BEDROCK,
1836
+ description: "Anthropic's most capable model available on Bedrock for enterprise workloads",
1837
+ capabilities: {
1838
+ vision: true,
1839
+ functionCalling: true,
1840
+ codeGeneration: true,
1841
+ reasoning: true,
1842
+ multimodal: true,
1843
+ streaming: true,
1844
+ jsonMode: false,
1845
+ },
1846
+ pricing: {
1847
+ inputCostPer1K: 0.015,
1848
+ outputCostPer1K: 0.075,
1849
+ currency: "USD",
1850
+ },
1851
+ performance: {
1852
+ speed: "medium",
1853
+ quality: "high",
1854
+ accuracy: "high",
1855
+ },
1856
+ limits: {
1857
+ maxContextTokens: 200000,
1858
+ maxOutputTokens: 64000,
1859
+ maxRequestsPerMinute: 50,
1860
+ },
1861
+ useCases: {
1862
+ coding: 10,
1863
+ creative: 10,
1864
+ analysis: 10,
1865
+ conversation: 9,
1866
+ reasoning: 10,
1867
+ translation: 9,
1868
+ summarization: 9,
1869
+ },
1870
+ aliases: ["bedrock-claude-4.5-opus", "bedrock-claude-flagship"],
1871
+ deprecated: false,
1872
+ isLocal: false,
1873
+ releaseDate: "2025-11-24",
1874
+ category: "reasoning",
1875
+ },
1876
+ [BedrockModels.LLAMA_4_MAVERICK_17B]: {
1877
+ id: BedrockModels.LLAMA_4_MAVERICK_17B,
1878
+ name: "Llama 4 Maverick (Bedrock)",
1879
+ provider: AIProviderName.BEDROCK,
1880
+ description: "Meta's latest Llama 4 model with vision on Bedrock",
1881
+ capabilities: {
1882
+ vision: true,
1883
+ functionCalling: true,
1884
+ codeGeneration: true,
1885
+ reasoning: true,
1886
+ multimodal: true,
1887
+ streaming: true,
1888
+ jsonMode: true,
1889
+ },
1890
+ pricing: {
1891
+ inputCostPer1K: 0.00019,
1892
+ outputCostPer1K: 0.00055,
163
1893
  currency: "USD",
164
1894
  },
165
1895
  performance: {
@@ -168,31 +1898,31 @@ export const MODEL_REGISTRY = {
168
1898
  accuracy: "high",
169
1899
  },
170
1900
  limits: {
171
- maxContextTokens: 1048576, // 1M tokens
1901
+ maxContextTokens: 131072,
172
1902
  maxOutputTokens: 8192,
173
- maxRequestsPerMinute: 1000,
1903
+ maxRequestsPerMinute: 200,
174
1904
  },
175
1905
  useCases: {
176
1906
  coding: 8,
177
- creative: 7,
178
- analysis: 9,
1907
+ creative: 8,
1908
+ analysis: 8,
179
1909
  conversation: 8,
180
1910
  reasoning: 8,
181
1911
  translation: 8,
182
- summarization: 9,
1912
+ summarization: 8,
183
1913
  },
184
- aliases: ["gemini-flash", "google-fast", "best-value"],
1914
+ aliases: ["bedrock-llama4", "bedrock-llama-maverick"],
185
1915
  deprecated: false,
186
- isLocal: false, // Cloud-based model
187
- releaseDate: "2024-12-11",
1916
+ isLocal: false,
1917
+ releaseDate: "2025-04-01",
188
1918
  category: "general",
189
1919
  },
190
- // Anthropic Models
191
- [AnthropicModels.CLAUDE_OPUS_4_5]: {
192
- id: AnthropicModels.CLAUDE_OPUS_4_5,
193
- name: "Claude Opus 4.5",
194
- provider: AIProviderName.ANTHROPIC,
195
- description: "Anthropic's most capable model with exceptional reasoning, coding, and multimodal capabilities",
1920
+ // Azure OpenAI GPT-5.1 Series (Latest - December 2025)
1921
+ [AzureOpenAIModels.GPT_5_1]: {
1922
+ id: AzureOpenAIModels.GPT_5_1,
1923
+ name: "GPT-5.1 (Azure)",
1924
+ provider: AIProviderName.AZURE,
1925
+ description: "Azure's latest GPT-5.1 flagship model with enhanced reasoning and multimodal capabilities",
196
1926
  capabilities: {
197
1927
  vision: true,
198
1928
  functionCalling: true,
@@ -200,11 +1930,11 @@ export const MODEL_REGISTRY = {
200
1930
  reasoning: true,
201
1931
  multimodal: true,
202
1932
  streaming: true,
203
- jsonMode: false,
1933
+ jsonMode: true,
204
1934
  },
205
1935
  pricing: {
206
1936
  inputCostPer1K: 0.015,
207
- outputCostPer1K: 0.075,
1937
+ outputCostPer1K: 0.045,
208
1938
  currency: "USD",
209
1939
  },
210
1940
  performance: {
@@ -213,35 +1943,30 @@ export const MODEL_REGISTRY = {
213
1943
  accuracy: "high",
214
1944
  },
215
1945
  limits: {
216
- maxContextTokens: 200000,
1946
+ maxContextTokens: 300000,
217
1947
  maxOutputTokens: 64000,
218
- maxRequestsPerMinute: 50,
1948
+ maxRequestsPerMinute: 100,
219
1949
  },
220
1950
  useCases: {
221
1951
  coding: 10,
222
1952
  creative: 10,
223
1953
  analysis: 10,
224
- conversation: 9,
1954
+ conversation: 10,
225
1955
  reasoning: 10,
226
1956
  translation: 9,
227
1957
  summarization: 9,
228
1958
  },
229
- aliases: [
230
- "claude-4.5-opus",
231
- "claude-opus-latest",
232
- "opus-4.5",
233
- "anthropic-flagship",
234
- ],
1959
+ aliases: ["azure-gpt-5.1", "gpt51-azure", "azure-flagship"],
235
1960
  deprecated: false,
236
1961
  isLocal: false,
237
- releaseDate: "2025-11-24",
1962
+ releaseDate: "2025-12-01",
238
1963
  category: "reasoning",
239
1964
  },
240
- [AnthropicModels.CLAUDE_SONNET_4_5]: {
241
- id: AnthropicModels.CLAUDE_SONNET_4_5,
242
- name: "Claude Sonnet 4.5",
243
- provider: AIProviderName.ANTHROPIC,
244
- description: "Balanced Claude model with excellent performance across all tasks including vision and reasoning",
1965
+ [AzureOpenAIModels.GPT_5_1_CHAT]: {
1966
+ id: AzureOpenAIModels.GPT_5_1_CHAT,
1967
+ name: "GPT-5.1 Chat (Azure)",
1968
+ provider: AIProviderName.AZURE,
1969
+ description: "Azure GPT-5.1 optimized for conversational interactions",
245
1970
  capabilities: {
246
1971
  vision: true,
247
1972
  functionCalling: true,
@@ -249,43 +1974,43 @@ export const MODEL_REGISTRY = {
249
1974
  reasoning: true,
250
1975
  multimodal: true,
251
1976
  streaming: true,
252
- jsonMode: false,
1977
+ jsonMode: true,
253
1978
  },
254
1979
  pricing: {
255
- inputCostPer1K: 0.003,
256
- outputCostPer1K: 0.015,
1980
+ inputCostPer1K: 0.012,
1981
+ outputCostPer1K: 0.036,
257
1982
  currency: "USD",
258
1983
  },
259
1984
  performance: {
260
- speed: "medium",
1985
+ speed: "fast",
261
1986
  quality: "high",
262
1987
  accuracy: "high",
263
1988
  },
264
1989
  limits: {
265
- maxContextTokens: 200000,
266
- maxOutputTokens: 64000,
267
- maxRequestsPerMinute: 100,
1990
+ maxContextTokens: 300000,
1991
+ maxOutputTokens: 32000,
1992
+ maxRequestsPerMinute: 150,
268
1993
  },
269
1994
  useCases: {
270
- coding: 10,
1995
+ coding: 8,
271
1996
  creative: 9,
272
1997
  analysis: 9,
273
- conversation: 9,
274
- reasoning: 10,
275
- translation: 8,
276
- summarization: 8,
1998
+ conversation: 10,
1999
+ reasoning: 9,
2000
+ translation: 9,
2001
+ summarization: 9,
277
2002
  },
278
- aliases: ["claude-4.5-sonnet", "claude-sonnet-latest", "sonnet-4.5"],
2003
+ aliases: ["azure-gpt-5.1-chat", "gpt51-chat-azure"],
279
2004
  deprecated: false,
280
2005
  isLocal: false,
281
- releaseDate: "2025-09-29",
282
- category: "coding",
2006
+ releaseDate: "2025-12-01",
2007
+ category: "general",
283
2008
  },
284
- [AnthropicModels.CLAUDE_4_5_HAIKU]: {
285
- id: AnthropicModels.CLAUDE_4_5_HAIKU,
286
- name: "Claude 4.5 Haiku",
287
- provider: AIProviderName.ANTHROPIC,
288
- description: "Latest fast and efficient Claude model with vision support",
2009
+ [AzureOpenAIModels.GPT_5_1_CODEX]: {
2010
+ id: AzureOpenAIModels.GPT_5_1_CODEX,
2011
+ name: "GPT-5.1 Codex (Azure)",
2012
+ provider: AIProviderName.AZURE,
2013
+ description: "Azure GPT-5.1 specialized for code generation and software development",
289
2014
  capabilities: {
290
2015
  vision: true,
291
2016
  functionCalling: true,
@@ -293,43 +2018,43 @@ export const MODEL_REGISTRY = {
293
2018
  reasoning: true,
294
2019
  multimodal: true,
295
2020
  streaming: true,
296
- jsonMode: false,
2021
+ jsonMode: true,
297
2022
  },
298
2023
  pricing: {
299
- inputCostPer1K: 0.001,
300
- outputCostPer1K: 0.005,
2024
+ inputCostPer1K: 0.012,
2025
+ outputCostPer1K: 0.036,
301
2026
  currency: "USD",
302
2027
  },
303
2028
  performance: {
304
- speed: "fast",
2029
+ speed: "medium",
305
2030
  quality: "high",
306
2031
  accuracy: "high",
307
2032
  },
308
2033
  limits: {
309
- maxContextTokens: 200000,
2034
+ maxContextTokens: 300000,
310
2035
  maxOutputTokens: 64000,
311
2036
  maxRequestsPerMinute: 100,
312
2037
  },
313
2038
  useCases: {
314
- coding: 8,
315
- creative: 8,
316
- analysis: 8,
317
- conversation: 9,
318
- reasoning: 8,
319
- translation: 8,
320
- summarization: 9,
2039
+ coding: 10,
2040
+ creative: 7,
2041
+ analysis: 9,
2042
+ conversation: 7,
2043
+ reasoning: 10,
2044
+ translation: 7,
2045
+ summarization: 8,
321
2046
  },
322
- aliases: ["claude-4.5-haiku", "claude-haiku-latest", "haiku-4.5"],
2047
+ aliases: ["azure-gpt-5.1-codex", "gpt51-codex-azure", "azure-code"],
323
2048
  deprecated: false,
324
2049
  isLocal: false,
325
- releaseDate: "2025-10-15",
326
- category: "general",
2050
+ releaseDate: "2025-12-01",
2051
+ category: "coding",
327
2052
  },
328
- [AnthropicModels.CLAUDE_3_5_SONNET]: {
329
- id: AnthropicModels.CLAUDE_3_5_SONNET,
330
- name: "Claude 3.5 Sonnet",
331
- provider: AIProviderName.ANTHROPIC,
332
- description: "Anthropic's most capable model with excellent reasoning and coding",
2053
+ [AzureOpenAIModels.GPT_5_1_CODEX_MINI]: {
2054
+ id: AzureOpenAIModels.GPT_5_1_CODEX_MINI,
2055
+ name: "GPT-5.1 Codex Mini (Azure)",
2056
+ provider: AIProviderName.AZURE,
2057
+ description: "Fast and efficient Azure code model for quick development tasks",
333
2058
  capabilities: {
334
2059
  vision: true,
335
2060
  functionCalling: true,
@@ -337,176 +2062,176 @@ export const MODEL_REGISTRY = {
337
2062
  reasoning: true,
338
2063
  multimodal: true,
339
2064
  streaming: true,
340
- jsonMode: false,
2065
+ jsonMode: true,
341
2066
  },
342
2067
  pricing: {
343
2068
  inputCostPer1K: 0.003,
344
- outputCostPer1K: 0.015,
2069
+ outputCostPer1K: 0.009,
345
2070
  currency: "USD",
346
2071
  },
347
2072
  performance: {
348
- speed: "medium",
2073
+ speed: "fast",
349
2074
  quality: "high",
350
2075
  accuracy: "high",
351
2076
  },
352
2077
  limits: {
353
2078
  maxContextTokens: 200000,
354
- maxOutputTokens: 8192,
355
- maxRequestsPerMinute: 50,
2079
+ maxOutputTokens: 32000,
2080
+ maxRequestsPerMinute: 300,
356
2081
  },
357
2082
  useCases: {
358
- coding: 10,
359
- creative: 9,
360
- analysis: 9,
361
- conversation: 9,
362
- reasoning: 10,
363
- translation: 8,
364
- summarization: 8,
2083
+ coding: 9,
2084
+ creative: 6,
2085
+ analysis: 8,
2086
+ conversation: 7,
2087
+ reasoning: 8,
2088
+ translation: 6,
2089
+ summarization: 7,
365
2090
  },
366
- aliases: [
367
- "claude-3.5-sonnet",
368
- "claude-sonnet",
369
- "best-coding",
370
- "claude-latest",
371
- ],
2091
+ aliases: ["azure-gpt-5.1-codex-mini", "gpt51-codex-mini-azure"],
372
2092
  deprecated: false,
373
- isLocal: false, // Cloud-based model
374
- releaseDate: "2024-10-22",
2093
+ isLocal: false,
2094
+ releaseDate: "2025-12-01",
375
2095
  category: "coding",
376
2096
  },
377
- [AnthropicModels.CLAUDE_3_5_HAIKU]: {
378
- id: AnthropicModels.CLAUDE_3_5_HAIKU,
379
- name: "Claude 3.5 Haiku",
380
- provider: AIProviderName.ANTHROPIC,
381
- description: "Fast and efficient Claude model for quick tasks",
2097
+ [AzureOpenAIModels.GPT_5_1_CODEX_MAX]: {
2098
+ id: AzureOpenAIModels.GPT_5_1_CODEX_MAX,
2099
+ name: "GPT-5.1 Codex Max (Azure)",
2100
+ provider: AIProviderName.AZURE,
2101
+ description: "Azure's most powerful code model for complex enterprise development",
382
2102
  capabilities: {
383
- vision: false,
2103
+ vision: true,
384
2104
  functionCalling: true,
385
2105
  codeGeneration: true,
386
2106
  reasoning: true,
387
- multimodal: false,
2107
+ multimodal: true,
388
2108
  streaming: true,
389
- jsonMode: false,
2109
+ jsonMode: true,
390
2110
  },
391
2111
  pricing: {
392
- inputCostPer1K: 0.001,
393
- outputCostPer1K: 0.005,
2112
+ inputCostPer1K: 0.025,
2113
+ outputCostPer1K: 0.075,
394
2114
  currency: "USD",
395
2115
  },
396
2116
  performance: {
397
- speed: "fast",
2117
+ speed: "slow",
398
2118
  quality: "high",
399
2119
  accuracy: "high",
400
2120
  },
401
2121
  limits: {
402
- maxContextTokens: 200000,
403
- maxOutputTokens: 8192,
404
- maxRequestsPerMinute: 100,
2122
+ maxContextTokens: 500000,
2123
+ maxOutputTokens: 128000,
2124
+ maxRequestsPerMinute: 50,
405
2125
  },
406
2126
  useCases: {
407
- coding: 8,
408
- creative: 7,
409
- analysis: 8,
410
- conversation: 8,
411
- reasoning: 8,
412
- translation: 8,
413
- summarization: 9,
2127
+ coding: 10,
2128
+ creative: 8,
2129
+ analysis: 10,
2130
+ conversation: 7,
2131
+ reasoning: 10,
2132
+ translation: 7,
2133
+ summarization: 8,
414
2134
  },
415
- aliases: ["claude-3.5-haiku", "claude-haiku", "claude-fast"],
2135
+ aliases: [
2136
+ "azure-gpt-5.1-codex-max",
2137
+ "gpt51-codex-max-azure",
2138
+ "azure-enterprise",
2139
+ ],
416
2140
  deprecated: false,
417
- isLocal: false, // Cloud-based model
418
- releaseDate: "2024-10-22",
419
- category: "general",
2141
+ isLocal: false,
2142
+ releaseDate: "2025-12-01",
2143
+ category: "coding",
420
2144
  },
421
- // Mistral Models
422
- "mistral-small-latest": {
423
- id: "mistral-small-latest",
424
- name: "Mistral Small",
425
- provider: AIProviderName.MISTRAL,
426
- description: "Efficient model for simple tasks and cost-sensitive applications",
2145
+ // Azure OpenAI GPT-5.0 Series (Azure-unique variants only)
2146
+ [AzureOpenAIModels.GPT_5_PRO]: {
2147
+ id: AzureOpenAIModels.GPT_5_PRO,
2148
+ name: "GPT-5 Pro (Azure)",
2149
+ provider: AIProviderName.AZURE,
2150
+ description: "Azure GPT-5 Pro with enhanced enterprise features",
427
2151
  capabilities: {
428
- vision: false,
2152
+ vision: true,
429
2153
  functionCalling: true,
430
2154
  codeGeneration: true,
431
2155
  reasoning: true,
432
- multimodal: false,
2156
+ multimodal: true,
433
2157
  streaming: true,
434
2158
  jsonMode: true,
435
2159
  },
436
2160
  pricing: {
437
- inputCostPer1K: 0.001,
438
- outputCostPer1K: 0.003,
2161
+ inputCostPer1K: 0.02,
2162
+ outputCostPer1K: 0.06,
439
2163
  currency: "USD",
440
2164
  },
441
2165
  performance: {
442
- speed: "fast",
443
- quality: "medium",
444
- accuracy: "medium",
2166
+ speed: "medium",
2167
+ quality: "high",
2168
+ accuracy: "high",
445
2169
  },
446
2170
  limits: {
447
- maxContextTokens: 32768,
448
- maxOutputTokens: 8192,
449
- maxRequestsPerMinute: 200,
2171
+ maxContextTokens: 256000,
2172
+ maxOutputTokens: 64000,
2173
+ maxRequestsPerMinute: 100,
450
2174
  },
451
2175
  useCases: {
452
- coding: 6,
453
- creative: 6,
454
- analysis: 7,
455
- conversation: 7,
456
- reasoning: 6,
457
- translation: 7,
458
- summarization: 7,
2176
+ coding: 10,
2177
+ creative: 10,
2178
+ analysis: 10,
2179
+ conversation: 9,
2180
+ reasoning: 10,
2181
+ translation: 9,
2182
+ summarization: 9,
459
2183
  },
460
- aliases: ["mistral-small", "mistral-cheap"],
2184
+ aliases: ["azure-gpt-5-pro", "gpt5-pro-azure"],
461
2185
  deprecated: false,
462
- isLocal: false, // Cloud-based model
463
- releaseDate: "2024-02-26",
464
- category: "general",
2186
+ isLocal: false,
2187
+ releaseDate: "2025-08-07",
2188
+ category: "reasoning",
465
2189
  },
466
- // Ollama Models (local)
467
- "llama3.2:latest": {
468
- id: "llama3.2:latest",
469
- name: "Llama 3.2 Latest",
470
- provider: AIProviderName.OLLAMA,
471
- description: "Local Llama model for private, offline AI generation",
2190
+ [AzureOpenAIModels.GPT_5_TURBO]: {
2191
+ id: AzureOpenAIModels.GPT_5_TURBO,
2192
+ name: "GPT-5 Turbo (Azure)",
2193
+ provider: AIProviderName.AZURE,
2194
+ description: "Azure GPT-5 Turbo optimized for fast responses",
472
2195
  capabilities: {
473
- vision: false,
474
- functionCalling: false,
2196
+ vision: true,
2197
+ functionCalling: true,
475
2198
  codeGeneration: true,
476
2199
  reasoning: true,
477
- multimodal: false,
2200
+ multimodal: true,
478
2201
  streaming: true,
479
- jsonMode: false,
2202
+ jsonMode: true,
480
2203
  },
481
2204
  pricing: {
482
- inputCostPer1K: 0, // Local execution
483
- outputCostPer1K: 0,
2205
+ inputCostPer1K: 0.008,
2206
+ outputCostPer1K: 0.024,
484
2207
  currency: "USD",
485
2208
  },
486
2209
  performance: {
487
- speed: "slow", // Depends on hardware
488
- quality: "medium",
489
- accuracy: "medium",
2210
+ speed: "fast",
2211
+ quality: "high",
2212
+ accuracy: "high",
490
2213
  },
491
2214
  limits: {
492
- maxContextTokens: 4096,
493
- maxOutputTokens: 2048,
2215
+ maxContextTokens: 200000,
2216
+ maxOutputTokens: 32768,
2217
+ maxRequestsPerMinute: 300,
494
2218
  },
495
2219
  useCases: {
496
- coding: 6,
497
- creative: 7,
498
- analysis: 6,
499
- conversation: 7,
500
- reasoning: 6,
501
- translation: 6,
502
- summarization: 6,
2220
+ coding: 9,
2221
+ creative: 9,
2222
+ analysis: 9,
2223
+ conversation: 9,
2224
+ reasoning: 9,
2225
+ translation: 9,
2226
+ summarization: 9,
503
2227
  },
504
- aliases: ["llama3.2", "llama", "local", "offline"],
2228
+ aliases: ["azure-gpt-5-turbo", "gpt5-turbo-azure"],
505
2229
  deprecated: false,
506
- isLocal: true, // Ollama runs locally
507
- releaseDate: "2024-09-25",
2230
+ isLocal: false,
2231
+ releaseDate: "2025-08-07",
508
2232
  category: "general",
509
2233
  },
2234
+ // Note: Azure models like O3, O4-mini, GPT-4o share IDs with OpenAI and use the OpenAI registry entries
510
2235
  };
511
2236
  /**
512
2237
  * Model aliases registry for quick resolution
@@ -522,60 +2247,87 @@ Object.values(MODEL_REGISTRY).forEach((model) => {
522
2247
  Object.entries(DEFAULT_MODEL_ALIASES).forEach(([k, v]) => {
523
2248
  MODEL_ALIASES[k.toLowerCase().replace(/_/g, "-")] = v;
524
2249
  });
525
- MODEL_ALIASES.local = "llama3.2:latest";
2250
+ MODEL_ALIASES.local = OllamaModels.LLAMA3_2_LATEST;
526
2251
  /**
527
2252
  * Use case to model mappings
528
2253
  */
529
2254
  export const USE_CASE_RECOMMENDATIONS = {
530
2255
  coding: [
531
- AnthropicModels.CLAUDE_3_5_SONNET,
532
- OpenAIModels.GPT_4O,
533
- GoogleAIModels.GEMINI_2_5_PRO,
2256
+ OpenAIModels.GPT_5_2_PRO,
2257
+ AnthropicModels.CLAUDE_OPUS_4_5,
2258
+ OpenAIModels.GPT_5_2,
2259
+ MistralModels.CODESTRAL_LATEST,
2260
+ AnthropicModels.CLAUDE_SONNET_4_5,
534
2261
  ],
535
2262
  creative: [
536
- AnthropicModels.CLAUDE_3_5_SONNET,
537
- OpenAIModels.GPT_4O,
2263
+ OpenAIModels.GPT_5_2,
2264
+ AnthropicModels.CLAUDE_OPUS_4_5,
2265
+ OpenAIModels.GPT_5,
538
2266
  GoogleAIModels.GEMINI_2_5_PRO,
539
2267
  ],
540
2268
  analysis: [
2269
+ OpenAIModels.GPT_5_2_PRO,
541
2270
  GoogleAIModels.GEMINI_2_5_PRO,
542
- AnthropicModels.CLAUDE_3_5_SONNET,
543
- OpenAIModels.GPT_4O,
2271
+ AnthropicModels.CLAUDE_OPUS_4_5,
2272
+ OpenAIModels.O3,
2273
+ BedrockModels.NOVA_PREMIER,
544
2274
  ],
545
2275
  conversation: [
2276
+ OpenAIModels.GPT_5_2_CHAT_LATEST,
2277
+ OpenAIModels.GPT_5,
2278
+ AnthropicModels.CLAUDE_SONNET_4_5,
546
2279
  OpenAIModels.GPT_4O,
547
- AnthropicModels.CLAUDE_3_5_SONNET,
548
- AnthropicModels.CLAUDE_3_5_HAIKU,
549
2280
  ],
550
2281
  reasoning: [
551
- AnthropicModels.CLAUDE_3_5_SONNET,
2282
+ OpenAIModels.GPT_5_2_PRO,
2283
+ OpenAIModels.GPT_5_2,
2284
+ OpenAIModels.O3,
2285
+ AnthropicModels.CLAUDE_OPUS_4_5,
552
2286
  GoogleAIModels.GEMINI_2_5_PRO,
553
- OpenAIModels.GPT_4O,
2287
+ OllamaModels.DEEPSEEK_R1_70B,
554
2288
  ],
555
2289
  translation: [
556
2290
  GoogleAIModels.GEMINI_2_5_PRO,
557
- OpenAIModels.GPT_4O,
558
- AnthropicModels.CLAUDE_3_5_HAIKU,
2291
+ MistralModels.MISTRAL_LARGE_LATEST,
2292
+ OpenAIModels.GPT_5,
559
2293
  ],
560
2294
  summarization: [
561
2295
  GoogleAIModels.GEMINI_2_5_FLASH,
562
- OpenAIModels.GPT_4O_MINI,
563
- AnthropicModels.CLAUDE_3_5_HAIKU,
2296
+ OpenAIModels.GPT_5_MINI,
2297
+ AnthropicModels.CLAUDE_4_5_HAIKU,
564
2298
  ],
565
2299
  "cost-effective": [
566
2300
  GoogleAIModels.GEMINI_2_5_FLASH,
567
2301
  OpenAIModels.GPT_4O_MINI,
568
- "mistral-small-latest",
2302
+ MistralModels.MISTRAL_SMALL_LATEST,
2303
+ BedrockModels.NOVA_LITE,
569
2304
  ],
570
2305
  "high-quality": [
571
- AnthropicModels.CLAUDE_3_5_SONNET,
572
- OpenAIModels.GPT_4O,
2306
+ OpenAIModels.GPT_5_2_PRO,
2307
+ OpenAIModels.GPT_5_2,
2308
+ AnthropicModels.CLAUDE_OPUS_4_5,
573
2309
  GoogleAIModels.GEMINI_2_5_PRO,
574
2310
  ],
575
2311
  fast: [
576
- OpenAIModels.GPT_4O_MINI,
2312
+ OpenAIModels.GPT_5_2_CHAT_LATEST,
2313
+ OpenAIModels.GPT_5_MINI,
577
2314
  GoogleAIModels.GEMINI_2_5_FLASH,
578
- AnthropicModels.CLAUDE_3_5_HAIKU,
2315
+ AnthropicModels.CLAUDE_4_5_HAIKU,
2316
+ OpenAIModels.O3_MINI,
2317
+ ],
2318
+ local: [
2319
+ OllamaModels.LLAMA4_LATEST,
2320
+ OllamaModels.DEEPSEEK_R1_70B,
2321
+ OllamaModels.QWEN3_72B,
2322
+ OllamaModels.LLAMA3_3_LATEST,
2323
+ ],
2324
+ multimodal: [
2325
+ OpenAIModels.GPT_5_2,
2326
+ OpenAIModels.GPT_5_2_PRO,
2327
+ AnthropicModels.CLAUDE_OPUS_4_5,
2328
+ GoogleAIModels.GEMINI_2_5_PRO,
2329
+ MistralModels.PIXTRAL_LARGE,
2330
+ BedrockModels.NOVA_PREMIER,
579
2331
  ],
580
2332
  };
581
2333
  /**