@hebo-ai/gateway 0.8.0-rc3 → 0.8.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -95,6 +95,18 @@ export declare const gpt51CodexMax: import("../../utils/preset").Preset<"openai/
|
|
|
95
95
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
96
96
|
providers: readonly ["openai", "azure"];
|
|
97
97
|
}>;
|
|
98
|
+
export declare const gpt51CodexMini: import("../../utils/preset").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
|
|
99
|
+
name: string;
|
|
100
|
+
created: string;
|
|
101
|
+
knowledge: string;
|
|
102
|
+
context: number;
|
|
103
|
+
modalities: {
|
|
104
|
+
input: readonly ["text", "image"];
|
|
105
|
+
output: readonly ["text"];
|
|
106
|
+
};
|
|
107
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
108
|
+
providers: readonly ["openai", "azure"];
|
|
109
|
+
}>;
|
|
98
110
|
export declare const gpt5Codex: import("../../utils/preset").Preset<"openai/gpt-5-codex", CatalogModel, {
|
|
99
111
|
name: string;
|
|
100
112
|
created: string;
|
|
@@ -165,6 +177,78 @@ export declare const gpt53Codex: import("../../utils/preset").Preset<"openai/gpt
|
|
|
165
177
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
166
178
|
providers: readonly ["openai", "azure"];
|
|
167
179
|
}>;
|
|
180
|
+
export declare const gpt53CodexSpark: import("../../utils/preset").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
|
|
181
|
+
name: string;
|
|
182
|
+
created: string;
|
|
183
|
+
knowledge: string;
|
|
184
|
+
context: number;
|
|
185
|
+
modalities: {
|
|
186
|
+
input: readonly ["text", "image"];
|
|
187
|
+
output: readonly ["text"];
|
|
188
|
+
};
|
|
189
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
190
|
+
providers: readonly ["openai", "azure"];
|
|
191
|
+
}>;
|
|
192
|
+
export declare const gpt53Chat: import("../../utils/preset").Preset<"openai/gpt-5.3-chat", CatalogModel, {
|
|
193
|
+
name: string;
|
|
194
|
+
created: string;
|
|
195
|
+
knowledge: string;
|
|
196
|
+
context: number;
|
|
197
|
+
modalities: {
|
|
198
|
+
input: readonly ["text", "image"];
|
|
199
|
+
output: readonly ["text"];
|
|
200
|
+
};
|
|
201
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
202
|
+
providers: readonly ["openai", "azure"];
|
|
203
|
+
}>;
|
|
204
|
+
export declare const gpt54: import("../../utils/preset").Preset<"openai/gpt-5.4", CatalogModel, {
|
|
205
|
+
name: string;
|
|
206
|
+
created: string;
|
|
207
|
+
knowledge: string;
|
|
208
|
+
context: number;
|
|
209
|
+
modalities: {
|
|
210
|
+
input: readonly ["text", "image"];
|
|
211
|
+
output: readonly ["text"];
|
|
212
|
+
};
|
|
213
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
214
|
+
providers: readonly ["openai", "azure"];
|
|
215
|
+
}>;
|
|
216
|
+
export declare const gpt54Mini: import("../../utils/preset").Preset<"openai/gpt-5.4-mini", CatalogModel, {
|
|
217
|
+
name: string;
|
|
218
|
+
created: string;
|
|
219
|
+
knowledge: string;
|
|
220
|
+
context: number;
|
|
221
|
+
modalities: {
|
|
222
|
+
input: readonly ["text", "image"];
|
|
223
|
+
output: readonly ["text"];
|
|
224
|
+
};
|
|
225
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
226
|
+
providers: readonly ["openai", "azure"];
|
|
227
|
+
}>;
|
|
228
|
+
export declare const gpt54Nano: import("../../utils/preset").Preset<"openai/gpt-5.4-nano", CatalogModel, {
|
|
229
|
+
name: string;
|
|
230
|
+
created: string;
|
|
231
|
+
knowledge: string;
|
|
232
|
+
context: number;
|
|
233
|
+
modalities: {
|
|
234
|
+
input: readonly ["text", "image"];
|
|
235
|
+
output: readonly ["text"];
|
|
236
|
+
};
|
|
237
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
238
|
+
providers: readonly ["openai", "azure"];
|
|
239
|
+
}>;
|
|
240
|
+
export declare const gpt54Pro: import("../../utils/preset").Preset<"openai/gpt-5.4-pro", CatalogModel, {
|
|
241
|
+
name: string;
|
|
242
|
+
created: string;
|
|
243
|
+
knowledge: string;
|
|
244
|
+
context: number;
|
|
245
|
+
modalities: {
|
|
246
|
+
input: readonly ["text", "image"];
|
|
247
|
+
output: readonly ["text"];
|
|
248
|
+
};
|
|
249
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
250
|
+
providers: readonly ["openai", "azure"];
|
|
251
|
+
}>;
|
|
168
252
|
export declare const textEmbedding3Small: import("../../utils/preset").Preset<"openai/text-embedding-3-small", CatalogModel, {
|
|
169
253
|
name: string;
|
|
170
254
|
created: string;
|
|
@@ -304,7 +388,7 @@ export declare const gptOss: {
|
|
|
304
388
|
}>];
|
|
305
389
|
};
|
|
306
390
|
export declare const gpt: {
|
|
307
|
-
readonly latest: readonly [import("../../utils/preset").Preset<"openai/gpt-5.
|
|
391
|
+
readonly latest: readonly [import("../../utils/preset").Preset<"openai/gpt-5.4", CatalogModel, {
|
|
308
392
|
name: string;
|
|
309
393
|
created: string;
|
|
310
394
|
knowledge: string;
|
|
@@ -315,7 +399,7 @@ export declare const gpt: {
|
|
|
315
399
|
};
|
|
316
400
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
317
401
|
providers: readonly ["openai", "azure"];
|
|
318
|
-
}>, import("../../utils/preset").Preset<"openai/gpt-5-mini", CatalogModel, {
|
|
402
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.4-mini", CatalogModel, {
|
|
319
403
|
name: string;
|
|
320
404
|
created: string;
|
|
321
405
|
knowledge: string;
|
|
@@ -326,7 +410,7 @@ export declare const gpt: {
|
|
|
326
410
|
};
|
|
327
411
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
328
412
|
providers: readonly ["openai", "azure"];
|
|
329
|
-
}>, import("../../utils/preset").Preset<"openai/gpt-5-nano", CatalogModel, {
|
|
413
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.4-nano", CatalogModel, {
|
|
330
414
|
name: string;
|
|
331
415
|
created: string;
|
|
332
416
|
knowledge: string;
|
|
@@ -426,6 +510,17 @@ export declare const gpt: {
|
|
|
426
510
|
};
|
|
427
511
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
428
512
|
providers: readonly ["openai", "azure"];
|
|
513
|
+
}> | import("../../utils/preset").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
|
|
514
|
+
name: string;
|
|
515
|
+
created: string;
|
|
516
|
+
knowledge: string;
|
|
517
|
+
context: number;
|
|
518
|
+
modalities: {
|
|
519
|
+
input: readonly ["text", "image"];
|
|
520
|
+
output: readonly ["text"];
|
|
521
|
+
};
|
|
522
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
523
|
+
providers: readonly ["openai", "azure"];
|
|
429
524
|
}> | import("../../utils/preset").Preset<"openai/gpt-5-codex", CatalogModel, {
|
|
430
525
|
name: string;
|
|
431
526
|
created: string;
|
|
@@ -490,6 +585,72 @@ export declare const gpt: {
|
|
|
490
585
|
};
|
|
491
586
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
492
587
|
providers: readonly ["openai", "azure"];
|
|
588
|
+
}> | import("../../utils/preset").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
|
|
589
|
+
name: string;
|
|
590
|
+
created: string;
|
|
591
|
+
knowledge: string;
|
|
592
|
+
context: number;
|
|
593
|
+
modalities: {
|
|
594
|
+
input: readonly ["text", "image"];
|
|
595
|
+
output: readonly ["text"];
|
|
596
|
+
};
|
|
597
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
598
|
+
providers: readonly ["openai", "azure"];
|
|
599
|
+
}> | import("../../utils/preset").Preset<"openai/gpt-5.3-chat", CatalogModel, {
|
|
600
|
+
name: string;
|
|
601
|
+
created: string;
|
|
602
|
+
knowledge: string;
|
|
603
|
+
context: number;
|
|
604
|
+
modalities: {
|
|
605
|
+
input: readonly ["text", "image"];
|
|
606
|
+
output: readonly ["text"];
|
|
607
|
+
};
|
|
608
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
609
|
+
providers: readonly ["openai", "azure"];
|
|
610
|
+
}> | import("../../utils/preset").Preset<"openai/gpt-5.4", CatalogModel, {
|
|
611
|
+
name: string;
|
|
612
|
+
created: string;
|
|
613
|
+
knowledge: string;
|
|
614
|
+
context: number;
|
|
615
|
+
modalities: {
|
|
616
|
+
input: readonly ["text", "image"];
|
|
617
|
+
output: readonly ["text"];
|
|
618
|
+
};
|
|
619
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
620
|
+
providers: readonly ["openai", "azure"];
|
|
621
|
+
}> | import("../../utils/preset").Preset<"openai/gpt-5.4-mini", CatalogModel, {
|
|
622
|
+
name: string;
|
|
623
|
+
created: string;
|
|
624
|
+
knowledge: string;
|
|
625
|
+
context: number;
|
|
626
|
+
modalities: {
|
|
627
|
+
input: readonly ["text", "image"];
|
|
628
|
+
output: readonly ["text"];
|
|
629
|
+
};
|
|
630
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
631
|
+
providers: readonly ["openai", "azure"];
|
|
632
|
+
}> | import("../../utils/preset").Preset<"openai/gpt-5.4-nano", CatalogModel, {
|
|
633
|
+
name: string;
|
|
634
|
+
created: string;
|
|
635
|
+
knowledge: string;
|
|
636
|
+
context: number;
|
|
637
|
+
modalities: {
|
|
638
|
+
input: readonly ["text", "image"];
|
|
639
|
+
output: readonly ["text"];
|
|
640
|
+
};
|
|
641
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
642
|
+
providers: readonly ["openai", "azure"];
|
|
643
|
+
}> | import("../../utils/preset").Preset<"openai/gpt-5.4-pro", CatalogModel, {
|
|
644
|
+
name: string;
|
|
645
|
+
created: string;
|
|
646
|
+
knowledge: string;
|
|
647
|
+
context: number;
|
|
648
|
+
modalities: {
|
|
649
|
+
input: readonly ["text", "image"];
|
|
650
|
+
output: readonly ["text"];
|
|
651
|
+
};
|
|
652
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
653
|
+
providers: readonly ["openai", "azure"];
|
|
493
654
|
}>)[];
|
|
494
655
|
readonly "v5.x": readonly [import("../../utils/preset").Preset<"openai/gpt-5", CatalogModel, {
|
|
495
656
|
name: string;
|
|
@@ -579,6 +740,17 @@ export declare const gpt: {
|
|
|
579
740
|
};
|
|
580
741
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
581
742
|
providers: readonly ["openai", "azure"];
|
|
743
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
|
|
744
|
+
name: string;
|
|
745
|
+
created: string;
|
|
746
|
+
knowledge: string;
|
|
747
|
+
context: number;
|
|
748
|
+
modalities: {
|
|
749
|
+
input: readonly ["text", "image"];
|
|
750
|
+
output: readonly ["text"];
|
|
751
|
+
};
|
|
752
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
753
|
+
providers: readonly ["openai", "azure"];
|
|
582
754
|
}>, import("../../utils/preset").Preset<"openai/gpt-5.2", CatalogModel, {
|
|
583
755
|
name: string;
|
|
584
756
|
created: string;
|
|
@@ -632,6 +804,72 @@ export declare const gpt: {
|
|
|
632
804
|
};
|
|
633
805
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
634
806
|
providers: readonly ["openai", "azure"];
|
|
807
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
|
|
808
|
+
name: string;
|
|
809
|
+
created: string;
|
|
810
|
+
knowledge: string;
|
|
811
|
+
context: number;
|
|
812
|
+
modalities: {
|
|
813
|
+
input: readonly ["text", "image"];
|
|
814
|
+
output: readonly ["text"];
|
|
815
|
+
};
|
|
816
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
817
|
+
providers: readonly ["openai", "azure"];
|
|
818
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.3-chat", CatalogModel, {
|
|
819
|
+
name: string;
|
|
820
|
+
created: string;
|
|
821
|
+
knowledge: string;
|
|
822
|
+
context: number;
|
|
823
|
+
modalities: {
|
|
824
|
+
input: readonly ["text", "image"];
|
|
825
|
+
output: readonly ["text"];
|
|
826
|
+
};
|
|
827
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
828
|
+
providers: readonly ["openai", "azure"];
|
|
829
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.4", CatalogModel, {
|
|
830
|
+
name: string;
|
|
831
|
+
created: string;
|
|
832
|
+
knowledge: string;
|
|
833
|
+
context: number;
|
|
834
|
+
modalities: {
|
|
835
|
+
input: readonly ["text", "image"];
|
|
836
|
+
output: readonly ["text"];
|
|
837
|
+
};
|
|
838
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
839
|
+
providers: readonly ["openai", "azure"];
|
|
840
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.4-mini", CatalogModel, {
|
|
841
|
+
name: string;
|
|
842
|
+
created: string;
|
|
843
|
+
knowledge: string;
|
|
844
|
+
context: number;
|
|
845
|
+
modalities: {
|
|
846
|
+
input: readonly ["text", "image"];
|
|
847
|
+
output: readonly ["text"];
|
|
848
|
+
};
|
|
849
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
850
|
+
providers: readonly ["openai", "azure"];
|
|
851
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.4-nano", CatalogModel, {
|
|
852
|
+
name: string;
|
|
853
|
+
created: string;
|
|
854
|
+
knowledge: string;
|
|
855
|
+
context: number;
|
|
856
|
+
modalities: {
|
|
857
|
+
input: readonly ["text", "image"];
|
|
858
|
+
output: readonly ["text"];
|
|
859
|
+
};
|
|
860
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
861
|
+
providers: readonly ["openai", "azure"];
|
|
862
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.4-pro", CatalogModel, {
|
|
863
|
+
name: string;
|
|
864
|
+
created: string;
|
|
865
|
+
knowledge: string;
|
|
866
|
+
context: number;
|
|
867
|
+
modalities: {
|
|
868
|
+
input: readonly ["text", "image"];
|
|
869
|
+
output: readonly ["text"];
|
|
870
|
+
};
|
|
871
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
872
|
+
providers: readonly ["openai", "azure"];
|
|
635
873
|
}>];
|
|
636
874
|
readonly v5: readonly [import("../../utils/preset").Preset<"openai/gpt-5", CatalogModel, {
|
|
637
875
|
name: string;
|
|
@@ -722,6 +960,17 @@ export declare const gpt: {
|
|
|
722
960
|
};
|
|
723
961
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
724
962
|
providers: readonly ["openai", "azure"];
|
|
963
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
|
|
964
|
+
name: string;
|
|
965
|
+
created: string;
|
|
966
|
+
knowledge: string;
|
|
967
|
+
context: number;
|
|
968
|
+
modalities: {
|
|
969
|
+
input: readonly ["text", "image"];
|
|
970
|
+
output: readonly ["text"];
|
|
971
|
+
};
|
|
972
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
973
|
+
providers: readonly ["openai", "azure"];
|
|
725
974
|
}>];
|
|
726
975
|
readonly "v5.2": readonly [import("../../utils/preset").Preset<"openai/gpt-5.2", CatalogModel, {
|
|
727
976
|
name: string;
|
|
@@ -777,6 +1026,73 @@ export declare const gpt: {
|
|
|
777
1026
|
};
|
|
778
1027
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
779
1028
|
providers: readonly ["openai", "azure"];
|
|
1029
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
|
|
1030
|
+
name: string;
|
|
1031
|
+
created: string;
|
|
1032
|
+
knowledge: string;
|
|
1033
|
+
context: number;
|
|
1034
|
+
modalities: {
|
|
1035
|
+
input: readonly ["text", "image"];
|
|
1036
|
+
output: readonly ["text"];
|
|
1037
|
+
};
|
|
1038
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1039
|
+
providers: readonly ["openai", "azure"];
|
|
1040
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.3-chat", CatalogModel, {
|
|
1041
|
+
name: string;
|
|
1042
|
+
created: string;
|
|
1043
|
+
knowledge: string;
|
|
1044
|
+
context: number;
|
|
1045
|
+
modalities: {
|
|
1046
|
+
input: readonly ["text", "image"];
|
|
1047
|
+
output: readonly ["text"];
|
|
1048
|
+
};
|
|
1049
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1050
|
+
providers: readonly ["openai", "azure"];
|
|
1051
|
+
}>];
|
|
1052
|
+
readonly "v5.4": readonly [import("../../utils/preset").Preset<"openai/gpt-5.4", CatalogModel, {
|
|
1053
|
+
name: string;
|
|
1054
|
+
created: string;
|
|
1055
|
+
knowledge: string;
|
|
1056
|
+
context: number;
|
|
1057
|
+
modalities: {
|
|
1058
|
+
input: readonly ["text", "image"];
|
|
1059
|
+
output: readonly ["text"];
|
|
1060
|
+
};
|
|
1061
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1062
|
+
providers: readonly ["openai", "azure"];
|
|
1063
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.4-mini", CatalogModel, {
|
|
1064
|
+
name: string;
|
|
1065
|
+
created: string;
|
|
1066
|
+
knowledge: string;
|
|
1067
|
+
context: number;
|
|
1068
|
+
modalities: {
|
|
1069
|
+
input: readonly ["text", "image"];
|
|
1070
|
+
output: readonly ["text"];
|
|
1071
|
+
};
|
|
1072
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1073
|
+
providers: readonly ["openai", "azure"];
|
|
1074
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.4-nano", CatalogModel, {
|
|
1075
|
+
name: string;
|
|
1076
|
+
created: string;
|
|
1077
|
+
knowledge: string;
|
|
1078
|
+
context: number;
|
|
1079
|
+
modalities: {
|
|
1080
|
+
input: readonly ["text", "image"];
|
|
1081
|
+
output: readonly ["text"];
|
|
1082
|
+
};
|
|
1083
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1084
|
+
providers: readonly ["openai", "azure"];
|
|
1085
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.4-pro", CatalogModel, {
|
|
1086
|
+
name: string;
|
|
1087
|
+
created: string;
|
|
1088
|
+
knowledge: string;
|
|
1089
|
+
context: number;
|
|
1090
|
+
modalities: {
|
|
1091
|
+
input: readonly ["text", "image"];
|
|
1092
|
+
output: readonly ["text"];
|
|
1093
|
+
};
|
|
1094
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1095
|
+
providers: readonly ["openai", "azure"];
|
|
780
1096
|
}>];
|
|
781
1097
|
readonly codex: readonly [import("../../utils/preset").Preset<"openai/gpt-5-codex", CatalogModel, {
|
|
782
1098
|
name: string;
|
|
@@ -811,6 +1127,17 @@ export declare const gpt: {
|
|
|
811
1127
|
};
|
|
812
1128
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
813
1129
|
providers: readonly ["openai", "azure"];
|
|
1130
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
|
|
1131
|
+
name: string;
|
|
1132
|
+
created: string;
|
|
1133
|
+
knowledge: string;
|
|
1134
|
+
context: number;
|
|
1135
|
+
modalities: {
|
|
1136
|
+
input: readonly ["text", "image"];
|
|
1137
|
+
output: readonly ["text"];
|
|
1138
|
+
};
|
|
1139
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1140
|
+
providers: readonly ["openai", "azure"];
|
|
814
1141
|
}>, import("../../utils/preset").Preset<"openai/gpt-5.2-codex", CatalogModel, {
|
|
815
1142
|
name: string;
|
|
816
1143
|
created: string;
|
|
@@ -831,6 +1158,17 @@ export declare const gpt: {
|
|
|
831
1158
|
};
|
|
832
1159
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
833
1160
|
providers: readonly ["openai", "azure"];
|
|
1161
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
|
|
1162
|
+
name: string;
|
|
1163
|
+
created: string;
|
|
1164
|
+
knowledge: string;
|
|
1165
|
+
context: number;
|
|
1166
|
+
modalities: {
|
|
1167
|
+
input: readonly ["text", "image"];
|
|
1168
|
+
output: readonly ["text"];
|
|
1169
|
+
};
|
|
1170
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1171
|
+
providers: readonly ["openai", "azure"];
|
|
834
1172
|
}>];
|
|
835
1173
|
readonly chat: readonly [import("../../utils/preset").Preset<"openai/gpt-5.1-chat", CatalogModel, {
|
|
836
1174
|
name: string;
|
|
@@ -854,6 +1192,17 @@ export declare const gpt: {
|
|
|
854
1192
|
};
|
|
855
1193
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
856
1194
|
providers: readonly ["openai", "azure"];
|
|
1195
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.3-chat", CatalogModel, {
|
|
1196
|
+
name: string;
|
|
1197
|
+
created: string;
|
|
1198
|
+
knowledge: string;
|
|
1199
|
+
context: number;
|
|
1200
|
+
modalities: {
|
|
1201
|
+
input: readonly ["text", "image"];
|
|
1202
|
+
output: readonly ["text"];
|
|
1203
|
+
};
|
|
1204
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1205
|
+
providers: readonly ["openai", "azure"];
|
|
857
1206
|
}>];
|
|
858
1207
|
readonly pro: readonly [import("../../utils/preset").Preset<"openai/gpt-5-pro", CatalogModel, {
|
|
859
1208
|
name: string;
|
|
@@ -877,6 +1226,17 @@ export declare const gpt: {
|
|
|
877
1226
|
};
|
|
878
1227
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
879
1228
|
providers: readonly ["openai", "azure"];
|
|
1229
|
+
}>, import("../../utils/preset").Preset<"openai/gpt-5.4-pro", CatalogModel, {
|
|
1230
|
+
name: string;
|
|
1231
|
+
created: string;
|
|
1232
|
+
knowledge: string;
|
|
1233
|
+
context: number;
|
|
1234
|
+
modalities: {
|
|
1235
|
+
input: readonly ["text", "image"];
|
|
1236
|
+
output: readonly ["text"];
|
|
1237
|
+
};
|
|
1238
|
+
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1239
|
+
providers: readonly ["openai", "azure"];
|
|
880
1240
|
}>];
|
|
881
1241
|
};
|
|
882
1242
|
export declare const textEmbeddings: {
|
|
@@ -105,6 +105,13 @@ export const gpt51CodexMax = presetFor()("openai/gpt-5.1-codex-max", {
|
|
|
105
105
|
knowledge: "2024-09",
|
|
106
106
|
context: 400000,
|
|
107
107
|
});
|
|
108
|
+
export const gpt51CodexMini = presetFor()("openai/gpt-5.1-codex-mini", {
|
|
109
|
+
...GPT_BASE,
|
|
110
|
+
name: "GPT-5.1 Codex Mini",
|
|
111
|
+
created: "2025-11-13",
|
|
112
|
+
knowledge: "2024-09",
|
|
113
|
+
context: 400000,
|
|
114
|
+
});
|
|
108
115
|
export const gpt5Codex = presetFor()("openai/gpt-5-codex", {
|
|
109
116
|
...GPT_BASE,
|
|
110
117
|
name: "GPT-5 Codex",
|
|
@@ -145,6 +152,48 @@ export const gpt53Codex = presetFor()("openai/gpt-5.3-codex", {
|
|
|
145
152
|
created: "2026-02-05",
|
|
146
153
|
knowledge: "2025-08",
|
|
147
154
|
});
|
|
155
|
+
export const gpt53CodexSpark = presetFor()("openai/gpt-5.3-codex-spark", {
|
|
156
|
+
...GPT_BASE,
|
|
157
|
+
name: "GPT-5.3 Codex Spark",
|
|
158
|
+
created: "2026-02-05",
|
|
159
|
+
knowledge: "2025-08",
|
|
160
|
+
context: 128000,
|
|
161
|
+
});
|
|
162
|
+
export const gpt53Chat = presetFor()("openai/gpt-5.3-chat", {
|
|
163
|
+
...GPT_BASE,
|
|
164
|
+
name: "GPT-5.3 Chat",
|
|
165
|
+
created: "2026-03-03",
|
|
166
|
+
knowledge: "2025-08",
|
|
167
|
+
context: 128000,
|
|
168
|
+
});
|
|
169
|
+
export const gpt54 = presetFor()("openai/gpt-5.4", {
|
|
170
|
+
...GPT_BASE,
|
|
171
|
+
name: "GPT-5.4",
|
|
172
|
+
created: "2026-03-05",
|
|
173
|
+
knowledge: "2025-08",
|
|
174
|
+
context: 1050000,
|
|
175
|
+
});
|
|
176
|
+
export const gpt54Mini = presetFor()("openai/gpt-5.4-mini", {
|
|
177
|
+
...GPT_BASE,
|
|
178
|
+
name: "GPT-5.4 Mini",
|
|
179
|
+
created: "2026-03-17",
|
|
180
|
+
knowledge: "2025-08",
|
|
181
|
+
context: 400000,
|
|
182
|
+
});
|
|
183
|
+
export const gpt54Nano = presetFor()("openai/gpt-5.4-nano", {
|
|
184
|
+
...GPT_BASE,
|
|
185
|
+
name: "GPT-5.4 Nano",
|
|
186
|
+
created: "2026-03-17",
|
|
187
|
+
knowledge: "2025-08",
|
|
188
|
+
context: 400000,
|
|
189
|
+
});
|
|
190
|
+
export const gpt54Pro = presetFor()("openai/gpt-5.4-pro", {
|
|
191
|
+
...GPT_PRO_BASE,
|
|
192
|
+
name: "GPT-5.4 Pro",
|
|
193
|
+
created: "2026-03-05",
|
|
194
|
+
knowledge: "2025-08",
|
|
195
|
+
context: 1050000,
|
|
196
|
+
});
|
|
148
197
|
export const textEmbedding3Small = presetFor()("openai/text-embedding-3-small", {
|
|
149
198
|
...EMBEDDINGS_BASE,
|
|
150
199
|
name: "Text Embedding 3 Small",
|
|
@@ -177,15 +226,30 @@ const gptOssGroups = {
|
|
|
177
226
|
};
|
|
178
227
|
const gptAtomic = {
|
|
179
228
|
v5: [gpt5, gpt5Mini, gpt5Nano, gpt5Pro],
|
|
180
|
-
"v5.1": [gpt51, gpt51Chat, gpt51Codex, gpt51CodexMax],
|
|
229
|
+
"v5.1": [gpt51, gpt51Chat, gpt51Codex, gpt51CodexMax, gpt51CodexMini],
|
|
181
230
|
"v5.2": [gpt52, gpt52Chat, gpt52Pro, gpt52Codex],
|
|
182
|
-
"v5.3": [gpt53Codex],
|
|
183
|
-
|
|
184
|
-
|
|
185
|
-
|
|
231
|
+
"v5.3": [gpt53Codex, gpt53CodexSpark, gpt53Chat],
|
|
232
|
+
"v5.4": [gpt54, gpt54Mini, gpt54Nano, gpt54Pro],
|
|
233
|
+
codex: [
|
|
234
|
+
gpt5Codex,
|
|
235
|
+
gpt51Codex,
|
|
236
|
+
gpt51CodexMax,
|
|
237
|
+
gpt51CodexMini,
|
|
238
|
+
gpt52Codex,
|
|
239
|
+
gpt53Codex,
|
|
240
|
+
gpt53CodexSpark,
|
|
241
|
+
],
|
|
242
|
+
chat: [gpt51Chat, gpt52Chat, gpt53Chat],
|
|
243
|
+
pro: [gpt5Pro, gpt52Pro, gpt54Pro],
|
|
186
244
|
};
|
|
187
245
|
const gptGroups = {
|
|
188
|
-
"v5.x": [
|
|
246
|
+
"v5.x": [
|
|
247
|
+
...gptAtomic["v5"],
|
|
248
|
+
...gptAtomic["v5.1"],
|
|
249
|
+
...gptAtomic["v5.2"],
|
|
250
|
+
...gptAtomic["v5.3"],
|
|
251
|
+
...gptAtomic["v5.4"],
|
|
252
|
+
],
|
|
189
253
|
};
|
|
190
254
|
const textEmbeddingsAtomic = {
|
|
191
255
|
v3: [textEmbedding3Small, textEmbedding3Large],
|
|
@@ -202,7 +266,7 @@ export const gptOss = {
|
|
|
202
266
|
export const gpt = {
|
|
203
267
|
...gptAtomic,
|
|
204
268
|
...gptGroups,
|
|
205
|
-
latest: [
|
|
269
|
+
latest: [gpt54, gpt54Mini, gpt54Nano],
|
|
206
270
|
all: Object.values(gptAtomic).flat(),
|
|
207
271
|
};
|
|
208
272
|
export const textEmbeddings = {
|
package/dist/models/types.d.ts
CHANGED
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import type { ProviderId } from "../providers/types";
|
|
2
|
-
export declare const CANONICAL_MODEL_IDS: readonly ["anthropic/claude-opus-4.6", "anthropic/claude-sonnet-4.6", "anthropic/claude-haiku-4.5", "anthropic/claude-sonnet-4.5", "anthropic/claude-opus-4.5", "anthropic/claude-opus-4.1", "anthropic/claude-opus-4", "anthropic/claude-sonnet-4", "anthropic/claude-sonnet-3.7", "anthropic/claude-sonnet-3.5", "anthropic/claude-haiku-3.5", "anthropic/claude-haiku-3", "openai/gpt-oss-20b", "openai/gpt-oss-120b", "openai/gpt-5", "openai/gpt-5-pro", "openai/gpt-5.2", "openai/gpt-5.2-chat", "openai/gpt-5.2-pro", "openai/gpt-5.2-codex", "openai/gpt-5.3-codex", "openai/gpt-5-mini", "openai/gpt-5-nano", "openai/gpt-5-codex", "openai/gpt-5.1-codex", "openai/gpt-5.1-codex-max", "openai/gpt-5.1-chat", "openai/gpt-5.1", "openai/text-embedding-3-small", "openai/text-embedding-3-large", "amazon/nova-micro", "amazon/nova-lite", "amazon/nova-pro", "amazon/nova-premier", "amazon/nova-2-lite", "amazon/nova-2-multimodal-embeddings", "google/gemini-2.5-flash-lite", "google/gemini-2.5-flash", "google/gemini-2.5-pro", "google/gemini-3-flash-preview", "google/gemini-3.1-flash-lite-preview", "google/gemini-3.1-pro-preview", "google/gemini-embedding-2-preview", "google/embedding-001", "meta/llama-3.1-8b", "meta/llama-3.1-70b", "meta/llama-3.1-405b", "meta/llama-3.2-1b", "meta/llama-3.2-3b", "meta/llama-3.2-11b", "meta/llama-3.2-90b", "meta/llama-3.3-70b", "meta/llama-4-scout", "meta/llama-4-maverick", "cohere/embed-v4.0", "cohere/embed-english-v3.0", "cohere/embed-english-light-v3.0", "cohere/embed-multilingual-v3.0", "cohere/embed-multilingual-light-v3.0", "cohere/command-a", "cohere/command-r7b", "cohere/command-a-translate", "cohere/command-a-reasoning", "cohere/command-a-vision", "cohere/command-r", "cohere/command-r-plus", "voyage/voyage-2-code", "voyage/voyage-2-law", "voyage/voyage-2-finance", "voyage/voyage-3-code", "voyage/voyage-3-large", "voyage/voyage-3.5-lite", "voyage/voyage-3.5", "voyage/voyage-4-lite", "voyage/voyage-4", "voyage/voyage-4-large"];
|
|
2
|
+
export declare const CANONICAL_MODEL_IDS: readonly ["anthropic/claude-opus-4.6", "anthropic/claude-sonnet-4.6", "anthropic/claude-haiku-4.5", "anthropic/claude-sonnet-4.5", "anthropic/claude-opus-4.5", "anthropic/claude-opus-4.1", "anthropic/claude-opus-4", "anthropic/claude-sonnet-4", "anthropic/claude-sonnet-3.7", "anthropic/claude-sonnet-3.5", "anthropic/claude-haiku-3.5", "anthropic/claude-haiku-3", "openai/gpt-oss-20b", "openai/gpt-oss-120b", "openai/gpt-5", "openai/gpt-5-pro", "openai/gpt-5.2", "openai/gpt-5.2-chat", "openai/gpt-5.2-pro", "openai/gpt-5.2-codex", "openai/gpt-5.3-codex", "openai/gpt-5.3-codex-spark", "openai/gpt-5.3-chat", "openai/gpt-5.4", "openai/gpt-5.4-mini", "openai/gpt-5.4-nano", "openai/gpt-5.4-pro", "openai/gpt-5-mini", "openai/gpt-5-nano", "openai/gpt-5-codex", "openai/gpt-5.1-codex", "openai/gpt-5.1-codex-max", "openai/gpt-5.1-codex-mini", "openai/gpt-5.1-chat", "openai/gpt-5.1", "openai/text-embedding-3-small", "openai/text-embedding-3-large", "amazon/nova-micro", "amazon/nova-lite", "amazon/nova-pro", "amazon/nova-premier", "amazon/nova-2-lite", "amazon/nova-2-multimodal-embeddings", "google/gemini-2.5-flash-lite", "google/gemini-2.5-flash", "google/gemini-2.5-pro", "google/gemini-3-flash-preview", "google/gemini-3.1-flash-lite-preview", "google/gemini-3.1-pro-preview", "google/gemini-embedding-2-preview", "google/embedding-001", "meta/llama-3.1-8b", "meta/llama-3.1-70b", "meta/llama-3.1-405b", "meta/llama-3.2-1b", "meta/llama-3.2-3b", "meta/llama-3.2-11b", "meta/llama-3.2-90b", "meta/llama-3.3-70b", "meta/llama-4-scout", "meta/llama-4-maverick", "cohere/embed-v4.0", "cohere/embed-english-v3.0", "cohere/embed-english-light-v3.0", "cohere/embed-multilingual-v3.0", "cohere/embed-multilingual-light-v3.0", "cohere/command-a", "cohere/command-r7b", "cohere/command-a-translate", "cohere/command-a-reasoning", "cohere/command-a-vision", "cohere/command-r", "cohere/command-r-plus", "voyage/voyage-2-code", "voyage/voyage-2-law", "voyage/voyage-2-finance", "voyage/voyage-3-code", "voyage/voyage-3-large", "voyage/voyage-3.5-lite", "voyage/voyage-3.5", "voyage/voyage-4-lite", "voyage/voyage-4", "voyage/voyage-4-large"];
|
|
3
3
|
export type CanonicalModelId = (typeof CANONICAL_MODEL_IDS)[number];
|
|
4
4
|
export type ModelId = CanonicalModelId | (string & {});
|
|
5
5
|
export type CatalogModel = {
|
package/dist/models/types.js
CHANGED
|
@@ -22,11 +22,18 @@ export const CANONICAL_MODEL_IDS = [
|
|
|
22
22
|
"openai/gpt-5.2-pro",
|
|
23
23
|
"openai/gpt-5.2-codex",
|
|
24
24
|
"openai/gpt-5.3-codex",
|
|
25
|
+
"openai/gpt-5.3-codex-spark",
|
|
26
|
+
"openai/gpt-5.3-chat",
|
|
27
|
+
"openai/gpt-5.4",
|
|
28
|
+
"openai/gpt-5.4-mini",
|
|
29
|
+
"openai/gpt-5.4-nano",
|
|
30
|
+
"openai/gpt-5.4-pro",
|
|
25
31
|
"openai/gpt-5-mini",
|
|
26
32
|
"openai/gpt-5-nano",
|
|
27
33
|
"openai/gpt-5-codex",
|
|
28
34
|
"openai/gpt-5.1-codex",
|
|
29
35
|
"openai/gpt-5.1-codex-max",
|
|
36
|
+
"openai/gpt-5.1-codex-mini",
|
|
30
37
|
"openai/gpt-5.1-chat",
|
|
31
38
|
"openai/gpt-5.1",
|
|
32
39
|
"openai/text-embedding-3-small",
|
package/package.json
CHANGED