@hebo-ai/gateway 0.8.0-rc3 → 0.8.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -95,6 +95,18 @@ export declare const gpt51CodexMax: import("../../utils/preset").Preset<"openai/
95
95
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
96
96
  providers: readonly ["openai", "azure"];
97
97
  }>;
98
+ export declare const gpt51CodexMini: import("../../utils/preset").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
99
+ name: string;
100
+ created: string;
101
+ knowledge: string;
102
+ context: number;
103
+ modalities: {
104
+ input: readonly ["text", "image"];
105
+ output: readonly ["text"];
106
+ };
107
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
108
+ providers: readonly ["openai", "azure"];
109
+ }>;
98
110
  export declare const gpt5Codex: import("../../utils/preset").Preset<"openai/gpt-5-codex", CatalogModel, {
99
111
  name: string;
100
112
  created: string;
@@ -165,6 +177,78 @@ export declare const gpt53Codex: import("../../utils/preset").Preset<"openai/gpt
165
177
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
166
178
  providers: readonly ["openai", "azure"];
167
179
  }>;
180
+ export declare const gpt53CodexSpark: import("../../utils/preset").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
181
+ name: string;
182
+ created: string;
183
+ knowledge: string;
184
+ context: number;
185
+ modalities: {
186
+ input: readonly ["text", "image"];
187
+ output: readonly ["text"];
188
+ };
189
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
190
+ providers: readonly ["openai", "azure"];
191
+ }>;
192
+ export declare const gpt53Chat: import("../../utils/preset").Preset<"openai/gpt-5.3-chat", CatalogModel, {
193
+ name: string;
194
+ created: string;
195
+ knowledge: string;
196
+ context: number;
197
+ modalities: {
198
+ input: readonly ["text", "image"];
199
+ output: readonly ["text"];
200
+ };
201
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
202
+ providers: readonly ["openai", "azure"];
203
+ }>;
204
+ export declare const gpt54: import("../../utils/preset").Preset<"openai/gpt-5.4", CatalogModel, {
205
+ name: string;
206
+ created: string;
207
+ knowledge: string;
208
+ context: number;
209
+ modalities: {
210
+ input: readonly ["text", "image"];
211
+ output: readonly ["text"];
212
+ };
213
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
214
+ providers: readonly ["openai", "azure"];
215
+ }>;
216
+ export declare const gpt54Mini: import("../../utils/preset").Preset<"openai/gpt-5.4-mini", CatalogModel, {
217
+ name: string;
218
+ created: string;
219
+ knowledge: string;
220
+ context: number;
221
+ modalities: {
222
+ input: readonly ["text", "image"];
223
+ output: readonly ["text"];
224
+ };
225
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
226
+ providers: readonly ["openai", "azure"];
227
+ }>;
228
+ export declare const gpt54Nano: import("../../utils/preset").Preset<"openai/gpt-5.4-nano", CatalogModel, {
229
+ name: string;
230
+ created: string;
231
+ knowledge: string;
232
+ context: number;
233
+ modalities: {
234
+ input: readonly ["text", "image"];
235
+ output: readonly ["text"];
236
+ };
237
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
238
+ providers: readonly ["openai", "azure"];
239
+ }>;
240
+ export declare const gpt54Pro: import("../../utils/preset").Preset<"openai/gpt-5.4-pro", CatalogModel, {
241
+ name: string;
242
+ created: string;
243
+ knowledge: string;
244
+ context: number;
245
+ modalities: {
246
+ input: readonly ["text", "image"];
247
+ output: readonly ["text"];
248
+ };
249
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
250
+ providers: readonly ["openai", "azure"];
251
+ }>;
168
252
  export declare const textEmbedding3Small: import("../../utils/preset").Preset<"openai/text-embedding-3-small", CatalogModel, {
169
253
  name: string;
170
254
  created: string;
@@ -304,7 +388,7 @@ export declare const gptOss: {
304
388
  }>];
305
389
  };
306
390
  export declare const gpt: {
307
- readonly latest: readonly [import("../../utils/preset").Preset<"openai/gpt-5.2", CatalogModel, {
391
+ readonly latest: readonly [import("../../utils/preset").Preset<"openai/gpt-5.4", CatalogModel, {
308
392
  name: string;
309
393
  created: string;
310
394
  knowledge: string;
@@ -315,7 +399,7 @@ export declare const gpt: {
315
399
  };
316
400
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
317
401
  providers: readonly ["openai", "azure"];
318
- }>, import("../../utils/preset").Preset<"openai/gpt-5-mini", CatalogModel, {
402
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.4-mini", CatalogModel, {
319
403
  name: string;
320
404
  created: string;
321
405
  knowledge: string;
@@ -326,7 +410,7 @@ export declare const gpt: {
326
410
  };
327
411
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
328
412
  providers: readonly ["openai", "azure"];
329
- }>, import("../../utils/preset").Preset<"openai/gpt-5-nano", CatalogModel, {
413
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.4-nano", CatalogModel, {
330
414
  name: string;
331
415
  created: string;
332
416
  knowledge: string;
@@ -426,6 +510,17 @@ export declare const gpt: {
426
510
  };
427
511
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
428
512
  providers: readonly ["openai", "azure"];
513
+ }> | import("../../utils/preset").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
514
+ name: string;
515
+ created: string;
516
+ knowledge: string;
517
+ context: number;
518
+ modalities: {
519
+ input: readonly ["text", "image"];
520
+ output: readonly ["text"];
521
+ };
522
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
523
+ providers: readonly ["openai", "azure"];
429
524
  }> | import("../../utils/preset").Preset<"openai/gpt-5-codex", CatalogModel, {
430
525
  name: string;
431
526
  created: string;
@@ -490,6 +585,72 @@ export declare const gpt: {
490
585
  };
491
586
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
492
587
  providers: readonly ["openai", "azure"];
588
+ }> | import("../../utils/preset").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
589
+ name: string;
590
+ created: string;
591
+ knowledge: string;
592
+ context: number;
593
+ modalities: {
594
+ input: readonly ["text", "image"];
595
+ output: readonly ["text"];
596
+ };
597
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
598
+ providers: readonly ["openai", "azure"];
599
+ }> | import("../../utils/preset").Preset<"openai/gpt-5.3-chat", CatalogModel, {
600
+ name: string;
601
+ created: string;
602
+ knowledge: string;
603
+ context: number;
604
+ modalities: {
605
+ input: readonly ["text", "image"];
606
+ output: readonly ["text"];
607
+ };
608
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
609
+ providers: readonly ["openai", "azure"];
610
+ }> | import("../../utils/preset").Preset<"openai/gpt-5.4", CatalogModel, {
611
+ name: string;
612
+ created: string;
613
+ knowledge: string;
614
+ context: number;
615
+ modalities: {
616
+ input: readonly ["text", "image"];
617
+ output: readonly ["text"];
618
+ };
619
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
620
+ providers: readonly ["openai", "azure"];
621
+ }> | import("../../utils/preset").Preset<"openai/gpt-5.4-mini", CatalogModel, {
622
+ name: string;
623
+ created: string;
624
+ knowledge: string;
625
+ context: number;
626
+ modalities: {
627
+ input: readonly ["text", "image"];
628
+ output: readonly ["text"];
629
+ };
630
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
631
+ providers: readonly ["openai", "azure"];
632
+ }> | import("../../utils/preset").Preset<"openai/gpt-5.4-nano", CatalogModel, {
633
+ name: string;
634
+ created: string;
635
+ knowledge: string;
636
+ context: number;
637
+ modalities: {
638
+ input: readonly ["text", "image"];
639
+ output: readonly ["text"];
640
+ };
641
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
642
+ providers: readonly ["openai", "azure"];
643
+ }> | import("../../utils/preset").Preset<"openai/gpt-5.4-pro", CatalogModel, {
644
+ name: string;
645
+ created: string;
646
+ knowledge: string;
647
+ context: number;
648
+ modalities: {
649
+ input: readonly ["text", "image"];
650
+ output: readonly ["text"];
651
+ };
652
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
653
+ providers: readonly ["openai", "azure"];
493
654
  }>)[];
494
655
  readonly "v5.x": readonly [import("../../utils/preset").Preset<"openai/gpt-5", CatalogModel, {
495
656
  name: string;
@@ -579,6 +740,17 @@ export declare const gpt: {
579
740
  };
580
741
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
581
742
  providers: readonly ["openai", "azure"];
743
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
744
+ name: string;
745
+ created: string;
746
+ knowledge: string;
747
+ context: number;
748
+ modalities: {
749
+ input: readonly ["text", "image"];
750
+ output: readonly ["text"];
751
+ };
752
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
753
+ providers: readonly ["openai", "azure"];
582
754
  }>, import("../../utils/preset").Preset<"openai/gpt-5.2", CatalogModel, {
583
755
  name: string;
584
756
  created: string;
@@ -632,6 +804,72 @@ export declare const gpt: {
632
804
  };
633
805
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
634
806
  providers: readonly ["openai", "azure"];
807
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
808
+ name: string;
809
+ created: string;
810
+ knowledge: string;
811
+ context: number;
812
+ modalities: {
813
+ input: readonly ["text", "image"];
814
+ output: readonly ["text"];
815
+ };
816
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
817
+ providers: readonly ["openai", "azure"];
818
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.3-chat", CatalogModel, {
819
+ name: string;
820
+ created: string;
821
+ knowledge: string;
822
+ context: number;
823
+ modalities: {
824
+ input: readonly ["text", "image"];
825
+ output: readonly ["text"];
826
+ };
827
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
828
+ providers: readonly ["openai", "azure"];
829
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.4", CatalogModel, {
830
+ name: string;
831
+ created: string;
832
+ knowledge: string;
833
+ context: number;
834
+ modalities: {
835
+ input: readonly ["text", "image"];
836
+ output: readonly ["text"];
837
+ };
838
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
839
+ providers: readonly ["openai", "azure"];
840
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.4-mini", CatalogModel, {
841
+ name: string;
842
+ created: string;
843
+ knowledge: string;
844
+ context: number;
845
+ modalities: {
846
+ input: readonly ["text", "image"];
847
+ output: readonly ["text"];
848
+ };
849
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
850
+ providers: readonly ["openai", "azure"];
851
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.4-nano", CatalogModel, {
852
+ name: string;
853
+ created: string;
854
+ knowledge: string;
855
+ context: number;
856
+ modalities: {
857
+ input: readonly ["text", "image"];
858
+ output: readonly ["text"];
859
+ };
860
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
861
+ providers: readonly ["openai", "azure"];
862
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.4-pro", CatalogModel, {
863
+ name: string;
864
+ created: string;
865
+ knowledge: string;
866
+ context: number;
867
+ modalities: {
868
+ input: readonly ["text", "image"];
869
+ output: readonly ["text"];
870
+ };
871
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
872
+ providers: readonly ["openai", "azure"];
635
873
  }>];
636
874
  readonly v5: readonly [import("../../utils/preset").Preset<"openai/gpt-5", CatalogModel, {
637
875
  name: string;
@@ -722,6 +960,17 @@ export declare const gpt: {
722
960
  };
723
961
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
724
962
  providers: readonly ["openai", "azure"];
963
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
964
+ name: string;
965
+ created: string;
966
+ knowledge: string;
967
+ context: number;
968
+ modalities: {
969
+ input: readonly ["text", "image"];
970
+ output: readonly ["text"];
971
+ };
972
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
973
+ providers: readonly ["openai", "azure"];
725
974
  }>];
726
975
  readonly "v5.2": readonly [import("../../utils/preset").Preset<"openai/gpt-5.2", CatalogModel, {
727
976
  name: string;
@@ -777,6 +1026,73 @@ export declare const gpt: {
777
1026
  };
778
1027
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
779
1028
  providers: readonly ["openai", "azure"];
1029
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
1030
+ name: string;
1031
+ created: string;
1032
+ knowledge: string;
1033
+ context: number;
1034
+ modalities: {
1035
+ input: readonly ["text", "image"];
1036
+ output: readonly ["text"];
1037
+ };
1038
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
1039
+ providers: readonly ["openai", "azure"];
1040
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.3-chat", CatalogModel, {
1041
+ name: string;
1042
+ created: string;
1043
+ knowledge: string;
1044
+ context: number;
1045
+ modalities: {
1046
+ input: readonly ["text", "image"];
1047
+ output: readonly ["text"];
1048
+ };
1049
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
1050
+ providers: readonly ["openai", "azure"];
1051
+ }>];
1052
+ readonly "v5.4": readonly [import("../../utils/preset").Preset<"openai/gpt-5.4", CatalogModel, {
1053
+ name: string;
1054
+ created: string;
1055
+ knowledge: string;
1056
+ context: number;
1057
+ modalities: {
1058
+ input: readonly ["text", "image"];
1059
+ output: readonly ["text"];
1060
+ };
1061
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
1062
+ providers: readonly ["openai", "azure"];
1063
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.4-mini", CatalogModel, {
1064
+ name: string;
1065
+ created: string;
1066
+ knowledge: string;
1067
+ context: number;
1068
+ modalities: {
1069
+ input: readonly ["text", "image"];
1070
+ output: readonly ["text"];
1071
+ };
1072
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
1073
+ providers: readonly ["openai", "azure"];
1074
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.4-nano", CatalogModel, {
1075
+ name: string;
1076
+ created: string;
1077
+ knowledge: string;
1078
+ context: number;
1079
+ modalities: {
1080
+ input: readonly ["text", "image"];
1081
+ output: readonly ["text"];
1082
+ };
1083
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
1084
+ providers: readonly ["openai", "azure"];
1085
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.4-pro", CatalogModel, {
1086
+ name: string;
1087
+ created: string;
1088
+ knowledge: string;
1089
+ context: number;
1090
+ modalities: {
1091
+ input: readonly ["text", "image"];
1092
+ output: readonly ["text"];
1093
+ };
1094
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
1095
+ providers: readonly ["openai", "azure"];
780
1096
  }>];
781
1097
  readonly codex: readonly [import("../../utils/preset").Preset<"openai/gpt-5-codex", CatalogModel, {
782
1098
  name: string;
@@ -811,6 +1127,17 @@ export declare const gpt: {
811
1127
  };
812
1128
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
813
1129
  providers: readonly ["openai", "azure"];
1130
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
1131
+ name: string;
1132
+ created: string;
1133
+ knowledge: string;
1134
+ context: number;
1135
+ modalities: {
1136
+ input: readonly ["text", "image"];
1137
+ output: readonly ["text"];
1138
+ };
1139
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
1140
+ providers: readonly ["openai", "azure"];
814
1141
  }>, import("../../utils/preset").Preset<"openai/gpt-5.2-codex", CatalogModel, {
815
1142
  name: string;
816
1143
  created: string;
@@ -831,6 +1158,17 @@ export declare const gpt: {
831
1158
  };
832
1159
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
833
1160
  providers: readonly ["openai", "azure"];
1161
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
1162
+ name: string;
1163
+ created: string;
1164
+ knowledge: string;
1165
+ context: number;
1166
+ modalities: {
1167
+ input: readonly ["text", "image"];
1168
+ output: readonly ["text"];
1169
+ };
1170
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
1171
+ providers: readonly ["openai", "azure"];
834
1172
  }>];
835
1173
  readonly chat: readonly [import("../../utils/preset").Preset<"openai/gpt-5.1-chat", CatalogModel, {
836
1174
  name: string;
@@ -854,6 +1192,17 @@ export declare const gpt: {
854
1192
  };
855
1193
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
856
1194
  providers: readonly ["openai", "azure"];
1195
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.3-chat", CatalogModel, {
1196
+ name: string;
1197
+ created: string;
1198
+ knowledge: string;
1199
+ context: number;
1200
+ modalities: {
1201
+ input: readonly ["text", "image"];
1202
+ output: readonly ["text"];
1203
+ };
1204
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
1205
+ providers: readonly ["openai", "azure"];
857
1206
  }>];
858
1207
  readonly pro: readonly [import("../../utils/preset").Preset<"openai/gpt-5-pro", CatalogModel, {
859
1208
  name: string;
@@ -877,6 +1226,17 @@ export declare const gpt: {
877
1226
  };
878
1227
  capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
879
1228
  providers: readonly ["openai", "azure"];
1229
+ }>, import("../../utils/preset").Preset<"openai/gpt-5.4-pro", CatalogModel, {
1230
+ name: string;
1231
+ created: string;
1232
+ knowledge: string;
1233
+ context: number;
1234
+ modalities: {
1235
+ input: readonly ["text", "image"];
1236
+ output: readonly ["text"];
1237
+ };
1238
+ capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
1239
+ providers: readonly ["openai", "azure"];
880
1240
  }>];
881
1241
  };
882
1242
  export declare const textEmbeddings: {
@@ -105,6 +105,13 @@ export const gpt51CodexMax = presetFor()("openai/gpt-5.1-codex-max", {
105
105
  knowledge: "2024-09",
106
106
  context: 400000,
107
107
  });
108
+ export const gpt51CodexMini = presetFor()("openai/gpt-5.1-codex-mini", {
109
+ ...GPT_BASE,
110
+ name: "GPT-5.1 Codex Mini",
111
+ created: "2025-11-13",
112
+ knowledge: "2024-09",
113
+ context: 400000,
114
+ });
108
115
  export const gpt5Codex = presetFor()("openai/gpt-5-codex", {
109
116
  ...GPT_BASE,
110
117
  name: "GPT-5 Codex",
@@ -145,6 +152,48 @@ export const gpt53Codex = presetFor()("openai/gpt-5.3-codex", {
145
152
  created: "2026-02-05",
146
153
  knowledge: "2025-08",
147
154
  });
155
+ export const gpt53CodexSpark = presetFor()("openai/gpt-5.3-codex-spark", {
156
+ ...GPT_BASE,
157
+ name: "GPT-5.3 Codex Spark",
158
+ created: "2026-02-05",
159
+ knowledge: "2025-08",
160
+ context: 128000,
161
+ });
162
+ export const gpt53Chat = presetFor()("openai/gpt-5.3-chat", {
163
+ ...GPT_BASE,
164
+ name: "GPT-5.3 Chat",
165
+ created: "2026-03-03",
166
+ knowledge: "2025-08",
167
+ context: 128000,
168
+ });
169
+ export const gpt54 = presetFor()("openai/gpt-5.4", {
170
+ ...GPT_BASE,
171
+ name: "GPT-5.4",
172
+ created: "2026-03-05",
173
+ knowledge: "2025-08",
174
+ context: 1050000,
175
+ });
176
+ export const gpt54Mini = presetFor()("openai/gpt-5.4-mini", {
177
+ ...GPT_BASE,
178
+ name: "GPT-5.4 Mini",
179
+ created: "2026-03-17",
180
+ knowledge: "2025-08",
181
+ context: 400000,
182
+ });
183
+ export const gpt54Nano = presetFor()("openai/gpt-5.4-nano", {
184
+ ...GPT_BASE,
185
+ name: "GPT-5.4 Nano",
186
+ created: "2026-03-17",
187
+ knowledge: "2025-08",
188
+ context: 400000,
189
+ });
190
+ export const gpt54Pro = presetFor()("openai/gpt-5.4-pro", {
191
+ ...GPT_PRO_BASE,
192
+ name: "GPT-5.4 Pro",
193
+ created: "2026-03-05",
194
+ knowledge: "2025-08",
195
+ context: 1050000,
196
+ });
148
197
  export const textEmbedding3Small = presetFor()("openai/text-embedding-3-small", {
149
198
  ...EMBEDDINGS_BASE,
150
199
  name: "Text Embedding 3 Small",
@@ -177,15 +226,30 @@ const gptOssGroups = {
177
226
  };
178
227
  const gptAtomic = {
179
228
  v5: [gpt5, gpt5Mini, gpt5Nano, gpt5Pro],
180
- "v5.1": [gpt51, gpt51Chat, gpt51Codex, gpt51CodexMax],
229
+ "v5.1": [gpt51, gpt51Chat, gpt51Codex, gpt51CodexMax, gpt51CodexMini],
181
230
  "v5.2": [gpt52, gpt52Chat, gpt52Pro, gpt52Codex],
182
- "v5.3": [gpt53Codex],
183
- codex: [gpt5Codex, gpt51Codex, gpt51CodexMax, gpt52Codex, gpt53Codex],
184
- chat: [gpt51Chat, gpt52Chat],
185
- pro: [gpt5Pro, gpt52Pro],
231
+ "v5.3": [gpt53Codex, gpt53CodexSpark, gpt53Chat],
232
+ "v5.4": [gpt54, gpt54Mini, gpt54Nano, gpt54Pro],
233
+ codex: [
234
+ gpt5Codex,
235
+ gpt51Codex,
236
+ gpt51CodexMax,
237
+ gpt51CodexMini,
238
+ gpt52Codex,
239
+ gpt53Codex,
240
+ gpt53CodexSpark,
241
+ ],
242
+ chat: [gpt51Chat, gpt52Chat, gpt53Chat],
243
+ pro: [gpt5Pro, gpt52Pro, gpt54Pro],
186
244
  };
187
245
  const gptGroups = {
188
- "v5.x": [...gptAtomic["v5"], ...gptAtomic["v5.1"], ...gptAtomic["v5.2"], ...gptAtomic["v5.3"]],
246
+ "v5.x": [
247
+ ...gptAtomic["v5"],
248
+ ...gptAtomic["v5.1"],
249
+ ...gptAtomic["v5.2"],
250
+ ...gptAtomic["v5.3"],
251
+ ...gptAtomic["v5.4"],
252
+ ],
189
253
  };
190
254
  const textEmbeddingsAtomic = {
191
255
  v3: [textEmbedding3Small, textEmbedding3Large],
@@ -202,7 +266,7 @@ export const gptOss = {
202
266
  export const gpt = {
203
267
  ...gptAtomic,
204
268
  ...gptGroups,
205
- latest: [gpt52, gpt5Mini, gpt5Nano],
269
+ latest: [gpt54, gpt54Mini, gpt54Nano],
206
270
  all: Object.values(gptAtomic).flat(),
207
271
  };
208
272
  export const textEmbeddings = {
@@ -1,5 +1,5 @@
1
1
  import type { ProviderId } from "../providers/types";
2
- export declare const CANONICAL_MODEL_IDS: readonly ["anthropic/claude-opus-4.6", "anthropic/claude-sonnet-4.6", "anthropic/claude-haiku-4.5", "anthropic/claude-sonnet-4.5", "anthropic/claude-opus-4.5", "anthropic/claude-opus-4.1", "anthropic/claude-opus-4", "anthropic/claude-sonnet-4", "anthropic/claude-sonnet-3.7", "anthropic/claude-sonnet-3.5", "anthropic/claude-haiku-3.5", "anthropic/claude-haiku-3", "openai/gpt-oss-20b", "openai/gpt-oss-120b", "openai/gpt-5", "openai/gpt-5-pro", "openai/gpt-5.2", "openai/gpt-5.2-chat", "openai/gpt-5.2-pro", "openai/gpt-5.2-codex", "openai/gpt-5.3-codex", "openai/gpt-5-mini", "openai/gpt-5-nano", "openai/gpt-5-codex", "openai/gpt-5.1-codex", "openai/gpt-5.1-codex-max", "openai/gpt-5.1-chat", "openai/gpt-5.1", "openai/text-embedding-3-small", "openai/text-embedding-3-large", "amazon/nova-micro", "amazon/nova-lite", "amazon/nova-pro", "amazon/nova-premier", "amazon/nova-2-lite", "amazon/nova-2-multimodal-embeddings", "google/gemini-2.5-flash-lite", "google/gemini-2.5-flash", "google/gemini-2.5-pro", "google/gemini-3-flash-preview", "google/gemini-3.1-flash-lite-preview", "google/gemini-3.1-pro-preview", "google/gemini-embedding-2-preview", "google/embedding-001", "meta/llama-3.1-8b", "meta/llama-3.1-70b", "meta/llama-3.1-405b", "meta/llama-3.2-1b", "meta/llama-3.2-3b", "meta/llama-3.2-11b", "meta/llama-3.2-90b", "meta/llama-3.3-70b", "meta/llama-4-scout", "meta/llama-4-maverick", "cohere/embed-v4.0", "cohere/embed-english-v3.0", "cohere/embed-english-light-v3.0", "cohere/embed-multilingual-v3.0", "cohere/embed-multilingual-light-v3.0", "cohere/command-a", "cohere/command-r7b", "cohere/command-a-translate", "cohere/command-a-reasoning", "cohere/command-a-vision", "cohere/command-r", "cohere/command-r-plus", "voyage/voyage-2-code", "voyage/voyage-2-law", "voyage/voyage-2-finance", "voyage/voyage-3-code", "voyage/voyage-3-large", "voyage/voyage-3.5-lite", "voyage/voyage-3.5", "voyage/voyage-4-lite", "voyage/voyage-4", "voyage/voyage-4-large"];
2
+ export declare const CANONICAL_MODEL_IDS: readonly ["anthropic/claude-opus-4.6", "anthropic/claude-sonnet-4.6", "anthropic/claude-haiku-4.5", "anthropic/claude-sonnet-4.5", "anthropic/claude-opus-4.5", "anthropic/claude-opus-4.1", "anthropic/claude-opus-4", "anthropic/claude-sonnet-4", "anthropic/claude-sonnet-3.7", "anthropic/claude-sonnet-3.5", "anthropic/claude-haiku-3.5", "anthropic/claude-haiku-3", "openai/gpt-oss-20b", "openai/gpt-oss-120b", "openai/gpt-5", "openai/gpt-5-pro", "openai/gpt-5.2", "openai/gpt-5.2-chat", "openai/gpt-5.2-pro", "openai/gpt-5.2-codex", "openai/gpt-5.3-codex", "openai/gpt-5.3-codex-spark", "openai/gpt-5.3-chat", "openai/gpt-5.4", "openai/gpt-5.4-mini", "openai/gpt-5.4-nano", "openai/gpt-5.4-pro", "openai/gpt-5-mini", "openai/gpt-5-nano", "openai/gpt-5-codex", "openai/gpt-5.1-codex", "openai/gpt-5.1-codex-max", "openai/gpt-5.1-codex-mini", "openai/gpt-5.1-chat", "openai/gpt-5.1", "openai/text-embedding-3-small", "openai/text-embedding-3-large", "amazon/nova-micro", "amazon/nova-lite", "amazon/nova-pro", "amazon/nova-premier", "amazon/nova-2-lite", "amazon/nova-2-multimodal-embeddings", "google/gemini-2.5-flash-lite", "google/gemini-2.5-flash", "google/gemini-2.5-pro", "google/gemini-3-flash-preview", "google/gemini-3.1-flash-lite-preview", "google/gemini-3.1-pro-preview", "google/gemini-embedding-2-preview", "google/embedding-001", "meta/llama-3.1-8b", "meta/llama-3.1-70b", "meta/llama-3.1-405b", "meta/llama-3.2-1b", "meta/llama-3.2-3b", "meta/llama-3.2-11b", "meta/llama-3.2-90b", "meta/llama-3.3-70b", "meta/llama-4-scout", "meta/llama-4-maverick", "cohere/embed-v4.0", "cohere/embed-english-v3.0", "cohere/embed-english-light-v3.0", "cohere/embed-multilingual-v3.0", "cohere/embed-multilingual-light-v3.0", "cohere/command-a", "cohere/command-r7b", "cohere/command-a-translate", "cohere/command-a-reasoning", "cohere/command-a-vision", "cohere/command-r", "cohere/command-r-plus", "voyage/voyage-2-code", "voyage/voyage-2-law", "voyage/voyage-2-finance", "voyage/voyage-3-code", "voyage/voyage-3-large", "voyage/voyage-3.5-lite", "voyage/voyage-3.5", "voyage/voyage-4-lite", "voyage/voyage-4", "voyage/voyage-4-large"];
3
3
  export type CanonicalModelId = (typeof CANONICAL_MODEL_IDS)[number];
4
4
  export type ModelId = CanonicalModelId | (string & {});
5
5
  export type CatalogModel = {
@@ -22,11 +22,18 @@ export const CANONICAL_MODEL_IDS = [
22
22
  "openai/gpt-5.2-pro",
23
23
  "openai/gpt-5.2-codex",
24
24
  "openai/gpt-5.3-codex",
25
+ "openai/gpt-5.3-codex-spark",
26
+ "openai/gpt-5.3-chat",
27
+ "openai/gpt-5.4",
28
+ "openai/gpt-5.4-mini",
29
+ "openai/gpt-5.4-nano",
30
+ "openai/gpt-5.4-pro",
25
31
  "openai/gpt-5-mini",
26
32
  "openai/gpt-5-nano",
27
33
  "openai/gpt-5-codex",
28
34
  "openai/gpt-5.1-codex",
29
35
  "openai/gpt-5.1-codex-max",
36
+ "openai/gpt-5.1-codex-mini",
30
37
  "openai/gpt-5.1-chat",
31
38
  "openai/gpt-5.1",
32
39
  "openai/text-embedding-3-small",
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@hebo-ai/gateway",
3
- "version": "0.8.0-rc3",
3
+ "version": "0.8.0",
4
4
  "description": "AI gateway as a framework. For full control over models, routing & lifecycle. OpenAI-compatible /chat/completions, /embeddings & /models.",
5
5
  "keywords": [
6
6
  "ai",