@hebo-ai/gateway 0.10.7 → 0.11.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +97 -84
- package/dist/config.js +21 -4
- package/dist/endpoints/chat-completions/handler.js +9 -5
- package/dist/endpoints/conversations/handler.js +3 -3
- package/dist/endpoints/embeddings/handler.js +2 -2
- package/dist/endpoints/messages/converters.js +17 -21
- package/dist/endpoints/messages/handler.js +5 -5
- package/dist/endpoints/responses/handler.js +9 -5
- package/dist/index.d.ts +1 -0
- package/dist/index.js +1 -0
- package/dist/models/alibaba/index.d.ts +2 -0
- package/dist/models/alibaba/index.js +2 -0
- package/dist/models/alibaba/middleware.d.ts +2 -0
- package/dist/models/alibaba/middleware.js +31 -0
- package/dist/models/alibaba/presets.d.ts +900 -0
- package/dist/models/alibaba/presets.js +252 -0
- package/dist/models/amazon/presets.d.ts +31 -31
- package/dist/models/anthropic/presets.d.ts +68 -68
- package/dist/models/cohere/presets.d.ts +38 -38
- package/dist/models/deepseek/index.d.ts +2 -0
- package/dist/models/deepseek/index.js +2 -0
- package/dist/models/deepseek/middleware.d.ts +2 -0
- package/dist/models/deepseek/middleware.js +25 -0
- package/dist/models/deepseek/presets.d.ts +51 -0
- package/dist/models/deepseek/presets.js +33 -0
- package/dist/models/google/presets.d.ts +94 -94
- package/dist/models/google/presets.js +3 -1
- package/dist/models/meta/presets.d.ts +84 -84
- package/dist/models/meta/presets.js +11 -12
- package/dist/models/minimax/index.d.ts +2 -0
- package/dist/models/minimax/index.js +2 -0
- package/dist/models/minimax/middleware.d.ts +2 -0
- package/dist/models/minimax/middleware.js +43 -0
- package/dist/models/minimax/presets.d.ts +99 -0
- package/dist/models/minimax/presets.js +47 -0
- package/dist/models/moonshot/index.d.ts +2 -0
- package/dist/models/moonshot/index.js +2 -0
- package/dist/models/moonshot/middleware.d.ts +2 -0
- package/dist/models/moonshot/middleware.js +32 -0
- package/dist/models/moonshot/presets.d.ts +104 -0
- package/dist/models/moonshot/presets.js +36 -0
- package/dist/models/openai/presets.d.ts +128 -128
- package/dist/models/openai/presets.js +1 -1
- package/dist/models/types.d.ts +1 -1
- package/dist/models/types.js +39 -0
- package/dist/models/voyage/presets.d.ts +46 -46
- package/dist/models/xai/index.d.ts +2 -0
- package/dist/models/xai/index.js +2 -0
- package/dist/models/xai/middleware.d.ts +2 -0
- package/dist/models/xai/middleware.js +37 -0
- package/dist/models/xai/presets.d.ts +210 -0
- package/dist/models/xai/presets.js +55 -0
- package/dist/models/zai/index.d.ts +2 -0
- package/dist/models/zai/index.js +2 -0
- package/dist/models/zai/middleware.d.ts +2 -0
- package/dist/models/zai/middleware.js +25 -0
- package/dist/models/zai/presets.d.ts +141 -0
- package/dist/models/zai/presets.js +41 -0
- package/dist/providers/alibaba/canonical.d.ts +3 -0
- package/dist/providers/alibaba/canonical.js +13 -0
- package/dist/providers/alibaba/index.d.ts +1 -0
- package/dist/providers/alibaba/index.js +1 -0
- package/dist/providers/bedrock/canonical.js +3 -0
- package/dist/providers/chutes/canonical.d.ts +3 -0
- package/dist/providers/chutes/canonical.js +14 -0
- package/dist/providers/chutes/index.d.ts +1 -0
- package/dist/providers/chutes/index.js +1 -0
- package/dist/providers/deepinfra/canonical.d.ts +3 -0
- package/dist/providers/deepinfra/canonical.js +40 -0
- package/dist/providers/deepinfra/index.d.ts +1 -0
- package/dist/providers/deepinfra/index.js +1 -0
- package/dist/providers/deepseek/canonical.d.ts +3 -0
- package/dist/providers/deepseek/canonical.js +9 -0
- package/dist/providers/deepseek/index.d.ts +1 -0
- package/dist/providers/deepseek/index.js +1 -0
- package/dist/providers/fireworks/canonical.d.ts +3 -0
- package/dist/providers/fireworks/canonical.js +27 -0
- package/dist/providers/fireworks/index.d.ts +2 -0
- package/dist/providers/fireworks/index.js +2 -0
- package/dist/providers/fireworks/middleware.d.ts +2 -0
- package/dist/providers/fireworks/middleware.js +35 -0
- package/dist/providers/groq/canonical.js +1 -1
- package/dist/providers/minimax/canonical.d.ts +3 -0
- package/dist/providers/minimax/canonical.js +9 -0
- package/dist/providers/minimax/index.d.ts +1 -0
- package/dist/providers/minimax/index.js +1 -0
- package/dist/providers/moonshot/canonical.d.ts +3 -0
- package/dist/providers/moonshot/canonical.js +6 -0
- package/dist/providers/moonshot/index.d.ts +1 -0
- package/dist/providers/moonshot/index.js +1 -0
- package/dist/providers/togetherai/canonical.d.ts +3 -0
- package/dist/providers/togetherai/canonical.js +24 -0
- package/dist/providers/togetherai/index.d.ts +1 -0
- package/dist/providers/togetherai/index.js +1 -0
- package/dist/providers/types.d.ts +1 -1
- package/dist/providers/types.js +10 -0
- package/dist/providers/vertex/canonical.js +5 -1
- package/dist/providers/xai/canonical.d.ts +3 -0
- package/dist/providers/xai/canonical.js +12 -0
- package/dist/providers/xai/index.d.ts +1 -0
- package/dist/providers/xai/index.js +1 -0
- package/dist/providers/zai/canonical.d.ts +3 -0
- package/dist/providers/zai/canonical.js +10 -0
- package/dist/providers/zai/index.d.ts +1 -0
- package/dist/providers/zai/index.js +1 -0
- package/dist/types.d.ts +30 -16
- package/dist/utils/index.d.ts +8 -0
- package/dist/utils/index.js +8 -0
- package/dist/utils/request.d.ts +2 -1
- package/dist/utils/request.js +35 -7
- package/package.json +133 -9
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import type { CatalogModel } from "../types";
|
|
2
|
-
export declare const gpt5: import("../../utils
|
|
2
|
+
export declare const gpt5: import("../../utils").Preset<"openai/gpt-5", CatalogModel, {
|
|
3
3
|
name: string;
|
|
4
4
|
created: string;
|
|
5
5
|
knowledge: string;
|
|
@@ -11,7 +11,7 @@ export declare const gpt5: import("../../utils/preset").Preset<"openai/gpt-5", C
|
|
|
11
11
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
12
12
|
providers: readonly ["openai", "azure"];
|
|
13
13
|
}>;
|
|
14
|
-
export declare const gpt5Pro: import("../../utils
|
|
14
|
+
export declare const gpt5Pro: import("../../utils").Preset<"openai/gpt-5-pro", CatalogModel, {
|
|
15
15
|
name: string;
|
|
16
16
|
created: string;
|
|
17
17
|
knowledge: string;
|
|
@@ -23,7 +23,7 @@ export declare const gpt5Pro: import("../../utils/preset").Preset<"openai/gpt-5-
|
|
|
23
23
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
24
24
|
providers: readonly ["openai", "azure"];
|
|
25
25
|
}>;
|
|
26
|
-
export declare const gpt5Mini: import("../../utils
|
|
26
|
+
export declare const gpt5Mini: import("../../utils").Preset<"openai/gpt-5-mini", CatalogModel, {
|
|
27
27
|
name: string;
|
|
28
28
|
created: string;
|
|
29
29
|
knowledge: string;
|
|
@@ -35,7 +35,7 @@ export declare const gpt5Mini: import("../../utils/preset").Preset<"openai/gpt-5
|
|
|
35
35
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
36
36
|
providers: readonly ["openai", "azure"];
|
|
37
37
|
}>;
|
|
38
|
-
export declare const gpt5Nano: import("../../utils
|
|
38
|
+
export declare const gpt5Nano: import("../../utils").Preset<"openai/gpt-5-nano", CatalogModel, {
|
|
39
39
|
name: string;
|
|
40
40
|
created: string;
|
|
41
41
|
knowledge: string;
|
|
@@ -47,7 +47,7 @@ export declare const gpt5Nano: import("../../utils/preset").Preset<"openai/gpt-5
|
|
|
47
47
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
48
48
|
providers: readonly ["openai", "azure"];
|
|
49
49
|
}>;
|
|
50
|
-
export declare const gpt51: import("../../utils
|
|
50
|
+
export declare const gpt51: import("../../utils").Preset<"openai/gpt-5.1", CatalogModel, {
|
|
51
51
|
name: string;
|
|
52
52
|
created: string;
|
|
53
53
|
knowledge: string;
|
|
@@ -59,7 +59,7 @@ export declare const gpt51: import("../../utils/preset").Preset<"openai/gpt-5.1"
|
|
|
59
59
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
60
60
|
providers: readonly ["openai", "azure"];
|
|
61
61
|
}>;
|
|
62
|
-
export declare const gpt51Chat: import("../../utils
|
|
62
|
+
export declare const gpt51Chat: import("../../utils").Preset<"openai/gpt-5.1-chat", CatalogModel, {
|
|
63
63
|
name: string;
|
|
64
64
|
created: string;
|
|
65
65
|
knowledge: string;
|
|
@@ -71,7 +71,7 @@ export declare const gpt51Chat: import("../../utils/preset").Preset<"openai/gpt-
|
|
|
71
71
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
72
72
|
providers: readonly ["openai", "azure"];
|
|
73
73
|
}>;
|
|
74
|
-
export declare const gpt51Codex: import("../../utils
|
|
74
|
+
export declare const gpt51Codex: import("../../utils").Preset<"openai/gpt-5.1-codex", CatalogModel, {
|
|
75
75
|
name: string;
|
|
76
76
|
created: string;
|
|
77
77
|
knowledge: string;
|
|
@@ -83,7 +83,7 @@ export declare const gpt51Codex: import("../../utils/preset").Preset<"openai/gpt
|
|
|
83
83
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
84
84
|
providers: readonly ["openai", "azure"];
|
|
85
85
|
}>;
|
|
86
|
-
export declare const gpt51CodexMax: import("../../utils
|
|
86
|
+
export declare const gpt51CodexMax: import("../../utils").Preset<"openai/gpt-5.1-codex-max", CatalogModel, {
|
|
87
87
|
name: string;
|
|
88
88
|
created: string;
|
|
89
89
|
knowledge: string;
|
|
@@ -95,7 +95,7 @@ export declare const gpt51CodexMax: import("../../utils/preset").Preset<"openai/
|
|
|
95
95
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
96
96
|
providers: readonly ["openai", "azure"];
|
|
97
97
|
}>;
|
|
98
|
-
export declare const gpt51CodexMini: import("../../utils
|
|
98
|
+
export declare const gpt51CodexMini: import("../../utils").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
|
|
99
99
|
name: string;
|
|
100
100
|
created: string;
|
|
101
101
|
knowledge: string;
|
|
@@ -107,7 +107,7 @@ export declare const gpt51CodexMini: import("../../utils/preset").Preset<"openai
|
|
|
107
107
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
108
108
|
providers: readonly ["openai", "azure"];
|
|
109
109
|
}>;
|
|
110
|
-
export declare const gpt5Codex: import("../../utils
|
|
110
|
+
export declare const gpt5Codex: import("../../utils").Preset<"openai/gpt-5-codex", CatalogModel, {
|
|
111
111
|
name: string;
|
|
112
112
|
created: string;
|
|
113
113
|
knowledge: string;
|
|
@@ -119,7 +119,7 @@ export declare const gpt5Codex: import("../../utils/preset").Preset<"openai/gpt-
|
|
|
119
119
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
120
120
|
providers: readonly ["openai", "azure"];
|
|
121
121
|
}>;
|
|
122
|
-
export declare const gpt52: import("../../utils
|
|
122
|
+
export declare const gpt52: import("../../utils").Preset<"openai/gpt-5.2", CatalogModel, {
|
|
123
123
|
name: string;
|
|
124
124
|
created: string;
|
|
125
125
|
knowledge: string;
|
|
@@ -131,7 +131,7 @@ export declare const gpt52: import("../../utils/preset").Preset<"openai/gpt-5.2"
|
|
|
131
131
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
132
132
|
providers: readonly ["openai", "azure"];
|
|
133
133
|
}>;
|
|
134
|
-
export declare const gpt52Chat: import("../../utils
|
|
134
|
+
export declare const gpt52Chat: import("../../utils").Preset<"openai/gpt-5.2-chat", CatalogModel, {
|
|
135
135
|
name: string;
|
|
136
136
|
created: string;
|
|
137
137
|
knowledge: string;
|
|
@@ -143,7 +143,7 @@ export declare const gpt52Chat: import("../../utils/preset").Preset<"openai/gpt-
|
|
|
143
143
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
144
144
|
providers: readonly ["openai", "azure"];
|
|
145
145
|
}>;
|
|
146
|
-
export declare const gpt52Pro: import("../../utils
|
|
146
|
+
export declare const gpt52Pro: import("../../utils").Preset<"openai/gpt-5.2-pro", CatalogModel, {
|
|
147
147
|
name: string;
|
|
148
148
|
created: string;
|
|
149
149
|
knowledge: string;
|
|
@@ -155,7 +155,7 @@ export declare const gpt52Pro: import("../../utils/preset").Preset<"openai/gpt-5
|
|
|
155
155
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
156
156
|
providers: readonly ["openai", "azure"];
|
|
157
157
|
}>;
|
|
158
|
-
export declare const gpt52Codex: import("../../utils
|
|
158
|
+
export declare const gpt52Codex: import("../../utils").Preset<"openai/gpt-5.2-codex", CatalogModel, {
|
|
159
159
|
name: string;
|
|
160
160
|
created: string;
|
|
161
161
|
knowledge: string;
|
|
@@ -166,7 +166,7 @@ export declare const gpt52Codex: import("../../utils/preset").Preset<"openai/gpt
|
|
|
166
166
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
167
167
|
providers: readonly ["openai", "azure"];
|
|
168
168
|
}>;
|
|
169
|
-
export declare const gpt53Codex: import("../../utils
|
|
169
|
+
export declare const gpt53Codex: import("../../utils").Preset<"openai/gpt-5.3-codex", CatalogModel, {
|
|
170
170
|
name: string;
|
|
171
171
|
created: string;
|
|
172
172
|
knowledge: string;
|
|
@@ -177,7 +177,7 @@ export declare const gpt53Codex: import("../../utils/preset").Preset<"openai/gpt
|
|
|
177
177
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
178
178
|
providers: readonly ["openai", "azure"];
|
|
179
179
|
}>;
|
|
180
|
-
export declare const gpt53CodexSpark: import("../../utils
|
|
180
|
+
export declare const gpt53CodexSpark: import("../../utils").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
|
|
181
181
|
name: string;
|
|
182
182
|
created: string;
|
|
183
183
|
knowledge: string;
|
|
@@ -189,7 +189,7 @@ export declare const gpt53CodexSpark: import("../../utils/preset").Preset<"opena
|
|
|
189
189
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
190
190
|
providers: readonly ["openai", "azure"];
|
|
191
191
|
}>;
|
|
192
|
-
export declare const gpt53Chat: import("../../utils
|
|
192
|
+
export declare const gpt53Chat: import("../../utils").Preset<"openai/gpt-5.3-chat", CatalogModel, {
|
|
193
193
|
name: string;
|
|
194
194
|
created: string;
|
|
195
195
|
knowledge: string;
|
|
@@ -201,7 +201,7 @@ export declare const gpt53Chat: import("../../utils/preset").Preset<"openai/gpt-
|
|
|
201
201
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
202
202
|
providers: readonly ["openai", "azure"];
|
|
203
203
|
}>;
|
|
204
|
-
export declare const gpt54: import("../../utils
|
|
204
|
+
export declare const gpt54: import("../../utils").Preset<"openai/gpt-5.4", CatalogModel, {
|
|
205
205
|
name: string;
|
|
206
206
|
created: string;
|
|
207
207
|
knowledge: string;
|
|
@@ -213,7 +213,7 @@ export declare const gpt54: import("../../utils/preset").Preset<"openai/gpt-5.4"
|
|
|
213
213
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
214
214
|
providers: readonly ["openai", "azure"];
|
|
215
215
|
}>;
|
|
216
|
-
export declare const gpt54Mini: import("../../utils
|
|
216
|
+
export declare const gpt54Mini: import("../../utils").Preset<"openai/gpt-5.4-mini", CatalogModel, {
|
|
217
217
|
name: string;
|
|
218
218
|
created: string;
|
|
219
219
|
knowledge: string;
|
|
@@ -225,7 +225,7 @@ export declare const gpt54Mini: import("../../utils/preset").Preset<"openai/gpt-
|
|
|
225
225
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
226
226
|
providers: readonly ["openai", "azure"];
|
|
227
227
|
}>;
|
|
228
|
-
export declare const gpt54Nano: import("../../utils
|
|
228
|
+
export declare const gpt54Nano: import("../../utils").Preset<"openai/gpt-5.4-nano", CatalogModel, {
|
|
229
229
|
name: string;
|
|
230
230
|
created: string;
|
|
231
231
|
knowledge: string;
|
|
@@ -237,7 +237,7 @@ export declare const gpt54Nano: import("../../utils/preset").Preset<"openai/gpt-
|
|
|
237
237
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
238
238
|
providers: readonly ["openai", "azure"];
|
|
239
239
|
}>;
|
|
240
|
-
export declare const gpt54Pro: import("../../utils
|
|
240
|
+
export declare const gpt54Pro: import("../../utils").Preset<"openai/gpt-5.4-pro", CatalogModel, {
|
|
241
241
|
name: string;
|
|
242
242
|
created: string;
|
|
243
243
|
knowledge: string;
|
|
@@ -249,7 +249,7 @@ export declare const gpt54Pro: import("../../utils/preset").Preset<"openai/gpt-5
|
|
|
249
249
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
250
250
|
providers: readonly ["openai", "azure"];
|
|
251
251
|
}>;
|
|
252
|
-
export declare const textEmbedding3Small: import("../../utils
|
|
252
|
+
export declare const textEmbedding3Small: import("../../utils").Preset<"openai/text-embedding-3-small", CatalogModel, {
|
|
253
253
|
name: string;
|
|
254
254
|
created: string;
|
|
255
255
|
context: number;
|
|
@@ -259,7 +259,7 @@ export declare const textEmbedding3Small: import("../../utils/preset").Preset<"o
|
|
|
259
259
|
};
|
|
260
260
|
providers: readonly ["openai", "azure"];
|
|
261
261
|
}>;
|
|
262
|
-
export declare const textEmbedding3Large: import("../../utils
|
|
262
|
+
export declare const textEmbedding3Large: import("../../utils").Preset<"openai/text-embedding-3-large", CatalogModel, {
|
|
263
263
|
name: string;
|
|
264
264
|
created: string;
|
|
265
265
|
context: number;
|
|
@@ -269,7 +269,7 @@ export declare const textEmbedding3Large: import("../../utils/preset").Preset<"o
|
|
|
269
269
|
};
|
|
270
270
|
providers: readonly ["openai", "azure"];
|
|
271
271
|
}>;
|
|
272
|
-
export declare const gptOss20b: import("../../utils
|
|
272
|
+
export declare const gptOss20b: import("../../utils").Preset<"openai/gpt-oss-20b", CatalogModel, {
|
|
273
273
|
name: string;
|
|
274
274
|
created: string;
|
|
275
275
|
knowledge: string;
|
|
@@ -279,9 +279,9 @@ export declare const gptOss20b: import("../../utils/preset").Preset<"openai/gpt-
|
|
|
279
279
|
};
|
|
280
280
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
281
281
|
context: number;
|
|
282
|
-
providers: readonly ["groq", "bedrock", "vertex"];
|
|
282
|
+
providers: readonly ["groq", "bedrock", "vertex", "togetherai", "deepinfra", "fireworks"];
|
|
283
283
|
}>;
|
|
284
|
-
export declare const gptOss120b: import("../../utils
|
|
284
|
+
export declare const gptOss120b: import("../../utils").Preset<"openai/gpt-oss-120b", CatalogModel, {
|
|
285
285
|
name: string;
|
|
286
286
|
created: string;
|
|
287
287
|
knowledge: string;
|
|
@@ -291,10 +291,10 @@ export declare const gptOss120b: import("../../utils/preset").Preset<"openai/gpt
|
|
|
291
291
|
};
|
|
292
292
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
293
293
|
context: number;
|
|
294
|
-
providers: readonly ["groq", "bedrock", "vertex"];
|
|
294
|
+
providers: readonly ["groq", "bedrock", "vertex", "togetherai", "deepinfra", "fireworks"];
|
|
295
295
|
}>;
|
|
296
296
|
export declare const gptOss: {
|
|
297
|
-
readonly latest: readonly [import("../../utils
|
|
297
|
+
readonly latest: readonly [import("../../utils").Preset<"openai/gpt-oss-20b", CatalogModel, {
|
|
298
298
|
name: string;
|
|
299
299
|
created: string;
|
|
300
300
|
knowledge: string;
|
|
@@ -304,8 +304,8 @@ export declare const gptOss: {
|
|
|
304
304
|
};
|
|
305
305
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
306
306
|
context: number;
|
|
307
|
-
providers: readonly ["groq", "bedrock", "vertex"];
|
|
308
|
-
}>, import("../../utils
|
|
307
|
+
providers: readonly ["groq", "bedrock", "vertex", "togetherai", "deepinfra", "fireworks"];
|
|
308
|
+
}>, import("../../utils").Preset<"openai/gpt-oss-120b", CatalogModel, {
|
|
309
309
|
name: string;
|
|
310
310
|
created: string;
|
|
311
311
|
knowledge: string;
|
|
@@ -315,9 +315,9 @@ export declare const gptOss: {
|
|
|
315
315
|
};
|
|
316
316
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
317
317
|
context: number;
|
|
318
|
-
providers: readonly ["groq", "bedrock", "vertex"];
|
|
318
|
+
providers: readonly ["groq", "bedrock", "vertex", "togetherai", "deepinfra", "fireworks"];
|
|
319
319
|
}>];
|
|
320
|
-
readonly all: (import("../../utils
|
|
320
|
+
readonly all: (import("../../utils").Preset<"openai/gpt-oss-20b", CatalogModel, {
|
|
321
321
|
name: string;
|
|
322
322
|
created: string;
|
|
323
323
|
knowledge: string;
|
|
@@ -327,8 +327,8 @@ export declare const gptOss: {
|
|
|
327
327
|
};
|
|
328
328
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
329
329
|
context: number;
|
|
330
|
-
providers: readonly ["groq", "bedrock", "vertex"];
|
|
331
|
-
}> | import("../../utils
|
|
330
|
+
providers: readonly ["groq", "bedrock", "vertex", "togetherai", "deepinfra", "fireworks"];
|
|
331
|
+
}> | import("../../utils").Preset<"openai/gpt-oss-120b", CatalogModel, {
|
|
332
332
|
name: string;
|
|
333
333
|
created: string;
|
|
334
334
|
knowledge: string;
|
|
@@ -338,9 +338,9 @@ export declare const gptOss: {
|
|
|
338
338
|
};
|
|
339
339
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
340
340
|
context: number;
|
|
341
|
-
providers: readonly ["groq", "bedrock", "vertex"];
|
|
341
|
+
providers: readonly ["groq", "bedrock", "vertex", "togetherai", "deepinfra", "fireworks"];
|
|
342
342
|
}>)[];
|
|
343
|
-
readonly "v1.x": readonly [import("../../utils
|
|
343
|
+
readonly "v1.x": readonly [import("../../utils").Preset<"openai/gpt-oss-20b", CatalogModel, {
|
|
344
344
|
name: string;
|
|
345
345
|
created: string;
|
|
346
346
|
knowledge: string;
|
|
@@ -350,8 +350,8 @@ export declare const gptOss: {
|
|
|
350
350
|
};
|
|
351
351
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
352
352
|
context: number;
|
|
353
|
-
providers: readonly ["groq", "bedrock", "vertex"];
|
|
354
|
-
}>, import("../../utils
|
|
353
|
+
providers: readonly ["groq", "bedrock", "vertex", "togetherai", "deepinfra", "fireworks"];
|
|
354
|
+
}>, import("../../utils").Preset<"openai/gpt-oss-120b", CatalogModel, {
|
|
355
355
|
name: string;
|
|
356
356
|
created: string;
|
|
357
357
|
knowledge: string;
|
|
@@ -361,9 +361,9 @@ export declare const gptOss: {
|
|
|
361
361
|
};
|
|
362
362
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
363
363
|
context: number;
|
|
364
|
-
providers: readonly ["groq", "bedrock", "vertex"];
|
|
364
|
+
providers: readonly ["groq", "bedrock", "vertex", "togetherai", "deepinfra", "fireworks"];
|
|
365
365
|
}>];
|
|
366
|
-
readonly v1: readonly [import("../../utils
|
|
366
|
+
readonly v1: readonly [import("../../utils").Preset<"openai/gpt-oss-20b", CatalogModel, {
|
|
367
367
|
name: string;
|
|
368
368
|
created: string;
|
|
369
369
|
knowledge: string;
|
|
@@ -373,8 +373,8 @@ export declare const gptOss: {
|
|
|
373
373
|
};
|
|
374
374
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
375
375
|
context: number;
|
|
376
|
-
providers: readonly ["groq", "bedrock", "vertex"];
|
|
377
|
-
}>, import("../../utils
|
|
376
|
+
providers: readonly ["groq", "bedrock", "vertex", "togetherai", "deepinfra", "fireworks"];
|
|
377
|
+
}>, import("../../utils").Preset<"openai/gpt-oss-120b", CatalogModel, {
|
|
378
378
|
name: string;
|
|
379
379
|
created: string;
|
|
380
380
|
knowledge: string;
|
|
@@ -384,11 +384,11 @@ export declare const gptOss: {
|
|
|
384
384
|
};
|
|
385
385
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
386
386
|
context: number;
|
|
387
|
-
providers: readonly ["groq", "bedrock", "vertex"];
|
|
387
|
+
providers: readonly ["groq", "bedrock", "vertex", "togetherai", "deepinfra", "fireworks"];
|
|
388
388
|
}>];
|
|
389
389
|
};
|
|
390
390
|
export declare const gpt: {
|
|
391
|
-
readonly latest: readonly [import("../../utils
|
|
391
|
+
readonly latest: readonly [import("../../utils").Preset<"openai/gpt-5.4", CatalogModel, {
|
|
392
392
|
name: string;
|
|
393
393
|
created: string;
|
|
394
394
|
knowledge: string;
|
|
@@ -399,7 +399,7 @@ export declare const gpt: {
|
|
|
399
399
|
};
|
|
400
400
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
401
401
|
providers: readonly ["openai", "azure"];
|
|
402
|
-
}>, import("../../utils
|
|
402
|
+
}>, import("../../utils").Preset<"openai/gpt-5.4-mini", CatalogModel, {
|
|
403
403
|
name: string;
|
|
404
404
|
created: string;
|
|
405
405
|
knowledge: string;
|
|
@@ -410,7 +410,7 @@ export declare const gpt: {
|
|
|
410
410
|
};
|
|
411
411
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
412
412
|
providers: readonly ["openai", "azure"];
|
|
413
|
-
}>, import("../../utils
|
|
413
|
+
}>, import("../../utils").Preset<"openai/gpt-5.4-nano", CatalogModel, {
|
|
414
414
|
name: string;
|
|
415
415
|
created: string;
|
|
416
416
|
knowledge: string;
|
|
@@ -422,7 +422,7 @@ export declare const gpt: {
|
|
|
422
422
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
423
423
|
providers: readonly ["openai", "azure"];
|
|
424
424
|
}>];
|
|
425
|
-
readonly all: (import("../../utils
|
|
425
|
+
readonly all: (import("../../utils").Preset<"openai/gpt-5", CatalogModel, {
|
|
426
426
|
name: string;
|
|
427
427
|
created: string;
|
|
428
428
|
knowledge: string;
|
|
@@ -433,7 +433,7 @@ export declare const gpt: {
|
|
|
433
433
|
};
|
|
434
434
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
435
435
|
providers: readonly ["openai", "azure"];
|
|
436
|
-
}> | import("../../utils
|
|
436
|
+
}> | import("../../utils").Preset<"openai/gpt-5-pro", CatalogModel, {
|
|
437
437
|
name: string;
|
|
438
438
|
created: string;
|
|
439
439
|
knowledge: string;
|
|
@@ -444,7 +444,7 @@ export declare const gpt: {
|
|
|
444
444
|
};
|
|
445
445
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
446
446
|
providers: readonly ["openai", "azure"];
|
|
447
|
-
}> | import("../../utils
|
|
447
|
+
}> | import("../../utils").Preset<"openai/gpt-5-mini", CatalogModel, {
|
|
448
448
|
name: string;
|
|
449
449
|
created: string;
|
|
450
450
|
knowledge: string;
|
|
@@ -455,7 +455,7 @@ export declare const gpt: {
|
|
|
455
455
|
};
|
|
456
456
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
457
457
|
providers: readonly ["openai", "azure"];
|
|
458
|
-
}> | import("../../utils
|
|
458
|
+
}> | import("../../utils").Preset<"openai/gpt-5-nano", CatalogModel, {
|
|
459
459
|
name: string;
|
|
460
460
|
created: string;
|
|
461
461
|
knowledge: string;
|
|
@@ -466,7 +466,7 @@ export declare const gpt: {
|
|
|
466
466
|
};
|
|
467
467
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
468
468
|
providers: readonly ["openai", "azure"];
|
|
469
|
-
}> | import("../../utils
|
|
469
|
+
}> | import("../../utils").Preset<"openai/gpt-5.1", CatalogModel, {
|
|
470
470
|
name: string;
|
|
471
471
|
created: string;
|
|
472
472
|
knowledge: string;
|
|
@@ -477,7 +477,7 @@ export declare const gpt: {
|
|
|
477
477
|
};
|
|
478
478
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
479
479
|
providers: readonly ["openai", "azure"];
|
|
480
|
-
}> | import("../../utils
|
|
480
|
+
}> | import("../../utils").Preset<"openai/gpt-5.1-chat", CatalogModel, {
|
|
481
481
|
name: string;
|
|
482
482
|
created: string;
|
|
483
483
|
knowledge: string;
|
|
@@ -488,7 +488,7 @@ export declare const gpt: {
|
|
|
488
488
|
};
|
|
489
489
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
490
490
|
providers: readonly ["openai", "azure"];
|
|
491
|
-
}> | import("../../utils
|
|
491
|
+
}> | import("../../utils").Preset<"openai/gpt-5.1-codex", CatalogModel, {
|
|
492
492
|
name: string;
|
|
493
493
|
created: string;
|
|
494
494
|
knowledge: string;
|
|
@@ -499,7 +499,7 @@ export declare const gpt: {
|
|
|
499
499
|
};
|
|
500
500
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
501
501
|
providers: readonly ["openai", "azure"];
|
|
502
|
-
}> | import("../../utils
|
|
502
|
+
}> | import("../../utils").Preset<"openai/gpt-5.1-codex-max", CatalogModel, {
|
|
503
503
|
name: string;
|
|
504
504
|
created: string;
|
|
505
505
|
knowledge: string;
|
|
@@ -510,7 +510,7 @@ export declare const gpt: {
|
|
|
510
510
|
};
|
|
511
511
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
512
512
|
providers: readonly ["openai", "azure"];
|
|
513
|
-
}> | import("../../utils
|
|
513
|
+
}> | import("../../utils").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
|
|
514
514
|
name: string;
|
|
515
515
|
created: string;
|
|
516
516
|
knowledge: string;
|
|
@@ -521,7 +521,7 @@ export declare const gpt: {
|
|
|
521
521
|
};
|
|
522
522
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
523
523
|
providers: readonly ["openai", "azure"];
|
|
524
|
-
}> | import("../../utils
|
|
524
|
+
}> | import("../../utils").Preset<"openai/gpt-5-codex", CatalogModel, {
|
|
525
525
|
name: string;
|
|
526
526
|
created: string;
|
|
527
527
|
knowledge: string;
|
|
@@ -532,7 +532,7 @@ export declare const gpt: {
|
|
|
532
532
|
};
|
|
533
533
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
534
534
|
providers: readonly ["openai", "azure"];
|
|
535
|
-
}> | import("../../utils
|
|
535
|
+
}> | import("../../utils").Preset<"openai/gpt-5.2", CatalogModel, {
|
|
536
536
|
name: string;
|
|
537
537
|
created: string;
|
|
538
538
|
knowledge: string;
|
|
@@ -543,7 +543,7 @@ export declare const gpt: {
|
|
|
543
543
|
};
|
|
544
544
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
545
545
|
providers: readonly ["openai", "azure"];
|
|
546
|
-
}> | import("../../utils
|
|
546
|
+
}> | import("../../utils").Preset<"openai/gpt-5.2-chat", CatalogModel, {
|
|
547
547
|
name: string;
|
|
548
548
|
created: string;
|
|
549
549
|
knowledge: string;
|
|
@@ -554,7 +554,7 @@ export declare const gpt: {
|
|
|
554
554
|
};
|
|
555
555
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
556
556
|
providers: readonly ["openai", "azure"];
|
|
557
|
-
}> | import("../../utils
|
|
557
|
+
}> | import("../../utils").Preset<"openai/gpt-5.2-pro", CatalogModel, {
|
|
558
558
|
name: string;
|
|
559
559
|
created: string;
|
|
560
560
|
knowledge: string;
|
|
@@ -565,7 +565,7 @@ export declare const gpt: {
|
|
|
565
565
|
};
|
|
566
566
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
567
567
|
providers: readonly ["openai", "azure"];
|
|
568
|
-
}> | import("../../utils
|
|
568
|
+
}> | import("../../utils").Preset<"openai/gpt-5.2-codex", CatalogModel, {
|
|
569
569
|
name: string;
|
|
570
570
|
created: string;
|
|
571
571
|
knowledge: string;
|
|
@@ -575,7 +575,7 @@ export declare const gpt: {
|
|
|
575
575
|
};
|
|
576
576
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
577
577
|
providers: readonly ["openai", "azure"];
|
|
578
|
-
}> | import("../../utils
|
|
578
|
+
}> | import("../../utils").Preset<"openai/gpt-5.3-codex", CatalogModel, {
|
|
579
579
|
name: string;
|
|
580
580
|
created: string;
|
|
581
581
|
knowledge: string;
|
|
@@ -585,7 +585,7 @@ export declare const gpt: {
|
|
|
585
585
|
};
|
|
586
586
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
587
587
|
providers: readonly ["openai", "azure"];
|
|
588
|
-
}> | import("../../utils
|
|
588
|
+
}> | import("../../utils").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
|
|
589
589
|
name: string;
|
|
590
590
|
created: string;
|
|
591
591
|
knowledge: string;
|
|
@@ -596,7 +596,7 @@ export declare const gpt: {
|
|
|
596
596
|
};
|
|
597
597
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
598
598
|
providers: readonly ["openai", "azure"];
|
|
599
|
-
}> | import("../../utils
|
|
599
|
+
}> | import("../../utils").Preset<"openai/gpt-5.3-chat", CatalogModel, {
|
|
600
600
|
name: string;
|
|
601
601
|
created: string;
|
|
602
602
|
knowledge: string;
|
|
@@ -607,7 +607,7 @@ export declare const gpt: {
|
|
|
607
607
|
};
|
|
608
608
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
609
609
|
providers: readonly ["openai", "azure"];
|
|
610
|
-
}> | import("../../utils
|
|
610
|
+
}> | import("../../utils").Preset<"openai/gpt-5.4", CatalogModel, {
|
|
611
611
|
name: string;
|
|
612
612
|
created: string;
|
|
613
613
|
knowledge: string;
|
|
@@ -618,7 +618,7 @@ export declare const gpt: {
|
|
|
618
618
|
};
|
|
619
619
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
620
620
|
providers: readonly ["openai", "azure"];
|
|
621
|
-
}> | import("../../utils
|
|
621
|
+
}> | import("../../utils").Preset<"openai/gpt-5.4-mini", CatalogModel, {
|
|
622
622
|
name: string;
|
|
623
623
|
created: string;
|
|
624
624
|
knowledge: string;
|
|
@@ -629,7 +629,7 @@ export declare const gpt: {
|
|
|
629
629
|
};
|
|
630
630
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
631
631
|
providers: readonly ["openai", "azure"];
|
|
632
|
-
}> | import("../../utils
|
|
632
|
+
}> | import("../../utils").Preset<"openai/gpt-5.4-nano", CatalogModel, {
|
|
633
633
|
name: string;
|
|
634
634
|
created: string;
|
|
635
635
|
knowledge: string;
|
|
@@ -640,7 +640,7 @@ export declare const gpt: {
|
|
|
640
640
|
};
|
|
641
641
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
642
642
|
providers: readonly ["openai", "azure"];
|
|
643
|
-
}> | import("../../utils
|
|
643
|
+
}> | import("../../utils").Preset<"openai/gpt-5.4-pro", CatalogModel, {
|
|
644
644
|
name: string;
|
|
645
645
|
created: string;
|
|
646
646
|
knowledge: string;
|
|
@@ -652,7 +652,7 @@ export declare const gpt: {
|
|
|
652
652
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
653
653
|
providers: readonly ["openai", "azure"];
|
|
654
654
|
}>)[];
|
|
655
|
-
readonly "v5.x": readonly [import("../../utils
|
|
655
|
+
readonly "v5.x": readonly [import("../../utils").Preset<"openai/gpt-5", CatalogModel, {
|
|
656
656
|
name: string;
|
|
657
657
|
created: string;
|
|
658
658
|
knowledge: string;
|
|
@@ -663,7 +663,7 @@ export declare const gpt: {
|
|
|
663
663
|
};
|
|
664
664
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
665
665
|
providers: readonly ["openai", "azure"];
|
|
666
|
-
}>, import("../../utils
|
|
666
|
+
}>, import("../../utils").Preset<"openai/gpt-5-mini", CatalogModel, {
|
|
667
667
|
name: string;
|
|
668
668
|
created: string;
|
|
669
669
|
knowledge: string;
|
|
@@ -674,7 +674,7 @@ export declare const gpt: {
|
|
|
674
674
|
};
|
|
675
675
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
676
676
|
providers: readonly ["openai", "azure"];
|
|
677
|
-
}>, import("../../utils
|
|
677
|
+
}>, import("../../utils").Preset<"openai/gpt-5-nano", CatalogModel, {
|
|
678
678
|
name: string;
|
|
679
679
|
created: string;
|
|
680
680
|
knowledge: string;
|
|
@@ -685,7 +685,7 @@ export declare const gpt: {
|
|
|
685
685
|
};
|
|
686
686
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
687
687
|
providers: readonly ["openai", "azure"];
|
|
688
|
-
}>, import("../../utils
|
|
688
|
+
}>, import("../../utils").Preset<"openai/gpt-5-pro", CatalogModel, {
|
|
689
689
|
name: string;
|
|
690
690
|
created: string;
|
|
691
691
|
knowledge: string;
|
|
@@ -696,7 +696,7 @@ export declare const gpt: {
|
|
|
696
696
|
};
|
|
697
697
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
698
698
|
providers: readonly ["openai", "azure"];
|
|
699
|
-
}>, import("../../utils
|
|
699
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1", CatalogModel, {
|
|
700
700
|
name: string;
|
|
701
701
|
created: string;
|
|
702
702
|
knowledge: string;
|
|
@@ -707,7 +707,7 @@ export declare const gpt: {
|
|
|
707
707
|
};
|
|
708
708
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
709
709
|
providers: readonly ["openai", "azure"];
|
|
710
|
-
}>, import("../../utils
|
|
710
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-chat", CatalogModel, {
|
|
711
711
|
name: string;
|
|
712
712
|
created: string;
|
|
713
713
|
knowledge: string;
|
|
@@ -718,7 +718,7 @@ export declare const gpt: {
|
|
|
718
718
|
};
|
|
719
719
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
720
720
|
providers: readonly ["openai", "azure"];
|
|
721
|
-
}>, import("../../utils
|
|
721
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-codex", CatalogModel, {
|
|
722
722
|
name: string;
|
|
723
723
|
created: string;
|
|
724
724
|
knowledge: string;
|
|
@@ -729,7 +729,7 @@ export declare const gpt: {
|
|
|
729
729
|
};
|
|
730
730
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
731
731
|
providers: readonly ["openai", "azure"];
|
|
732
|
-
}>, import("../../utils
|
|
732
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-codex-max", CatalogModel, {
|
|
733
733
|
name: string;
|
|
734
734
|
created: string;
|
|
735
735
|
knowledge: string;
|
|
@@ -740,7 +740,7 @@ export declare const gpt: {
|
|
|
740
740
|
};
|
|
741
741
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
742
742
|
providers: readonly ["openai", "azure"];
|
|
743
|
-
}>, import("../../utils
|
|
743
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
|
|
744
744
|
name: string;
|
|
745
745
|
created: string;
|
|
746
746
|
knowledge: string;
|
|
@@ -751,7 +751,7 @@ export declare const gpt: {
|
|
|
751
751
|
};
|
|
752
752
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
753
753
|
providers: readonly ["openai", "azure"];
|
|
754
|
-
}>, import("../../utils
|
|
754
|
+
}>, import("../../utils").Preset<"openai/gpt-5.2", CatalogModel, {
|
|
755
755
|
name: string;
|
|
756
756
|
created: string;
|
|
757
757
|
knowledge: string;
|
|
@@ -762,7 +762,7 @@ export declare const gpt: {
|
|
|
762
762
|
};
|
|
763
763
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
764
764
|
providers: readonly ["openai", "azure"];
|
|
765
|
-
}>, import("../../utils
|
|
765
|
+
}>, import("../../utils").Preset<"openai/gpt-5.2-chat", CatalogModel, {
|
|
766
766
|
name: string;
|
|
767
767
|
created: string;
|
|
768
768
|
knowledge: string;
|
|
@@ -773,7 +773,7 @@ export declare const gpt: {
|
|
|
773
773
|
};
|
|
774
774
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
775
775
|
providers: readonly ["openai", "azure"];
|
|
776
|
-
}>, import("../../utils
|
|
776
|
+
}>, import("../../utils").Preset<"openai/gpt-5.2-pro", CatalogModel, {
|
|
777
777
|
name: string;
|
|
778
778
|
created: string;
|
|
779
779
|
knowledge: string;
|
|
@@ -784,7 +784,7 @@ export declare const gpt: {
|
|
|
784
784
|
};
|
|
785
785
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
786
786
|
providers: readonly ["openai", "azure"];
|
|
787
|
-
}>, import("../../utils
|
|
787
|
+
}>, import("../../utils").Preset<"openai/gpt-5.2-codex", CatalogModel, {
|
|
788
788
|
name: string;
|
|
789
789
|
created: string;
|
|
790
790
|
knowledge: string;
|
|
@@ -794,7 +794,7 @@ export declare const gpt: {
|
|
|
794
794
|
};
|
|
795
795
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
796
796
|
providers: readonly ["openai", "azure"];
|
|
797
|
-
}>, import("../../utils
|
|
797
|
+
}>, import("../../utils").Preset<"openai/gpt-5.3-codex", CatalogModel, {
|
|
798
798
|
name: string;
|
|
799
799
|
created: string;
|
|
800
800
|
knowledge: string;
|
|
@@ -804,7 +804,7 @@ export declare const gpt: {
|
|
|
804
804
|
};
|
|
805
805
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
806
806
|
providers: readonly ["openai", "azure"];
|
|
807
|
-
}>, import("../../utils
|
|
807
|
+
}>, import("../../utils").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
|
|
808
808
|
name: string;
|
|
809
809
|
created: string;
|
|
810
810
|
knowledge: string;
|
|
@@ -815,7 +815,7 @@ export declare const gpt: {
|
|
|
815
815
|
};
|
|
816
816
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
817
817
|
providers: readonly ["openai", "azure"];
|
|
818
|
-
}>, import("../../utils
|
|
818
|
+
}>, import("../../utils").Preset<"openai/gpt-5.3-chat", CatalogModel, {
|
|
819
819
|
name: string;
|
|
820
820
|
created: string;
|
|
821
821
|
knowledge: string;
|
|
@@ -826,7 +826,7 @@ export declare const gpt: {
|
|
|
826
826
|
};
|
|
827
827
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
828
828
|
providers: readonly ["openai", "azure"];
|
|
829
|
-
}>, import("../../utils
|
|
829
|
+
}>, import("../../utils").Preset<"openai/gpt-5.4", CatalogModel, {
|
|
830
830
|
name: string;
|
|
831
831
|
created: string;
|
|
832
832
|
knowledge: string;
|
|
@@ -837,7 +837,7 @@ export declare const gpt: {
|
|
|
837
837
|
};
|
|
838
838
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
839
839
|
providers: readonly ["openai", "azure"];
|
|
840
|
-
}>, import("../../utils
|
|
840
|
+
}>, import("../../utils").Preset<"openai/gpt-5.4-mini", CatalogModel, {
|
|
841
841
|
name: string;
|
|
842
842
|
created: string;
|
|
843
843
|
knowledge: string;
|
|
@@ -848,7 +848,7 @@ export declare const gpt: {
|
|
|
848
848
|
};
|
|
849
849
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
850
850
|
providers: readonly ["openai", "azure"];
|
|
851
|
-
}>, import("../../utils
|
|
851
|
+
}>, import("../../utils").Preset<"openai/gpt-5.4-nano", CatalogModel, {
|
|
852
852
|
name: string;
|
|
853
853
|
created: string;
|
|
854
854
|
knowledge: string;
|
|
@@ -859,7 +859,7 @@ export declare const gpt: {
|
|
|
859
859
|
};
|
|
860
860
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
861
861
|
providers: readonly ["openai", "azure"];
|
|
862
|
-
}>, import("../../utils
|
|
862
|
+
}>, import("../../utils").Preset<"openai/gpt-5.4-pro", CatalogModel, {
|
|
863
863
|
name: string;
|
|
864
864
|
created: string;
|
|
865
865
|
knowledge: string;
|
|
@@ -871,7 +871,7 @@ export declare const gpt: {
|
|
|
871
871
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
872
872
|
providers: readonly ["openai", "azure"];
|
|
873
873
|
}>];
|
|
874
|
-
readonly v5: readonly [import("../../utils
|
|
874
|
+
readonly v5: readonly [import("../../utils").Preset<"openai/gpt-5", CatalogModel, {
|
|
875
875
|
name: string;
|
|
876
876
|
created: string;
|
|
877
877
|
knowledge: string;
|
|
@@ -882,7 +882,7 @@ export declare const gpt: {
|
|
|
882
882
|
};
|
|
883
883
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
884
884
|
providers: readonly ["openai", "azure"];
|
|
885
|
-
}>, import("../../utils
|
|
885
|
+
}>, import("../../utils").Preset<"openai/gpt-5-mini", CatalogModel, {
|
|
886
886
|
name: string;
|
|
887
887
|
created: string;
|
|
888
888
|
knowledge: string;
|
|
@@ -893,7 +893,7 @@ export declare const gpt: {
|
|
|
893
893
|
};
|
|
894
894
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
895
895
|
providers: readonly ["openai", "azure"];
|
|
896
|
-
}>, import("../../utils
|
|
896
|
+
}>, import("../../utils").Preset<"openai/gpt-5-nano", CatalogModel, {
|
|
897
897
|
name: string;
|
|
898
898
|
created: string;
|
|
899
899
|
knowledge: string;
|
|
@@ -904,7 +904,7 @@ export declare const gpt: {
|
|
|
904
904
|
};
|
|
905
905
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
906
906
|
providers: readonly ["openai", "azure"];
|
|
907
|
-
}>, import("../../utils
|
|
907
|
+
}>, import("../../utils").Preset<"openai/gpt-5-pro", CatalogModel, {
|
|
908
908
|
name: string;
|
|
909
909
|
created: string;
|
|
910
910
|
knowledge: string;
|
|
@@ -916,7 +916,7 @@ export declare const gpt: {
|
|
|
916
916
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
917
917
|
providers: readonly ["openai", "azure"];
|
|
918
918
|
}>];
|
|
919
|
-
readonly "v5.1": readonly [import("../../utils
|
|
919
|
+
readonly "v5.1": readonly [import("../../utils").Preset<"openai/gpt-5.1", CatalogModel, {
|
|
920
920
|
name: string;
|
|
921
921
|
created: string;
|
|
922
922
|
knowledge: string;
|
|
@@ -927,7 +927,7 @@ export declare const gpt: {
|
|
|
927
927
|
};
|
|
928
928
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
929
929
|
providers: readonly ["openai", "azure"];
|
|
930
|
-
}>, import("../../utils
|
|
930
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-chat", CatalogModel, {
|
|
931
931
|
name: string;
|
|
932
932
|
created: string;
|
|
933
933
|
knowledge: string;
|
|
@@ -938,7 +938,7 @@ export declare const gpt: {
|
|
|
938
938
|
};
|
|
939
939
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
940
940
|
providers: readonly ["openai", "azure"];
|
|
941
|
-
}>, import("../../utils
|
|
941
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-codex", CatalogModel, {
|
|
942
942
|
name: string;
|
|
943
943
|
created: string;
|
|
944
944
|
knowledge: string;
|
|
@@ -949,7 +949,7 @@ export declare const gpt: {
|
|
|
949
949
|
};
|
|
950
950
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
951
951
|
providers: readonly ["openai", "azure"];
|
|
952
|
-
}>, import("../../utils
|
|
952
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-codex-max", CatalogModel, {
|
|
953
953
|
name: string;
|
|
954
954
|
created: string;
|
|
955
955
|
knowledge: string;
|
|
@@ -960,7 +960,7 @@ export declare const gpt: {
|
|
|
960
960
|
};
|
|
961
961
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
962
962
|
providers: readonly ["openai", "azure"];
|
|
963
|
-
}>, import("../../utils
|
|
963
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
|
|
964
964
|
name: string;
|
|
965
965
|
created: string;
|
|
966
966
|
knowledge: string;
|
|
@@ -972,7 +972,7 @@ export declare const gpt: {
|
|
|
972
972
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
973
973
|
providers: readonly ["openai", "azure"];
|
|
974
974
|
}>];
|
|
975
|
-
readonly "v5.2": readonly [import("../../utils
|
|
975
|
+
readonly "v5.2": readonly [import("../../utils").Preset<"openai/gpt-5.2", CatalogModel, {
|
|
976
976
|
name: string;
|
|
977
977
|
created: string;
|
|
978
978
|
knowledge: string;
|
|
@@ -983,7 +983,7 @@ export declare const gpt: {
|
|
|
983
983
|
};
|
|
984
984
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
985
985
|
providers: readonly ["openai", "azure"];
|
|
986
|
-
}>, import("../../utils
|
|
986
|
+
}>, import("../../utils").Preset<"openai/gpt-5.2-chat", CatalogModel, {
|
|
987
987
|
name: string;
|
|
988
988
|
created: string;
|
|
989
989
|
knowledge: string;
|
|
@@ -994,7 +994,7 @@ export declare const gpt: {
|
|
|
994
994
|
};
|
|
995
995
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
996
996
|
providers: readonly ["openai", "azure"];
|
|
997
|
-
}>, import("../../utils
|
|
997
|
+
}>, import("../../utils").Preset<"openai/gpt-5.2-pro", CatalogModel, {
|
|
998
998
|
name: string;
|
|
999
999
|
created: string;
|
|
1000
1000
|
knowledge: string;
|
|
@@ -1005,7 +1005,7 @@ export declare const gpt: {
|
|
|
1005
1005
|
};
|
|
1006
1006
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1007
1007
|
providers: readonly ["openai", "azure"];
|
|
1008
|
-
}>, import("../../utils
|
|
1008
|
+
}>, import("../../utils").Preset<"openai/gpt-5.2-codex", CatalogModel, {
|
|
1009
1009
|
name: string;
|
|
1010
1010
|
created: string;
|
|
1011
1011
|
knowledge: string;
|
|
@@ -1016,7 +1016,7 @@ export declare const gpt: {
|
|
|
1016
1016
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1017
1017
|
providers: readonly ["openai", "azure"];
|
|
1018
1018
|
}>];
|
|
1019
|
-
readonly "v5.3": readonly [import("../../utils
|
|
1019
|
+
readonly "v5.3": readonly [import("../../utils").Preset<"openai/gpt-5.3-codex", CatalogModel, {
|
|
1020
1020
|
name: string;
|
|
1021
1021
|
created: string;
|
|
1022
1022
|
knowledge: string;
|
|
@@ -1026,7 +1026,7 @@ export declare const gpt: {
|
|
|
1026
1026
|
};
|
|
1027
1027
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1028
1028
|
providers: readonly ["openai", "azure"];
|
|
1029
|
-
}>, import("../../utils
|
|
1029
|
+
}>, import("../../utils").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
|
|
1030
1030
|
name: string;
|
|
1031
1031
|
created: string;
|
|
1032
1032
|
knowledge: string;
|
|
@@ -1037,7 +1037,7 @@ export declare const gpt: {
|
|
|
1037
1037
|
};
|
|
1038
1038
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1039
1039
|
providers: readonly ["openai", "azure"];
|
|
1040
|
-
}>, import("../../utils
|
|
1040
|
+
}>, import("../../utils").Preset<"openai/gpt-5.3-chat", CatalogModel, {
|
|
1041
1041
|
name: string;
|
|
1042
1042
|
created: string;
|
|
1043
1043
|
knowledge: string;
|
|
@@ -1049,7 +1049,7 @@ export declare const gpt: {
|
|
|
1049
1049
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1050
1050
|
providers: readonly ["openai", "azure"];
|
|
1051
1051
|
}>];
|
|
1052
|
-
readonly "v5.4": readonly [import("../../utils
|
|
1052
|
+
readonly "v5.4": readonly [import("../../utils").Preset<"openai/gpt-5.4", CatalogModel, {
|
|
1053
1053
|
name: string;
|
|
1054
1054
|
created: string;
|
|
1055
1055
|
knowledge: string;
|
|
@@ -1060,7 +1060,7 @@ export declare const gpt: {
|
|
|
1060
1060
|
};
|
|
1061
1061
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1062
1062
|
providers: readonly ["openai", "azure"];
|
|
1063
|
-
}>, import("../../utils
|
|
1063
|
+
}>, import("../../utils").Preset<"openai/gpt-5.4-mini", CatalogModel, {
|
|
1064
1064
|
name: string;
|
|
1065
1065
|
created: string;
|
|
1066
1066
|
knowledge: string;
|
|
@@ -1071,7 +1071,7 @@ export declare const gpt: {
|
|
|
1071
1071
|
};
|
|
1072
1072
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1073
1073
|
providers: readonly ["openai", "azure"];
|
|
1074
|
-
}>, import("../../utils
|
|
1074
|
+
}>, import("../../utils").Preset<"openai/gpt-5.4-nano", CatalogModel, {
|
|
1075
1075
|
name: string;
|
|
1076
1076
|
created: string;
|
|
1077
1077
|
knowledge: string;
|
|
@@ -1082,7 +1082,7 @@ export declare const gpt: {
|
|
|
1082
1082
|
};
|
|
1083
1083
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1084
1084
|
providers: readonly ["openai", "azure"];
|
|
1085
|
-
}>, import("../../utils
|
|
1085
|
+
}>, import("../../utils").Preset<"openai/gpt-5.4-pro", CatalogModel, {
|
|
1086
1086
|
name: string;
|
|
1087
1087
|
created: string;
|
|
1088
1088
|
knowledge: string;
|
|
@@ -1094,7 +1094,7 @@ export declare const gpt: {
|
|
|
1094
1094
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1095
1095
|
providers: readonly ["openai", "azure"];
|
|
1096
1096
|
}>];
|
|
1097
|
-
readonly codex: readonly [import("../../utils
|
|
1097
|
+
readonly codex: readonly [import("../../utils").Preset<"openai/gpt-5-codex", CatalogModel, {
|
|
1098
1098
|
name: string;
|
|
1099
1099
|
created: string;
|
|
1100
1100
|
knowledge: string;
|
|
@@ -1105,7 +1105,7 @@ export declare const gpt: {
|
|
|
1105
1105
|
};
|
|
1106
1106
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1107
1107
|
providers: readonly ["openai", "azure"];
|
|
1108
|
-
}>, import("../../utils
|
|
1108
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-codex", CatalogModel, {
|
|
1109
1109
|
name: string;
|
|
1110
1110
|
created: string;
|
|
1111
1111
|
knowledge: string;
|
|
@@ -1116,7 +1116,7 @@ export declare const gpt: {
|
|
|
1116
1116
|
};
|
|
1117
1117
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1118
1118
|
providers: readonly ["openai", "azure"];
|
|
1119
|
-
}>, import("../../utils
|
|
1119
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-codex-max", CatalogModel, {
|
|
1120
1120
|
name: string;
|
|
1121
1121
|
created: string;
|
|
1122
1122
|
knowledge: string;
|
|
@@ -1127,7 +1127,7 @@ export declare const gpt: {
|
|
|
1127
1127
|
};
|
|
1128
1128
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1129
1129
|
providers: readonly ["openai", "azure"];
|
|
1130
|
-
}>, import("../../utils
|
|
1130
|
+
}>, import("../../utils").Preset<"openai/gpt-5.1-codex-mini", CatalogModel, {
|
|
1131
1131
|
name: string;
|
|
1132
1132
|
created: string;
|
|
1133
1133
|
knowledge: string;
|
|
@@ -1138,7 +1138,7 @@ export declare const gpt: {
|
|
|
1138
1138
|
};
|
|
1139
1139
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1140
1140
|
providers: readonly ["openai", "azure"];
|
|
1141
|
-
}>, import("../../utils
|
|
1141
|
+
}>, import("../../utils").Preset<"openai/gpt-5.2-codex", CatalogModel, {
|
|
1142
1142
|
name: string;
|
|
1143
1143
|
created: string;
|
|
1144
1144
|
knowledge: string;
|
|
@@ -1148,7 +1148,7 @@ export declare const gpt: {
|
|
|
1148
1148
|
};
|
|
1149
1149
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1150
1150
|
providers: readonly ["openai", "azure"];
|
|
1151
|
-
}>, import("../../utils
|
|
1151
|
+
}>, import("../../utils").Preset<"openai/gpt-5.3-codex", CatalogModel, {
|
|
1152
1152
|
name: string;
|
|
1153
1153
|
created: string;
|
|
1154
1154
|
knowledge: string;
|
|
@@ -1158,7 +1158,7 @@ export declare const gpt: {
|
|
|
1158
1158
|
};
|
|
1159
1159
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1160
1160
|
providers: readonly ["openai", "azure"];
|
|
1161
|
-
}>, import("../../utils
|
|
1161
|
+
}>, import("../../utils").Preset<"openai/gpt-5.3-codex-spark", CatalogModel, {
|
|
1162
1162
|
name: string;
|
|
1163
1163
|
created: string;
|
|
1164
1164
|
knowledge: string;
|
|
@@ -1170,7 +1170,7 @@ export declare const gpt: {
|
|
|
1170
1170
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1171
1171
|
providers: readonly ["openai", "azure"];
|
|
1172
1172
|
}>];
|
|
1173
|
-
readonly chat: readonly [import("../../utils
|
|
1173
|
+
readonly chat: readonly [import("../../utils").Preset<"openai/gpt-5.1-chat", CatalogModel, {
|
|
1174
1174
|
name: string;
|
|
1175
1175
|
created: string;
|
|
1176
1176
|
knowledge: string;
|
|
@@ -1181,7 +1181,7 @@ export declare const gpt: {
|
|
|
1181
1181
|
};
|
|
1182
1182
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1183
1183
|
providers: readonly ["openai", "azure"];
|
|
1184
|
-
}>, import("../../utils
|
|
1184
|
+
}>, import("../../utils").Preset<"openai/gpt-5.2-chat", CatalogModel, {
|
|
1185
1185
|
name: string;
|
|
1186
1186
|
created: string;
|
|
1187
1187
|
knowledge: string;
|
|
@@ -1192,7 +1192,7 @@ export declare const gpt: {
|
|
|
1192
1192
|
};
|
|
1193
1193
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1194
1194
|
providers: readonly ["openai", "azure"];
|
|
1195
|
-
}>, import("../../utils
|
|
1195
|
+
}>, import("../../utils").Preset<"openai/gpt-5.3-chat", CatalogModel, {
|
|
1196
1196
|
name: string;
|
|
1197
1197
|
created: string;
|
|
1198
1198
|
knowledge: string;
|
|
@@ -1204,7 +1204,7 @@ export declare const gpt: {
|
|
|
1204
1204
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1205
1205
|
providers: readonly ["openai", "azure"];
|
|
1206
1206
|
}>];
|
|
1207
|
-
readonly pro: readonly [import("../../utils
|
|
1207
|
+
readonly pro: readonly [import("../../utils").Preset<"openai/gpt-5-pro", CatalogModel, {
|
|
1208
1208
|
name: string;
|
|
1209
1209
|
created: string;
|
|
1210
1210
|
knowledge: string;
|
|
@@ -1215,7 +1215,7 @@ export declare const gpt: {
|
|
|
1215
1215
|
};
|
|
1216
1216
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1217
1217
|
providers: readonly ["openai", "azure"];
|
|
1218
|
-
}>, import("../../utils
|
|
1218
|
+
}>, import("../../utils").Preset<"openai/gpt-5.2-pro", CatalogModel, {
|
|
1219
1219
|
name: string;
|
|
1220
1220
|
created: string;
|
|
1221
1221
|
knowledge: string;
|
|
@@ -1226,7 +1226,7 @@ export declare const gpt: {
|
|
|
1226
1226
|
};
|
|
1227
1227
|
capabilities: readonly ["attachments", "reasoning", "tool_call", "structured_output", "temperature"];
|
|
1228
1228
|
providers: readonly ["openai", "azure"];
|
|
1229
|
-
}>, import("../../utils
|
|
1229
|
+
}>, import("../../utils").Preset<"openai/gpt-5.4-pro", CatalogModel, {
|
|
1230
1230
|
name: string;
|
|
1231
1231
|
created: string;
|
|
1232
1232
|
knowledge: string;
|
|
@@ -1240,7 +1240,7 @@ export declare const gpt: {
|
|
|
1240
1240
|
}>];
|
|
1241
1241
|
};
|
|
1242
1242
|
export declare const textEmbeddings: {
|
|
1243
|
-
readonly latest: readonly [import("../../utils
|
|
1243
|
+
readonly latest: readonly [import("../../utils").Preset<"openai/text-embedding-3-small", CatalogModel, {
|
|
1244
1244
|
name: string;
|
|
1245
1245
|
created: string;
|
|
1246
1246
|
context: number;
|
|
@@ -1249,7 +1249,7 @@ export declare const textEmbeddings: {
|
|
|
1249
1249
|
output: readonly ["embedding"];
|
|
1250
1250
|
};
|
|
1251
1251
|
providers: readonly ["openai", "azure"];
|
|
1252
|
-
}>, import("../../utils
|
|
1252
|
+
}>, import("../../utils").Preset<"openai/text-embedding-3-large", CatalogModel, {
|
|
1253
1253
|
name: string;
|
|
1254
1254
|
created: string;
|
|
1255
1255
|
context: number;
|
|
@@ -1259,7 +1259,7 @@ export declare const textEmbeddings: {
|
|
|
1259
1259
|
};
|
|
1260
1260
|
providers: readonly ["openai", "azure"];
|
|
1261
1261
|
}>];
|
|
1262
|
-
readonly all: (import("../../utils
|
|
1262
|
+
readonly all: (import("../../utils").Preset<"openai/text-embedding-3-small", CatalogModel, {
|
|
1263
1263
|
name: string;
|
|
1264
1264
|
created: string;
|
|
1265
1265
|
context: number;
|
|
@@ -1268,7 +1268,7 @@ export declare const textEmbeddings: {
|
|
|
1268
1268
|
output: readonly ["embedding"];
|
|
1269
1269
|
};
|
|
1270
1270
|
providers: readonly ["openai", "azure"];
|
|
1271
|
-
}> | import("../../utils
|
|
1271
|
+
}> | import("../../utils").Preset<"openai/text-embedding-3-large", CatalogModel, {
|
|
1272
1272
|
name: string;
|
|
1273
1273
|
created: string;
|
|
1274
1274
|
context: number;
|
|
@@ -1278,7 +1278,7 @@ export declare const textEmbeddings: {
|
|
|
1278
1278
|
};
|
|
1279
1279
|
providers: readonly ["openai", "azure"];
|
|
1280
1280
|
}>)[];
|
|
1281
|
-
readonly "v3.x": readonly [import("../../utils
|
|
1281
|
+
readonly "v3.x": readonly [import("../../utils").Preset<"openai/text-embedding-3-small", CatalogModel, {
|
|
1282
1282
|
name: string;
|
|
1283
1283
|
created: string;
|
|
1284
1284
|
context: number;
|
|
@@ -1287,7 +1287,7 @@ export declare const textEmbeddings: {
|
|
|
1287
1287
|
output: readonly ["embedding"];
|
|
1288
1288
|
};
|
|
1289
1289
|
providers: readonly ["openai", "azure"];
|
|
1290
|
-
}>, import("../../utils
|
|
1290
|
+
}>, import("../../utils").Preset<"openai/text-embedding-3-large", CatalogModel, {
|
|
1291
1291
|
name: string;
|
|
1292
1292
|
created: string;
|
|
1293
1293
|
context: number;
|
|
@@ -1297,7 +1297,7 @@ export declare const textEmbeddings: {
|
|
|
1297
1297
|
};
|
|
1298
1298
|
providers: readonly ["openai", "azure"];
|
|
1299
1299
|
}>];
|
|
1300
|
-
readonly v3: readonly [import("../../utils
|
|
1300
|
+
readonly v3: readonly [import("../../utils").Preset<"openai/text-embedding-3-small", CatalogModel, {
|
|
1301
1301
|
name: string;
|
|
1302
1302
|
created: string;
|
|
1303
1303
|
context: number;
|
|
@@ -1306,7 +1306,7 @@ export declare const textEmbeddings: {
|
|
|
1306
1306
|
output: readonly ["embedding"];
|
|
1307
1307
|
};
|
|
1308
1308
|
providers: readonly ["openai", "azure"];
|
|
1309
|
-
}>, import("../../utils
|
|
1309
|
+
}>, import("../../utils").Preset<"openai/text-embedding-3-large", CatalogModel, {
|
|
1310
1310
|
name: string;
|
|
1311
1311
|
created: string;
|
|
1312
1312
|
context: number;
|