@ai-sdk/openai 2.0.0-beta.1 → 2.0.0-beta.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +10 -0
- package/dist/index.d.mts +6 -25
- package/dist/index.d.ts +6 -25
- package/dist/index.js +320 -320
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +16 -16
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +28 -203
- package/dist/internal/index.d.ts +28 -203
- package/dist/internal/index.js +320 -320
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +16 -16
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +2 -2
package/dist/index.js
CHANGED
|
@@ -31,7 +31,7 @@ var import_provider_utils12 = require("@ai-sdk/provider-utils");
|
|
|
31
31
|
// src/openai-chat-language-model.ts
|
|
32
32
|
var import_provider3 = require("@ai-sdk/provider");
|
|
33
33
|
var import_provider_utils5 = require("@ai-sdk/provider-utils");
|
|
34
|
-
var
|
|
34
|
+
var import_v45 = require("zod/v4");
|
|
35
35
|
|
|
36
36
|
// src/convert-to-openai-chat-messages.ts
|
|
37
37
|
var import_provider = require("@ai-sdk/provider");
|
|
@@ -242,15 +242,15 @@ function mapOpenAIFinishReason(finishReason) {
|
|
|
242
242
|
}
|
|
243
243
|
|
|
244
244
|
// src/openai-chat-options.ts
|
|
245
|
-
var
|
|
246
|
-
var openaiProviderOptions =
|
|
245
|
+
var import_v4 = require("zod/v4");
|
|
246
|
+
var openaiProviderOptions = import_v4.z.object({
|
|
247
247
|
/**
|
|
248
248
|
* Modify the likelihood of specified tokens appearing in the completion.
|
|
249
249
|
*
|
|
250
250
|
* Accepts a JSON object that maps tokens (specified by their token ID in
|
|
251
251
|
* the GPT tokenizer) to an associated bias value from -100 to 100.
|
|
252
252
|
*/
|
|
253
|
-
logitBias:
|
|
253
|
+
logitBias: import_v4.z.record(import_v4.z.coerce.number(), import_v4.z.number()).optional(),
|
|
254
254
|
/**
|
|
255
255
|
* Return the log probabilities of the tokens.
|
|
256
256
|
*
|
|
@@ -260,63 +260,63 @@ var openaiProviderOptions = import_zod.z.object({
|
|
|
260
260
|
* Setting to a number will return the log probabilities of the top n
|
|
261
261
|
* tokens that were generated.
|
|
262
262
|
*/
|
|
263
|
-
logprobs:
|
|
263
|
+
logprobs: import_v4.z.union([import_v4.z.boolean(), import_v4.z.number()]).optional(),
|
|
264
264
|
/**
|
|
265
265
|
* Whether to enable parallel function calling during tool use. Default to true.
|
|
266
266
|
*/
|
|
267
|
-
parallelToolCalls:
|
|
267
|
+
parallelToolCalls: import_v4.z.boolean().optional(),
|
|
268
268
|
/**
|
|
269
269
|
* A unique identifier representing your end-user, which can help OpenAI to
|
|
270
270
|
* monitor and detect abuse.
|
|
271
271
|
*/
|
|
272
|
-
user:
|
|
272
|
+
user: import_v4.z.string().optional(),
|
|
273
273
|
/**
|
|
274
274
|
* Reasoning effort for reasoning models. Defaults to `medium`.
|
|
275
275
|
*/
|
|
276
|
-
reasoningEffort:
|
|
276
|
+
reasoningEffort: import_v4.z.enum(["low", "medium", "high"]).optional(),
|
|
277
277
|
/**
|
|
278
278
|
* Maximum number of completion tokens to generate. Useful for reasoning models.
|
|
279
279
|
*/
|
|
280
|
-
maxCompletionTokens:
|
|
280
|
+
maxCompletionTokens: import_v4.z.number().optional(),
|
|
281
281
|
/**
|
|
282
282
|
* Whether to enable persistence in responses API.
|
|
283
283
|
*/
|
|
284
|
-
store:
|
|
284
|
+
store: import_v4.z.boolean().optional(),
|
|
285
285
|
/**
|
|
286
286
|
* Metadata to associate with the request.
|
|
287
287
|
*/
|
|
288
|
-
metadata:
|
|
288
|
+
metadata: import_v4.z.record(import_v4.z.string().max(64), import_v4.z.string().max(512)).optional(),
|
|
289
289
|
/**
|
|
290
290
|
* Parameters for prediction mode.
|
|
291
291
|
*/
|
|
292
|
-
prediction:
|
|
292
|
+
prediction: import_v4.z.record(import_v4.z.string(), import_v4.z.any()).optional(),
|
|
293
293
|
/**
|
|
294
294
|
* Whether to use structured outputs.
|
|
295
295
|
*
|
|
296
296
|
* @default true
|
|
297
297
|
*/
|
|
298
|
-
structuredOutputs:
|
|
298
|
+
structuredOutputs: import_v4.z.boolean().optional(),
|
|
299
299
|
/**
|
|
300
300
|
* Service tier for the request. Set to 'flex' for 50% cheaper processing
|
|
301
301
|
* at the cost of increased latency. Only available for o3 and o4-mini models.
|
|
302
302
|
*
|
|
303
303
|
* @default 'auto'
|
|
304
304
|
*/
|
|
305
|
-
serviceTier:
|
|
305
|
+
serviceTier: import_v4.z.enum(["auto", "flex"]).optional()
|
|
306
306
|
});
|
|
307
307
|
|
|
308
308
|
// src/openai-error.ts
|
|
309
|
-
var
|
|
309
|
+
var import_v42 = require("zod/v4");
|
|
310
310
|
var import_provider_utils2 = require("@ai-sdk/provider-utils");
|
|
311
|
-
var openaiErrorDataSchema =
|
|
312
|
-
error:
|
|
313
|
-
message:
|
|
311
|
+
var openaiErrorDataSchema = import_v42.z.object({
|
|
312
|
+
error: import_v42.z.object({
|
|
313
|
+
message: import_v42.z.string(),
|
|
314
314
|
// The additional information below is handled loosely to support
|
|
315
315
|
// OpenAI-compatible providers that have slightly different error
|
|
316
316
|
// responses:
|
|
317
|
-
type:
|
|
318
|
-
param:
|
|
319
|
-
code:
|
|
317
|
+
type: import_v42.z.string().nullish(),
|
|
318
|
+
param: import_v42.z.any().nullish(),
|
|
319
|
+
code: import_v42.z.union([import_v42.z.string(), import_v42.z.number()]).nullish()
|
|
320
320
|
})
|
|
321
321
|
});
|
|
322
322
|
var openaiFailedResponseHandler = (0, import_provider_utils2.createJsonErrorResponseHandler)({
|
|
@@ -329,70 +329,70 @@ var import_provider2 = require("@ai-sdk/provider");
|
|
|
329
329
|
|
|
330
330
|
// src/tool/file-search.ts
|
|
331
331
|
var import_provider_utils3 = require("@ai-sdk/provider-utils");
|
|
332
|
-
var
|
|
333
|
-
var fileSearchArgsSchema =
|
|
332
|
+
var import_v43 = require("zod/v4");
|
|
333
|
+
var fileSearchArgsSchema = import_v43.z.object({
|
|
334
334
|
/**
|
|
335
335
|
* List of vector store IDs to search through. If not provided, searches all available vector stores.
|
|
336
336
|
*/
|
|
337
|
-
vectorStoreIds:
|
|
337
|
+
vectorStoreIds: import_v43.z.array(import_v43.z.string()).optional(),
|
|
338
338
|
/**
|
|
339
339
|
* Maximum number of search results to return. Defaults to 10.
|
|
340
340
|
*/
|
|
341
|
-
maxResults:
|
|
341
|
+
maxResults: import_v43.z.number().optional(),
|
|
342
342
|
/**
|
|
343
343
|
* Type of search to perform. Defaults to 'auto'.
|
|
344
344
|
*/
|
|
345
|
-
searchType:
|
|
345
|
+
searchType: import_v43.z.enum(["auto", "keyword", "semantic"]).optional()
|
|
346
346
|
});
|
|
347
347
|
var fileSearch = (0, import_provider_utils3.createProviderDefinedToolFactory)({
|
|
348
348
|
id: "openai.file_search",
|
|
349
349
|
name: "file_search",
|
|
350
|
-
inputSchema:
|
|
351
|
-
query:
|
|
350
|
+
inputSchema: import_v43.z.object({
|
|
351
|
+
query: import_v43.z.string()
|
|
352
352
|
})
|
|
353
353
|
});
|
|
354
354
|
|
|
355
355
|
// src/tool/web-search-preview.ts
|
|
356
356
|
var import_provider_utils4 = require("@ai-sdk/provider-utils");
|
|
357
|
-
var
|
|
358
|
-
var webSearchPreviewArgsSchema =
|
|
357
|
+
var import_v44 = require("zod/v4");
|
|
358
|
+
var webSearchPreviewArgsSchema = import_v44.z.object({
|
|
359
359
|
/**
|
|
360
360
|
* Search context size to use for the web search.
|
|
361
361
|
* - high: Most comprehensive context, highest cost, slower response
|
|
362
362
|
* - medium: Balanced context, cost, and latency (default)
|
|
363
363
|
* - low: Least context, lowest cost, fastest response
|
|
364
364
|
*/
|
|
365
|
-
searchContextSize:
|
|
365
|
+
searchContextSize: import_v44.z.enum(["low", "medium", "high"]).optional(),
|
|
366
366
|
/**
|
|
367
367
|
* User location information to provide geographically relevant search results.
|
|
368
368
|
*/
|
|
369
|
-
userLocation:
|
|
369
|
+
userLocation: import_v44.z.object({
|
|
370
370
|
/**
|
|
371
371
|
* Type of location (always 'approximate')
|
|
372
372
|
*/
|
|
373
|
-
type:
|
|
373
|
+
type: import_v44.z.literal("approximate"),
|
|
374
374
|
/**
|
|
375
375
|
* Two-letter ISO country code (e.g., 'US', 'GB')
|
|
376
376
|
*/
|
|
377
|
-
country:
|
|
377
|
+
country: import_v44.z.string().optional(),
|
|
378
378
|
/**
|
|
379
379
|
* City name (free text, e.g., 'Minneapolis')
|
|
380
380
|
*/
|
|
381
|
-
city:
|
|
381
|
+
city: import_v44.z.string().optional(),
|
|
382
382
|
/**
|
|
383
383
|
* Region name (free text, e.g., 'Minnesota')
|
|
384
384
|
*/
|
|
385
|
-
region:
|
|
385
|
+
region: import_v44.z.string().optional(),
|
|
386
386
|
/**
|
|
387
387
|
* IANA timezone (e.g., 'America/Chicago')
|
|
388
388
|
*/
|
|
389
|
-
timezone:
|
|
389
|
+
timezone: import_v44.z.string().optional()
|
|
390
390
|
}).optional()
|
|
391
391
|
});
|
|
392
392
|
var webSearchPreview = (0, import_provider_utils4.createProviderDefinedToolFactory)({
|
|
393
393
|
id: "openai.web_search_preview",
|
|
394
394
|
name: "web_search_preview",
|
|
395
|
-
inputSchema:
|
|
395
|
+
inputSchema: import_v44.z.object({})
|
|
396
396
|
});
|
|
397
397
|
|
|
398
398
|
// src/openai-prepare-tools.ts
|
|
@@ -943,97 +943,97 @@ var OpenAIChatLanguageModel = class {
|
|
|
943
943
|
};
|
|
944
944
|
}
|
|
945
945
|
};
|
|
946
|
-
var openaiTokenUsageSchema =
|
|
947
|
-
prompt_tokens:
|
|
948
|
-
completion_tokens:
|
|
949
|
-
total_tokens:
|
|
950
|
-
prompt_tokens_details:
|
|
951
|
-
cached_tokens:
|
|
946
|
+
var openaiTokenUsageSchema = import_v45.z.object({
|
|
947
|
+
prompt_tokens: import_v45.z.number().nullish(),
|
|
948
|
+
completion_tokens: import_v45.z.number().nullish(),
|
|
949
|
+
total_tokens: import_v45.z.number().nullish(),
|
|
950
|
+
prompt_tokens_details: import_v45.z.object({
|
|
951
|
+
cached_tokens: import_v45.z.number().nullish()
|
|
952
952
|
}).nullish(),
|
|
953
|
-
completion_tokens_details:
|
|
954
|
-
reasoning_tokens:
|
|
955
|
-
accepted_prediction_tokens:
|
|
956
|
-
rejected_prediction_tokens:
|
|
953
|
+
completion_tokens_details: import_v45.z.object({
|
|
954
|
+
reasoning_tokens: import_v45.z.number().nullish(),
|
|
955
|
+
accepted_prediction_tokens: import_v45.z.number().nullish(),
|
|
956
|
+
rejected_prediction_tokens: import_v45.z.number().nullish()
|
|
957
957
|
}).nullish()
|
|
958
958
|
}).nullish();
|
|
959
|
-
var openaiChatResponseSchema =
|
|
960
|
-
id:
|
|
961
|
-
created:
|
|
962
|
-
model:
|
|
963
|
-
choices:
|
|
964
|
-
|
|
965
|
-
message:
|
|
966
|
-
role:
|
|
967
|
-
content:
|
|
968
|
-
tool_calls:
|
|
969
|
-
|
|
970
|
-
id:
|
|
971
|
-
type:
|
|
972
|
-
function:
|
|
973
|
-
name:
|
|
974
|
-
arguments:
|
|
959
|
+
var openaiChatResponseSchema = import_v45.z.object({
|
|
960
|
+
id: import_v45.z.string().nullish(),
|
|
961
|
+
created: import_v45.z.number().nullish(),
|
|
962
|
+
model: import_v45.z.string().nullish(),
|
|
963
|
+
choices: import_v45.z.array(
|
|
964
|
+
import_v45.z.object({
|
|
965
|
+
message: import_v45.z.object({
|
|
966
|
+
role: import_v45.z.literal("assistant").nullish(),
|
|
967
|
+
content: import_v45.z.string().nullish(),
|
|
968
|
+
tool_calls: import_v45.z.array(
|
|
969
|
+
import_v45.z.object({
|
|
970
|
+
id: import_v45.z.string().nullish(),
|
|
971
|
+
type: import_v45.z.literal("function"),
|
|
972
|
+
function: import_v45.z.object({
|
|
973
|
+
name: import_v45.z.string(),
|
|
974
|
+
arguments: import_v45.z.string()
|
|
975
975
|
})
|
|
976
976
|
})
|
|
977
977
|
).nullish()
|
|
978
978
|
}),
|
|
979
|
-
index:
|
|
980
|
-
logprobs:
|
|
981
|
-
content:
|
|
982
|
-
|
|
983
|
-
token:
|
|
984
|
-
logprob:
|
|
985
|
-
top_logprobs:
|
|
986
|
-
|
|
987
|
-
token:
|
|
988
|
-
logprob:
|
|
979
|
+
index: import_v45.z.number(),
|
|
980
|
+
logprobs: import_v45.z.object({
|
|
981
|
+
content: import_v45.z.array(
|
|
982
|
+
import_v45.z.object({
|
|
983
|
+
token: import_v45.z.string(),
|
|
984
|
+
logprob: import_v45.z.number(),
|
|
985
|
+
top_logprobs: import_v45.z.array(
|
|
986
|
+
import_v45.z.object({
|
|
987
|
+
token: import_v45.z.string(),
|
|
988
|
+
logprob: import_v45.z.number()
|
|
989
989
|
})
|
|
990
990
|
)
|
|
991
991
|
})
|
|
992
992
|
).nullish()
|
|
993
993
|
}).nullish(),
|
|
994
|
-
finish_reason:
|
|
994
|
+
finish_reason: import_v45.z.string().nullish()
|
|
995
995
|
})
|
|
996
996
|
),
|
|
997
997
|
usage: openaiTokenUsageSchema
|
|
998
998
|
});
|
|
999
|
-
var openaiChatChunkSchema =
|
|
1000
|
-
|
|
1001
|
-
id:
|
|
1002
|
-
created:
|
|
1003
|
-
model:
|
|
1004
|
-
choices:
|
|
1005
|
-
|
|
1006
|
-
delta:
|
|
1007
|
-
role:
|
|
1008
|
-
content:
|
|
1009
|
-
tool_calls:
|
|
1010
|
-
|
|
1011
|
-
index:
|
|
1012
|
-
id:
|
|
1013
|
-
type:
|
|
1014
|
-
function:
|
|
1015
|
-
name:
|
|
1016
|
-
arguments:
|
|
999
|
+
var openaiChatChunkSchema = import_v45.z.union([
|
|
1000
|
+
import_v45.z.object({
|
|
1001
|
+
id: import_v45.z.string().nullish(),
|
|
1002
|
+
created: import_v45.z.number().nullish(),
|
|
1003
|
+
model: import_v45.z.string().nullish(),
|
|
1004
|
+
choices: import_v45.z.array(
|
|
1005
|
+
import_v45.z.object({
|
|
1006
|
+
delta: import_v45.z.object({
|
|
1007
|
+
role: import_v45.z.enum(["assistant"]).nullish(),
|
|
1008
|
+
content: import_v45.z.string().nullish(),
|
|
1009
|
+
tool_calls: import_v45.z.array(
|
|
1010
|
+
import_v45.z.object({
|
|
1011
|
+
index: import_v45.z.number(),
|
|
1012
|
+
id: import_v45.z.string().nullish(),
|
|
1013
|
+
type: import_v45.z.literal("function").nullish(),
|
|
1014
|
+
function: import_v45.z.object({
|
|
1015
|
+
name: import_v45.z.string().nullish(),
|
|
1016
|
+
arguments: import_v45.z.string().nullish()
|
|
1017
1017
|
})
|
|
1018
1018
|
})
|
|
1019
1019
|
).nullish()
|
|
1020
1020
|
}).nullish(),
|
|
1021
|
-
logprobs:
|
|
1022
|
-
content:
|
|
1023
|
-
|
|
1024
|
-
token:
|
|
1025
|
-
logprob:
|
|
1026
|
-
top_logprobs:
|
|
1027
|
-
|
|
1028
|
-
token:
|
|
1029
|
-
logprob:
|
|
1021
|
+
logprobs: import_v45.z.object({
|
|
1022
|
+
content: import_v45.z.array(
|
|
1023
|
+
import_v45.z.object({
|
|
1024
|
+
token: import_v45.z.string(),
|
|
1025
|
+
logprob: import_v45.z.number(),
|
|
1026
|
+
top_logprobs: import_v45.z.array(
|
|
1027
|
+
import_v45.z.object({
|
|
1028
|
+
token: import_v45.z.string(),
|
|
1029
|
+
logprob: import_v45.z.number()
|
|
1030
1030
|
})
|
|
1031
1031
|
)
|
|
1032
1032
|
})
|
|
1033
1033
|
).nullish()
|
|
1034
1034
|
}).nullish(),
|
|
1035
|
-
finish_reason:
|
|
1036
|
-
index:
|
|
1035
|
+
finish_reason: import_v45.z.string().nullish(),
|
|
1036
|
+
index: import_v45.z.number()
|
|
1037
1037
|
})
|
|
1038
1038
|
),
|
|
1039
1039
|
usage: openaiTokenUsageSchema
|
|
@@ -1088,7 +1088,7 @@ var reasoningModels = {
|
|
|
1088
1088
|
|
|
1089
1089
|
// src/openai-completion-language-model.ts
|
|
1090
1090
|
var import_provider_utils6 = require("@ai-sdk/provider-utils");
|
|
1091
|
-
var
|
|
1091
|
+
var import_v47 = require("zod/v4");
|
|
1092
1092
|
|
|
1093
1093
|
// src/convert-to-openai-completion-prompt.ts
|
|
1094
1094
|
var import_provider4 = require("@ai-sdk/provider");
|
|
@@ -1166,12 +1166,12 @@ ${user}:`]
|
|
|
1166
1166
|
}
|
|
1167
1167
|
|
|
1168
1168
|
// src/openai-completion-options.ts
|
|
1169
|
-
var
|
|
1170
|
-
var openaiCompletionProviderOptions =
|
|
1169
|
+
var import_v46 = require("zod/v4");
|
|
1170
|
+
var openaiCompletionProviderOptions = import_v46.z.object({
|
|
1171
1171
|
/**
|
|
1172
1172
|
Echo back the prompt in addition to the completion.
|
|
1173
1173
|
*/
|
|
1174
|
-
echo:
|
|
1174
|
+
echo: import_v46.z.boolean().optional(),
|
|
1175
1175
|
/**
|
|
1176
1176
|
Modify the likelihood of specified tokens appearing in the completion.
|
|
1177
1177
|
|
|
@@ -1186,16 +1186,16 @@ var openaiCompletionProviderOptions = import_zod6.z.object({
|
|
|
1186
1186
|
As an example, you can pass {"50256": -100} to prevent the <|endoftext|>
|
|
1187
1187
|
token from being generated.
|
|
1188
1188
|
*/
|
|
1189
|
-
logitBias:
|
|
1189
|
+
logitBias: import_v46.z.record(import_v46.z.string(), import_v46.z.number()).optional(),
|
|
1190
1190
|
/**
|
|
1191
1191
|
The suffix that comes after a completion of inserted text.
|
|
1192
1192
|
*/
|
|
1193
|
-
suffix:
|
|
1193
|
+
suffix: import_v46.z.string().optional(),
|
|
1194
1194
|
/**
|
|
1195
1195
|
A unique identifier representing your end-user, which can help OpenAI to
|
|
1196
1196
|
monitor and detect abuse. Learn more.
|
|
1197
1197
|
*/
|
|
1198
|
-
user:
|
|
1198
|
+
user: import_v46.z.string().optional(),
|
|
1199
1199
|
/**
|
|
1200
1200
|
Return the log probabilities of the tokens. Including logprobs will increase
|
|
1201
1201
|
the response size and can slow down response times. However, it can
|
|
@@ -1205,7 +1205,7 @@ var openaiCompletionProviderOptions = import_zod6.z.object({
|
|
|
1205
1205
|
Setting to a number will return the log probabilities of the top n
|
|
1206
1206
|
tokens that were generated.
|
|
1207
1207
|
*/
|
|
1208
|
-
logprobs:
|
|
1208
|
+
logprobs: import_v46.z.union([import_v46.z.boolean(), import_v46.z.number()]).optional()
|
|
1209
1209
|
});
|
|
1210
1210
|
|
|
1211
1211
|
// src/openai-completion-language-model.ts
|
|
@@ -1437,42 +1437,42 @@ var OpenAICompletionLanguageModel = class {
|
|
|
1437
1437
|
};
|
|
1438
1438
|
}
|
|
1439
1439
|
};
|
|
1440
|
-
var usageSchema =
|
|
1441
|
-
prompt_tokens:
|
|
1442
|
-
completion_tokens:
|
|
1443
|
-
total_tokens:
|
|
1440
|
+
var usageSchema = import_v47.z.object({
|
|
1441
|
+
prompt_tokens: import_v47.z.number(),
|
|
1442
|
+
completion_tokens: import_v47.z.number(),
|
|
1443
|
+
total_tokens: import_v47.z.number()
|
|
1444
1444
|
});
|
|
1445
|
-
var openaiCompletionResponseSchema =
|
|
1446
|
-
id:
|
|
1447
|
-
created:
|
|
1448
|
-
model:
|
|
1449
|
-
choices:
|
|
1450
|
-
|
|
1451
|
-
text:
|
|
1452
|
-
finish_reason:
|
|
1453
|
-
logprobs:
|
|
1454
|
-
tokens:
|
|
1455
|
-
token_logprobs:
|
|
1456
|
-
top_logprobs:
|
|
1445
|
+
var openaiCompletionResponseSchema = import_v47.z.object({
|
|
1446
|
+
id: import_v47.z.string().nullish(),
|
|
1447
|
+
created: import_v47.z.number().nullish(),
|
|
1448
|
+
model: import_v47.z.string().nullish(),
|
|
1449
|
+
choices: import_v47.z.array(
|
|
1450
|
+
import_v47.z.object({
|
|
1451
|
+
text: import_v47.z.string(),
|
|
1452
|
+
finish_reason: import_v47.z.string(),
|
|
1453
|
+
logprobs: import_v47.z.object({
|
|
1454
|
+
tokens: import_v47.z.array(import_v47.z.string()),
|
|
1455
|
+
token_logprobs: import_v47.z.array(import_v47.z.number()),
|
|
1456
|
+
top_logprobs: import_v47.z.array(import_v47.z.record(import_v47.z.string(), import_v47.z.number())).nullish()
|
|
1457
1457
|
}).nullish()
|
|
1458
1458
|
})
|
|
1459
1459
|
),
|
|
1460
1460
|
usage: usageSchema.nullish()
|
|
1461
1461
|
});
|
|
1462
|
-
var openaiCompletionChunkSchema =
|
|
1463
|
-
|
|
1464
|
-
id:
|
|
1465
|
-
created:
|
|
1466
|
-
model:
|
|
1467
|
-
choices:
|
|
1468
|
-
|
|
1469
|
-
text:
|
|
1470
|
-
finish_reason:
|
|
1471
|
-
index:
|
|
1472
|
-
logprobs:
|
|
1473
|
-
tokens:
|
|
1474
|
-
token_logprobs:
|
|
1475
|
-
top_logprobs:
|
|
1462
|
+
var openaiCompletionChunkSchema = import_v47.z.union([
|
|
1463
|
+
import_v47.z.object({
|
|
1464
|
+
id: import_v47.z.string().nullish(),
|
|
1465
|
+
created: import_v47.z.number().nullish(),
|
|
1466
|
+
model: import_v47.z.string().nullish(),
|
|
1467
|
+
choices: import_v47.z.array(
|
|
1468
|
+
import_v47.z.object({
|
|
1469
|
+
text: import_v47.z.string(),
|
|
1470
|
+
finish_reason: import_v47.z.string().nullish(),
|
|
1471
|
+
index: import_v47.z.number(),
|
|
1472
|
+
logprobs: import_v47.z.object({
|
|
1473
|
+
tokens: import_v47.z.array(import_v47.z.string()),
|
|
1474
|
+
token_logprobs: import_v47.z.array(import_v47.z.number()),
|
|
1475
|
+
top_logprobs: import_v47.z.array(import_v47.z.record(import_v47.z.string(), import_v47.z.number())).nullish()
|
|
1476
1476
|
}).nullish()
|
|
1477
1477
|
})
|
|
1478
1478
|
),
|
|
@@ -1484,21 +1484,21 @@ var openaiCompletionChunkSchema = import_zod7.z.union([
|
|
|
1484
1484
|
// src/openai-embedding-model.ts
|
|
1485
1485
|
var import_provider5 = require("@ai-sdk/provider");
|
|
1486
1486
|
var import_provider_utils7 = require("@ai-sdk/provider-utils");
|
|
1487
|
-
var
|
|
1487
|
+
var import_v49 = require("zod/v4");
|
|
1488
1488
|
|
|
1489
1489
|
// src/openai-embedding-options.ts
|
|
1490
|
-
var
|
|
1491
|
-
var openaiEmbeddingProviderOptions =
|
|
1490
|
+
var import_v48 = require("zod/v4");
|
|
1491
|
+
var openaiEmbeddingProviderOptions = import_v48.z.object({
|
|
1492
1492
|
/**
|
|
1493
1493
|
The number of dimensions the resulting output embeddings should have.
|
|
1494
1494
|
Only supported in text-embedding-3 and later models.
|
|
1495
1495
|
*/
|
|
1496
|
-
dimensions:
|
|
1496
|
+
dimensions: import_v48.z.number().optional(),
|
|
1497
1497
|
/**
|
|
1498
1498
|
A unique identifier representing your end-user, which can help OpenAI to
|
|
1499
1499
|
monitor and detect abuse. Learn more.
|
|
1500
1500
|
*/
|
|
1501
|
-
user:
|
|
1501
|
+
user: import_v48.z.string().optional()
|
|
1502
1502
|
});
|
|
1503
1503
|
|
|
1504
1504
|
// src/openai-embedding-model.ts
|
|
@@ -1564,14 +1564,14 @@ var OpenAIEmbeddingModel = class {
|
|
|
1564
1564
|
};
|
|
1565
1565
|
}
|
|
1566
1566
|
};
|
|
1567
|
-
var openaiTextEmbeddingResponseSchema =
|
|
1568
|
-
data:
|
|
1569
|
-
usage:
|
|
1567
|
+
var openaiTextEmbeddingResponseSchema = import_v49.z.object({
|
|
1568
|
+
data: import_v49.z.array(import_v49.z.object({ embedding: import_v49.z.array(import_v49.z.number()) })),
|
|
1569
|
+
usage: import_v49.z.object({ prompt_tokens: import_v49.z.number() }).nullish()
|
|
1570
1570
|
});
|
|
1571
1571
|
|
|
1572
1572
|
// src/openai-image-model.ts
|
|
1573
1573
|
var import_provider_utils8 = require("@ai-sdk/provider-utils");
|
|
1574
|
-
var
|
|
1574
|
+
var import_v410 = require("zod/v4");
|
|
1575
1575
|
|
|
1576
1576
|
// src/openai-image-settings.ts
|
|
1577
1577
|
var modelMaxImagesPerCall = {
|
|
@@ -1659,9 +1659,9 @@ var OpenAIImageModel = class {
|
|
|
1659
1659
|
};
|
|
1660
1660
|
}
|
|
1661
1661
|
};
|
|
1662
|
-
var openaiImageResponseSchema =
|
|
1663
|
-
data:
|
|
1664
|
-
|
|
1662
|
+
var openaiImageResponseSchema = import_v410.z.object({
|
|
1663
|
+
data: import_v410.z.array(
|
|
1664
|
+
import_v410.z.object({ b64_json: import_v410.z.string(), revised_prompt: import_v410.z.string().optional() })
|
|
1665
1665
|
)
|
|
1666
1666
|
});
|
|
1667
1667
|
|
|
@@ -1673,33 +1673,33 @@ var openaiTools = {
|
|
|
1673
1673
|
|
|
1674
1674
|
// src/openai-transcription-model.ts
|
|
1675
1675
|
var import_provider_utils9 = require("@ai-sdk/provider-utils");
|
|
1676
|
-
var
|
|
1676
|
+
var import_v412 = require("zod/v4");
|
|
1677
1677
|
|
|
1678
1678
|
// src/openai-transcription-options.ts
|
|
1679
|
-
var
|
|
1680
|
-
var openAITranscriptionProviderOptions =
|
|
1679
|
+
var import_v411 = require("zod/v4");
|
|
1680
|
+
var openAITranscriptionProviderOptions = import_v411.z.object({
|
|
1681
1681
|
/**
|
|
1682
1682
|
* Additional information to include in the transcription response.
|
|
1683
1683
|
*/
|
|
1684
|
-
include:
|
|
1684
|
+
include: import_v411.z.array(import_v411.z.string()).optional(),
|
|
1685
1685
|
/**
|
|
1686
1686
|
* The language of the input audio in ISO-639-1 format.
|
|
1687
1687
|
*/
|
|
1688
|
-
language:
|
|
1688
|
+
language: import_v411.z.string().optional(),
|
|
1689
1689
|
/**
|
|
1690
1690
|
* An optional text to guide the model's style or continue a previous audio segment.
|
|
1691
1691
|
*/
|
|
1692
|
-
prompt:
|
|
1692
|
+
prompt: import_v411.z.string().optional(),
|
|
1693
1693
|
/**
|
|
1694
1694
|
* The sampling temperature, between 0 and 1.
|
|
1695
1695
|
* @default 0
|
|
1696
1696
|
*/
|
|
1697
|
-
temperature:
|
|
1697
|
+
temperature: import_v411.z.number().min(0).max(1).default(0).optional(),
|
|
1698
1698
|
/**
|
|
1699
1699
|
* The timestamp granularities to populate for this transcription.
|
|
1700
1700
|
* @default ['segment']
|
|
1701
1701
|
*/
|
|
1702
|
-
timestampGranularities:
|
|
1702
|
+
timestampGranularities: import_v411.z.array(import_v411.z.enum(["word", "segment"])).default(["segment"]).optional()
|
|
1703
1703
|
});
|
|
1704
1704
|
|
|
1705
1705
|
// src/openai-transcription-model.ts
|
|
@@ -1847,22 +1847,22 @@ var OpenAITranscriptionModel = class {
|
|
|
1847
1847
|
};
|
|
1848
1848
|
}
|
|
1849
1849
|
};
|
|
1850
|
-
var openaiTranscriptionResponseSchema =
|
|
1851
|
-
text:
|
|
1852
|
-
language:
|
|
1853
|
-
duration:
|
|
1854
|
-
words:
|
|
1855
|
-
|
|
1856
|
-
word:
|
|
1857
|
-
start:
|
|
1858
|
-
end:
|
|
1850
|
+
var openaiTranscriptionResponseSchema = import_v412.z.object({
|
|
1851
|
+
text: import_v412.z.string(),
|
|
1852
|
+
language: import_v412.z.string().nullish(),
|
|
1853
|
+
duration: import_v412.z.number().nullish(),
|
|
1854
|
+
words: import_v412.z.array(
|
|
1855
|
+
import_v412.z.object({
|
|
1856
|
+
word: import_v412.z.string(),
|
|
1857
|
+
start: import_v412.z.number(),
|
|
1858
|
+
end: import_v412.z.number()
|
|
1859
1859
|
})
|
|
1860
1860
|
).nullish()
|
|
1861
1861
|
});
|
|
1862
1862
|
|
|
1863
1863
|
// src/responses/openai-responses-language-model.ts
|
|
1864
1864
|
var import_provider_utils10 = require("@ai-sdk/provider-utils");
|
|
1865
|
-
var
|
|
1865
|
+
var import_v413 = require("zod/v4");
|
|
1866
1866
|
|
|
1867
1867
|
// src/responses/convert-to-openai-responses-messages.ts
|
|
1868
1868
|
var import_provider6 = require("@ai-sdk/provider");
|
|
@@ -2253,59 +2253,59 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2253
2253
|
body,
|
|
2254
2254
|
failedResponseHandler: openaiFailedResponseHandler,
|
|
2255
2255
|
successfulResponseHandler: (0, import_provider_utils10.createJsonResponseHandler)(
|
|
2256
|
-
|
|
2257
|
-
id:
|
|
2258
|
-
created_at:
|
|
2259
|
-
model:
|
|
2260
|
-
output:
|
|
2261
|
-
|
|
2262
|
-
|
|
2263
|
-
type:
|
|
2264
|
-
role:
|
|
2265
|
-
content:
|
|
2266
|
-
|
|
2267
|
-
type:
|
|
2268
|
-
text:
|
|
2269
|
-
annotations:
|
|
2270
|
-
|
|
2271
|
-
type:
|
|
2272
|
-
start_index:
|
|
2273
|
-
end_index:
|
|
2274
|
-
url:
|
|
2275
|
-
title:
|
|
2256
|
+
import_v413.z.object({
|
|
2257
|
+
id: import_v413.z.string(),
|
|
2258
|
+
created_at: import_v413.z.number(),
|
|
2259
|
+
model: import_v413.z.string(),
|
|
2260
|
+
output: import_v413.z.array(
|
|
2261
|
+
import_v413.z.discriminatedUnion("type", [
|
|
2262
|
+
import_v413.z.object({
|
|
2263
|
+
type: import_v413.z.literal("message"),
|
|
2264
|
+
role: import_v413.z.literal("assistant"),
|
|
2265
|
+
content: import_v413.z.array(
|
|
2266
|
+
import_v413.z.object({
|
|
2267
|
+
type: import_v413.z.literal("output_text"),
|
|
2268
|
+
text: import_v413.z.string(),
|
|
2269
|
+
annotations: import_v413.z.array(
|
|
2270
|
+
import_v413.z.object({
|
|
2271
|
+
type: import_v413.z.literal("url_citation"),
|
|
2272
|
+
start_index: import_v413.z.number(),
|
|
2273
|
+
end_index: import_v413.z.number(),
|
|
2274
|
+
url: import_v413.z.string(),
|
|
2275
|
+
title: import_v413.z.string()
|
|
2276
2276
|
})
|
|
2277
2277
|
)
|
|
2278
2278
|
})
|
|
2279
2279
|
)
|
|
2280
2280
|
}),
|
|
2281
|
-
|
|
2282
|
-
type:
|
|
2283
|
-
call_id:
|
|
2284
|
-
name:
|
|
2285
|
-
arguments:
|
|
2281
|
+
import_v413.z.object({
|
|
2282
|
+
type: import_v413.z.literal("function_call"),
|
|
2283
|
+
call_id: import_v413.z.string(),
|
|
2284
|
+
name: import_v413.z.string(),
|
|
2285
|
+
arguments: import_v413.z.string()
|
|
2286
2286
|
}),
|
|
2287
|
-
|
|
2288
|
-
type:
|
|
2289
|
-
id:
|
|
2290
|
-
status:
|
|
2287
|
+
import_v413.z.object({
|
|
2288
|
+
type: import_v413.z.literal("web_search_call"),
|
|
2289
|
+
id: import_v413.z.string(),
|
|
2290
|
+
status: import_v413.z.string().optional()
|
|
2291
2291
|
}),
|
|
2292
|
-
|
|
2293
|
-
type:
|
|
2294
|
-
id:
|
|
2295
|
-
status:
|
|
2292
|
+
import_v413.z.object({
|
|
2293
|
+
type: import_v413.z.literal("computer_call"),
|
|
2294
|
+
id: import_v413.z.string(),
|
|
2295
|
+
status: import_v413.z.string().optional()
|
|
2296
2296
|
}),
|
|
2297
|
-
|
|
2298
|
-
type:
|
|
2299
|
-
summary:
|
|
2300
|
-
|
|
2301
|
-
type:
|
|
2302
|
-
text:
|
|
2297
|
+
import_v413.z.object({
|
|
2298
|
+
type: import_v413.z.literal("reasoning"),
|
|
2299
|
+
summary: import_v413.z.array(
|
|
2300
|
+
import_v413.z.object({
|
|
2301
|
+
type: import_v413.z.literal("summary_text"),
|
|
2302
|
+
text: import_v413.z.string()
|
|
2303
2303
|
})
|
|
2304
2304
|
)
|
|
2305
2305
|
})
|
|
2306
2306
|
])
|
|
2307
2307
|
),
|
|
2308
|
-
incomplete_details:
|
|
2308
|
+
incomplete_details: import_v413.z.object({ reason: import_v413.z.string() }).nullable(),
|
|
2309
2309
|
usage: usageSchema2
|
|
2310
2310
|
})
|
|
2311
2311
|
),
|
|
@@ -2646,124 +2646,124 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2646
2646
|
};
|
|
2647
2647
|
}
|
|
2648
2648
|
};
|
|
2649
|
-
var usageSchema2 =
|
|
2650
|
-
input_tokens:
|
|
2651
|
-
input_tokens_details:
|
|
2652
|
-
output_tokens:
|
|
2653
|
-
output_tokens_details:
|
|
2649
|
+
var usageSchema2 = import_v413.z.object({
|
|
2650
|
+
input_tokens: import_v413.z.number(),
|
|
2651
|
+
input_tokens_details: import_v413.z.object({ cached_tokens: import_v413.z.number().nullish() }).nullish(),
|
|
2652
|
+
output_tokens: import_v413.z.number(),
|
|
2653
|
+
output_tokens_details: import_v413.z.object({ reasoning_tokens: import_v413.z.number().nullish() }).nullish()
|
|
2654
2654
|
});
|
|
2655
|
-
var textDeltaChunkSchema =
|
|
2656
|
-
type:
|
|
2657
|
-
item_id:
|
|
2658
|
-
delta:
|
|
2655
|
+
var textDeltaChunkSchema = import_v413.z.object({
|
|
2656
|
+
type: import_v413.z.literal("response.output_text.delta"),
|
|
2657
|
+
item_id: import_v413.z.string(),
|
|
2658
|
+
delta: import_v413.z.string()
|
|
2659
2659
|
});
|
|
2660
|
-
var responseFinishedChunkSchema =
|
|
2661
|
-
type:
|
|
2662
|
-
response:
|
|
2663
|
-
incomplete_details:
|
|
2660
|
+
var responseFinishedChunkSchema = import_v413.z.object({
|
|
2661
|
+
type: import_v413.z.enum(["response.completed", "response.incomplete"]),
|
|
2662
|
+
response: import_v413.z.object({
|
|
2663
|
+
incomplete_details: import_v413.z.object({ reason: import_v413.z.string() }).nullish(),
|
|
2664
2664
|
usage: usageSchema2
|
|
2665
2665
|
})
|
|
2666
2666
|
});
|
|
2667
|
-
var responseCreatedChunkSchema =
|
|
2668
|
-
type:
|
|
2669
|
-
response:
|
|
2670
|
-
id:
|
|
2671
|
-
created_at:
|
|
2672
|
-
model:
|
|
2667
|
+
var responseCreatedChunkSchema = import_v413.z.object({
|
|
2668
|
+
type: import_v413.z.literal("response.created"),
|
|
2669
|
+
response: import_v413.z.object({
|
|
2670
|
+
id: import_v413.z.string(),
|
|
2671
|
+
created_at: import_v413.z.number(),
|
|
2672
|
+
model: import_v413.z.string()
|
|
2673
2673
|
})
|
|
2674
2674
|
});
|
|
2675
|
-
var responseOutputItemAddedSchema =
|
|
2676
|
-
type:
|
|
2677
|
-
output_index:
|
|
2678
|
-
item:
|
|
2679
|
-
|
|
2680
|
-
type:
|
|
2681
|
-
id:
|
|
2675
|
+
var responseOutputItemAddedSchema = import_v413.z.object({
|
|
2676
|
+
type: import_v413.z.literal("response.output_item.added"),
|
|
2677
|
+
output_index: import_v413.z.number(),
|
|
2678
|
+
item: import_v413.z.discriminatedUnion("type", [
|
|
2679
|
+
import_v413.z.object({
|
|
2680
|
+
type: import_v413.z.literal("message"),
|
|
2681
|
+
id: import_v413.z.string()
|
|
2682
2682
|
}),
|
|
2683
|
-
|
|
2684
|
-
type:
|
|
2685
|
-
id:
|
|
2683
|
+
import_v413.z.object({
|
|
2684
|
+
type: import_v413.z.literal("reasoning"),
|
|
2685
|
+
id: import_v413.z.string()
|
|
2686
2686
|
}),
|
|
2687
|
-
|
|
2688
|
-
type:
|
|
2689
|
-
id:
|
|
2690
|
-
call_id:
|
|
2691
|
-
name:
|
|
2692
|
-
arguments:
|
|
2687
|
+
import_v413.z.object({
|
|
2688
|
+
type: import_v413.z.literal("function_call"),
|
|
2689
|
+
id: import_v413.z.string(),
|
|
2690
|
+
call_id: import_v413.z.string(),
|
|
2691
|
+
name: import_v413.z.string(),
|
|
2692
|
+
arguments: import_v413.z.string()
|
|
2693
2693
|
}),
|
|
2694
|
-
|
|
2695
|
-
type:
|
|
2696
|
-
id:
|
|
2697
|
-
status:
|
|
2694
|
+
import_v413.z.object({
|
|
2695
|
+
type: import_v413.z.literal("web_search_call"),
|
|
2696
|
+
id: import_v413.z.string(),
|
|
2697
|
+
status: import_v413.z.string()
|
|
2698
2698
|
}),
|
|
2699
|
-
|
|
2700
|
-
type:
|
|
2701
|
-
id:
|
|
2702
|
-
status:
|
|
2699
|
+
import_v413.z.object({
|
|
2700
|
+
type: import_v413.z.literal("computer_call"),
|
|
2701
|
+
id: import_v413.z.string(),
|
|
2702
|
+
status: import_v413.z.string()
|
|
2703
2703
|
})
|
|
2704
2704
|
])
|
|
2705
2705
|
});
|
|
2706
|
-
var responseOutputItemDoneSchema =
|
|
2707
|
-
type:
|
|
2708
|
-
output_index:
|
|
2709
|
-
item:
|
|
2710
|
-
|
|
2711
|
-
type:
|
|
2712
|
-
id:
|
|
2706
|
+
var responseOutputItemDoneSchema = import_v413.z.object({
|
|
2707
|
+
type: import_v413.z.literal("response.output_item.done"),
|
|
2708
|
+
output_index: import_v413.z.number(),
|
|
2709
|
+
item: import_v413.z.discriminatedUnion("type", [
|
|
2710
|
+
import_v413.z.object({
|
|
2711
|
+
type: import_v413.z.literal("message"),
|
|
2712
|
+
id: import_v413.z.string()
|
|
2713
2713
|
}),
|
|
2714
|
-
|
|
2715
|
-
type:
|
|
2716
|
-
id:
|
|
2714
|
+
import_v413.z.object({
|
|
2715
|
+
type: import_v413.z.literal("reasoning"),
|
|
2716
|
+
id: import_v413.z.string()
|
|
2717
2717
|
}),
|
|
2718
|
-
|
|
2719
|
-
type:
|
|
2720
|
-
id:
|
|
2721
|
-
call_id:
|
|
2722
|
-
name:
|
|
2723
|
-
arguments:
|
|
2724
|
-
status:
|
|
2718
|
+
import_v413.z.object({
|
|
2719
|
+
type: import_v413.z.literal("function_call"),
|
|
2720
|
+
id: import_v413.z.string(),
|
|
2721
|
+
call_id: import_v413.z.string(),
|
|
2722
|
+
name: import_v413.z.string(),
|
|
2723
|
+
arguments: import_v413.z.string(),
|
|
2724
|
+
status: import_v413.z.literal("completed")
|
|
2725
2725
|
}),
|
|
2726
|
-
|
|
2727
|
-
type:
|
|
2728
|
-
id:
|
|
2729
|
-
status:
|
|
2726
|
+
import_v413.z.object({
|
|
2727
|
+
type: import_v413.z.literal("web_search_call"),
|
|
2728
|
+
id: import_v413.z.string(),
|
|
2729
|
+
status: import_v413.z.literal("completed")
|
|
2730
2730
|
}),
|
|
2731
|
-
|
|
2732
|
-
type:
|
|
2733
|
-
id:
|
|
2734
|
-
status:
|
|
2731
|
+
import_v413.z.object({
|
|
2732
|
+
type: import_v413.z.literal("computer_call"),
|
|
2733
|
+
id: import_v413.z.string(),
|
|
2734
|
+
status: import_v413.z.literal("completed")
|
|
2735
2735
|
})
|
|
2736
2736
|
])
|
|
2737
2737
|
});
|
|
2738
|
-
var responseFunctionCallArgumentsDeltaSchema =
|
|
2739
|
-
type:
|
|
2740
|
-
item_id:
|
|
2741
|
-
output_index:
|
|
2742
|
-
delta:
|
|
2738
|
+
var responseFunctionCallArgumentsDeltaSchema = import_v413.z.object({
|
|
2739
|
+
type: import_v413.z.literal("response.function_call_arguments.delta"),
|
|
2740
|
+
item_id: import_v413.z.string(),
|
|
2741
|
+
output_index: import_v413.z.number(),
|
|
2742
|
+
delta: import_v413.z.string()
|
|
2743
2743
|
});
|
|
2744
|
-
var responseAnnotationAddedSchema =
|
|
2745
|
-
type:
|
|
2746
|
-
annotation:
|
|
2747
|
-
type:
|
|
2748
|
-
url:
|
|
2749
|
-
title:
|
|
2744
|
+
var responseAnnotationAddedSchema = import_v413.z.object({
|
|
2745
|
+
type: import_v413.z.literal("response.output_text.annotation.added"),
|
|
2746
|
+
annotation: import_v413.z.object({
|
|
2747
|
+
type: import_v413.z.literal("url_citation"),
|
|
2748
|
+
url: import_v413.z.string(),
|
|
2749
|
+
title: import_v413.z.string()
|
|
2750
2750
|
})
|
|
2751
2751
|
});
|
|
2752
|
-
var responseReasoningSummaryTextDeltaSchema =
|
|
2753
|
-
type:
|
|
2754
|
-
item_id:
|
|
2755
|
-
output_index:
|
|
2756
|
-
summary_index:
|
|
2757
|
-
delta:
|
|
2752
|
+
var responseReasoningSummaryTextDeltaSchema = import_v413.z.object({
|
|
2753
|
+
type: import_v413.z.literal("response.reasoning_summary_text.delta"),
|
|
2754
|
+
item_id: import_v413.z.string(),
|
|
2755
|
+
output_index: import_v413.z.number(),
|
|
2756
|
+
summary_index: import_v413.z.number(),
|
|
2757
|
+
delta: import_v413.z.string()
|
|
2758
2758
|
});
|
|
2759
|
-
var responseReasoningSummaryPartDoneSchema =
|
|
2760
|
-
type:
|
|
2761
|
-
item_id:
|
|
2762
|
-
output_index:
|
|
2763
|
-
summary_index:
|
|
2764
|
-
part:
|
|
2759
|
+
var responseReasoningSummaryPartDoneSchema = import_v413.z.object({
|
|
2760
|
+
type: import_v413.z.literal("response.reasoning_summary_part.done"),
|
|
2761
|
+
item_id: import_v413.z.string(),
|
|
2762
|
+
output_index: import_v413.z.number(),
|
|
2763
|
+
summary_index: import_v413.z.number(),
|
|
2764
|
+
part: import_v413.z.unknown().nullish()
|
|
2765
2765
|
});
|
|
2766
|
-
var openaiResponsesChunkSchema =
|
|
2766
|
+
var openaiResponsesChunkSchema = import_v413.z.union([
|
|
2767
2767
|
textDeltaChunkSchema,
|
|
2768
2768
|
responseFinishedChunkSchema,
|
|
2769
2769
|
responseCreatedChunkSchema,
|
|
@@ -2773,7 +2773,7 @@ var openaiResponsesChunkSchema = import_zod13.z.union([
|
|
|
2773
2773
|
responseAnnotationAddedSchema,
|
|
2774
2774
|
responseReasoningSummaryTextDeltaSchema,
|
|
2775
2775
|
responseReasoningSummaryPartDoneSchema,
|
|
2776
|
-
|
|
2776
|
+
import_v413.z.object({ type: import_v413.z.string() }).passthrough()
|
|
2777
2777
|
// fallback for unknown chunks
|
|
2778
2778
|
]);
|
|
2779
2779
|
function isTextDeltaChunk(chunk) {
|
|
@@ -2824,25 +2824,25 @@ function getResponsesModelConfig(modelId) {
|
|
|
2824
2824
|
function supportsFlexProcessing2(modelId) {
|
|
2825
2825
|
return modelId.startsWith("o3") || modelId.startsWith("o4-mini");
|
|
2826
2826
|
}
|
|
2827
|
-
var openaiResponsesProviderOptionsSchema =
|
|
2828
|
-
metadata:
|
|
2829
|
-
parallelToolCalls:
|
|
2830
|
-
previousResponseId:
|
|
2831
|
-
store:
|
|
2832
|
-
user:
|
|
2833
|
-
reasoningEffort:
|
|
2834
|
-
strictSchemas:
|
|
2835
|
-
instructions:
|
|
2836
|
-
reasoningSummary:
|
|
2837
|
-
serviceTier:
|
|
2827
|
+
var openaiResponsesProviderOptionsSchema = import_v413.z.object({
|
|
2828
|
+
metadata: import_v413.z.any().nullish(),
|
|
2829
|
+
parallelToolCalls: import_v413.z.boolean().nullish(),
|
|
2830
|
+
previousResponseId: import_v413.z.string().nullish(),
|
|
2831
|
+
store: import_v413.z.boolean().nullish(),
|
|
2832
|
+
user: import_v413.z.string().nullish(),
|
|
2833
|
+
reasoningEffort: import_v413.z.string().nullish(),
|
|
2834
|
+
strictSchemas: import_v413.z.boolean().nullish(),
|
|
2835
|
+
instructions: import_v413.z.string().nullish(),
|
|
2836
|
+
reasoningSummary: import_v413.z.string().nullish(),
|
|
2837
|
+
serviceTier: import_v413.z.enum(["auto", "flex"]).nullish()
|
|
2838
2838
|
});
|
|
2839
2839
|
|
|
2840
2840
|
// src/openai-speech-model.ts
|
|
2841
2841
|
var import_provider_utils11 = require("@ai-sdk/provider-utils");
|
|
2842
|
-
var
|
|
2843
|
-
var OpenAIProviderOptionsSchema =
|
|
2844
|
-
instructions:
|
|
2845
|
-
speed:
|
|
2842
|
+
var import_v414 = require("zod/v4");
|
|
2843
|
+
var OpenAIProviderOptionsSchema = import_v414.z.object({
|
|
2844
|
+
instructions: import_v414.z.string().nullish(),
|
|
2845
|
+
speed: import_v414.z.number().min(0.25).max(4).default(1).nullish()
|
|
2846
2846
|
});
|
|
2847
2847
|
var OpenAISpeechModel = class {
|
|
2848
2848
|
constructor(modelId, config) {
|