@ai-sdk/openai 3.0.0-beta.77 → 3.0.0-beta.78
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +6 -0
- package/dist/index.js +79 -2
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +79 -2
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.js +78 -1
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +78 -1
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +1 -1
package/dist/index.mjs
CHANGED
|
@@ -3332,7 +3332,18 @@ var openaiResponsesModelIds = [
|
|
|
3332
3332
|
var openaiResponsesProviderOptionsSchema = lazySchema17(
|
|
3333
3333
|
() => zodSchema17(
|
|
3334
3334
|
z19.object({
|
|
3335
|
+
/**
|
|
3336
|
+
* The ID of the OpenAI Conversation to continue.
|
|
3337
|
+
* You must create a conversation first via the OpenAI API.
|
|
3338
|
+
* Cannot be used in conjunction with `previousResponseId`.
|
|
3339
|
+
* Defaults to `undefined`.
|
|
3340
|
+
* @see https://platform.openai.com/docs/api-reference/conversations/create
|
|
3341
|
+
*/
|
|
3335
3342
|
conversation: z19.string().nullish(),
|
|
3343
|
+
/**
|
|
3344
|
+
* The set of extra fields to include in the response (advanced, usually not needed).
|
|
3345
|
+
* Example values: 'reasoning.encrypted_content', 'file_search_call.results', 'message.output_text.logprobs'.
|
|
3346
|
+
*/
|
|
3336
3347
|
include: z19.array(
|
|
3337
3348
|
z19.enum([
|
|
3338
3349
|
"reasoning.encrypted_content",
|
|
@@ -3341,9 +3352,16 @@ var openaiResponsesProviderOptionsSchema = lazySchema17(
|
|
|
3341
3352
|
"message.output_text.logprobs"
|
|
3342
3353
|
])
|
|
3343
3354
|
).nullish(),
|
|
3355
|
+
/**
|
|
3356
|
+
* Instructions for the model.
|
|
3357
|
+
* They can be used to change the system or developer message when continuing a conversation using the `previousResponseId` option.
|
|
3358
|
+
* Defaults to `undefined`.
|
|
3359
|
+
*/
|
|
3344
3360
|
instructions: z19.string().nullish(),
|
|
3345
3361
|
/**
|
|
3346
|
-
* Return the log probabilities of the tokens.
|
|
3362
|
+
* Return the log probabilities of the tokens. Including logprobs will increase
|
|
3363
|
+
* the response size and can slow down response times. However, it can
|
|
3364
|
+
* be useful to better understand how the model is behaving.
|
|
3347
3365
|
*
|
|
3348
3366
|
* Setting to true will return the log probabilities of the tokens that
|
|
3349
3367
|
* were generated.
|
|
@@ -3361,9 +3379,22 @@ var openaiResponsesProviderOptionsSchema = lazySchema17(
|
|
|
3361
3379
|
* Any further attempts to call a tool by the model will be ignored.
|
|
3362
3380
|
*/
|
|
3363
3381
|
maxToolCalls: z19.number().nullish(),
|
|
3382
|
+
/**
|
|
3383
|
+
* Additional metadata to store with the generation.
|
|
3384
|
+
*/
|
|
3364
3385
|
metadata: z19.any().nullish(),
|
|
3386
|
+
/**
|
|
3387
|
+
* Whether to use parallel tool calls. Defaults to `true`.
|
|
3388
|
+
*/
|
|
3365
3389
|
parallelToolCalls: z19.boolean().nullish(),
|
|
3390
|
+
/**
|
|
3391
|
+
* The ID of the previous response. You can use it to continue a conversation.
|
|
3392
|
+
* Defaults to `undefined`.
|
|
3393
|
+
*/
|
|
3366
3394
|
previousResponseId: z19.string().nullish(),
|
|
3395
|
+
/**
|
|
3396
|
+
* Sets a cache key to tie this prompt to cached prefixes for better caching performance.
|
|
3397
|
+
*/
|
|
3367
3398
|
promptCacheKey: z19.string().nullish(),
|
|
3368
3399
|
/**
|
|
3369
3400
|
* The retention policy for the prompt cache.
|
|
@@ -3374,14 +3405,60 @@ var openaiResponsesProviderOptionsSchema = lazySchema17(
|
|
|
3374
3405
|
* @default 'in_memory'
|
|
3375
3406
|
*/
|
|
3376
3407
|
promptCacheRetention: z19.enum(["in_memory", "24h"]).nullish(),
|
|
3408
|
+
/**
|
|
3409
|
+
* Reasoning effort for reasoning models. Defaults to `medium`. If you use
|
|
3410
|
+
* `providerOptions` to set the `reasoningEffort` option, this model setting will be ignored.
|
|
3411
|
+
* Valid values: 'none' | 'minimal' | 'low' | 'medium' | 'high'
|
|
3412
|
+
*
|
|
3413
|
+
* Note: The 'none' type for `reasoningEffort` is only available for OpenAI's GPT-5.1
|
|
3414
|
+
* models. Setting `reasoningEffort` to 'none' with other models will result in
|
|
3415
|
+
* an error.
|
|
3416
|
+
*/
|
|
3377
3417
|
reasoningEffort: z19.string().nullish(),
|
|
3418
|
+
/**
|
|
3419
|
+
* Controls reasoning summary output from the model.
|
|
3420
|
+
* Set to "auto" to automatically receive the richest level available,
|
|
3421
|
+
* or "detailed" for comprehensive summaries.
|
|
3422
|
+
*/
|
|
3378
3423
|
reasoningSummary: z19.string().nullish(),
|
|
3424
|
+
/**
|
|
3425
|
+
* The identifier for safety monitoring and tracking.
|
|
3426
|
+
*/
|
|
3379
3427
|
safetyIdentifier: z19.string().nullish(),
|
|
3428
|
+
/**
|
|
3429
|
+
* Service tier for the request.
|
|
3430
|
+
* Set to 'flex' for 50% cheaper processing at the cost of increased latency (available for o3, o4-mini, and gpt-5 models).
|
|
3431
|
+
* Set to 'priority' for faster processing with Enterprise access (available for gpt-4, gpt-5, gpt-5-mini, o3, o4-mini; gpt-5-nano is not supported).
|
|
3432
|
+
*
|
|
3433
|
+
* Defaults to 'auto'.
|
|
3434
|
+
*/
|
|
3380
3435
|
serviceTier: z19.enum(["auto", "flex", "priority", "default"]).nullish(),
|
|
3436
|
+
/**
|
|
3437
|
+
* Whether to store the generation. Defaults to `true`.
|
|
3438
|
+
*/
|
|
3381
3439
|
store: z19.boolean().nullish(),
|
|
3440
|
+
/**
|
|
3441
|
+
* Whether to use strict JSON schema validation.
|
|
3442
|
+
* Defaults to `true`.
|
|
3443
|
+
*/
|
|
3382
3444
|
strictJsonSchema: z19.boolean().nullish(),
|
|
3445
|
+
/**
|
|
3446
|
+
* Controls the verbosity of the model's responses. Lower values ('low') will result
|
|
3447
|
+
* in more concise responses, while higher values ('high') will result in more verbose responses.
|
|
3448
|
+
* Valid values: 'low', 'medium', 'high'.
|
|
3449
|
+
*/
|
|
3383
3450
|
textVerbosity: z19.enum(["low", "medium", "high"]).nullish(),
|
|
3451
|
+
/**
|
|
3452
|
+
* Controls output truncation. 'auto' (default) performs truncation automatically;
|
|
3453
|
+
* 'disabled' turns truncation off.
|
|
3454
|
+
*/
|
|
3384
3455
|
truncation: z19.enum(["auto", "disabled"]).nullish(),
|
|
3456
|
+
/**
|
|
3457
|
+
* A unique identifier representing your end-user, which can help OpenAI to
|
|
3458
|
+
* monitor and detect abuse.
|
|
3459
|
+
* Defaults to `undefined`.
|
|
3460
|
+
* @see https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids
|
|
3461
|
+
*/
|
|
3385
3462
|
user: z19.string().nullish()
|
|
3386
3463
|
})
|
|
3387
3464
|
)
|
|
@@ -5254,7 +5331,7 @@ var OpenAITranscriptionModel = class {
|
|
|
5254
5331
|
};
|
|
5255
5332
|
|
|
5256
5333
|
// src/version.ts
|
|
5257
|
-
var VERSION = true ? "3.0.0-beta.
|
|
5334
|
+
var VERSION = true ? "3.0.0-beta.78" : "0.0.0-test";
|
|
5258
5335
|
|
|
5259
5336
|
// src/openai-provider.ts
|
|
5260
5337
|
function createOpenAI(options = {}) {
|