@ai-sdk/openai 3.0.0-beta.77 → 3.0.0-beta.78

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3264,7 +3264,18 @@ var openaiResponsesModelIds = [
3264
3264
  var openaiResponsesProviderOptionsSchema = lazySchema14(
3265
3265
  () => zodSchema14(
3266
3266
  z16.object({
3267
+ /**
3268
+ * The ID of the OpenAI Conversation to continue.
3269
+ * You must create a conversation first via the OpenAI API.
3270
+ * Cannot be used in conjunction with `previousResponseId`.
3271
+ * Defaults to `undefined`.
3272
+ * @see https://platform.openai.com/docs/api-reference/conversations/create
3273
+ */
3267
3274
  conversation: z16.string().nullish(),
3275
+ /**
3276
+ * The set of extra fields to include in the response (advanced, usually not needed).
3277
+ * Example values: 'reasoning.encrypted_content', 'file_search_call.results', 'message.output_text.logprobs'.
3278
+ */
3268
3279
  include: z16.array(
3269
3280
  z16.enum([
3270
3281
  "reasoning.encrypted_content",
@@ -3273,9 +3284,16 @@ var openaiResponsesProviderOptionsSchema = lazySchema14(
3273
3284
  "message.output_text.logprobs"
3274
3285
  ])
3275
3286
  ).nullish(),
3287
+ /**
3288
+ * Instructions for the model.
3289
+ * They can be used to change the system or developer message when continuing a conversation using the `previousResponseId` option.
3290
+ * Defaults to `undefined`.
3291
+ */
3276
3292
  instructions: z16.string().nullish(),
3277
3293
  /**
3278
- * Return the log probabilities of the tokens.
3294
+ * Return the log probabilities of the tokens. Including logprobs will increase
3295
+ * the response size and can slow down response times. However, it can
3296
+ * be useful to better understand how the model is behaving.
3279
3297
  *
3280
3298
  * Setting to true will return the log probabilities of the tokens that
3281
3299
  * were generated.
@@ -3293,9 +3311,22 @@ var openaiResponsesProviderOptionsSchema = lazySchema14(
3293
3311
  * Any further attempts to call a tool by the model will be ignored.
3294
3312
  */
3295
3313
  maxToolCalls: z16.number().nullish(),
3314
+ /**
3315
+ * Additional metadata to store with the generation.
3316
+ */
3296
3317
  metadata: z16.any().nullish(),
3318
+ /**
3319
+ * Whether to use parallel tool calls. Defaults to `true`.
3320
+ */
3297
3321
  parallelToolCalls: z16.boolean().nullish(),
3322
+ /**
3323
+ * The ID of the previous response. You can use it to continue a conversation.
3324
+ * Defaults to `undefined`.
3325
+ */
3298
3326
  previousResponseId: z16.string().nullish(),
3327
+ /**
3328
+ * Sets a cache key to tie this prompt to cached prefixes for better caching performance.
3329
+ */
3299
3330
  promptCacheKey: z16.string().nullish(),
3300
3331
  /**
3301
3332
  * The retention policy for the prompt cache.
@@ -3306,14 +3337,60 @@ var openaiResponsesProviderOptionsSchema = lazySchema14(
3306
3337
  * @default 'in_memory'
3307
3338
  */
3308
3339
  promptCacheRetention: z16.enum(["in_memory", "24h"]).nullish(),
3340
+ /**
3341
+ * Reasoning effort for reasoning models. Defaults to `medium`. If you use
3342
+ * `providerOptions` to set the `reasoningEffort` option, this model setting will be ignored.
3343
+ * Valid values: 'none' | 'minimal' | 'low' | 'medium' | 'high'
3344
+ *
3345
+ * Note: The 'none' type for `reasoningEffort` is only available for OpenAI's GPT-5.1
3346
+ * models. Setting `reasoningEffort` to 'none' with other models will result in
3347
+ * an error.
3348
+ */
3309
3349
  reasoningEffort: z16.string().nullish(),
3350
+ /**
3351
+ * Controls reasoning summary output from the model.
3352
+ * Set to "auto" to automatically receive the richest level available,
3353
+ * or "detailed" for comprehensive summaries.
3354
+ */
3310
3355
  reasoningSummary: z16.string().nullish(),
3356
+ /**
3357
+ * The identifier for safety monitoring and tracking.
3358
+ */
3311
3359
  safetyIdentifier: z16.string().nullish(),
3360
+ /**
3361
+ * Service tier for the request.
3362
+ * Set to 'flex' for 50% cheaper processing at the cost of increased latency (available for o3, o4-mini, and gpt-5 models).
3363
+ * Set to 'priority' for faster processing with Enterprise access (available for gpt-4, gpt-5, gpt-5-mini, o3, o4-mini; gpt-5-nano is not supported).
3364
+ *
3365
+ * Defaults to 'auto'.
3366
+ */
3312
3367
  serviceTier: z16.enum(["auto", "flex", "priority", "default"]).nullish(),
3368
+ /**
3369
+ * Whether to store the generation. Defaults to `true`.
3370
+ */
3313
3371
  store: z16.boolean().nullish(),
3372
+ /**
3373
+ * Whether to use strict JSON schema validation.
3374
+ * Defaults to `true`.
3375
+ */
3314
3376
  strictJsonSchema: z16.boolean().nullish(),
3377
+ /**
3378
+ * Controls the verbosity of the model's responses. Lower values ('low') will result
3379
+ * in more concise responses, while higher values ('high') will result in more verbose responses.
3380
+ * Valid values: 'low', 'medium', 'high'.
3381
+ */
3315
3382
  textVerbosity: z16.enum(["low", "medium", "high"]).nullish(),
3383
+ /**
3384
+ * Controls output truncation. 'auto' (default) performs truncation automatically;
3385
+ * 'disabled' turns truncation off.
3386
+ */
3316
3387
  truncation: z16.enum(["auto", "disabled"]).nullish(),
3388
+ /**
3389
+ * A unique identifier representing your end-user, which can help OpenAI to
3390
+ * monitor and detect abuse.
3391
+ * Defaults to `undefined`.
3392
+ * @see https://platform.openai.com/docs/guides/safety-best-practices/end-user-ids
3393
+ */
3317
3394
  user: z16.string().nullish()
3318
3395
  })
3319
3396
  )