@ai-sdk/openai 3.0.0-beta.31 → 3.0.0-beta.33

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2142,6 +2142,7 @@ import {
2142
2142
  } from "@ai-sdk/provider";
2143
2143
  import {
2144
2144
  convertToBase64 as convertToBase642,
2145
+ isNonNullable,
2145
2146
  parseProviderOptions as parseProviderOptions6,
2146
2147
  validateTypes
2147
2148
  } from "@ai-sdk/provider-utils";
@@ -2422,22 +2423,33 @@ async function convertToOpenAIResponsesInput({
2422
2423
  break;
2423
2424
  case "content":
2424
2425
  contentValue = output.value.map((item) => {
2426
+ var _a2;
2425
2427
  switch (item.type) {
2426
2428
  case "text": {
2427
2429
  return { type: "input_text", text: item.text };
2428
2430
  }
2429
- case "media": {
2430
- return item.mediaType.startsWith("image/") ? {
2431
+ case "image-data": {
2432
+ return {
2431
2433
  type: "input_image",
2432
2434
  image_url: `data:${item.mediaType};base64,${item.data}`
2433
- } : {
2435
+ };
2436
+ }
2437
+ case "file-data": {
2438
+ return {
2434
2439
  type: "input_file",
2435
- filename: "data",
2440
+ filename: (_a2 = item.filename) != null ? _a2 : "data",
2436
2441
  file_data: `data:${item.mediaType};base64,${item.data}`
2437
2442
  };
2438
2443
  }
2444
+ default: {
2445
+ warnings.push({
2446
+ type: "other",
2447
+ message: `unsupported tool content part type: ${item.type}`
2448
+ });
2449
+ return void 0;
2450
+ }
2439
2451
  }
2440
- });
2452
+ }).filter(isNonNullable);
2441
2453
  break;
2442
2454
  }
2443
2455
  input.push({
@@ -3005,6 +3017,7 @@ var openaiResponsesProviderOptionsSchema = lazySchema13(
3005
3017
  store: z15.boolean().nullish(),
3006
3018
  strictJsonSchema: z15.boolean().nullish(),
3007
3019
  textVerbosity: z15.enum(["low", "medium", "high"]).nullish(),
3020
+ truncation: z15.enum(["auto", "disabled"]).nullish(),
3008
3021
  user: z15.string().nullish()
3009
3022
  })
3010
3023
  )
@@ -3533,6 +3546,7 @@ var OpenAIResponsesLanguageModel = class {
3533
3546
  prompt_cache_key: openaiOptions == null ? void 0 : openaiOptions.promptCacheKey,
3534
3547
  safety_identifier: openaiOptions == null ? void 0 : openaiOptions.safetyIdentifier,
3535
3548
  top_logprobs: topLogprobs,
3549
+ truncation: openaiOptions == null ? void 0 : openaiOptions.truncation,
3536
3550
  // model-specific settings:
3537
3551
  ...modelConfig.isReasoningModel && ((openaiOptions == null ? void 0 : openaiOptions.reasoningEffort) != null || (openaiOptions == null ? void 0 : openaiOptions.reasoningSummary) != null) && {
3538
3552
  reasoning: {
@@ -3543,9 +3557,6 @@ var OpenAIResponsesLanguageModel = class {
3543
3557
  summary: openaiOptions.reasoningSummary
3544
3558
  }
3545
3559
  }
3546
- },
3547
- ...modelConfig.requiredAutoTruncation && {
3548
- truncation: "auto"
3549
3560
  }
3550
3561
  };
3551
3562
  if (modelConfig.isReasoningModel) {
@@ -4401,7 +4412,6 @@ function getResponsesModelConfig(modelId) {
4401
4412
  const supportsFlexProcessing2 = modelId.startsWith("o3") || modelId.startsWith("o4-mini") || modelId.startsWith("gpt-5") && !modelId.startsWith("gpt-5-chat");
4402
4413
  const supportsPriorityProcessing2 = modelId.startsWith("gpt-4") || modelId.startsWith("gpt-5-mini") || modelId.startsWith("gpt-5") && !modelId.startsWith("gpt-5-nano") && !modelId.startsWith("gpt-5-chat") || modelId.startsWith("o3") || modelId.startsWith("o4-mini");
4403
4414
  const defaults = {
4404
- requiredAutoTruncation: false,
4405
4415
  systemMessageMode: "system",
4406
4416
  supportsFlexProcessing: supportsFlexProcessing2,
4407
4417
  supportsPriorityProcessing: supportsPriorityProcessing2