@ai-sdk/openai 3.0.0-beta.32 → 3.0.0-beta.33

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,11 @@
1
1
  # @ai-sdk/openai
2
2
 
3
+ ## 3.0.0-beta.33
4
+
5
+ ### Patch Changes
6
+
7
+ - 7756857: fix(provider/openai): add truncation parameter support for Responses API
8
+
3
9
  ## 3.0.0-beta.32
4
10
 
5
11
  ### Patch Changes
package/dist/index.d.mts CHANGED
@@ -328,6 +328,7 @@ declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazyS
328
328
  store?: boolean | null | undefined;
329
329
  strictJsonSchema?: boolean | null | undefined;
330
330
  textVerbosity?: "low" | "medium" | "high" | null | undefined;
331
+ truncation?: "auto" | "disabled" | null | undefined;
331
332
  user?: string | null | undefined;
332
333
  }>;
333
334
  type OpenAIResponsesProviderOptions = InferSchema<typeof openaiResponsesProviderOptionsSchema>;
package/dist/index.d.ts CHANGED
@@ -328,6 +328,7 @@ declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazyS
328
328
  store?: boolean | null | undefined;
329
329
  strictJsonSchema?: boolean | null | undefined;
330
330
  textVerbosity?: "low" | "medium" | "high" | null | undefined;
331
+ truncation?: "auto" | "disabled" | null | undefined;
331
332
  user?: string | null | undefined;
332
333
  }>;
333
334
  type OpenAIResponsesProviderOptions = InferSchema<typeof openaiResponsesProviderOptionsSchema>;
package/dist/index.js CHANGED
@@ -2933,6 +2933,7 @@ var openaiResponsesProviderOptionsSchema = (0, import_provider_utils22.lazySchem
2933
2933
  store: import_v417.z.boolean().nullish(),
2934
2934
  strictJsonSchema: import_v417.z.boolean().nullish(),
2935
2935
  textVerbosity: import_v417.z.enum(["low", "medium", "high"]).nullish(),
2936
+ truncation: import_v417.z.enum(["auto", "disabled"]).nullish(),
2936
2937
  user: import_v417.z.string().nullish()
2937
2938
  })
2938
2939
  )
@@ -3211,6 +3212,7 @@ var OpenAIResponsesLanguageModel = class {
3211
3212
  prompt_cache_key: openaiOptions == null ? void 0 : openaiOptions.promptCacheKey,
3212
3213
  safety_identifier: openaiOptions == null ? void 0 : openaiOptions.safetyIdentifier,
3213
3214
  top_logprobs: topLogprobs,
3215
+ truncation: openaiOptions == null ? void 0 : openaiOptions.truncation,
3214
3216
  // model-specific settings:
3215
3217
  ...modelConfig.isReasoningModel && ((openaiOptions == null ? void 0 : openaiOptions.reasoningEffort) != null || (openaiOptions == null ? void 0 : openaiOptions.reasoningSummary) != null) && {
3216
3218
  reasoning: {
@@ -3221,9 +3223,6 @@ var OpenAIResponsesLanguageModel = class {
3221
3223
  summary: openaiOptions.reasoningSummary
3222
3224
  }
3223
3225
  }
3224
- },
3225
- ...modelConfig.requiredAutoTruncation && {
3226
- truncation: "auto"
3227
3226
  }
3228
3227
  };
3229
3228
  if (modelConfig.isReasoningModel) {
@@ -4079,7 +4078,6 @@ function getResponsesModelConfig(modelId) {
4079
4078
  const supportsFlexProcessing2 = modelId.startsWith("o3") || modelId.startsWith("o4-mini") || modelId.startsWith("gpt-5") && !modelId.startsWith("gpt-5-chat");
4080
4079
  const supportsPriorityProcessing2 = modelId.startsWith("gpt-4") || modelId.startsWith("gpt-5-mini") || modelId.startsWith("gpt-5") && !modelId.startsWith("gpt-5-nano") && !modelId.startsWith("gpt-5-chat") || modelId.startsWith("o3") || modelId.startsWith("o4-mini");
4081
4080
  const defaults = {
4082
- requiredAutoTruncation: false,
4083
4081
  systemMessageMode: "system",
4084
4082
  supportsFlexProcessing: supportsFlexProcessing2,
4085
4083
  supportsPriorityProcessing: supportsPriorityProcessing2
@@ -4476,7 +4474,7 @@ var OpenAITranscriptionModel = class {
4476
4474
  };
4477
4475
 
4478
4476
  // src/version.ts
4479
- var VERSION = true ? "3.0.0-beta.32" : "0.0.0-test";
4477
+ var VERSION = true ? "3.0.0-beta.33" : "0.0.0-test";
4480
4478
 
4481
4479
  // src/openai-provider.ts
4482
4480
  function createOpenAI(options = {}) {