@ai-sdk/openai 3.0.0-beta.32 → 3.0.0-beta.33
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +6 -0
- package/dist/index.d.mts +1 -0
- package/dist/index.d.ts +1 -0
- package/dist/index.js +3 -5
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +3 -5
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.js +2 -4
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +2 -4
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +1 -1
package/dist/internal/index.js
CHANGED
|
@@ -3001,6 +3001,7 @@ var openaiResponsesProviderOptionsSchema = (0, import_provider_utils22.lazySchem
|
|
|
3001
3001
|
store: import_v415.z.boolean().nullish(),
|
|
3002
3002
|
strictJsonSchema: import_v415.z.boolean().nullish(),
|
|
3003
3003
|
textVerbosity: import_v415.z.enum(["low", "medium", "high"]).nullish(),
|
|
3004
|
+
truncation: import_v415.z.enum(["auto", "disabled"]).nullish(),
|
|
3004
3005
|
user: import_v415.z.string().nullish()
|
|
3005
3006
|
})
|
|
3006
3007
|
)
|
|
@@ -3507,6 +3508,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
3507
3508
|
prompt_cache_key: openaiOptions == null ? void 0 : openaiOptions.promptCacheKey,
|
|
3508
3509
|
safety_identifier: openaiOptions == null ? void 0 : openaiOptions.safetyIdentifier,
|
|
3509
3510
|
top_logprobs: topLogprobs,
|
|
3511
|
+
truncation: openaiOptions == null ? void 0 : openaiOptions.truncation,
|
|
3510
3512
|
// model-specific settings:
|
|
3511
3513
|
...modelConfig.isReasoningModel && ((openaiOptions == null ? void 0 : openaiOptions.reasoningEffort) != null || (openaiOptions == null ? void 0 : openaiOptions.reasoningSummary) != null) && {
|
|
3512
3514
|
reasoning: {
|
|
@@ -3517,9 +3519,6 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
3517
3519
|
summary: openaiOptions.reasoningSummary
|
|
3518
3520
|
}
|
|
3519
3521
|
}
|
|
3520
|
-
},
|
|
3521
|
-
...modelConfig.requiredAutoTruncation && {
|
|
3522
|
-
truncation: "auto"
|
|
3523
3522
|
}
|
|
3524
3523
|
};
|
|
3525
3524
|
if (modelConfig.isReasoningModel) {
|
|
@@ -4375,7 +4374,6 @@ function getResponsesModelConfig(modelId) {
|
|
|
4375
4374
|
const supportsFlexProcessing2 = modelId.startsWith("o3") || modelId.startsWith("o4-mini") || modelId.startsWith("gpt-5") && !modelId.startsWith("gpt-5-chat");
|
|
4376
4375
|
const supportsPriorityProcessing2 = modelId.startsWith("gpt-4") || modelId.startsWith("gpt-5-mini") || modelId.startsWith("gpt-5") && !modelId.startsWith("gpt-5-nano") && !modelId.startsWith("gpt-5-chat") || modelId.startsWith("o3") || modelId.startsWith("o4-mini");
|
|
4377
4376
|
const defaults = {
|
|
4378
|
-
requiredAutoTruncation: false,
|
|
4379
4377
|
systemMessageMode: "system",
|
|
4380
4378
|
supportsFlexProcessing: supportsFlexProcessing2,
|
|
4381
4379
|
supportsPriorityProcessing: supportsPriorityProcessing2
|