@ai-sdk/openai 3.0.0-beta.59 → 3.0.0-beta.60

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,11 @@
1
1
  # @ai-sdk/openai
2
2
 
3
+ ## 3.0.0-beta.60
4
+
5
+ ### Patch Changes
6
+
7
+ - 0877683: feat(provider/openai): support conversations api
8
+
3
9
  ## 3.0.0-beta.59
4
10
 
5
11
  ### Patch Changes
package/dist/index.d.mts CHANGED
@@ -382,6 +382,7 @@ declare const openaiTools: {
382
382
 
383
383
  type OpenAIResponsesModelId = 'chatgpt-4o-latest' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo' | 'gpt-4-0613' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-turbo' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1' | 'gpt-4' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini' | 'gpt-4o' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.1-codex-mini' | 'gpt-5.1-codex' | 'gpt-5-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5-codex' | 'gpt-5-mini-2025-08-07' | 'gpt-5-mini' | 'gpt-5-nano-2025-08-07' | 'gpt-5-nano' | 'gpt-5-pro-2025-10-06' | 'gpt-5-pro' | 'gpt-5' | 'o1-2024-12-17' | 'o1' | 'o3-2025-04-16' | 'o3-mini-2025-01-31' | 'o3-mini' | 'o3' | (string & {});
384
384
  declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
385
+ conversation?: string | null | undefined;
385
386
  include?: ("file_search_call.results" | "message.output_text.logprobs" | "reasoning.encrypted_content")[] | null | undefined;
386
387
  instructions?: string | null | undefined;
387
388
  logprobs?: number | boolean | undefined;
package/dist/index.d.ts CHANGED
@@ -382,6 +382,7 @@ declare const openaiTools: {
382
382
 
383
383
  type OpenAIResponsesModelId = 'chatgpt-4o-latest' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo' | 'gpt-4-0613' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-turbo' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1' | 'gpt-4' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini' | 'gpt-4o' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.1-codex-mini' | 'gpt-5.1-codex' | 'gpt-5-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5-codex' | 'gpt-5-mini-2025-08-07' | 'gpt-5-mini' | 'gpt-5-nano-2025-08-07' | 'gpt-5-nano' | 'gpt-5-pro-2025-10-06' | 'gpt-5-pro' | 'gpt-5' | 'o1-2024-12-17' | 'o1' | 'o3-2025-04-16' | 'o3-mini-2025-01-31' | 'o3-mini' | 'o3' | (string & {});
384
384
  declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
385
+ conversation?: string | null | undefined;
385
386
  include?: ("file_search_call.results" | "message.output_text.logprobs" | "reasoning.encrypted_content")[] | null | undefined;
386
387
  instructions?: string | null | undefined;
387
388
  logprobs?: number | boolean | undefined;
package/dist/index.js CHANGED
@@ -3157,6 +3157,7 @@ var openaiResponsesModelIds = [
3157
3157
  var openaiResponsesProviderOptionsSchema = (0, import_provider_utils23.lazySchema)(
3158
3158
  () => (0, import_provider_utils23.zodSchema)(
3159
3159
  import_v418.z.object({
3160
+ conversation: import_v418.z.string().nullish(),
3160
3161
  include: import_v418.z.array(
3161
3162
  import_v418.z.enum([
3162
3163
  "reasoning.encrypted_content",
@@ -3440,6 +3441,13 @@ var OpenAIResponsesLanguageModel = class {
3440
3441
  providerOptions,
3441
3442
  schema: openaiResponsesProviderOptionsSchema
3442
3443
  });
3444
+ if ((openaiOptions == null ? void 0 : openaiOptions.conversation) && (openaiOptions == null ? void 0 : openaiOptions.previousResponseId)) {
3445
+ warnings.push({
3446
+ type: "unsupported-setting",
3447
+ setting: "conversation",
3448
+ details: "conversation and previousResponseId cannot be used together"
3449
+ });
3450
+ }
3443
3451
  const { input, warnings: inputWarnings } = await convertToOpenAIResponsesInput({
3444
3452
  prompt,
3445
3453
  systemMessageMode: modelConfig.systemMessageMode,
@@ -3502,6 +3510,7 @@ var OpenAIResponsesLanguageModel = class {
3502
3510
  }
3503
3511
  },
3504
3512
  // provider options:
3513
+ conversation: openaiOptions == null ? void 0 : openaiOptions.conversation,
3505
3514
  max_tool_calls: openaiOptions == null ? void 0 : openaiOptions.maxToolCalls,
3506
3515
  metadata: openaiOptions == null ? void 0 : openaiOptions.metadata,
3507
3516
  parallel_tool_calls: openaiOptions == null ? void 0 : openaiOptions.parallelToolCalls,
@@ -4961,7 +4970,7 @@ var OpenAITranscriptionModel = class {
4961
4970
  };
4962
4971
 
4963
4972
  // src/version.ts
4964
- var VERSION = true ? "3.0.0-beta.59" : "0.0.0-test";
4973
+ var VERSION = true ? "3.0.0-beta.60" : "0.0.0-test";
4965
4974
 
4966
4975
  // src/openai-provider.ts
4967
4976
  function createOpenAI(options = {}) {