@ai-sdk/openai 3.0.25 → 3.0.27

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3,7 +3,7 @@ import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
3
3
  import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
4
4
 
5
5
  type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'o4-mini' | 'o4-mini-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.2' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | (string & {});
6
- declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<{
6
+ declare const openaiLanguageModelChatOptions: _ai_sdk_provider_utils.LazySchema<{
7
7
  logitBias?: Record<number, number> | undefined;
8
8
  logprobs?: number | boolean | undefined;
9
9
  parallelToolCalls?: boolean | undefined;
@@ -22,7 +22,7 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<
22
22
  systemMessageMode?: "remove" | "system" | "developer" | undefined;
23
23
  forceReasoning?: boolean | undefined;
24
24
  }>;
25
- type OpenAIChatLanguageModelOptions = InferSchema<typeof openaiChatLanguageModelOptions>;
25
+ type OpenAILanguageModelChatOptions = InferSchema<typeof openaiLanguageModelChatOptions>;
26
26
 
27
27
  type OpenAIChatConfig = {
28
28
  provider: string;
@@ -48,14 +48,14 @@ declare class OpenAIChatLanguageModel implements LanguageModelV3 {
48
48
  }
49
49
 
50
50
  type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
51
- declare const openaiCompletionProviderOptions: _ai_sdk_provider_utils.LazySchema<{
51
+ declare const openaiLanguageModelCompletionOptions: _ai_sdk_provider_utils.LazySchema<{
52
52
  echo?: boolean | undefined;
53
53
  logitBias?: Record<string, number> | undefined;
54
54
  suffix?: string | undefined;
55
55
  user?: string | undefined;
56
56
  logprobs?: number | boolean | undefined;
57
57
  }>;
58
- type OpenAICompletionProviderOptions = InferSchema<typeof openaiCompletionProviderOptions>;
58
+ type OpenAILanguageModelCompletionOptions = InferSchema<typeof openaiLanguageModelCompletionOptions>;
59
59
 
60
60
  type OpenAICompletionConfig = {
61
61
  provider: string;
@@ -100,11 +100,11 @@ type OpenAIConfig = {
100
100
  };
101
101
 
102
102
  type OpenAIEmbeddingModelId = 'text-embedding-3-small' | 'text-embedding-3-large' | 'text-embedding-ada-002' | (string & {});
103
- declare const openaiEmbeddingProviderOptions: _ai_sdk_provider_utils.LazySchema<{
103
+ declare const openaiEmbeddingModelOptions: _ai_sdk_provider_utils.LazySchema<{
104
104
  dimensions?: number | undefined;
105
105
  user?: string | undefined;
106
106
  }>;
107
- type OpenAIEmbeddingProviderOptions = InferSchema<typeof openaiEmbeddingProviderOptions>;
107
+ type OpenAIEmbeddingModelOptions = InferSchema<typeof openaiEmbeddingModelOptions>;
108
108
 
109
109
  declare class OpenAIEmbeddingModel implements EmbeddingModelV3 {
110
110
  readonly specificationVersion = "v3";
@@ -137,18 +137,18 @@ declare class OpenAIImageModel implements ImageModelV3 {
137
137
  }
138
138
 
139
139
  type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
140
- declare const openAITranscriptionProviderOptions: _ai_sdk_provider_utils.LazySchema<{
140
+ declare const openAITranscriptionModelOptions: _ai_sdk_provider_utils.LazySchema<{
141
141
  include?: string[] | undefined;
142
142
  language?: string | undefined;
143
143
  prompt?: string | undefined;
144
144
  temperature?: number | undefined;
145
145
  timestampGranularities?: ("word" | "segment")[] | undefined;
146
146
  }>;
147
- type OpenAITranscriptionProviderOptions = InferSchema<typeof openAITranscriptionProviderOptions>;
147
+ type OpenAITranscriptionModelOptions = InferSchema<typeof openAITranscriptionModelOptions>;
148
148
 
149
149
  type OpenAITranscriptionCallOptions = Omit<TranscriptionModelV3CallOptions, 'providerOptions'> & {
150
150
  providerOptions?: {
151
- openai?: OpenAITranscriptionProviderOptions;
151
+ openai?: OpenAITranscriptionModelOptions;
152
152
  };
153
153
  };
154
154
  interface OpenAITranscriptionModelConfig extends OpenAIConfig {
@@ -167,11 +167,11 @@ declare class OpenAITranscriptionModel implements TranscriptionModelV3 {
167
167
  }
168
168
 
169
169
  type OpenAISpeechModelId = 'tts-1' | 'tts-1-hd' | 'gpt-4o-mini-tts' | (string & {});
170
- declare const openaiSpeechProviderOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
170
+ declare const openaiSpeechModelOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
171
171
  instructions?: string | null | undefined;
172
172
  speed?: number | null | undefined;
173
173
  }>;
174
- type OpenAISpeechCallOptions = InferSchema<typeof openaiSpeechProviderOptionsSchema>;
174
+ type OpenAISpeechModelOptions = InferSchema<typeof openaiSpeechModelOptionsSchema>;
175
175
 
176
176
  interface OpenAISpeechModelConfig extends OpenAIConfig {
177
177
  _internal?: {
@@ -1083,4 +1083,4 @@ declare const webSearchPreview: _ai_sdk_provider_utils.ProviderToolFactoryWithOu
1083
1083
  };
1084
1084
  }>;
1085
1085
 
1086
- export { type ApplyPatchOperation, OpenAIChatLanguageModel, type OpenAIChatLanguageModelOptions, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, OpenAIResponsesLanguageModel, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, type ResponsesProviderMetadata, type ResponsesReasoningProviderMetadata, type ResponsesSourceDocumentProviderMetadata, type ResponsesTextProviderMetadata, applyPatch, applyPatchArgsSchema, applyPatchInputSchema, applyPatchOutputSchema, applyPatchToolFactory, codeInterpreter, codeInterpreterArgsSchema, codeInterpreterInputSchema, codeInterpreterOutputSchema, codeInterpreterToolFactory, fileSearch, fileSearchArgsSchema, fileSearchOutputSchema, hasDefaultResponseFormat, imageGeneration, imageGenerationArgsSchema, imageGenerationOutputSchema, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiChatLanguageModelOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiSpeechProviderOptionsSchema, webSearchPreview, webSearchPreviewArgsSchema, webSearchPreviewInputSchema };
1086
+ export { type ApplyPatchOperation, OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingModelOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAILanguageModelChatOptions, type OpenAILanguageModelCompletionOptions, OpenAIResponsesLanguageModel, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAISpeechModelOptions, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionModelOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, type ResponsesProviderMetadata, type ResponsesReasoningProviderMetadata, type ResponsesSourceDocumentProviderMetadata, type ResponsesTextProviderMetadata, applyPatch, applyPatchArgsSchema, applyPatchInputSchema, applyPatchOutputSchema, applyPatchToolFactory, codeInterpreter, codeInterpreterArgsSchema, codeInterpreterInputSchema, codeInterpreterOutputSchema, codeInterpreterToolFactory, fileSearch, fileSearchArgsSchema, fileSearchOutputSchema, hasDefaultResponseFormat, imageGeneration, imageGenerationArgsSchema, imageGenerationOutputSchema, modelMaxImagesPerCall, openAITranscriptionModelOptions, openaiEmbeddingModelOptions, openaiLanguageModelChatOptions, openaiLanguageModelCompletionOptions, openaiSpeechModelOptionsSchema, webSearchPreview, webSearchPreviewArgsSchema, webSearchPreviewInputSchema };
@@ -3,7 +3,7 @@ import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
3
3
  import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
4
4
 
5
5
  type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'o4-mini' | 'o4-mini-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.2' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | (string & {});
6
- declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<{
6
+ declare const openaiLanguageModelChatOptions: _ai_sdk_provider_utils.LazySchema<{
7
7
  logitBias?: Record<number, number> | undefined;
8
8
  logprobs?: number | boolean | undefined;
9
9
  parallelToolCalls?: boolean | undefined;
@@ -22,7 +22,7 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<
22
22
  systemMessageMode?: "remove" | "system" | "developer" | undefined;
23
23
  forceReasoning?: boolean | undefined;
24
24
  }>;
25
- type OpenAIChatLanguageModelOptions = InferSchema<typeof openaiChatLanguageModelOptions>;
25
+ type OpenAILanguageModelChatOptions = InferSchema<typeof openaiLanguageModelChatOptions>;
26
26
 
27
27
  type OpenAIChatConfig = {
28
28
  provider: string;
@@ -48,14 +48,14 @@ declare class OpenAIChatLanguageModel implements LanguageModelV3 {
48
48
  }
49
49
 
50
50
  type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
51
- declare const openaiCompletionProviderOptions: _ai_sdk_provider_utils.LazySchema<{
51
+ declare const openaiLanguageModelCompletionOptions: _ai_sdk_provider_utils.LazySchema<{
52
52
  echo?: boolean | undefined;
53
53
  logitBias?: Record<string, number> | undefined;
54
54
  suffix?: string | undefined;
55
55
  user?: string | undefined;
56
56
  logprobs?: number | boolean | undefined;
57
57
  }>;
58
- type OpenAICompletionProviderOptions = InferSchema<typeof openaiCompletionProviderOptions>;
58
+ type OpenAILanguageModelCompletionOptions = InferSchema<typeof openaiLanguageModelCompletionOptions>;
59
59
 
60
60
  type OpenAICompletionConfig = {
61
61
  provider: string;
@@ -100,11 +100,11 @@ type OpenAIConfig = {
100
100
  };
101
101
 
102
102
  type OpenAIEmbeddingModelId = 'text-embedding-3-small' | 'text-embedding-3-large' | 'text-embedding-ada-002' | (string & {});
103
- declare const openaiEmbeddingProviderOptions: _ai_sdk_provider_utils.LazySchema<{
103
+ declare const openaiEmbeddingModelOptions: _ai_sdk_provider_utils.LazySchema<{
104
104
  dimensions?: number | undefined;
105
105
  user?: string | undefined;
106
106
  }>;
107
- type OpenAIEmbeddingProviderOptions = InferSchema<typeof openaiEmbeddingProviderOptions>;
107
+ type OpenAIEmbeddingModelOptions = InferSchema<typeof openaiEmbeddingModelOptions>;
108
108
 
109
109
  declare class OpenAIEmbeddingModel implements EmbeddingModelV3 {
110
110
  readonly specificationVersion = "v3";
@@ -137,18 +137,18 @@ declare class OpenAIImageModel implements ImageModelV3 {
137
137
  }
138
138
 
139
139
  type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
140
- declare const openAITranscriptionProviderOptions: _ai_sdk_provider_utils.LazySchema<{
140
+ declare const openAITranscriptionModelOptions: _ai_sdk_provider_utils.LazySchema<{
141
141
  include?: string[] | undefined;
142
142
  language?: string | undefined;
143
143
  prompt?: string | undefined;
144
144
  temperature?: number | undefined;
145
145
  timestampGranularities?: ("word" | "segment")[] | undefined;
146
146
  }>;
147
- type OpenAITranscriptionProviderOptions = InferSchema<typeof openAITranscriptionProviderOptions>;
147
+ type OpenAITranscriptionModelOptions = InferSchema<typeof openAITranscriptionModelOptions>;
148
148
 
149
149
  type OpenAITranscriptionCallOptions = Omit<TranscriptionModelV3CallOptions, 'providerOptions'> & {
150
150
  providerOptions?: {
151
- openai?: OpenAITranscriptionProviderOptions;
151
+ openai?: OpenAITranscriptionModelOptions;
152
152
  };
153
153
  };
154
154
  interface OpenAITranscriptionModelConfig extends OpenAIConfig {
@@ -167,11 +167,11 @@ declare class OpenAITranscriptionModel implements TranscriptionModelV3 {
167
167
  }
168
168
 
169
169
  type OpenAISpeechModelId = 'tts-1' | 'tts-1-hd' | 'gpt-4o-mini-tts' | (string & {});
170
- declare const openaiSpeechProviderOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
170
+ declare const openaiSpeechModelOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
171
171
  instructions?: string | null | undefined;
172
172
  speed?: number | null | undefined;
173
173
  }>;
174
- type OpenAISpeechCallOptions = InferSchema<typeof openaiSpeechProviderOptionsSchema>;
174
+ type OpenAISpeechModelOptions = InferSchema<typeof openaiSpeechModelOptionsSchema>;
175
175
 
176
176
  interface OpenAISpeechModelConfig extends OpenAIConfig {
177
177
  _internal?: {
@@ -1083,4 +1083,4 @@ declare const webSearchPreview: _ai_sdk_provider_utils.ProviderToolFactoryWithOu
1083
1083
  };
1084
1084
  }>;
1085
1085
 
1086
- export { type ApplyPatchOperation, OpenAIChatLanguageModel, type OpenAIChatLanguageModelOptions, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, OpenAIResponsesLanguageModel, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, type ResponsesProviderMetadata, type ResponsesReasoningProviderMetadata, type ResponsesSourceDocumentProviderMetadata, type ResponsesTextProviderMetadata, applyPatch, applyPatchArgsSchema, applyPatchInputSchema, applyPatchOutputSchema, applyPatchToolFactory, codeInterpreter, codeInterpreterArgsSchema, codeInterpreterInputSchema, codeInterpreterOutputSchema, codeInterpreterToolFactory, fileSearch, fileSearchArgsSchema, fileSearchOutputSchema, hasDefaultResponseFormat, imageGeneration, imageGenerationArgsSchema, imageGenerationOutputSchema, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiChatLanguageModelOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiSpeechProviderOptionsSchema, webSearchPreview, webSearchPreviewArgsSchema, webSearchPreviewInputSchema };
1086
+ export { type ApplyPatchOperation, OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingModelOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAILanguageModelChatOptions, type OpenAILanguageModelCompletionOptions, OpenAIResponsesLanguageModel, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAISpeechModelOptions, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionModelOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, type ResponsesProviderMetadata, type ResponsesReasoningProviderMetadata, type ResponsesSourceDocumentProviderMetadata, type ResponsesTextProviderMetadata, applyPatch, applyPatchArgsSchema, applyPatchInputSchema, applyPatchOutputSchema, applyPatchToolFactory, codeInterpreter, codeInterpreterArgsSchema, codeInterpreterInputSchema, codeInterpreterOutputSchema, codeInterpreterToolFactory, fileSearch, fileSearchArgsSchema, fileSearchOutputSchema, hasDefaultResponseFormat, imageGeneration, imageGenerationArgsSchema, imageGenerationOutputSchema, modelMaxImagesPerCall, openAITranscriptionModelOptions, openaiEmbeddingModelOptions, openaiLanguageModelChatOptions, openaiLanguageModelCompletionOptions, openaiSpeechModelOptionsSchema, webSearchPreview, webSearchPreviewArgsSchema, webSearchPreviewInputSchema };
@@ -45,11 +45,11 @@ __export(internal_exports, {
45
45
  imageGenerationArgsSchema: () => imageGenerationArgsSchema,
46
46
  imageGenerationOutputSchema: () => imageGenerationOutputSchema,
47
47
  modelMaxImagesPerCall: () => modelMaxImagesPerCall,
48
- openAITranscriptionProviderOptions: () => openAITranscriptionProviderOptions,
49
- openaiChatLanguageModelOptions: () => openaiChatLanguageModelOptions,
50
- openaiCompletionProviderOptions: () => openaiCompletionProviderOptions,
51
- openaiEmbeddingProviderOptions: () => openaiEmbeddingProviderOptions,
52
- openaiSpeechProviderOptionsSchema: () => openaiSpeechProviderOptionsSchema,
48
+ openAITranscriptionModelOptions: () => openAITranscriptionModelOptions,
49
+ openaiEmbeddingModelOptions: () => openaiEmbeddingModelOptions,
50
+ openaiLanguageModelChatOptions: () => openaiLanguageModelChatOptions,
51
+ openaiLanguageModelCompletionOptions: () => openaiLanguageModelCompletionOptions,
52
+ openaiSpeechModelOptionsSchema: () => openaiSpeechModelOptionsSchema,
53
53
  webSearchPreview: () => webSearchPreview,
54
54
  webSearchPreviewArgsSchema: () => webSearchPreviewArgsSchema,
55
55
  webSearchPreviewInputSchema: () => webSearchPreviewInputSchema
@@ -494,7 +494,7 @@ var openaiChatChunkSchema = (0, import_provider_utils3.lazySchema)(
494
494
  // src/chat/openai-chat-options.ts
495
495
  var import_provider_utils4 = require("@ai-sdk/provider-utils");
496
496
  var import_v43 = require("zod/v4");
497
- var openaiChatLanguageModelOptions = (0, import_provider_utils4.lazySchema)(
497
+ var openaiLanguageModelChatOptions = (0, import_provider_utils4.lazySchema)(
498
498
  () => (0, import_provider_utils4.zodSchema)(
499
499
  import_v43.z.object({
500
500
  /**
@@ -705,7 +705,7 @@ var OpenAIChatLanguageModel = class {
705
705
  const openaiOptions = (_a = await (0, import_provider_utils5.parseProviderOptions)({
706
706
  provider: "openai",
707
707
  providerOptions,
708
- schema: openaiChatLanguageModelOptions
708
+ schema: openaiLanguageModelChatOptions
709
709
  })) != null ? _a : {};
710
710
  const modelCapabilities = getOpenAILanguageModelCapabilities(this.modelId);
711
711
  const isReasoningModel = (_b = openaiOptions.forceReasoning) != null ? _b : modelCapabilities.isReasoningModel;
@@ -1362,7 +1362,7 @@ var openaiCompletionChunkSchema = (0, import_provider_utils6.lazySchema)(
1362
1362
  // src/completion/openai-completion-options.ts
1363
1363
  var import_provider_utils7 = require("@ai-sdk/provider-utils");
1364
1364
  var import_v45 = require("zod/v4");
1365
- var openaiCompletionProviderOptions = (0, import_provider_utils7.lazySchema)(
1365
+ var openaiLanguageModelCompletionOptions = (0, import_provider_utils7.lazySchema)(
1366
1366
  () => (0, import_provider_utils7.zodSchema)(
1367
1367
  import_v45.z.object({
1368
1368
  /**
@@ -1443,12 +1443,12 @@ var OpenAICompletionLanguageModel = class {
1443
1443
  ...await (0, import_provider_utils8.parseProviderOptions)({
1444
1444
  provider: "openai",
1445
1445
  providerOptions,
1446
- schema: openaiCompletionProviderOptions
1446
+ schema: openaiLanguageModelCompletionOptions
1447
1447
  }),
1448
1448
  ...await (0, import_provider_utils8.parseProviderOptions)({
1449
1449
  provider: this.providerOptionsName,
1450
1450
  providerOptions,
1451
- schema: openaiCompletionProviderOptions
1451
+ schema: openaiLanguageModelCompletionOptions
1452
1452
  })
1453
1453
  };
1454
1454
  if (topK != null) {
@@ -1643,7 +1643,7 @@ var import_provider_utils11 = require("@ai-sdk/provider-utils");
1643
1643
  // src/embedding/openai-embedding-options.ts
1644
1644
  var import_provider_utils9 = require("@ai-sdk/provider-utils");
1645
1645
  var import_v46 = require("zod/v4");
1646
- var openaiEmbeddingProviderOptions = (0, import_provider_utils9.lazySchema)(
1646
+ var openaiEmbeddingModelOptions = (0, import_provider_utils9.lazySchema)(
1647
1647
  () => (0, import_provider_utils9.zodSchema)(
1648
1648
  import_v46.z.object({
1649
1649
  /**
@@ -1702,7 +1702,7 @@ var OpenAIEmbeddingModel = class {
1702
1702
  const openaiOptions = (_a = await (0, import_provider_utils11.parseProviderOptions)({
1703
1703
  provider: "openai",
1704
1704
  providerOptions,
1705
- schema: openaiEmbeddingProviderOptions
1705
+ schema: openaiEmbeddingModelOptions
1706
1706
  })) != null ? _a : {};
1707
1707
  const {
1708
1708
  responseHeaders,
@@ -1995,7 +1995,7 @@ var openaiTranscriptionResponseSchema = (0, import_provider_utils14.lazySchema)(
1995
1995
  // src/transcription/openai-transcription-options.ts
1996
1996
  var import_provider_utils15 = require("@ai-sdk/provider-utils");
1997
1997
  var import_v410 = require("zod/v4");
1998
- var openAITranscriptionProviderOptions = (0, import_provider_utils15.lazySchema)(
1998
+ var openAITranscriptionModelOptions = (0, import_provider_utils15.lazySchema)(
1999
1999
  () => (0, import_provider_utils15.zodSchema)(
2000
2000
  import_v410.z.object({
2001
2001
  /**
@@ -2102,7 +2102,7 @@ var OpenAITranscriptionModel = class {
2102
2102
  const openAIOptions = await (0, import_provider_utils16.parseProviderOptions)({
2103
2103
  provider: "openai",
2104
2104
  providerOptions,
2105
- schema: openAITranscriptionProviderOptions
2105
+ schema: openAITranscriptionModelOptions
2106
2106
  });
2107
2107
  const formData = new FormData();
2108
2108
  const blob = audio instanceof Uint8Array ? new Blob([audio]) : new Blob([(0, import_provider_utils16.convertBase64ToUint8Array)(audio)]);
@@ -2197,7 +2197,7 @@ var import_provider_utils18 = require("@ai-sdk/provider-utils");
2197
2197
  // src/speech/openai-speech-options.ts
2198
2198
  var import_provider_utils17 = require("@ai-sdk/provider-utils");
2199
2199
  var import_v411 = require("zod/v4");
2200
- var openaiSpeechProviderOptionsSchema = (0, import_provider_utils17.lazySchema)(
2200
+ var openaiSpeechModelOptionsSchema = (0, import_provider_utils17.lazySchema)(
2201
2201
  () => (0, import_provider_utils17.zodSchema)(
2202
2202
  import_v411.z.object({
2203
2203
  instructions: import_v411.z.string().nullish(),
@@ -2229,7 +2229,7 @@ var OpenAISpeechModel = class {
2229
2229
  const openAIOptions = await (0, import_provider_utils18.parseProviderOptions)({
2230
2230
  provider: "openai",
2231
2231
  providerOptions,
2232
- schema: openaiSpeechProviderOptionsSchema
2232
+ schema: openaiSpeechModelOptionsSchema
2233
2233
  });
2234
2234
  const requestBody = {
2235
2235
  model: this.modelId,
@@ -3647,7 +3647,7 @@ var openaiResponsesModelIds = [
3647
3647
  "gpt-5-chat-latest",
3648
3648
  ...openaiResponsesReasoningModelIds
3649
3649
  ];
3650
- var openaiResponsesProviderOptionsSchema = (0, import_provider_utils24.lazySchema)(
3650
+ var openaiLanguageModelResponsesOptionsSchema = (0, import_provider_utils24.lazySchema)(
3651
3651
  () => (0, import_provider_utils24.zodSchema)(
3652
3652
  import_v417.z.object({
3653
3653
  /**
@@ -4342,13 +4342,13 @@ var OpenAIResponsesLanguageModel = class {
4342
4342
  let openaiOptions = await (0, import_provider_utils32.parseProviderOptions)({
4343
4343
  provider: providerOptionsName,
4344
4344
  providerOptions,
4345
- schema: openaiResponsesProviderOptionsSchema
4345
+ schema: openaiLanguageModelResponsesOptionsSchema
4346
4346
  });
4347
4347
  if (openaiOptions == null && providerOptionsName !== "openai") {
4348
4348
  openaiOptions = await (0, import_provider_utils32.parseProviderOptions)({
4349
4349
  provider: "openai",
4350
4350
  providerOptions,
4351
- schema: openaiResponsesProviderOptionsSchema
4351
+ schema: openaiLanguageModelResponsesOptionsSchema
4352
4352
  });
4353
4353
  }
4354
4354
  const isReasoningModel = (_a = openaiOptions == null ? void 0 : openaiOptions.forceReasoning) != null ? _a : modelCapabilities.isReasoningModel;
@@ -5742,11 +5742,11 @@ function escapeJSONDelta(delta) {
5742
5742
  imageGenerationArgsSchema,
5743
5743
  imageGenerationOutputSchema,
5744
5744
  modelMaxImagesPerCall,
5745
- openAITranscriptionProviderOptions,
5746
- openaiChatLanguageModelOptions,
5747
- openaiCompletionProviderOptions,
5748
- openaiEmbeddingProviderOptions,
5749
- openaiSpeechProviderOptionsSchema,
5745
+ openAITranscriptionModelOptions,
5746
+ openaiEmbeddingModelOptions,
5747
+ openaiLanguageModelChatOptions,
5748
+ openaiLanguageModelCompletionOptions,
5749
+ openaiSpeechModelOptionsSchema,
5750
5750
  webSearchPreview,
5751
5751
  webSearchPreviewArgsSchema,
5752
5752
  webSearchPreviewInputSchema