@ai-sdk/openai 3.0.26 → 3.0.28

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,18 @@
1
1
  # @ai-sdk/openai
2
2
 
3
+ ## 3.0.28
4
+
5
+ ### Patch Changes
6
+
7
+ - Updated dependencies [4024a3a]
8
+ - @ai-sdk/provider-utils@4.0.15
9
+
10
+ ## 3.0.27
11
+
12
+ ### Patch Changes
13
+
14
+ - 99fbed8: feat: normalize provider specific model options type names and ensure they are exported
15
+
3
16
  ## 3.0.26
4
17
 
5
18
  ### Patch Changes
package/dist/index.d.mts CHANGED
@@ -4,7 +4,7 @@ import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
4
4
  import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
5
5
 
6
6
  type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'o4-mini' | 'o4-mini-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.2' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | (string & {});
7
- declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<{
7
+ declare const openaiLanguageModelChatOptions: _ai_sdk_provider_utils.LazySchema<{
8
8
  logitBias?: Record<number, number> | undefined;
9
9
  logprobs?: number | boolean | undefined;
10
10
  parallelToolCalls?: boolean | undefined;
@@ -23,11 +23,24 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<
23
23
  systemMessageMode?: "remove" | "system" | "developer" | undefined;
24
24
  forceReasoning?: boolean | undefined;
25
25
  }>;
26
- type OpenAIChatLanguageModelOptions = InferSchema<typeof openaiChatLanguageModelOptions>;
26
+ type OpenAILanguageModelChatOptions = InferSchema<typeof openaiLanguageModelChatOptions>;
27
27
 
28
28
  type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
29
+ declare const openaiLanguageModelCompletionOptions: _ai_sdk_provider_utils.LazySchema<{
30
+ echo?: boolean | undefined;
31
+ logitBias?: Record<string, number> | undefined;
32
+ suffix?: string | undefined;
33
+ user?: string | undefined;
34
+ logprobs?: number | boolean | undefined;
35
+ }>;
36
+ type OpenAILanguageModelCompletionOptions = InferSchema<typeof openaiLanguageModelCompletionOptions>;
29
37
 
30
38
  type OpenAIEmbeddingModelId = 'text-embedding-3-small' | 'text-embedding-3-large' | 'text-embedding-ada-002' | (string & {});
39
+ declare const openaiEmbeddingModelOptions: _ai_sdk_provider_utils.LazySchema<{
40
+ dimensions?: number | undefined;
41
+ user?: string | undefined;
42
+ }>;
43
+ type OpenAIEmbeddingModelOptions = InferSchema<typeof openaiEmbeddingModelOptions>;
31
44
 
32
45
  type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | 'gpt-image-1.5' | (string & {});
33
46
 
@@ -785,7 +798,7 @@ declare const openaiTools: {
785
798
  };
786
799
 
787
800
  type OpenAIResponsesModelId = 'chatgpt-4o-latest' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo' | 'gpt-4-0613' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-turbo' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1' | 'gpt-4' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini' | 'gpt-4o' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.1-codex-mini' | 'gpt-5.1-codex' | 'gpt-5.1-codex-max' | 'gpt-5.2' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | 'gpt-5-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5-codex' | 'gpt-5-mini-2025-08-07' | 'gpt-5-mini' | 'gpt-5-nano-2025-08-07' | 'gpt-5-nano' | 'gpt-5-pro-2025-10-06' | 'gpt-5-pro' | 'gpt-5' | 'o1-2024-12-17' | 'o1' | 'o3-2025-04-16' | 'o3-mini-2025-01-31' | 'o3-mini' | 'o3' | (string & {});
788
- declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
801
+ declare const openaiLanguageModelResponsesOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
789
802
  conversation?: string | null | undefined;
790
803
  include?: ("file_search_call.results" | "message.output_text.logprobs" | "reasoning.encrypted_content")[] | null | undefined;
791
804
  instructions?: string | null | undefined;
@@ -808,11 +821,24 @@ declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazyS
808
821
  systemMessageMode?: "remove" | "system" | "developer" | undefined;
809
822
  forceReasoning?: boolean | undefined;
810
823
  }>;
811
- type OpenAIResponsesProviderOptions = InferSchema<typeof openaiResponsesProviderOptionsSchema>;
824
+ type OpenAILanguageModelResponsesOptions = InferSchema<typeof openaiLanguageModelResponsesOptionsSchema>;
812
825
 
813
826
  type OpenAISpeechModelId = 'tts-1' | 'tts-1-hd' | 'gpt-4o-mini-tts' | (string & {});
827
+ declare const openaiSpeechModelOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
828
+ instructions?: string | null | undefined;
829
+ speed?: number | null | undefined;
830
+ }>;
831
+ type OpenAISpeechModelOptions = InferSchema<typeof openaiSpeechModelOptionsSchema>;
814
832
 
815
833
  type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
834
+ declare const openAITranscriptionModelOptions: _ai_sdk_provider_utils.LazySchema<{
835
+ include?: string[] | undefined;
836
+ language?: string | undefined;
837
+ prompt?: string | undefined;
838
+ temperature?: number | undefined;
839
+ timestampGranularities?: ("word" | "segment")[] | undefined;
840
+ }>;
841
+ type OpenAITranscriptionModelOptions = InferSchema<typeof openAITranscriptionModelOptions>;
816
842
 
817
843
  interface OpenAIProvider extends ProviderV3 {
818
844
  (modelId: OpenAIResponsesModelId): LanguageModelV3;
@@ -954,4 +980,4 @@ type OpenaiResponsesSourceDocumentProviderMetadata = {
954
980
 
955
981
  declare const VERSION: string;
956
982
 
957
- export { type OpenAIChatLanguageModelOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAIResponsesProviderOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
983
+ export { type OpenAILanguageModelChatOptions as OpenAIChatLanguageModelOptions, type OpenAIEmbeddingModelOptions, type OpenAILanguageModelChatOptions, type OpenAILanguageModelCompletionOptions, type OpenAILanguageModelResponsesOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAILanguageModelResponsesOptions as OpenAIResponsesProviderOptions, type OpenAISpeechModelOptions, type OpenAITranscriptionModelOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
package/dist/index.d.ts CHANGED
@@ -4,7 +4,7 @@ import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
4
4
  import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
5
5
 
6
6
  type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'o4-mini' | 'o4-mini-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.2' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | (string & {});
7
- declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<{
7
+ declare const openaiLanguageModelChatOptions: _ai_sdk_provider_utils.LazySchema<{
8
8
  logitBias?: Record<number, number> | undefined;
9
9
  logprobs?: number | boolean | undefined;
10
10
  parallelToolCalls?: boolean | undefined;
@@ -23,11 +23,24 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<
23
23
  systemMessageMode?: "remove" | "system" | "developer" | undefined;
24
24
  forceReasoning?: boolean | undefined;
25
25
  }>;
26
- type OpenAIChatLanguageModelOptions = InferSchema<typeof openaiChatLanguageModelOptions>;
26
+ type OpenAILanguageModelChatOptions = InferSchema<typeof openaiLanguageModelChatOptions>;
27
27
 
28
28
  type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
29
+ declare const openaiLanguageModelCompletionOptions: _ai_sdk_provider_utils.LazySchema<{
30
+ echo?: boolean | undefined;
31
+ logitBias?: Record<string, number> | undefined;
32
+ suffix?: string | undefined;
33
+ user?: string | undefined;
34
+ logprobs?: number | boolean | undefined;
35
+ }>;
36
+ type OpenAILanguageModelCompletionOptions = InferSchema<typeof openaiLanguageModelCompletionOptions>;
29
37
 
30
38
  type OpenAIEmbeddingModelId = 'text-embedding-3-small' | 'text-embedding-3-large' | 'text-embedding-ada-002' | (string & {});
39
+ declare const openaiEmbeddingModelOptions: _ai_sdk_provider_utils.LazySchema<{
40
+ dimensions?: number | undefined;
41
+ user?: string | undefined;
42
+ }>;
43
+ type OpenAIEmbeddingModelOptions = InferSchema<typeof openaiEmbeddingModelOptions>;
31
44
 
32
45
  type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | 'gpt-image-1.5' | (string & {});
33
46
 
@@ -785,7 +798,7 @@ declare const openaiTools: {
785
798
  };
786
799
 
787
800
  type OpenAIResponsesModelId = 'chatgpt-4o-latest' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo' | 'gpt-4-0613' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-turbo' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1' | 'gpt-4' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini' | 'gpt-4o' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.1-codex-mini' | 'gpt-5.1-codex' | 'gpt-5.1-codex-max' | 'gpt-5.2' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | 'gpt-5-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5-codex' | 'gpt-5-mini-2025-08-07' | 'gpt-5-mini' | 'gpt-5-nano-2025-08-07' | 'gpt-5-nano' | 'gpt-5-pro-2025-10-06' | 'gpt-5-pro' | 'gpt-5' | 'o1-2024-12-17' | 'o1' | 'o3-2025-04-16' | 'o3-mini-2025-01-31' | 'o3-mini' | 'o3' | (string & {});
788
- declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
801
+ declare const openaiLanguageModelResponsesOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
789
802
  conversation?: string | null | undefined;
790
803
  include?: ("file_search_call.results" | "message.output_text.logprobs" | "reasoning.encrypted_content")[] | null | undefined;
791
804
  instructions?: string | null | undefined;
@@ -808,11 +821,24 @@ declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazyS
808
821
  systemMessageMode?: "remove" | "system" | "developer" | undefined;
809
822
  forceReasoning?: boolean | undefined;
810
823
  }>;
811
- type OpenAIResponsesProviderOptions = InferSchema<typeof openaiResponsesProviderOptionsSchema>;
824
+ type OpenAILanguageModelResponsesOptions = InferSchema<typeof openaiLanguageModelResponsesOptionsSchema>;
812
825
 
813
826
  type OpenAISpeechModelId = 'tts-1' | 'tts-1-hd' | 'gpt-4o-mini-tts' | (string & {});
827
+ declare const openaiSpeechModelOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
828
+ instructions?: string | null | undefined;
829
+ speed?: number | null | undefined;
830
+ }>;
831
+ type OpenAISpeechModelOptions = InferSchema<typeof openaiSpeechModelOptionsSchema>;
814
832
 
815
833
  type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
834
+ declare const openAITranscriptionModelOptions: _ai_sdk_provider_utils.LazySchema<{
835
+ include?: string[] | undefined;
836
+ language?: string | undefined;
837
+ prompt?: string | undefined;
838
+ temperature?: number | undefined;
839
+ timestampGranularities?: ("word" | "segment")[] | undefined;
840
+ }>;
841
+ type OpenAITranscriptionModelOptions = InferSchema<typeof openAITranscriptionModelOptions>;
816
842
 
817
843
  interface OpenAIProvider extends ProviderV3 {
818
844
  (modelId: OpenAIResponsesModelId): LanguageModelV3;
@@ -954,4 +980,4 @@ type OpenaiResponsesSourceDocumentProviderMetadata = {
954
980
 
955
981
  declare const VERSION: string;
956
982
 
957
- export { type OpenAIChatLanguageModelOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAIResponsesProviderOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
983
+ export { type OpenAILanguageModelChatOptions as OpenAIChatLanguageModelOptions, type OpenAIEmbeddingModelOptions, type OpenAILanguageModelChatOptions, type OpenAILanguageModelCompletionOptions, type OpenAILanguageModelResponsesOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAILanguageModelResponsesOptions as OpenAIResponsesProviderOptions, type OpenAISpeechModelOptions, type OpenAITranscriptionModelOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
package/dist/index.js CHANGED
@@ -467,7 +467,7 @@ var openaiChatChunkSchema = (0, import_provider_utils3.lazySchema)(
467
467
  // src/chat/openai-chat-options.ts
468
468
  var import_provider_utils4 = require("@ai-sdk/provider-utils");
469
469
  var import_v43 = require("zod/v4");
470
- var openaiChatLanguageModelOptions = (0, import_provider_utils4.lazySchema)(
470
+ var openaiLanguageModelChatOptions = (0, import_provider_utils4.lazySchema)(
471
471
  () => (0, import_provider_utils4.zodSchema)(
472
472
  import_v43.z.object({
473
473
  /**
@@ -678,7 +678,7 @@ var OpenAIChatLanguageModel = class {
678
678
  const openaiOptions = (_a = await (0, import_provider_utils5.parseProviderOptions)({
679
679
  provider: "openai",
680
680
  providerOptions,
681
- schema: openaiChatLanguageModelOptions
681
+ schema: openaiLanguageModelChatOptions
682
682
  })) != null ? _a : {};
683
683
  const modelCapabilities = getOpenAILanguageModelCapabilities(this.modelId);
684
684
  const isReasoningModel = (_b = openaiOptions.forceReasoning) != null ? _b : modelCapabilities.isReasoningModel;
@@ -1335,7 +1335,7 @@ var openaiCompletionChunkSchema = (0, import_provider_utils6.lazySchema)(
1335
1335
  // src/completion/openai-completion-options.ts
1336
1336
  var import_provider_utils7 = require("@ai-sdk/provider-utils");
1337
1337
  var import_v45 = require("zod/v4");
1338
- var openaiCompletionProviderOptions = (0, import_provider_utils7.lazySchema)(
1338
+ var openaiLanguageModelCompletionOptions = (0, import_provider_utils7.lazySchema)(
1339
1339
  () => (0, import_provider_utils7.zodSchema)(
1340
1340
  import_v45.z.object({
1341
1341
  /**
@@ -1416,12 +1416,12 @@ var OpenAICompletionLanguageModel = class {
1416
1416
  ...await (0, import_provider_utils8.parseProviderOptions)({
1417
1417
  provider: "openai",
1418
1418
  providerOptions,
1419
- schema: openaiCompletionProviderOptions
1419
+ schema: openaiLanguageModelCompletionOptions
1420
1420
  }),
1421
1421
  ...await (0, import_provider_utils8.parseProviderOptions)({
1422
1422
  provider: this.providerOptionsName,
1423
1423
  providerOptions,
1424
- schema: openaiCompletionProviderOptions
1424
+ schema: openaiLanguageModelCompletionOptions
1425
1425
  })
1426
1426
  };
1427
1427
  if (topK != null) {
@@ -1616,7 +1616,7 @@ var import_provider_utils11 = require("@ai-sdk/provider-utils");
1616
1616
  // src/embedding/openai-embedding-options.ts
1617
1617
  var import_provider_utils9 = require("@ai-sdk/provider-utils");
1618
1618
  var import_v46 = require("zod/v4");
1619
- var openaiEmbeddingProviderOptions = (0, import_provider_utils9.lazySchema)(
1619
+ var openaiEmbeddingModelOptions = (0, import_provider_utils9.lazySchema)(
1620
1620
  () => (0, import_provider_utils9.zodSchema)(
1621
1621
  import_v46.z.object({
1622
1622
  /**
@@ -1675,7 +1675,7 @@ var OpenAIEmbeddingModel = class {
1675
1675
  const openaiOptions = (_a = await (0, import_provider_utils11.parseProviderOptions)({
1676
1676
  provider: "openai",
1677
1677
  providerOptions,
1678
- schema: openaiEmbeddingProviderOptions
1678
+ schema: openaiEmbeddingModelOptions
1679
1679
  })) != null ? _a : {};
1680
1680
  const {
1681
1681
  responseHeaders,
@@ -3656,7 +3656,7 @@ var openaiResponsesModelIds = [
3656
3656
  "gpt-5-chat-latest",
3657
3657
  ...openaiResponsesReasoningModelIds
3658
3658
  ];
3659
- var openaiResponsesProviderOptionsSchema = (0, import_provider_utils25.lazySchema)(
3659
+ var openaiLanguageModelResponsesOptionsSchema = (0, import_provider_utils25.lazySchema)(
3660
3660
  () => (0, import_provider_utils25.zodSchema)(
3661
3661
  import_v420.z.object({
3662
3662
  /**
@@ -4059,13 +4059,13 @@ var OpenAIResponsesLanguageModel = class {
4059
4059
  let openaiOptions = await (0, import_provider_utils27.parseProviderOptions)({
4060
4060
  provider: providerOptionsName,
4061
4061
  providerOptions,
4062
- schema: openaiResponsesProviderOptionsSchema
4062
+ schema: openaiLanguageModelResponsesOptionsSchema
4063
4063
  });
4064
4064
  if (openaiOptions == null && providerOptionsName !== "openai") {
4065
4065
  openaiOptions = await (0, import_provider_utils27.parseProviderOptions)({
4066
4066
  provider: "openai",
4067
4067
  providerOptions,
4068
- schema: openaiResponsesProviderOptionsSchema
4068
+ schema: openaiLanguageModelResponsesOptionsSchema
4069
4069
  });
4070
4070
  }
4071
4071
  const isReasoningModel = (_a = openaiOptions == null ? void 0 : openaiOptions.forceReasoning) != null ? _a : modelCapabilities.isReasoningModel;
@@ -5439,7 +5439,7 @@ var import_provider_utils29 = require("@ai-sdk/provider-utils");
5439
5439
  // src/speech/openai-speech-options.ts
5440
5440
  var import_provider_utils28 = require("@ai-sdk/provider-utils");
5441
5441
  var import_v421 = require("zod/v4");
5442
- var openaiSpeechProviderOptionsSchema = (0, import_provider_utils28.lazySchema)(
5442
+ var openaiSpeechModelOptionsSchema = (0, import_provider_utils28.lazySchema)(
5443
5443
  () => (0, import_provider_utils28.zodSchema)(
5444
5444
  import_v421.z.object({
5445
5445
  instructions: import_v421.z.string().nullish(),
@@ -5471,7 +5471,7 @@ var OpenAISpeechModel = class {
5471
5471
  const openAIOptions = await (0, import_provider_utils29.parseProviderOptions)({
5472
5472
  provider: "openai",
5473
5473
  providerOptions,
5474
- schema: openaiSpeechProviderOptionsSchema
5474
+ schema: openaiSpeechModelOptionsSchema
5475
5475
  });
5476
5476
  const requestBody = {
5477
5477
  model: this.modelId,
@@ -5589,7 +5589,7 @@ var openaiTranscriptionResponseSchema = (0, import_provider_utils30.lazySchema)(
5589
5589
  // src/transcription/openai-transcription-options.ts
5590
5590
  var import_provider_utils31 = require("@ai-sdk/provider-utils");
5591
5591
  var import_v423 = require("zod/v4");
5592
- var openAITranscriptionProviderOptions = (0, import_provider_utils31.lazySchema)(
5592
+ var openAITranscriptionModelOptions = (0, import_provider_utils31.lazySchema)(
5593
5593
  () => (0, import_provider_utils31.zodSchema)(
5594
5594
  import_v423.z.object({
5595
5595
  /**
@@ -5696,7 +5696,7 @@ var OpenAITranscriptionModel = class {
5696
5696
  const openAIOptions = await (0, import_provider_utils32.parseProviderOptions)({
5697
5697
  provider: "openai",
5698
5698
  providerOptions,
5699
- schema: openAITranscriptionProviderOptions
5699
+ schema: openAITranscriptionModelOptions
5700
5700
  });
5701
5701
  const formData = new FormData();
5702
5702
  const blob = audio instanceof Uint8Array ? new Blob([audio]) : new Blob([(0, import_provider_utils32.convertBase64ToUint8Array)(audio)]);
@@ -5786,7 +5786,7 @@ var OpenAITranscriptionModel = class {
5786
5786
  };
5787
5787
 
5788
5788
  // src/version.ts
5789
- var VERSION = true ? "3.0.26" : "0.0.0-test";
5789
+ var VERSION = true ? "3.0.28" : "0.0.0-test";
5790
5790
 
5791
5791
  // src/openai-provider.ts
5792
5792
  function createOpenAI(options = {}) {