@ai-sdk/openai 2.1.0-beta.2 → 2.1.0-beta.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,9 +1,9 @@
1
- import { LanguageModelV2, EmbeddingModelV3, ImageModelV2, TranscriptionModelV2CallOptions, TranscriptionModelV2, SpeechModelV2 } from '@ai-sdk/provider';
1
+ import { LanguageModelV2, EmbeddingModelV3, ImageModelV3, TranscriptionModelV2CallOptions, TranscriptionModelV2, SpeechModelV2 } from '@ai-sdk/provider';
2
2
  import { FetchFunction } from '@ai-sdk/provider-utils';
3
3
  import { z } from 'zod/v4';
4
4
 
5
5
  type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | (string & {});
6
- declare const openaiProviderOptions: z.ZodObject<{
6
+ declare const openaiChatLanguageModelOptions: z.ZodObject<{
7
7
  logitBias: z.ZodOptional<z.ZodRecord<z.ZodCoercedNumber<string>, z.ZodNumber>>;
8
8
  logprobs: z.ZodOptional<z.ZodUnion<readonly [z.ZodBoolean, z.ZodNumber]>>;
9
9
  parallelToolCalls: z.ZodOptional<z.ZodBoolean>;
@@ -33,7 +33,7 @@ declare const openaiProviderOptions: z.ZodObject<{
33
33
  promptCacheKey: z.ZodOptional<z.ZodString>;
34
34
  safetyIdentifier: z.ZodOptional<z.ZodString>;
35
35
  }, z.core.$strip>;
36
- type OpenAIProviderOptions = z.infer<typeof openaiProviderOptions>;
36
+ type OpenAIChatLanguageModelOptions = z.infer<typeof openaiChatLanguageModelOptions>;
37
37
 
38
38
  type OpenAIChatConfig = {
39
39
  provider: string;
@@ -137,14 +137,14 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
137
137
  currentDate?: () => Date;
138
138
  };
139
139
  }
140
- declare class OpenAIImageModel implements ImageModelV2 {
140
+ declare class OpenAIImageModel implements ImageModelV3 {
141
141
  readonly modelId: OpenAIImageModelId;
142
142
  private readonly config;
143
- readonly specificationVersion = "v2";
143
+ readonly specificationVersion = "v3";
144
144
  get maxImagesPerCall(): number;
145
145
  get provider(): string;
146
146
  constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
147
- doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV2['doGenerate']>>>;
147
+ doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
148
148
  }
149
149
 
150
150
  type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
@@ -247,4 +247,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
247
247
  }, z.core.$strip>;
248
248
  type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
249
249
 
250
- export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
250
+ export { OpenAIChatLanguageModel, type OpenAIChatLanguageModelOptions, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiChatLanguageModelOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions };
@@ -1,9 +1,9 @@
1
- import { LanguageModelV2, EmbeddingModelV3, ImageModelV2, TranscriptionModelV2CallOptions, TranscriptionModelV2, SpeechModelV2 } from '@ai-sdk/provider';
1
+ import { LanguageModelV2, EmbeddingModelV3, ImageModelV3, TranscriptionModelV2CallOptions, TranscriptionModelV2, SpeechModelV2 } from '@ai-sdk/provider';
2
2
  import { FetchFunction } from '@ai-sdk/provider-utils';
3
3
  import { z } from 'zod/v4';
4
4
 
5
5
  type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | (string & {});
6
- declare const openaiProviderOptions: z.ZodObject<{
6
+ declare const openaiChatLanguageModelOptions: z.ZodObject<{
7
7
  logitBias: z.ZodOptional<z.ZodRecord<z.ZodCoercedNumber<string>, z.ZodNumber>>;
8
8
  logprobs: z.ZodOptional<z.ZodUnion<readonly [z.ZodBoolean, z.ZodNumber]>>;
9
9
  parallelToolCalls: z.ZodOptional<z.ZodBoolean>;
@@ -33,7 +33,7 @@ declare const openaiProviderOptions: z.ZodObject<{
33
33
  promptCacheKey: z.ZodOptional<z.ZodString>;
34
34
  safetyIdentifier: z.ZodOptional<z.ZodString>;
35
35
  }, z.core.$strip>;
36
- type OpenAIProviderOptions = z.infer<typeof openaiProviderOptions>;
36
+ type OpenAIChatLanguageModelOptions = z.infer<typeof openaiChatLanguageModelOptions>;
37
37
 
38
38
  type OpenAIChatConfig = {
39
39
  provider: string;
@@ -137,14 +137,14 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
137
137
  currentDate?: () => Date;
138
138
  };
139
139
  }
140
- declare class OpenAIImageModel implements ImageModelV2 {
140
+ declare class OpenAIImageModel implements ImageModelV3 {
141
141
  readonly modelId: OpenAIImageModelId;
142
142
  private readonly config;
143
- readonly specificationVersion = "v2";
143
+ readonly specificationVersion = "v3";
144
144
  get maxImagesPerCall(): number;
145
145
  get provider(): string;
146
146
  constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
147
- doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV2['doGenerate']>>>;
147
+ doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
148
148
  }
149
149
 
150
150
  type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
@@ -247,4 +247,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
247
247
  }, z.core.$strip>;
248
248
  type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
249
249
 
250
- export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
250
+ export { OpenAIChatLanguageModel, type OpenAIChatLanguageModelOptions, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiChatLanguageModelOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions };
@@ -30,9 +30,9 @@ __export(internal_exports, {
30
30
  hasDefaultResponseFormat: () => hasDefaultResponseFormat,
31
31
  modelMaxImagesPerCall: () => modelMaxImagesPerCall,
32
32
  openAITranscriptionProviderOptions: () => openAITranscriptionProviderOptions,
33
+ openaiChatLanguageModelOptions: () => openaiChatLanguageModelOptions,
33
34
  openaiCompletionProviderOptions: () => openaiCompletionProviderOptions,
34
- openaiEmbeddingProviderOptions: () => openaiEmbeddingProviderOptions,
35
- openaiProviderOptions: () => openaiProviderOptions
35
+ openaiEmbeddingProviderOptions: () => openaiEmbeddingProviderOptions
36
36
  });
37
37
  module.exports = __toCommonJS(internal_exports);
38
38
 
@@ -270,7 +270,7 @@ function mapOpenAIFinishReason(finishReason) {
270
270
 
271
271
  // src/chat/openai-chat-options.ts
272
272
  var import_v42 = require("zod/v4");
273
- var openaiProviderOptions = import_v42.z.object({
273
+ var openaiChatLanguageModelOptions = import_v42.z.object({
274
274
  /**
275
275
  * Modify the likelihood of specified tokens appearing in the completion.
276
276
  *
@@ -452,7 +452,7 @@ var OpenAIChatLanguageModel = class {
452
452
  const openaiOptions = (_a = await (0, import_provider_utils3.parseProviderOptions)({
453
453
  provider: "openai",
454
454
  providerOptions,
455
- schema: openaiProviderOptions
455
+ schema: openaiChatLanguageModelOptions
456
456
  })) != null ? _a : {};
457
457
  const structuredOutputs = (_b = openaiOptions.structuredOutputs) != null ? _b : true;
458
458
  if (topK != null) {
@@ -1606,7 +1606,7 @@ var OpenAIImageModel = class {
1606
1606
  constructor(modelId, config) {
1607
1607
  this.modelId = modelId;
1608
1608
  this.config = config;
1609
- this.specificationVersion = "v2";
1609
+ this.specificationVersion = "v3";
1610
1610
  }
1611
1611
  get maxImagesPerCall() {
1612
1612
  var _a;
@@ -3816,8 +3816,8 @@ var openaiResponsesProviderOptionsSchema = import_v418.z.object({
3816
3816
  hasDefaultResponseFormat,
3817
3817
  modelMaxImagesPerCall,
3818
3818
  openAITranscriptionProviderOptions,
3819
+ openaiChatLanguageModelOptions,
3819
3820
  openaiCompletionProviderOptions,
3820
- openaiEmbeddingProviderOptions,
3821
- openaiProviderOptions
3821
+ openaiEmbeddingProviderOptions
3822
3822
  });
3823
3823
  //# sourceMappingURL=index.js.map