@ai-sdk/openai 2.1.0-beta.9 → 3.0.0-beta.18

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,40 +1,26 @@
1
- import { LanguageModelV3, EmbeddingModelV3, ImageModelV3, TranscriptionModelV2CallOptions, TranscriptionModelV2, SpeechModelV2 } from '@ai-sdk/provider';
1
+ import { LanguageModelV3, EmbeddingModelV3, ImageModelV3, TranscriptionModelV3CallOptions, TranscriptionModelV3, SpeechModelV3 } from '@ai-sdk/provider';
2
2
  import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
3
- import { FetchFunction } from '@ai-sdk/provider-utils';
4
- import { z } from 'zod/v4';
3
+ import { InferValidator, FetchFunction } from '@ai-sdk/provider-utils';
5
4
 
6
5
  type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | (string & {});
7
- declare const openaiChatLanguageModelOptions: z.ZodObject<{
8
- logitBias: z.ZodOptional<z.ZodRecord<z.ZodCoercedNumber<string>, z.ZodNumber>>;
9
- logprobs: z.ZodOptional<z.ZodUnion<readonly [z.ZodBoolean, z.ZodNumber]>>;
10
- parallelToolCalls: z.ZodOptional<z.ZodBoolean>;
11
- user: z.ZodOptional<z.ZodString>;
12
- reasoningEffort: z.ZodOptional<z.ZodEnum<{
13
- minimal: "minimal";
14
- low: "low";
15
- medium: "medium";
16
- high: "high";
17
- }>>;
18
- maxCompletionTokens: z.ZodOptional<z.ZodNumber>;
19
- store: z.ZodOptional<z.ZodBoolean>;
20
- metadata: z.ZodOptional<z.ZodRecord<z.ZodString, z.ZodString>>;
21
- prediction: z.ZodOptional<z.ZodRecord<z.ZodString, z.ZodAny>>;
22
- structuredOutputs: z.ZodOptional<z.ZodBoolean>;
23
- serviceTier: z.ZodOptional<z.ZodEnum<{
24
- auto: "auto";
25
- flex: "flex";
26
- priority: "priority";
27
- }>>;
28
- strictJsonSchema: z.ZodOptional<z.ZodBoolean>;
29
- textVerbosity: z.ZodOptional<z.ZodEnum<{
30
- low: "low";
31
- medium: "medium";
32
- high: "high";
33
- }>>;
34
- promptCacheKey: z.ZodOptional<z.ZodString>;
35
- safetyIdentifier: z.ZodOptional<z.ZodString>;
36
- }, z.core.$strip>;
37
- type OpenAIChatLanguageModelOptions = z.infer<typeof openaiChatLanguageModelOptions>;
6
+ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazyValidator<{
7
+ logitBias?: Record<number, number> | undefined;
8
+ logprobs?: number | boolean | undefined;
9
+ parallelToolCalls?: boolean | undefined;
10
+ user?: string | undefined;
11
+ reasoningEffort?: "minimal" | "low" | "medium" | "high" | undefined;
12
+ maxCompletionTokens?: number | undefined;
13
+ store?: boolean | undefined;
14
+ metadata?: Record<string, string> | undefined;
15
+ prediction?: Record<string, any> | undefined;
16
+ structuredOutputs?: boolean | undefined;
17
+ serviceTier?: "auto" | "flex" | "priority" | undefined;
18
+ strictJsonSchema?: boolean | undefined;
19
+ textVerbosity?: "low" | "medium" | "high" | undefined;
20
+ promptCacheKey?: string | undefined;
21
+ safetyIdentifier?: string | undefined;
22
+ }>;
23
+ type OpenAIChatLanguageModelOptions = InferValidator<typeof openaiChatLanguageModelOptions>;
38
24
 
39
25
  type OpenAIChatConfig = {
40
26
  provider: string;
@@ -60,14 +46,14 @@ declare class OpenAIChatLanguageModel implements LanguageModelV3 {
60
46
  }
61
47
 
62
48
  type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
63
- declare const openaiCompletionProviderOptions: z.ZodObject<{
64
- echo: z.ZodOptional<z.ZodBoolean>;
65
- logitBias: z.ZodOptional<z.ZodRecord<z.ZodString, z.ZodNumber>>;
66
- suffix: z.ZodOptional<z.ZodString>;
67
- user: z.ZodOptional<z.ZodString>;
68
- logprobs: z.ZodOptional<z.ZodUnion<readonly [z.ZodBoolean, z.ZodNumber]>>;
69
- }, z.core.$strip>;
70
- type OpenAICompletionProviderOptions = z.infer<typeof openaiCompletionProviderOptions>;
49
+ declare const openaiCompletionProviderOptions: _ai_sdk_provider_utils.LazyValidator<{
50
+ echo?: boolean | undefined;
51
+ logitBias?: Record<string, number> | undefined;
52
+ suffix?: string | undefined;
53
+ user?: string | undefined;
54
+ logprobs?: number | boolean | undefined;
55
+ }>;
56
+ type OpenAICompletionProviderOptions = InferValidator<typeof openaiCompletionProviderOptions>;
71
57
 
72
58
  type OpenAICompletionConfig = {
73
59
  provider: string;
@@ -112,11 +98,11 @@ type OpenAIConfig = {
112
98
  };
113
99
 
114
100
  type OpenAIEmbeddingModelId = 'text-embedding-3-small' | 'text-embedding-3-large' | 'text-embedding-ada-002' | (string & {});
115
- declare const openaiEmbeddingProviderOptions: z.ZodObject<{
116
- dimensions: z.ZodOptional<z.ZodNumber>;
117
- user: z.ZodOptional<z.ZodString>;
118
- }, z.core.$strip>;
119
- type OpenAIEmbeddingProviderOptions = z.infer<typeof openaiEmbeddingProviderOptions>;
101
+ declare const openaiEmbeddingProviderOptions: _ai_sdk_provider_utils.LazyValidator<{
102
+ dimensions?: number | undefined;
103
+ user?: string | undefined;
104
+ }>;
105
+ type OpenAIEmbeddingProviderOptions = InferValidator<typeof openaiEmbeddingProviderOptions>;
120
106
 
121
107
  declare class OpenAIEmbeddingModel implements EmbeddingModelV3<string> {
122
108
  readonly specificationVersion = "v3";
@@ -129,7 +115,7 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV3<string> {
129
115
  doEmbed({ values, headers, abortSignal, providerOptions, }: Parameters<EmbeddingModelV3<string>['doEmbed']>[0]): Promise<Awaited<ReturnType<EmbeddingModelV3<string>['doEmbed']>>>;
130
116
  }
131
117
 
132
- type OpenAIImageModelId = 'gpt-image-1' | 'dall-e-3' | 'dall-e-2' | (string & {});
118
+ type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | (string & {});
133
119
  declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
134
120
  declare const hasDefaultResponseFormat: Set<string>;
135
121
 
@@ -149,19 +135,16 @@ declare class OpenAIImageModel implements ImageModelV3 {
149
135
  }
150
136
 
151
137
  type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
152
- declare const openAITranscriptionProviderOptions: z.ZodObject<{
153
- include: z.ZodOptional<z.ZodArray<z.ZodString>>;
154
- language: z.ZodOptional<z.ZodString>;
155
- prompt: z.ZodOptional<z.ZodString>;
156
- temperature: z.ZodOptional<z.ZodDefault<z.ZodNumber>>;
157
- timestampGranularities: z.ZodOptional<z.ZodDefault<z.ZodArray<z.ZodEnum<{
158
- word: "word";
159
- segment: "segment";
160
- }>>>>;
161
- }, z.core.$strip>;
162
- type OpenAITranscriptionProviderOptions = z.infer<typeof openAITranscriptionProviderOptions>;
138
+ declare const openAITranscriptionProviderOptions: _ai_sdk_provider_utils.LazyValidator<{
139
+ include?: string[] | undefined;
140
+ language?: string | undefined;
141
+ prompt?: string | undefined;
142
+ temperature?: number | undefined;
143
+ timestampGranularities?: ("word" | "segment")[] | undefined;
144
+ }>;
145
+ type OpenAITranscriptionProviderOptions = InferValidator<typeof openAITranscriptionProviderOptions>;
163
146
 
164
- type OpenAITranscriptionCallOptions = Omit<TranscriptionModelV2CallOptions, 'providerOptions'> & {
147
+ type OpenAITranscriptionCallOptions = Omit<TranscriptionModelV3CallOptions, 'providerOptions'> & {
165
148
  providerOptions?: {
166
149
  openai?: OpenAITranscriptionProviderOptions;
167
150
  };
@@ -171,39 +154,39 @@ interface OpenAITranscriptionModelConfig extends OpenAIConfig {
171
154
  currentDate?: () => Date;
172
155
  };
173
156
  }
174
- declare class OpenAITranscriptionModel implements TranscriptionModelV2 {
157
+ declare class OpenAITranscriptionModel implements TranscriptionModelV3 {
175
158
  readonly modelId: OpenAITranscriptionModelId;
176
159
  private readonly config;
177
- readonly specificationVersion = "v2";
160
+ readonly specificationVersion = "v3";
178
161
  get provider(): string;
179
162
  constructor(modelId: OpenAITranscriptionModelId, config: OpenAITranscriptionModelConfig);
180
163
  private getArgs;
181
- doGenerate(options: OpenAITranscriptionCallOptions): Promise<Awaited<ReturnType<TranscriptionModelV2['doGenerate']>>>;
164
+ doGenerate(options: OpenAITranscriptionCallOptions): Promise<Awaited<ReturnType<TranscriptionModelV3['doGenerate']>>>;
182
165
  }
183
166
 
184
167
  type OpenAISpeechModelId = 'tts-1' | 'tts-1-hd' | 'gpt-4o-mini-tts' | (string & {});
168
+ declare const openaiSpeechProviderOptionsSchema: _ai_sdk_provider_utils.LazyValidator<{
169
+ instructions?: string | null | undefined;
170
+ speed?: number | null | undefined;
171
+ }>;
172
+ type OpenAISpeechCallOptions = InferValidator<typeof openaiSpeechProviderOptionsSchema>;
185
173
 
186
- declare const OpenAIProviderOptionsSchema: z.ZodObject<{
187
- instructions: z.ZodOptional<z.ZodNullable<z.ZodString>>;
188
- speed: z.ZodOptional<z.ZodNullable<z.ZodDefault<z.ZodNumber>>>;
189
- }, z.core.$strip>;
190
- type OpenAISpeechCallOptions = z.infer<typeof OpenAIProviderOptionsSchema>;
191
174
  interface OpenAISpeechModelConfig extends OpenAIConfig {
192
175
  _internal?: {
193
176
  currentDate?: () => Date;
194
177
  };
195
178
  }
196
- declare class OpenAISpeechModel implements SpeechModelV2 {
179
+ declare class OpenAISpeechModel implements SpeechModelV3 {
197
180
  readonly modelId: OpenAISpeechModelId;
198
181
  private readonly config;
199
- readonly specificationVersion = "v2";
182
+ readonly specificationVersion = "v3";
200
183
  get provider(): string;
201
184
  constructor(modelId: OpenAISpeechModelId, config: OpenAISpeechModelConfig);
202
185
  private getArgs;
203
- doGenerate(options: Parameters<SpeechModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<SpeechModelV2['doGenerate']>>>;
186
+ doGenerate(options: Parameters<SpeechModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<SpeechModelV3['doGenerate']>>>;
204
187
  }
205
188
 
206
- type OpenAIResponsesModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5-codex' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | (string & {});
189
+ type OpenAIResponsesModelId = 'chatgpt-4o-latest' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo' | 'gpt-4-0613' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-turbo' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1' | 'gpt-4' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini' | 'gpt-4o' | 'gpt-5-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5-codex' | 'gpt-5-mini-2025-08-07' | 'gpt-5-mini' | 'gpt-5-nano-2025-08-07' | 'gpt-5-nano' | 'gpt-5-pro-2025-10-06' | 'gpt-5-pro' | 'gpt-5' | 'o1-2024-12-17' | 'o1' | 'o3-2025-04-16' | 'o3-mini-2025-01-31' | 'o3-mini' | 'o3' | (string & {});
207
190
 
208
191
  declare class OpenAIResponsesLanguageModel implements LanguageModelV3 {
209
192
  readonly specificationVersion = "v3";
@@ -216,56 +199,25 @@ declare class OpenAIResponsesLanguageModel implements LanguageModelV3 {
216
199
  doGenerate(options: Parameters<LanguageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doGenerate']>>>;
217
200
  doStream(options: Parameters<LanguageModelV3['doStream']>[0]): Promise<Awaited<ReturnType<LanguageModelV3['doStream']>>>;
218
201
  }
219
- declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
220
- include: z.ZodOptional<z.ZodNullable<z.ZodArray<z.ZodEnum<{
221
- "file_search_call.results": "file_search_call.results";
222
- "message.output_text.logprobs": "message.output_text.logprobs";
223
- "reasoning.encrypted_content": "reasoning.encrypted_content";
224
- }>>>>;
225
- instructions: z.ZodOptional<z.ZodNullable<z.ZodString>>;
226
- logprobs: z.ZodOptional<z.ZodUnion<readonly [z.ZodBoolean, z.ZodNumber]>>;
227
- maxToolCalls: z.ZodOptional<z.ZodNullable<z.ZodNumber>>;
228
- metadata: z.ZodOptional<z.ZodNullable<z.ZodAny>>;
229
- parallelToolCalls: z.ZodOptional<z.ZodNullable<z.ZodBoolean>>;
230
- previousResponseId: z.ZodOptional<z.ZodNullable<z.ZodString>>;
231
- promptCacheKey: z.ZodOptional<z.ZodNullable<z.ZodString>>;
232
- reasoningEffort: z.ZodOptional<z.ZodNullable<z.ZodString>>;
233
- reasoningSummary: z.ZodOptional<z.ZodNullable<z.ZodString>>;
234
- safetyIdentifier: z.ZodOptional<z.ZodNullable<z.ZodString>>;
235
- serviceTier: z.ZodOptional<z.ZodNullable<z.ZodEnum<{
236
- auto: "auto";
237
- flex: "flex";
238
- priority: "priority";
239
- }>>>;
240
- store: z.ZodOptional<z.ZodNullable<z.ZodBoolean>>;
241
- strictJsonSchema: z.ZodOptional<z.ZodNullable<z.ZodBoolean>>;
242
- textVerbosity: z.ZodOptional<z.ZodNullable<z.ZodEnum<{
243
- low: "low";
244
- medium: "medium";
245
- high: "high";
246
- }>>>;
247
- user: z.ZodOptional<z.ZodNullable<z.ZodString>>;
248
- }, z.core.$strip>;
249
- type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
250
202
 
251
- declare const codeInterpreterInputSchema: z.ZodObject<{
252
- code: z.ZodOptional<z.ZodNullable<z.ZodString>>;
253
- containerId: z.ZodString;
254
- }, z.core.$strip>;
255
- declare const codeInterpreterOutputSchema: z.ZodObject<{
256
- outputs: z.ZodOptional<z.ZodNullable<z.ZodArray<z.ZodDiscriminatedUnion<[z.ZodObject<{
257
- type: z.ZodLiteral<"logs">;
258
- logs: z.ZodString;
259
- }, z.core.$strip>, z.ZodObject<{
260
- type: z.ZodLiteral<"image">;
261
- url: z.ZodString;
262
- }, z.core.$strip>]>>>>;
263
- }, z.core.$strip>;
264
- declare const codeInterpreterArgsSchema: z.ZodObject<{
265
- container: z.ZodOptional<z.ZodUnion<readonly [z.ZodString, z.ZodObject<{
266
- fileIds: z.ZodOptional<z.ZodArray<z.ZodString>>;
267
- }, z.core.$strip>]>>;
268
- }, z.core.$strip>;
203
+ declare const codeInterpreterInputSchema: _ai_sdk_provider_utils.LazySchema<{
204
+ containerId: string;
205
+ code?: string | null | undefined;
206
+ }>;
207
+ declare const codeInterpreterOutputSchema: _ai_sdk_provider_utils.LazySchema<{
208
+ outputs?: ({
209
+ type: "logs";
210
+ logs: string;
211
+ } | {
212
+ type: "image";
213
+ url: string;
214
+ })[] | null | undefined;
215
+ }>;
216
+ declare const codeInterpreterArgsSchema: _ai_sdk_provider_utils.LazySchema<{
217
+ container?: string | {
218
+ fileIds?: string[] | undefined;
219
+ } | undefined;
220
+ }>;
269
221
  type CodeInterpreterArgs = {
270
222
  /**
271
223
  * The code interpreter container.
@@ -364,36 +316,25 @@ type OpenAIResponsesFileSearchToolCompoundFilter = {
364
316
  filters: Array<OpenAIResponsesFileSearchToolComparisonFilter | OpenAIResponsesFileSearchToolCompoundFilter>;
365
317
  };
366
318
 
367
- declare const fileSearchArgsSchema: z.ZodObject<{
368
- vectorStoreIds: z.ZodArray<z.ZodString>;
369
- maxNumResults: z.ZodOptional<z.ZodNumber>;
370
- ranking: z.ZodOptional<z.ZodObject<{
371
- ranker: z.ZodOptional<z.ZodString>;
372
- scoreThreshold: z.ZodOptional<z.ZodNumber>;
373
- }, z.core.$strip>>;
374
- filters: z.ZodOptional<z.ZodUnion<readonly [z.ZodObject<{
375
- key: z.ZodString;
376
- type: z.ZodEnum<{
377
- lt: "lt";
378
- ne: "ne";
379
- eq: "eq";
380
- gt: "gt";
381
- gte: "gte";
382
- lte: "lte";
383
- }>;
384
- value: z.ZodUnion<readonly [z.ZodString, z.ZodNumber, z.ZodBoolean]>;
385
- }, z.core.$strip>, z.ZodType<any, unknown, z.core.$ZodTypeInternals<any, unknown>>]>>;
386
- }, z.core.$strip>;
387
- declare const fileSearchOutputSchema: z.ZodObject<{
388
- queries: z.ZodArray<z.ZodString>;
389
- results: z.ZodNullable<z.ZodArray<z.ZodObject<{
390
- attributes: z.ZodRecord<z.ZodString, z.ZodUnknown>;
391
- fileId: z.ZodString;
392
- filename: z.ZodString;
393
- score: z.ZodNumber;
394
- text: z.ZodString;
395
- }, z.core.$strip>>>;
396
- }, z.core.$strip>;
319
+ declare const fileSearchArgsSchema: _ai_sdk_provider_utils.LazySchema<{
320
+ vectorStoreIds: string[];
321
+ maxNumResults?: number | undefined;
322
+ ranking?: {
323
+ ranker?: string | undefined;
324
+ scoreThreshold?: number | undefined;
325
+ } | undefined;
326
+ filters?: any;
327
+ }>;
328
+ declare const fileSearchOutputSchema: _ai_sdk_provider_utils.LazySchema<{
329
+ queries: string[];
330
+ results: {
331
+ attributes: Record<string, unknown>;
332
+ fileId: string;
333
+ filename: string;
334
+ score: number;
335
+ text: string;
336
+ }[] | null;
337
+ }>;
397
338
  declare const fileSearch: _ai_sdk_provider_utils.ProviderDefinedToolFactoryWithOutputSchema<{}, {
398
339
  /**
399
340
  * The search query to execute.
@@ -458,4 +399,85 @@ declare const fileSearch: _ai_sdk_provider_utils.ProviderDefinedToolFactoryWithO
458
399
  filters?: OpenAIResponsesFileSearchToolComparisonFilter | OpenAIResponsesFileSearchToolCompoundFilter;
459
400
  }>;
460
401
 
461
- export { OpenAIChatLanguageModel, type OpenAIChatLanguageModelOptions, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, codeInterpreter, codeInterpreterArgsSchema, codeInterpreterInputSchema, codeInterpreterOutputSchema, codeInterpreterToolFactory, fileSearch, fileSearchArgsSchema, fileSearchOutputSchema, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiChatLanguageModelOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions };
402
+ declare const imageGenerationArgsSchema: _ai_sdk_provider_utils.LazySchema<{
403
+ background?: "auto" | "opaque" | "transparent" | undefined;
404
+ inputFidelity?: "low" | "high" | undefined;
405
+ inputImageMask?: {
406
+ fileId?: string | undefined;
407
+ imageUrl?: string | undefined;
408
+ } | undefined;
409
+ model?: string | undefined;
410
+ moderation?: "auto" | undefined;
411
+ outputCompression?: number | undefined;
412
+ outputFormat?: "png" | "jpeg" | "webp" | undefined;
413
+ partialImages?: number | undefined;
414
+ quality?: "auto" | "low" | "medium" | "high" | undefined;
415
+ size?: "auto" | "1024x1024" | "1024x1536" | "1536x1024" | undefined;
416
+ }>;
417
+ declare const imageGenerationOutputSchema: _ai_sdk_provider_utils.LazySchema<{
418
+ result: string;
419
+ }>;
420
+ type ImageGenerationArgs = {
421
+ /**
422
+ * Background type for the generated image. Default is 'auto'.
423
+ */
424
+ background?: 'auto' | 'opaque' | 'transparent';
425
+ /**
426
+ * Input fidelity for the generated image. Default is 'low'.
427
+ */
428
+ inputFidelity?: 'low' | 'high';
429
+ /**
430
+ * Optional mask for inpainting.
431
+ * Contains image_url (string, optional) and file_id (string, optional).
432
+ */
433
+ inputImageMask?: {
434
+ /**
435
+ * File ID for the mask image.
436
+ */
437
+ fileId?: string;
438
+ /**
439
+ * Base64-encoded mask image.
440
+ */
441
+ imageUrl?: string;
442
+ };
443
+ /**
444
+ * The image generation model to use. Default: gpt-image-1.
445
+ */
446
+ model?: string;
447
+ /**
448
+ * Moderation level for the generated image. Default: auto.
449
+ */
450
+ moderation?: 'auto';
451
+ /**
452
+ * Compression level for the output image. Default: 100.
453
+ */
454
+ outputCompression?: number;
455
+ /**
456
+ * The output format of the generated image. One of png, webp, or jpeg.
457
+ * Default: png
458
+ */
459
+ outputFormat?: 'png' | 'jpeg' | 'webp';
460
+ /**
461
+ * Number of partial images to generate in streaming mode, from 0 (default value) to 3.
462
+ */
463
+ partialImages?: number;
464
+ /**
465
+ * The quality of the generated image.
466
+ * One of low, medium, high, or auto. Default: auto.
467
+ */
468
+ quality?: 'auto' | 'low' | 'medium' | 'high';
469
+ /**
470
+ * The size of the generated image.
471
+ * One of 1024x1024, 1024x1536, 1536x1024, or auto.
472
+ * Default: auto.
473
+ */
474
+ size?: 'auto' | '1024x1024' | '1024x1536' | '1536x1024';
475
+ };
476
+ declare const imageGeneration: (args?: ImageGenerationArgs) => _ai_sdk_provider_utils.Tool<{}, {
477
+ /**
478
+ * The generated image encoded in base64.
479
+ */
480
+ result: string;
481
+ }>;
482
+
483
+ export { OpenAIChatLanguageModel, type OpenAIChatLanguageModelOptions, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, OpenAIResponsesLanguageModel, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, codeInterpreter, codeInterpreterArgsSchema, codeInterpreterInputSchema, codeInterpreterOutputSchema, codeInterpreterToolFactory, fileSearch, fileSearchArgsSchema, fileSearchOutputSchema, hasDefaultResponseFormat, imageGeneration, imageGenerationArgsSchema, imageGenerationOutputSchema, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiChatLanguageModelOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiSpeechProviderOptionsSchema };