@ai-sdk/openai 2.0.0-canary.17 → 2.0.0-canary.18

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -227,13 +227,6 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV2<string> {
227
227
  type OpenAIImageModelId = 'gpt-image-1' | 'dall-e-3' | 'dall-e-2' | (string & {});
228
228
  declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
229
229
  declare const hasDefaultResponseFormat: Set<string>;
230
- interface OpenAIImageSettings {
231
- /**
232
- Override the maximum number of images per call (default is dependent on the
233
- model, or 1 for an unknown model).
234
- */
235
- maxImagesPerCall?: number;
236
- }
237
230
 
238
231
  interface OpenAIImageModelConfig extends OpenAIConfig {
239
232
  _internal?: {
@@ -242,12 +235,11 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
242
235
  }
243
236
  declare class OpenAIImageModel implements ImageModelV2 {
244
237
  readonly modelId: OpenAIImageModelId;
245
- private readonly settings;
246
238
  private readonly config;
247
239
  readonly specificationVersion = "v2";
248
240
  get maxImagesPerCall(): number;
249
241
  get provider(): string;
250
- constructor(modelId: OpenAIImageModelId, settings: OpenAIImageSettings, config: OpenAIImageModelConfig);
242
+ constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
251
243
  doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV2['doGenerate']>>>;
252
244
  }
253
245
 
@@ -384,4 +376,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
384
376
  }>;
385
377
  type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
386
378
 
387
- export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIImageSettings, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
379
+ export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
@@ -227,13 +227,6 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV2<string> {
227
227
  type OpenAIImageModelId = 'gpt-image-1' | 'dall-e-3' | 'dall-e-2' | (string & {});
228
228
  declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
229
229
  declare const hasDefaultResponseFormat: Set<string>;
230
- interface OpenAIImageSettings {
231
- /**
232
- Override the maximum number of images per call (default is dependent on the
233
- model, or 1 for an unknown model).
234
- */
235
- maxImagesPerCall?: number;
236
- }
237
230
 
238
231
  interface OpenAIImageModelConfig extends OpenAIConfig {
239
232
  _internal?: {
@@ -242,12 +235,11 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
242
235
  }
243
236
  declare class OpenAIImageModel implements ImageModelV2 {
244
237
  readonly modelId: OpenAIImageModelId;
245
- private readonly settings;
246
238
  private readonly config;
247
239
  readonly specificationVersion = "v2";
248
240
  get maxImagesPerCall(): number;
249
241
  get provider(): string;
250
- constructor(modelId: OpenAIImageModelId, settings: OpenAIImageSettings, config: OpenAIImageModelConfig);
242
+ constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
251
243
  doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV2['doGenerate']>>>;
252
244
  }
253
245
 
@@ -384,4 +376,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
384
376
  }>;
385
377
  type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
386
378
 
387
- export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIImageSettings, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
379
+ export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
@@ -1434,15 +1434,14 @@ var hasDefaultResponseFormat = /* @__PURE__ */ new Set(["gpt-image-1"]);
1434
1434
 
1435
1435
  // src/openai-image-model.ts
1436
1436
  var OpenAIImageModel = class {
1437
- constructor(modelId, settings, config) {
1437
+ constructor(modelId, config) {
1438
1438
  this.modelId = modelId;
1439
- this.settings = settings;
1440
1439
  this.config = config;
1441
1440
  this.specificationVersion = "v2";
1442
1441
  }
1443
1442
  get maxImagesPerCall() {
1444
- var _a, _b;
1445
- return (_b = (_a = this.settings.maxImagesPerCall) != null ? _a : modelMaxImagesPerCall[this.modelId]) != null ? _b : 1;
1443
+ var _a;
1444
+ return (_a = modelMaxImagesPerCall[this.modelId]) != null ? _a : 1;
1446
1445
  }
1447
1446
  get provider() {
1448
1447
  return this.config.provider;