@ai-sdk/openai 2.0.0-canary.16 → 2.0.0-canary.18

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -83,7 +83,6 @@ type OpenAIProviderOptions = z.infer<typeof openaiProviderOptions>;
83
83
 
84
84
  type OpenAIChatConfig = {
85
85
  provider: string;
86
- compatibility: 'strict' | 'compatible';
87
86
  headers: () => Record<string, string | undefined>;
88
87
  url: (options: {
89
88
  modelId: string;
@@ -162,7 +161,6 @@ type OpenAICompletionProviderOptions = z.infer<typeof openaiCompletionProviderOp
162
161
 
163
162
  type OpenAICompletionConfig = {
164
163
  provider: string;
165
- compatibility: 'strict' | 'compatible';
166
164
  headers: () => Record<string, string | undefined>;
167
165
  url: (options: {
168
166
  modelId: string;
@@ -229,13 +227,6 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV2<string> {
229
227
  type OpenAIImageModelId = 'gpt-image-1' | 'dall-e-3' | 'dall-e-2' | (string & {});
230
228
  declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
231
229
  declare const hasDefaultResponseFormat: Set<string>;
232
- interface OpenAIImageSettings {
233
- /**
234
- Override the maximum number of images per call (default is dependent on the
235
- model, or 1 for an unknown model).
236
- */
237
- maxImagesPerCall?: number;
238
- }
239
230
 
240
231
  interface OpenAIImageModelConfig extends OpenAIConfig {
241
232
  _internal?: {
@@ -244,12 +235,11 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
244
235
  }
245
236
  declare class OpenAIImageModel implements ImageModelV2 {
246
237
  readonly modelId: OpenAIImageModelId;
247
- private readonly settings;
248
238
  private readonly config;
249
239
  readonly specificationVersion = "v2";
250
240
  get maxImagesPerCall(): number;
251
241
  get provider(): string;
252
- constructor(modelId: OpenAIImageModelId, settings: OpenAIImageSettings, config: OpenAIImageModelConfig);
242
+ constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
253
243
  doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV2['doGenerate']>>>;
254
244
  }
255
245
 
@@ -386,4 +376,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
386
376
  }>;
387
377
  type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
388
378
 
389
- export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIImageSettings, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
379
+ export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
@@ -83,7 +83,6 @@ type OpenAIProviderOptions = z.infer<typeof openaiProviderOptions>;
83
83
 
84
84
  type OpenAIChatConfig = {
85
85
  provider: string;
86
- compatibility: 'strict' | 'compatible';
87
86
  headers: () => Record<string, string | undefined>;
88
87
  url: (options: {
89
88
  modelId: string;
@@ -162,7 +161,6 @@ type OpenAICompletionProviderOptions = z.infer<typeof openaiCompletionProviderOp
162
161
 
163
162
  type OpenAICompletionConfig = {
164
163
  provider: string;
165
- compatibility: 'strict' | 'compatible';
166
164
  headers: () => Record<string, string | undefined>;
167
165
  url: (options: {
168
166
  modelId: string;
@@ -229,13 +227,6 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV2<string> {
229
227
  type OpenAIImageModelId = 'gpt-image-1' | 'dall-e-3' | 'dall-e-2' | (string & {});
230
228
  declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
231
229
  declare const hasDefaultResponseFormat: Set<string>;
232
- interface OpenAIImageSettings {
233
- /**
234
- Override the maximum number of images per call (default is dependent on the
235
- model, or 1 for an unknown model).
236
- */
237
- maxImagesPerCall?: number;
238
- }
239
230
 
240
231
  interface OpenAIImageModelConfig extends OpenAIConfig {
241
232
  _internal?: {
@@ -244,12 +235,11 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
244
235
  }
245
236
  declare class OpenAIImageModel implements ImageModelV2 {
246
237
  readonly modelId: OpenAIImageModelId;
247
- private readonly settings;
248
238
  private readonly config;
249
239
  readonly specificationVersion = "v2";
250
240
  get maxImagesPerCall(): number;
251
241
  get provider(): string;
252
- constructor(modelId: OpenAIImageModelId, settings: OpenAIImageSettings, config: OpenAIImageModelConfig);
242
+ constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
253
243
  doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV2['doGenerate']>>>;
254
244
  }
255
245
 
@@ -386,4 +376,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
386
376
  }>;
387
377
  type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
388
378
 
389
- export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIImageSettings, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
379
+ export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
@@ -625,8 +625,9 @@ var OpenAIChatLanguageModel = class {
625
625
  const body = {
626
626
  ...args,
627
627
  stream: true,
628
- // only include stream_options when in strict compatibility mode:
629
- stream_options: this.config.compatibility === "strict" ? { include_usage: true } : void 0
628
+ stream_options: {
629
+ include_usage: true
630
+ }
630
631
  };
631
632
  const { responseHeaders, value: response } = await (0, import_provider_utils3.postJsonToApi)({
632
633
  url: this.config.url({
@@ -1202,8 +1203,9 @@ var OpenAICompletionLanguageModel = class {
1202
1203
  const body = {
1203
1204
  ...args,
1204
1205
  stream: true,
1205
- // only include stream_options when in strict compatibility mode:
1206
- stream_options: this.config.compatibility === "strict" ? { include_usage: true } : void 0
1206
+ stream_options: {
1207
+ include_usage: true
1208
+ }
1207
1209
  };
1208
1210
  const { responseHeaders, value: response } = await (0, import_provider_utils4.postJsonToApi)({
1209
1211
  url: this.config.url({
@@ -1432,15 +1434,14 @@ var hasDefaultResponseFormat = /* @__PURE__ */ new Set(["gpt-image-1"]);
1432
1434
 
1433
1435
  // src/openai-image-model.ts
1434
1436
  var OpenAIImageModel = class {
1435
- constructor(modelId, settings, config) {
1437
+ constructor(modelId, config) {
1436
1438
  this.modelId = modelId;
1437
- this.settings = settings;
1438
1439
  this.config = config;
1439
1440
  this.specificationVersion = "v2";
1440
1441
  }
1441
1442
  get maxImagesPerCall() {
1442
- var _a, _b;
1443
- return (_b = (_a = this.settings.maxImagesPerCall) != null ? _a : modelMaxImagesPerCall[this.modelId]) != null ? _b : 1;
1443
+ var _a;
1444
+ return (_a = modelMaxImagesPerCall[this.modelId]) != null ? _a : 1;
1444
1445
  }
1445
1446
  get provider() {
1446
1447
  return this.config.provider;