@ai-sdk/openai 2.0.0-canary.17 → 2.0.0-canary.19
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +49 -0
- package/dist/index.d.mts +3 -9
- package/dist/index.d.ts +3 -9
- package/dist/index.js +4 -5
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +4 -5
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +2 -10
- package/dist/internal/index.d.ts +2 -10
- package/dist/internal/index.js +3 -4
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +3 -4
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +2 -2
|
@@ -227,13 +227,6 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV2<string> {
|
|
|
227
227
|
type OpenAIImageModelId = 'gpt-image-1' | 'dall-e-3' | 'dall-e-2' | (string & {});
|
|
228
228
|
declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
|
|
229
229
|
declare const hasDefaultResponseFormat: Set<string>;
|
|
230
|
-
interface OpenAIImageSettings {
|
|
231
|
-
/**
|
|
232
|
-
Override the maximum number of images per call (default is dependent on the
|
|
233
|
-
model, or 1 for an unknown model).
|
|
234
|
-
*/
|
|
235
|
-
maxImagesPerCall?: number;
|
|
236
|
-
}
|
|
237
230
|
|
|
238
231
|
interface OpenAIImageModelConfig extends OpenAIConfig {
|
|
239
232
|
_internal?: {
|
|
@@ -242,12 +235,11 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
|
|
|
242
235
|
}
|
|
243
236
|
declare class OpenAIImageModel implements ImageModelV2 {
|
|
244
237
|
readonly modelId: OpenAIImageModelId;
|
|
245
|
-
private readonly settings;
|
|
246
238
|
private readonly config;
|
|
247
239
|
readonly specificationVersion = "v2";
|
|
248
240
|
get maxImagesPerCall(): number;
|
|
249
241
|
get provider(): string;
|
|
250
|
-
constructor(modelId: OpenAIImageModelId,
|
|
242
|
+
constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
|
|
251
243
|
doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV2['doGenerate']>>>;
|
|
252
244
|
}
|
|
253
245
|
|
|
@@ -384,4 +376,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
|
|
|
384
376
|
}>;
|
|
385
377
|
type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
|
|
386
378
|
|
|
387
|
-
export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type
|
|
379
|
+
export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
|
package/dist/internal/index.d.ts
CHANGED
|
@@ -227,13 +227,6 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV2<string> {
|
|
|
227
227
|
type OpenAIImageModelId = 'gpt-image-1' | 'dall-e-3' | 'dall-e-2' | (string & {});
|
|
228
228
|
declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
|
|
229
229
|
declare const hasDefaultResponseFormat: Set<string>;
|
|
230
|
-
interface OpenAIImageSettings {
|
|
231
|
-
/**
|
|
232
|
-
Override the maximum number of images per call (default is dependent on the
|
|
233
|
-
model, or 1 for an unknown model).
|
|
234
|
-
*/
|
|
235
|
-
maxImagesPerCall?: number;
|
|
236
|
-
}
|
|
237
230
|
|
|
238
231
|
interface OpenAIImageModelConfig extends OpenAIConfig {
|
|
239
232
|
_internal?: {
|
|
@@ -242,12 +235,11 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
|
|
|
242
235
|
}
|
|
243
236
|
declare class OpenAIImageModel implements ImageModelV2 {
|
|
244
237
|
readonly modelId: OpenAIImageModelId;
|
|
245
|
-
private readonly settings;
|
|
246
238
|
private readonly config;
|
|
247
239
|
readonly specificationVersion = "v2";
|
|
248
240
|
get maxImagesPerCall(): number;
|
|
249
241
|
get provider(): string;
|
|
250
|
-
constructor(modelId: OpenAIImageModelId,
|
|
242
|
+
constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
|
|
251
243
|
doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV2['doGenerate']>>>;
|
|
252
244
|
}
|
|
253
245
|
|
|
@@ -384,4 +376,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
|
|
|
384
376
|
}>;
|
|
385
377
|
type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
|
|
386
378
|
|
|
387
|
-
export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type
|
|
379
|
+
export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
|
package/dist/internal/index.js
CHANGED
|
@@ -1434,15 +1434,14 @@ var hasDefaultResponseFormat = /* @__PURE__ */ new Set(["gpt-image-1"]);
|
|
|
1434
1434
|
|
|
1435
1435
|
// src/openai-image-model.ts
|
|
1436
1436
|
var OpenAIImageModel = class {
|
|
1437
|
-
constructor(modelId,
|
|
1437
|
+
constructor(modelId, config) {
|
|
1438
1438
|
this.modelId = modelId;
|
|
1439
|
-
this.settings = settings;
|
|
1440
1439
|
this.config = config;
|
|
1441
1440
|
this.specificationVersion = "v2";
|
|
1442
1441
|
}
|
|
1443
1442
|
get maxImagesPerCall() {
|
|
1444
|
-
var _a
|
|
1445
|
-
return (
|
|
1443
|
+
var _a;
|
|
1444
|
+
return (_a = modelMaxImagesPerCall[this.modelId]) != null ? _a : 1;
|
|
1446
1445
|
}
|
|
1447
1446
|
get provider() {
|
|
1448
1447
|
return this.config.provider;
|