@ai-sdk/openai 2.0.0-canary.16 → 2.0.0-canary.18
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +50 -0
- package/dist/index.d.mts +4 -17
- package/dist/index.d.ts +4 -17
- package/dist/index.js +13 -18
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +13 -18
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +2 -12
- package/dist/internal/index.d.ts +2 -12
- package/dist/internal/index.js +9 -8
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +9 -8
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +2 -2
|
@@ -83,7 +83,6 @@ type OpenAIProviderOptions = z.infer<typeof openaiProviderOptions>;
|
|
|
83
83
|
|
|
84
84
|
type OpenAIChatConfig = {
|
|
85
85
|
provider: string;
|
|
86
|
-
compatibility: 'strict' | 'compatible';
|
|
87
86
|
headers: () => Record<string, string | undefined>;
|
|
88
87
|
url: (options: {
|
|
89
88
|
modelId: string;
|
|
@@ -162,7 +161,6 @@ type OpenAICompletionProviderOptions = z.infer<typeof openaiCompletionProviderOp
|
|
|
162
161
|
|
|
163
162
|
type OpenAICompletionConfig = {
|
|
164
163
|
provider: string;
|
|
165
|
-
compatibility: 'strict' | 'compatible';
|
|
166
164
|
headers: () => Record<string, string | undefined>;
|
|
167
165
|
url: (options: {
|
|
168
166
|
modelId: string;
|
|
@@ -229,13 +227,6 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV2<string> {
|
|
|
229
227
|
type OpenAIImageModelId = 'gpt-image-1' | 'dall-e-3' | 'dall-e-2' | (string & {});
|
|
230
228
|
declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
|
|
231
229
|
declare const hasDefaultResponseFormat: Set<string>;
|
|
232
|
-
interface OpenAIImageSettings {
|
|
233
|
-
/**
|
|
234
|
-
Override the maximum number of images per call (default is dependent on the
|
|
235
|
-
model, or 1 for an unknown model).
|
|
236
|
-
*/
|
|
237
|
-
maxImagesPerCall?: number;
|
|
238
|
-
}
|
|
239
230
|
|
|
240
231
|
interface OpenAIImageModelConfig extends OpenAIConfig {
|
|
241
232
|
_internal?: {
|
|
@@ -244,12 +235,11 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
|
|
|
244
235
|
}
|
|
245
236
|
declare class OpenAIImageModel implements ImageModelV2 {
|
|
246
237
|
readonly modelId: OpenAIImageModelId;
|
|
247
|
-
private readonly settings;
|
|
248
238
|
private readonly config;
|
|
249
239
|
readonly specificationVersion = "v2";
|
|
250
240
|
get maxImagesPerCall(): number;
|
|
251
241
|
get provider(): string;
|
|
252
|
-
constructor(modelId: OpenAIImageModelId,
|
|
242
|
+
constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
|
|
253
243
|
doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV2['doGenerate']>>>;
|
|
254
244
|
}
|
|
255
245
|
|
|
@@ -386,4 +376,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
|
|
|
386
376
|
}>;
|
|
387
377
|
type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
|
|
388
378
|
|
|
389
|
-
export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type
|
|
379
|
+
export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
|
package/dist/internal/index.d.ts
CHANGED
|
@@ -83,7 +83,6 @@ type OpenAIProviderOptions = z.infer<typeof openaiProviderOptions>;
|
|
|
83
83
|
|
|
84
84
|
type OpenAIChatConfig = {
|
|
85
85
|
provider: string;
|
|
86
|
-
compatibility: 'strict' | 'compatible';
|
|
87
86
|
headers: () => Record<string, string | undefined>;
|
|
88
87
|
url: (options: {
|
|
89
88
|
modelId: string;
|
|
@@ -162,7 +161,6 @@ type OpenAICompletionProviderOptions = z.infer<typeof openaiCompletionProviderOp
|
|
|
162
161
|
|
|
163
162
|
type OpenAICompletionConfig = {
|
|
164
163
|
provider: string;
|
|
165
|
-
compatibility: 'strict' | 'compatible';
|
|
166
164
|
headers: () => Record<string, string | undefined>;
|
|
167
165
|
url: (options: {
|
|
168
166
|
modelId: string;
|
|
@@ -229,13 +227,6 @@ declare class OpenAIEmbeddingModel implements EmbeddingModelV2<string> {
|
|
|
229
227
|
type OpenAIImageModelId = 'gpt-image-1' | 'dall-e-3' | 'dall-e-2' | (string & {});
|
|
230
228
|
declare const modelMaxImagesPerCall: Record<OpenAIImageModelId, number>;
|
|
231
229
|
declare const hasDefaultResponseFormat: Set<string>;
|
|
232
|
-
interface OpenAIImageSettings {
|
|
233
|
-
/**
|
|
234
|
-
Override the maximum number of images per call (default is dependent on the
|
|
235
|
-
model, or 1 for an unknown model).
|
|
236
|
-
*/
|
|
237
|
-
maxImagesPerCall?: number;
|
|
238
|
-
}
|
|
239
230
|
|
|
240
231
|
interface OpenAIImageModelConfig extends OpenAIConfig {
|
|
241
232
|
_internal?: {
|
|
@@ -244,12 +235,11 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
|
|
|
244
235
|
}
|
|
245
236
|
declare class OpenAIImageModel implements ImageModelV2 {
|
|
246
237
|
readonly modelId: OpenAIImageModelId;
|
|
247
|
-
private readonly settings;
|
|
248
238
|
private readonly config;
|
|
249
239
|
readonly specificationVersion = "v2";
|
|
250
240
|
get maxImagesPerCall(): number;
|
|
251
241
|
get provider(): string;
|
|
252
|
-
constructor(modelId: OpenAIImageModelId,
|
|
242
|
+
constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
|
|
253
243
|
doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV2['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV2['doGenerate']>>>;
|
|
254
244
|
}
|
|
255
245
|
|
|
@@ -386,4 +376,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
|
|
|
386
376
|
}>;
|
|
387
377
|
type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
|
|
388
378
|
|
|
389
|
-
export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type
|
|
379
|
+
export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, type OpenAIProviderOptions, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions, openaiProviderOptions };
|
package/dist/internal/index.js
CHANGED
|
@@ -625,8 +625,9 @@ var OpenAIChatLanguageModel = class {
|
|
|
625
625
|
const body = {
|
|
626
626
|
...args,
|
|
627
627
|
stream: true,
|
|
628
|
-
|
|
629
|
-
|
|
628
|
+
stream_options: {
|
|
629
|
+
include_usage: true
|
|
630
|
+
}
|
|
630
631
|
};
|
|
631
632
|
const { responseHeaders, value: response } = await (0, import_provider_utils3.postJsonToApi)({
|
|
632
633
|
url: this.config.url({
|
|
@@ -1202,8 +1203,9 @@ var OpenAICompletionLanguageModel = class {
|
|
|
1202
1203
|
const body = {
|
|
1203
1204
|
...args,
|
|
1204
1205
|
stream: true,
|
|
1205
|
-
|
|
1206
|
-
|
|
1206
|
+
stream_options: {
|
|
1207
|
+
include_usage: true
|
|
1208
|
+
}
|
|
1207
1209
|
};
|
|
1208
1210
|
const { responseHeaders, value: response } = await (0, import_provider_utils4.postJsonToApi)({
|
|
1209
1211
|
url: this.config.url({
|
|
@@ -1432,15 +1434,14 @@ var hasDefaultResponseFormat = /* @__PURE__ */ new Set(["gpt-image-1"]);
|
|
|
1432
1434
|
|
|
1433
1435
|
// src/openai-image-model.ts
|
|
1434
1436
|
var OpenAIImageModel = class {
|
|
1435
|
-
constructor(modelId,
|
|
1437
|
+
constructor(modelId, config) {
|
|
1436
1438
|
this.modelId = modelId;
|
|
1437
|
-
this.settings = settings;
|
|
1438
1439
|
this.config = config;
|
|
1439
1440
|
this.specificationVersion = "v2";
|
|
1440
1441
|
}
|
|
1441
1442
|
get maxImagesPerCall() {
|
|
1442
|
-
var _a
|
|
1443
|
-
return (
|
|
1443
|
+
var _a;
|
|
1444
|
+
return (_a = modelMaxImagesPerCall[this.modelId]) != null ? _a : 1;
|
|
1444
1445
|
}
|
|
1445
1446
|
get provider() {
|
|
1446
1447
|
return this.config.provider;
|