@ai-sdk/openai 2.1.0-beta.2 → 2.1.0-beta.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +32 -0
- package/dist/index.d.mts +36 -5
- package/dist/index.d.ts +36 -5
- package/dist/index.js +4 -4
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +4 -4
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +7 -7
- package/dist/internal/index.d.ts +7 -7
- package/dist/internal/index.js +7 -7
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +5 -5
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +3 -3
|
@@ -1,9 +1,9 @@
|
|
|
1
|
-
import { LanguageModelV2, EmbeddingModelV3,
|
|
1
|
+
import { LanguageModelV2, EmbeddingModelV3, ImageModelV3, TranscriptionModelV2CallOptions, TranscriptionModelV2, SpeechModelV2 } from '@ai-sdk/provider';
|
|
2
2
|
import { FetchFunction } from '@ai-sdk/provider-utils';
|
|
3
3
|
import { z } from 'zod/v4';
|
|
4
4
|
|
|
5
5
|
type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | (string & {});
|
|
6
|
-
declare const
|
|
6
|
+
declare const openaiChatLanguageModelOptions: z.ZodObject<{
|
|
7
7
|
logitBias: z.ZodOptional<z.ZodRecord<z.ZodCoercedNumber<string>, z.ZodNumber>>;
|
|
8
8
|
logprobs: z.ZodOptional<z.ZodUnion<readonly [z.ZodBoolean, z.ZodNumber]>>;
|
|
9
9
|
parallelToolCalls: z.ZodOptional<z.ZodBoolean>;
|
|
@@ -33,7 +33,7 @@ declare const openaiProviderOptions: z.ZodObject<{
|
|
|
33
33
|
promptCacheKey: z.ZodOptional<z.ZodString>;
|
|
34
34
|
safetyIdentifier: z.ZodOptional<z.ZodString>;
|
|
35
35
|
}, z.core.$strip>;
|
|
36
|
-
type
|
|
36
|
+
type OpenAIChatLanguageModelOptions = z.infer<typeof openaiChatLanguageModelOptions>;
|
|
37
37
|
|
|
38
38
|
type OpenAIChatConfig = {
|
|
39
39
|
provider: string;
|
|
@@ -137,14 +137,14 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
|
|
|
137
137
|
currentDate?: () => Date;
|
|
138
138
|
};
|
|
139
139
|
}
|
|
140
|
-
declare class OpenAIImageModel implements
|
|
140
|
+
declare class OpenAIImageModel implements ImageModelV3 {
|
|
141
141
|
readonly modelId: OpenAIImageModelId;
|
|
142
142
|
private readonly config;
|
|
143
|
-
readonly specificationVersion = "
|
|
143
|
+
readonly specificationVersion = "v3";
|
|
144
144
|
get maxImagesPerCall(): number;
|
|
145
145
|
get provider(): string;
|
|
146
146
|
constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
|
|
147
|
-
doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<
|
|
147
|
+
doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
|
|
148
148
|
}
|
|
149
149
|
|
|
150
150
|
type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
|
|
@@ -247,4 +247,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
|
|
|
247
247
|
}, z.core.$strip>;
|
|
248
248
|
type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
|
|
249
249
|
|
|
250
|
-
export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId,
|
|
250
|
+
export { OpenAIChatLanguageModel, type OpenAIChatLanguageModelOptions, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiChatLanguageModelOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions };
|
package/dist/internal/index.d.ts
CHANGED
|
@@ -1,9 +1,9 @@
|
|
|
1
|
-
import { LanguageModelV2, EmbeddingModelV3,
|
|
1
|
+
import { LanguageModelV2, EmbeddingModelV3, ImageModelV3, TranscriptionModelV2CallOptions, TranscriptionModelV2, SpeechModelV2 } from '@ai-sdk/provider';
|
|
2
2
|
import { FetchFunction } from '@ai-sdk/provider-utils';
|
|
3
3
|
import { z } from 'zod/v4';
|
|
4
4
|
|
|
5
5
|
type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | (string & {});
|
|
6
|
-
declare const
|
|
6
|
+
declare const openaiChatLanguageModelOptions: z.ZodObject<{
|
|
7
7
|
logitBias: z.ZodOptional<z.ZodRecord<z.ZodCoercedNumber<string>, z.ZodNumber>>;
|
|
8
8
|
logprobs: z.ZodOptional<z.ZodUnion<readonly [z.ZodBoolean, z.ZodNumber]>>;
|
|
9
9
|
parallelToolCalls: z.ZodOptional<z.ZodBoolean>;
|
|
@@ -33,7 +33,7 @@ declare const openaiProviderOptions: z.ZodObject<{
|
|
|
33
33
|
promptCacheKey: z.ZodOptional<z.ZodString>;
|
|
34
34
|
safetyIdentifier: z.ZodOptional<z.ZodString>;
|
|
35
35
|
}, z.core.$strip>;
|
|
36
|
-
type
|
|
36
|
+
type OpenAIChatLanguageModelOptions = z.infer<typeof openaiChatLanguageModelOptions>;
|
|
37
37
|
|
|
38
38
|
type OpenAIChatConfig = {
|
|
39
39
|
provider: string;
|
|
@@ -137,14 +137,14 @@ interface OpenAIImageModelConfig extends OpenAIConfig {
|
|
|
137
137
|
currentDate?: () => Date;
|
|
138
138
|
};
|
|
139
139
|
}
|
|
140
|
-
declare class OpenAIImageModel implements
|
|
140
|
+
declare class OpenAIImageModel implements ImageModelV3 {
|
|
141
141
|
readonly modelId: OpenAIImageModelId;
|
|
142
142
|
private readonly config;
|
|
143
|
-
readonly specificationVersion = "
|
|
143
|
+
readonly specificationVersion = "v3";
|
|
144
144
|
get maxImagesPerCall(): number;
|
|
145
145
|
get provider(): string;
|
|
146
146
|
constructor(modelId: OpenAIImageModelId, config: OpenAIImageModelConfig);
|
|
147
|
-
doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<
|
|
147
|
+
doGenerate({ prompt, n, size, aspectRatio, seed, providerOptions, headers, abortSignal, }: Parameters<ImageModelV3['doGenerate']>[0]): Promise<Awaited<ReturnType<ImageModelV3['doGenerate']>>>;
|
|
148
148
|
}
|
|
149
149
|
|
|
150
150
|
type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
|
|
@@ -247,4 +247,4 @@ declare const openaiResponsesProviderOptionsSchema: z.ZodObject<{
|
|
|
247
247
|
}, z.core.$strip>;
|
|
248
248
|
type OpenAIResponsesProviderOptions = z.infer<typeof openaiResponsesProviderOptionsSchema>;
|
|
249
249
|
|
|
250
|
-
export { OpenAIChatLanguageModel, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId,
|
|
250
|
+
export { OpenAIChatLanguageModel, type OpenAIChatLanguageModelOptions, type OpenAIChatModelId, OpenAICompletionLanguageModel, type OpenAICompletionModelId, type OpenAICompletionProviderOptions, OpenAIEmbeddingModel, type OpenAIEmbeddingModelId, type OpenAIEmbeddingProviderOptions, OpenAIImageModel, type OpenAIImageModelId, OpenAIResponsesLanguageModel, type OpenAIResponsesProviderOptions, type OpenAISpeechCallOptions, OpenAISpeechModel, type OpenAISpeechModelId, type OpenAITranscriptionCallOptions, OpenAITranscriptionModel, type OpenAITranscriptionModelId, type OpenAITranscriptionProviderOptions, hasDefaultResponseFormat, modelMaxImagesPerCall, openAITranscriptionProviderOptions, openaiChatLanguageModelOptions, openaiCompletionProviderOptions, openaiEmbeddingProviderOptions };
|
package/dist/internal/index.js
CHANGED
|
@@ -30,9 +30,9 @@ __export(internal_exports, {
|
|
|
30
30
|
hasDefaultResponseFormat: () => hasDefaultResponseFormat,
|
|
31
31
|
modelMaxImagesPerCall: () => modelMaxImagesPerCall,
|
|
32
32
|
openAITranscriptionProviderOptions: () => openAITranscriptionProviderOptions,
|
|
33
|
+
openaiChatLanguageModelOptions: () => openaiChatLanguageModelOptions,
|
|
33
34
|
openaiCompletionProviderOptions: () => openaiCompletionProviderOptions,
|
|
34
|
-
openaiEmbeddingProviderOptions: () => openaiEmbeddingProviderOptions
|
|
35
|
-
openaiProviderOptions: () => openaiProviderOptions
|
|
35
|
+
openaiEmbeddingProviderOptions: () => openaiEmbeddingProviderOptions
|
|
36
36
|
});
|
|
37
37
|
module.exports = __toCommonJS(internal_exports);
|
|
38
38
|
|
|
@@ -270,7 +270,7 @@ function mapOpenAIFinishReason(finishReason) {
|
|
|
270
270
|
|
|
271
271
|
// src/chat/openai-chat-options.ts
|
|
272
272
|
var import_v42 = require("zod/v4");
|
|
273
|
-
var
|
|
273
|
+
var openaiChatLanguageModelOptions = import_v42.z.object({
|
|
274
274
|
/**
|
|
275
275
|
* Modify the likelihood of specified tokens appearing in the completion.
|
|
276
276
|
*
|
|
@@ -452,7 +452,7 @@ var OpenAIChatLanguageModel = class {
|
|
|
452
452
|
const openaiOptions = (_a = await (0, import_provider_utils3.parseProviderOptions)({
|
|
453
453
|
provider: "openai",
|
|
454
454
|
providerOptions,
|
|
455
|
-
schema:
|
|
455
|
+
schema: openaiChatLanguageModelOptions
|
|
456
456
|
})) != null ? _a : {};
|
|
457
457
|
const structuredOutputs = (_b = openaiOptions.structuredOutputs) != null ? _b : true;
|
|
458
458
|
if (topK != null) {
|
|
@@ -1606,7 +1606,7 @@ var OpenAIImageModel = class {
|
|
|
1606
1606
|
constructor(modelId, config) {
|
|
1607
1607
|
this.modelId = modelId;
|
|
1608
1608
|
this.config = config;
|
|
1609
|
-
this.specificationVersion = "
|
|
1609
|
+
this.specificationVersion = "v3";
|
|
1610
1610
|
}
|
|
1611
1611
|
get maxImagesPerCall() {
|
|
1612
1612
|
var _a;
|
|
@@ -3816,8 +3816,8 @@ var openaiResponsesProviderOptionsSchema = import_v418.z.object({
|
|
|
3816
3816
|
hasDefaultResponseFormat,
|
|
3817
3817
|
modelMaxImagesPerCall,
|
|
3818
3818
|
openAITranscriptionProviderOptions,
|
|
3819
|
+
openaiChatLanguageModelOptions,
|
|
3819
3820
|
openaiCompletionProviderOptions,
|
|
3820
|
-
openaiEmbeddingProviderOptions
|
|
3821
|
-
openaiProviderOptions
|
|
3821
|
+
openaiEmbeddingProviderOptions
|
|
3822
3822
|
});
|
|
3823
3823
|
//# sourceMappingURL=index.js.map
|