@ai-sdk/openai 3.0.26 → 3.0.27
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +6 -0
- package/dist/index.d.mts +31 -5
- package/dist/index.d.ts +31 -5
- package/dist/index.js +15 -15
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +15 -15
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +12 -12
- package/dist/internal/index.d.ts +12 -12
- package/dist/internal/index.js +24 -24
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +19 -19
- package/dist/internal/index.mjs.map +1 -1
- package/docs/03-openai.mdx +49 -38
- package/package.json +1 -1
- package/src/chat/openai-chat-language-model.ts +2 -2
- package/src/chat/openai-chat-options.ts +3 -3
- package/src/completion/openai-completion-language-model.ts +3 -3
- package/src/completion/openai-completion-options.ts +3 -3
- package/src/embedding/openai-embedding-model.ts +2 -2
- package/src/embedding/openai-embedding-options.ts +3 -3
- package/src/index.ts +14 -2
- package/src/responses/openai-responses-language-model.ts +3 -3
- package/src/responses/openai-responses-options.ts +3 -3
- package/src/speech/openai-speech-model.ts +2 -2
- package/src/speech/openai-speech-options.ts +3 -3
- package/src/transcription/openai-transcription-model.ts +4 -4
- package/src/transcription/openai-transcription-options.ts +3 -3
package/CHANGELOG.md
CHANGED
package/dist/index.d.mts
CHANGED
|
@@ -4,7 +4,7 @@ import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
|
|
|
4
4
|
import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
|
|
5
5
|
|
|
6
6
|
type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'o4-mini' | 'o4-mini-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.2' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | (string & {});
|
|
7
|
-
declare const
|
|
7
|
+
declare const openaiLanguageModelChatOptions: _ai_sdk_provider_utils.LazySchema<{
|
|
8
8
|
logitBias?: Record<number, number> | undefined;
|
|
9
9
|
logprobs?: number | boolean | undefined;
|
|
10
10
|
parallelToolCalls?: boolean | undefined;
|
|
@@ -23,11 +23,24 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<
|
|
|
23
23
|
systemMessageMode?: "remove" | "system" | "developer" | undefined;
|
|
24
24
|
forceReasoning?: boolean | undefined;
|
|
25
25
|
}>;
|
|
26
|
-
type
|
|
26
|
+
type OpenAILanguageModelChatOptions = InferSchema<typeof openaiLanguageModelChatOptions>;
|
|
27
27
|
|
|
28
28
|
type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
|
|
29
|
+
declare const openaiLanguageModelCompletionOptions: _ai_sdk_provider_utils.LazySchema<{
|
|
30
|
+
echo?: boolean | undefined;
|
|
31
|
+
logitBias?: Record<string, number> | undefined;
|
|
32
|
+
suffix?: string | undefined;
|
|
33
|
+
user?: string | undefined;
|
|
34
|
+
logprobs?: number | boolean | undefined;
|
|
35
|
+
}>;
|
|
36
|
+
type OpenAILanguageModelCompletionOptions = InferSchema<typeof openaiLanguageModelCompletionOptions>;
|
|
29
37
|
|
|
30
38
|
type OpenAIEmbeddingModelId = 'text-embedding-3-small' | 'text-embedding-3-large' | 'text-embedding-ada-002' | (string & {});
|
|
39
|
+
declare const openaiEmbeddingModelOptions: _ai_sdk_provider_utils.LazySchema<{
|
|
40
|
+
dimensions?: number | undefined;
|
|
41
|
+
user?: string | undefined;
|
|
42
|
+
}>;
|
|
43
|
+
type OpenAIEmbeddingModelOptions = InferSchema<typeof openaiEmbeddingModelOptions>;
|
|
31
44
|
|
|
32
45
|
type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | 'gpt-image-1.5' | (string & {});
|
|
33
46
|
|
|
@@ -785,7 +798,7 @@ declare const openaiTools: {
|
|
|
785
798
|
};
|
|
786
799
|
|
|
787
800
|
type OpenAIResponsesModelId = 'chatgpt-4o-latest' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo' | 'gpt-4-0613' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-turbo' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1' | 'gpt-4' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini' | 'gpt-4o' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.1-codex-mini' | 'gpt-5.1-codex' | 'gpt-5.1-codex-max' | 'gpt-5.2' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | 'gpt-5-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5-codex' | 'gpt-5-mini-2025-08-07' | 'gpt-5-mini' | 'gpt-5-nano-2025-08-07' | 'gpt-5-nano' | 'gpt-5-pro-2025-10-06' | 'gpt-5-pro' | 'gpt-5' | 'o1-2024-12-17' | 'o1' | 'o3-2025-04-16' | 'o3-mini-2025-01-31' | 'o3-mini' | 'o3' | (string & {});
|
|
788
|
-
declare const
|
|
801
|
+
declare const openaiLanguageModelResponsesOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
789
802
|
conversation?: string | null | undefined;
|
|
790
803
|
include?: ("file_search_call.results" | "message.output_text.logprobs" | "reasoning.encrypted_content")[] | null | undefined;
|
|
791
804
|
instructions?: string | null | undefined;
|
|
@@ -808,11 +821,24 @@ declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazyS
|
|
|
808
821
|
systemMessageMode?: "remove" | "system" | "developer" | undefined;
|
|
809
822
|
forceReasoning?: boolean | undefined;
|
|
810
823
|
}>;
|
|
811
|
-
type
|
|
824
|
+
type OpenAILanguageModelResponsesOptions = InferSchema<typeof openaiLanguageModelResponsesOptionsSchema>;
|
|
812
825
|
|
|
813
826
|
type OpenAISpeechModelId = 'tts-1' | 'tts-1-hd' | 'gpt-4o-mini-tts' | (string & {});
|
|
827
|
+
declare const openaiSpeechModelOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
828
|
+
instructions?: string | null | undefined;
|
|
829
|
+
speed?: number | null | undefined;
|
|
830
|
+
}>;
|
|
831
|
+
type OpenAISpeechModelOptions = InferSchema<typeof openaiSpeechModelOptionsSchema>;
|
|
814
832
|
|
|
815
833
|
type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
|
|
834
|
+
declare const openAITranscriptionModelOptions: _ai_sdk_provider_utils.LazySchema<{
|
|
835
|
+
include?: string[] | undefined;
|
|
836
|
+
language?: string | undefined;
|
|
837
|
+
prompt?: string | undefined;
|
|
838
|
+
temperature?: number | undefined;
|
|
839
|
+
timestampGranularities?: ("word" | "segment")[] | undefined;
|
|
840
|
+
}>;
|
|
841
|
+
type OpenAITranscriptionModelOptions = InferSchema<typeof openAITranscriptionModelOptions>;
|
|
816
842
|
|
|
817
843
|
interface OpenAIProvider extends ProviderV3 {
|
|
818
844
|
(modelId: OpenAIResponsesModelId): LanguageModelV3;
|
|
@@ -954,4 +980,4 @@ type OpenaiResponsesSourceDocumentProviderMetadata = {
|
|
|
954
980
|
|
|
955
981
|
declare const VERSION: string;
|
|
956
982
|
|
|
957
|
-
export { type OpenAIChatLanguageModelOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAIResponsesProviderOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
|
|
983
|
+
export { type OpenAILanguageModelChatOptions as OpenAIChatLanguageModelOptions, type OpenAIEmbeddingModelOptions, type OpenAILanguageModelChatOptions, type OpenAILanguageModelCompletionOptions, type OpenAILanguageModelResponsesOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAILanguageModelResponsesOptions as OpenAIResponsesProviderOptions, type OpenAISpeechModelOptions, type OpenAITranscriptionModelOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
|
package/dist/index.d.ts
CHANGED
|
@@ -4,7 +4,7 @@ import * as _ai_sdk_provider_utils from '@ai-sdk/provider-utils';
|
|
|
4
4
|
import { InferSchema, FetchFunction } from '@ai-sdk/provider-utils';
|
|
5
5
|
|
|
6
6
|
type OpenAIChatModelId = 'o1' | 'o1-2024-12-17' | 'o3-mini' | 'o3-mini-2025-01-31' | 'o3' | 'o3-2025-04-16' | 'o4-mini' | 'o4-mini-2025-04-16' | 'gpt-4.1' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4o' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini' | 'gpt-4o-mini-2024-07-18' | 'gpt-4-turbo' | 'gpt-4-turbo-2024-04-09' | 'gpt-4' | 'gpt-4-0613' | 'gpt-4.5-preview' | 'gpt-4.5-preview-2025-02-27' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo' | 'gpt-3.5-turbo-1106' | 'chatgpt-4o-latest' | 'gpt-5' | 'gpt-5-2025-08-07' | 'gpt-5-mini' | 'gpt-5-mini-2025-08-07' | 'gpt-5-nano' | 'gpt-5-nano-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.2' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | (string & {});
|
|
7
|
-
declare const
|
|
7
|
+
declare const openaiLanguageModelChatOptions: _ai_sdk_provider_utils.LazySchema<{
|
|
8
8
|
logitBias?: Record<number, number> | undefined;
|
|
9
9
|
logprobs?: number | boolean | undefined;
|
|
10
10
|
parallelToolCalls?: boolean | undefined;
|
|
@@ -23,11 +23,24 @@ declare const openaiChatLanguageModelOptions: _ai_sdk_provider_utils.LazySchema<
|
|
|
23
23
|
systemMessageMode?: "remove" | "system" | "developer" | undefined;
|
|
24
24
|
forceReasoning?: boolean | undefined;
|
|
25
25
|
}>;
|
|
26
|
-
type
|
|
26
|
+
type OpenAILanguageModelChatOptions = InferSchema<typeof openaiLanguageModelChatOptions>;
|
|
27
27
|
|
|
28
28
|
type OpenAICompletionModelId = 'gpt-3.5-turbo-instruct' | (string & {});
|
|
29
|
+
declare const openaiLanguageModelCompletionOptions: _ai_sdk_provider_utils.LazySchema<{
|
|
30
|
+
echo?: boolean | undefined;
|
|
31
|
+
logitBias?: Record<string, number> | undefined;
|
|
32
|
+
suffix?: string | undefined;
|
|
33
|
+
user?: string | undefined;
|
|
34
|
+
logprobs?: number | boolean | undefined;
|
|
35
|
+
}>;
|
|
36
|
+
type OpenAILanguageModelCompletionOptions = InferSchema<typeof openaiLanguageModelCompletionOptions>;
|
|
29
37
|
|
|
30
38
|
type OpenAIEmbeddingModelId = 'text-embedding-3-small' | 'text-embedding-3-large' | 'text-embedding-ada-002' | (string & {});
|
|
39
|
+
declare const openaiEmbeddingModelOptions: _ai_sdk_provider_utils.LazySchema<{
|
|
40
|
+
dimensions?: number | undefined;
|
|
41
|
+
user?: string | undefined;
|
|
42
|
+
}>;
|
|
43
|
+
type OpenAIEmbeddingModelOptions = InferSchema<typeof openaiEmbeddingModelOptions>;
|
|
31
44
|
|
|
32
45
|
type OpenAIImageModelId = 'dall-e-3' | 'dall-e-2' | 'gpt-image-1' | 'gpt-image-1-mini' | 'gpt-image-1.5' | (string & {});
|
|
33
46
|
|
|
@@ -785,7 +798,7 @@ declare const openaiTools: {
|
|
|
785
798
|
};
|
|
786
799
|
|
|
787
800
|
type OpenAIResponsesModelId = 'chatgpt-4o-latest' | 'gpt-3.5-turbo-0125' | 'gpt-3.5-turbo-1106' | 'gpt-3.5-turbo' | 'gpt-4-0613' | 'gpt-4-turbo-2024-04-09' | 'gpt-4-turbo' | 'gpt-4.1-2025-04-14' | 'gpt-4.1-mini-2025-04-14' | 'gpt-4.1-mini' | 'gpt-4.1-nano-2025-04-14' | 'gpt-4.1-nano' | 'gpt-4.1' | 'gpt-4' | 'gpt-4o-2024-05-13' | 'gpt-4o-2024-08-06' | 'gpt-4o-2024-11-20' | 'gpt-4o-mini-2024-07-18' | 'gpt-4o-mini' | 'gpt-4o' | 'gpt-5.1' | 'gpt-5.1-chat-latest' | 'gpt-5.1-codex-mini' | 'gpt-5.1-codex' | 'gpt-5.1-codex-max' | 'gpt-5.2' | 'gpt-5.2-chat-latest' | 'gpt-5.2-pro' | 'gpt-5-2025-08-07' | 'gpt-5-chat-latest' | 'gpt-5-codex' | 'gpt-5-mini-2025-08-07' | 'gpt-5-mini' | 'gpt-5-nano-2025-08-07' | 'gpt-5-nano' | 'gpt-5-pro-2025-10-06' | 'gpt-5-pro' | 'gpt-5' | 'o1-2024-12-17' | 'o1' | 'o3-2025-04-16' | 'o3-mini-2025-01-31' | 'o3-mini' | 'o3' | (string & {});
|
|
788
|
-
declare const
|
|
801
|
+
declare const openaiLanguageModelResponsesOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
789
802
|
conversation?: string | null | undefined;
|
|
790
803
|
include?: ("file_search_call.results" | "message.output_text.logprobs" | "reasoning.encrypted_content")[] | null | undefined;
|
|
791
804
|
instructions?: string | null | undefined;
|
|
@@ -808,11 +821,24 @@ declare const openaiResponsesProviderOptionsSchema: _ai_sdk_provider_utils.LazyS
|
|
|
808
821
|
systemMessageMode?: "remove" | "system" | "developer" | undefined;
|
|
809
822
|
forceReasoning?: boolean | undefined;
|
|
810
823
|
}>;
|
|
811
|
-
type
|
|
824
|
+
type OpenAILanguageModelResponsesOptions = InferSchema<typeof openaiLanguageModelResponsesOptionsSchema>;
|
|
812
825
|
|
|
813
826
|
type OpenAISpeechModelId = 'tts-1' | 'tts-1-hd' | 'gpt-4o-mini-tts' | (string & {});
|
|
827
|
+
declare const openaiSpeechModelOptionsSchema: _ai_sdk_provider_utils.LazySchema<{
|
|
828
|
+
instructions?: string | null | undefined;
|
|
829
|
+
speed?: number | null | undefined;
|
|
830
|
+
}>;
|
|
831
|
+
type OpenAISpeechModelOptions = InferSchema<typeof openaiSpeechModelOptionsSchema>;
|
|
814
832
|
|
|
815
833
|
type OpenAITranscriptionModelId = 'whisper-1' | 'gpt-4o-mini-transcribe' | 'gpt-4o-transcribe' | (string & {});
|
|
834
|
+
declare const openAITranscriptionModelOptions: _ai_sdk_provider_utils.LazySchema<{
|
|
835
|
+
include?: string[] | undefined;
|
|
836
|
+
language?: string | undefined;
|
|
837
|
+
prompt?: string | undefined;
|
|
838
|
+
temperature?: number | undefined;
|
|
839
|
+
timestampGranularities?: ("word" | "segment")[] | undefined;
|
|
840
|
+
}>;
|
|
841
|
+
type OpenAITranscriptionModelOptions = InferSchema<typeof openAITranscriptionModelOptions>;
|
|
816
842
|
|
|
817
843
|
interface OpenAIProvider extends ProviderV3 {
|
|
818
844
|
(modelId: OpenAIResponsesModelId): LanguageModelV3;
|
|
@@ -954,4 +980,4 @@ type OpenaiResponsesSourceDocumentProviderMetadata = {
|
|
|
954
980
|
|
|
955
981
|
declare const VERSION: string;
|
|
956
982
|
|
|
957
|
-
export { type OpenAIChatLanguageModelOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAIResponsesProviderOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
|
|
983
|
+
export { type OpenAILanguageModelChatOptions as OpenAIChatLanguageModelOptions, type OpenAIEmbeddingModelOptions, type OpenAILanguageModelChatOptions, type OpenAILanguageModelCompletionOptions, type OpenAILanguageModelResponsesOptions, type OpenAIProvider, type OpenAIProviderSettings, type OpenAILanguageModelResponsesOptions as OpenAIResponsesProviderOptions, type OpenAISpeechModelOptions, type OpenAITranscriptionModelOptions, type OpenaiResponsesProviderMetadata, type OpenaiResponsesReasoningProviderMetadata, type OpenaiResponsesSourceDocumentProviderMetadata, type OpenaiResponsesTextProviderMetadata, VERSION, createOpenAI, openai };
|
package/dist/index.js
CHANGED
|
@@ -467,7 +467,7 @@ var openaiChatChunkSchema = (0, import_provider_utils3.lazySchema)(
|
|
|
467
467
|
// src/chat/openai-chat-options.ts
|
|
468
468
|
var import_provider_utils4 = require("@ai-sdk/provider-utils");
|
|
469
469
|
var import_v43 = require("zod/v4");
|
|
470
|
-
var
|
|
470
|
+
var openaiLanguageModelChatOptions = (0, import_provider_utils4.lazySchema)(
|
|
471
471
|
() => (0, import_provider_utils4.zodSchema)(
|
|
472
472
|
import_v43.z.object({
|
|
473
473
|
/**
|
|
@@ -678,7 +678,7 @@ var OpenAIChatLanguageModel = class {
|
|
|
678
678
|
const openaiOptions = (_a = await (0, import_provider_utils5.parseProviderOptions)({
|
|
679
679
|
provider: "openai",
|
|
680
680
|
providerOptions,
|
|
681
|
-
schema:
|
|
681
|
+
schema: openaiLanguageModelChatOptions
|
|
682
682
|
})) != null ? _a : {};
|
|
683
683
|
const modelCapabilities = getOpenAILanguageModelCapabilities(this.modelId);
|
|
684
684
|
const isReasoningModel = (_b = openaiOptions.forceReasoning) != null ? _b : modelCapabilities.isReasoningModel;
|
|
@@ -1335,7 +1335,7 @@ var openaiCompletionChunkSchema = (0, import_provider_utils6.lazySchema)(
|
|
|
1335
1335
|
// src/completion/openai-completion-options.ts
|
|
1336
1336
|
var import_provider_utils7 = require("@ai-sdk/provider-utils");
|
|
1337
1337
|
var import_v45 = require("zod/v4");
|
|
1338
|
-
var
|
|
1338
|
+
var openaiLanguageModelCompletionOptions = (0, import_provider_utils7.lazySchema)(
|
|
1339
1339
|
() => (0, import_provider_utils7.zodSchema)(
|
|
1340
1340
|
import_v45.z.object({
|
|
1341
1341
|
/**
|
|
@@ -1416,12 +1416,12 @@ var OpenAICompletionLanguageModel = class {
|
|
|
1416
1416
|
...await (0, import_provider_utils8.parseProviderOptions)({
|
|
1417
1417
|
provider: "openai",
|
|
1418
1418
|
providerOptions,
|
|
1419
|
-
schema:
|
|
1419
|
+
schema: openaiLanguageModelCompletionOptions
|
|
1420
1420
|
}),
|
|
1421
1421
|
...await (0, import_provider_utils8.parseProviderOptions)({
|
|
1422
1422
|
provider: this.providerOptionsName,
|
|
1423
1423
|
providerOptions,
|
|
1424
|
-
schema:
|
|
1424
|
+
schema: openaiLanguageModelCompletionOptions
|
|
1425
1425
|
})
|
|
1426
1426
|
};
|
|
1427
1427
|
if (topK != null) {
|
|
@@ -1616,7 +1616,7 @@ var import_provider_utils11 = require("@ai-sdk/provider-utils");
|
|
|
1616
1616
|
// src/embedding/openai-embedding-options.ts
|
|
1617
1617
|
var import_provider_utils9 = require("@ai-sdk/provider-utils");
|
|
1618
1618
|
var import_v46 = require("zod/v4");
|
|
1619
|
-
var
|
|
1619
|
+
var openaiEmbeddingModelOptions = (0, import_provider_utils9.lazySchema)(
|
|
1620
1620
|
() => (0, import_provider_utils9.zodSchema)(
|
|
1621
1621
|
import_v46.z.object({
|
|
1622
1622
|
/**
|
|
@@ -1675,7 +1675,7 @@ var OpenAIEmbeddingModel = class {
|
|
|
1675
1675
|
const openaiOptions = (_a = await (0, import_provider_utils11.parseProviderOptions)({
|
|
1676
1676
|
provider: "openai",
|
|
1677
1677
|
providerOptions,
|
|
1678
|
-
schema:
|
|
1678
|
+
schema: openaiEmbeddingModelOptions
|
|
1679
1679
|
})) != null ? _a : {};
|
|
1680
1680
|
const {
|
|
1681
1681
|
responseHeaders,
|
|
@@ -3656,7 +3656,7 @@ var openaiResponsesModelIds = [
|
|
|
3656
3656
|
"gpt-5-chat-latest",
|
|
3657
3657
|
...openaiResponsesReasoningModelIds
|
|
3658
3658
|
];
|
|
3659
|
-
var
|
|
3659
|
+
var openaiLanguageModelResponsesOptionsSchema = (0, import_provider_utils25.lazySchema)(
|
|
3660
3660
|
() => (0, import_provider_utils25.zodSchema)(
|
|
3661
3661
|
import_v420.z.object({
|
|
3662
3662
|
/**
|
|
@@ -4059,13 +4059,13 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
4059
4059
|
let openaiOptions = await (0, import_provider_utils27.parseProviderOptions)({
|
|
4060
4060
|
provider: providerOptionsName,
|
|
4061
4061
|
providerOptions,
|
|
4062
|
-
schema:
|
|
4062
|
+
schema: openaiLanguageModelResponsesOptionsSchema
|
|
4063
4063
|
});
|
|
4064
4064
|
if (openaiOptions == null && providerOptionsName !== "openai") {
|
|
4065
4065
|
openaiOptions = await (0, import_provider_utils27.parseProviderOptions)({
|
|
4066
4066
|
provider: "openai",
|
|
4067
4067
|
providerOptions,
|
|
4068
|
-
schema:
|
|
4068
|
+
schema: openaiLanguageModelResponsesOptionsSchema
|
|
4069
4069
|
});
|
|
4070
4070
|
}
|
|
4071
4071
|
const isReasoningModel = (_a = openaiOptions == null ? void 0 : openaiOptions.forceReasoning) != null ? _a : modelCapabilities.isReasoningModel;
|
|
@@ -5439,7 +5439,7 @@ var import_provider_utils29 = require("@ai-sdk/provider-utils");
|
|
|
5439
5439
|
// src/speech/openai-speech-options.ts
|
|
5440
5440
|
var import_provider_utils28 = require("@ai-sdk/provider-utils");
|
|
5441
5441
|
var import_v421 = require("zod/v4");
|
|
5442
|
-
var
|
|
5442
|
+
var openaiSpeechModelOptionsSchema = (0, import_provider_utils28.lazySchema)(
|
|
5443
5443
|
() => (0, import_provider_utils28.zodSchema)(
|
|
5444
5444
|
import_v421.z.object({
|
|
5445
5445
|
instructions: import_v421.z.string().nullish(),
|
|
@@ -5471,7 +5471,7 @@ var OpenAISpeechModel = class {
|
|
|
5471
5471
|
const openAIOptions = await (0, import_provider_utils29.parseProviderOptions)({
|
|
5472
5472
|
provider: "openai",
|
|
5473
5473
|
providerOptions,
|
|
5474
|
-
schema:
|
|
5474
|
+
schema: openaiSpeechModelOptionsSchema
|
|
5475
5475
|
});
|
|
5476
5476
|
const requestBody = {
|
|
5477
5477
|
model: this.modelId,
|
|
@@ -5589,7 +5589,7 @@ var openaiTranscriptionResponseSchema = (0, import_provider_utils30.lazySchema)(
|
|
|
5589
5589
|
// src/transcription/openai-transcription-options.ts
|
|
5590
5590
|
var import_provider_utils31 = require("@ai-sdk/provider-utils");
|
|
5591
5591
|
var import_v423 = require("zod/v4");
|
|
5592
|
-
var
|
|
5592
|
+
var openAITranscriptionModelOptions = (0, import_provider_utils31.lazySchema)(
|
|
5593
5593
|
() => (0, import_provider_utils31.zodSchema)(
|
|
5594
5594
|
import_v423.z.object({
|
|
5595
5595
|
/**
|
|
@@ -5696,7 +5696,7 @@ var OpenAITranscriptionModel = class {
|
|
|
5696
5696
|
const openAIOptions = await (0, import_provider_utils32.parseProviderOptions)({
|
|
5697
5697
|
provider: "openai",
|
|
5698
5698
|
providerOptions,
|
|
5699
|
-
schema:
|
|
5699
|
+
schema: openAITranscriptionModelOptions
|
|
5700
5700
|
});
|
|
5701
5701
|
const formData = new FormData();
|
|
5702
5702
|
const blob = audio instanceof Uint8Array ? new Blob([audio]) : new Blob([(0, import_provider_utils32.convertBase64ToUint8Array)(audio)]);
|
|
@@ -5786,7 +5786,7 @@ var OpenAITranscriptionModel = class {
|
|
|
5786
5786
|
};
|
|
5787
5787
|
|
|
5788
5788
|
// src/version.ts
|
|
5789
|
-
var VERSION = true ? "3.0.
|
|
5789
|
+
var VERSION = true ? "3.0.27" : "0.0.0-test";
|
|
5790
5790
|
|
|
5791
5791
|
// src/openai-provider.ts
|
|
5792
5792
|
function createOpenAI(options = {}) {
|