@ai-sdk/openai 3.0.25 → 3.0.27
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +14 -0
- package/dist/index.d.mts +31 -5
- package/dist/index.d.ts +31 -5
- package/dist/index.js +15 -15
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +15 -15
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +12 -12
- package/dist/internal/index.d.ts +12 -12
- package/dist/internal/index.js +24 -24
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +19 -19
- package/dist/internal/index.mjs.map +1 -1
- package/docs/03-openai.mdx +49 -38
- package/package.json +3 -3
- package/src/chat/openai-chat-language-model.ts +2 -2
- package/src/chat/openai-chat-options.ts +3 -3
- package/src/completion/openai-completion-language-model.ts +3 -3
- package/src/completion/openai-completion-options.ts +3 -3
- package/src/embedding/openai-embedding-model.ts +2 -2
- package/src/embedding/openai-embedding-options.ts +3 -3
- package/src/index.ts +14 -2
- package/src/responses/openai-responses-language-model.ts +3 -3
- package/src/responses/openai-responses-options.ts +3 -3
- package/src/speech/openai-speech-model.ts +2 -2
- package/src/speech/openai-speech-options.ts +3 -3
- package/src/transcription/openai-transcription-model.ts +4 -4
- package/src/transcription/openai-transcription-options.ts +3 -3
package/dist/internal/index.mjs
CHANGED
|
@@ -448,7 +448,7 @@ var openaiChatChunkSchema = lazySchema(
|
|
|
448
448
|
// src/chat/openai-chat-options.ts
|
|
449
449
|
import { lazySchema as lazySchema2, zodSchema as zodSchema2 } from "@ai-sdk/provider-utils";
|
|
450
450
|
import { z as z3 } from "zod/v4";
|
|
451
|
-
var
|
|
451
|
+
var openaiLanguageModelChatOptions = lazySchema2(
|
|
452
452
|
() => zodSchema2(
|
|
453
453
|
z3.object({
|
|
454
454
|
/**
|
|
@@ -661,7 +661,7 @@ var OpenAIChatLanguageModel = class {
|
|
|
661
661
|
const openaiOptions = (_a = await parseProviderOptions({
|
|
662
662
|
provider: "openai",
|
|
663
663
|
providerOptions,
|
|
664
|
-
schema:
|
|
664
|
+
schema: openaiLanguageModelChatOptions
|
|
665
665
|
})) != null ? _a : {};
|
|
666
666
|
const modelCapabilities = getOpenAILanguageModelCapabilities(this.modelId);
|
|
667
667
|
const isReasoningModel = (_b = openaiOptions.forceReasoning) != null ? _b : modelCapabilities.isReasoningModel;
|
|
@@ -1327,7 +1327,7 @@ var openaiCompletionChunkSchema = lazySchema3(
|
|
|
1327
1327
|
// src/completion/openai-completion-options.ts
|
|
1328
1328
|
import { lazySchema as lazySchema4, zodSchema as zodSchema4 } from "@ai-sdk/provider-utils";
|
|
1329
1329
|
import { z as z5 } from "zod/v4";
|
|
1330
|
-
var
|
|
1330
|
+
var openaiLanguageModelCompletionOptions = lazySchema4(
|
|
1331
1331
|
() => zodSchema4(
|
|
1332
1332
|
z5.object({
|
|
1333
1333
|
/**
|
|
@@ -1408,12 +1408,12 @@ var OpenAICompletionLanguageModel = class {
|
|
|
1408
1408
|
...await parseProviderOptions2({
|
|
1409
1409
|
provider: "openai",
|
|
1410
1410
|
providerOptions,
|
|
1411
|
-
schema:
|
|
1411
|
+
schema: openaiLanguageModelCompletionOptions
|
|
1412
1412
|
}),
|
|
1413
1413
|
...await parseProviderOptions2({
|
|
1414
1414
|
provider: this.providerOptionsName,
|
|
1415
1415
|
providerOptions,
|
|
1416
|
-
schema:
|
|
1416
|
+
schema: openaiLanguageModelCompletionOptions
|
|
1417
1417
|
})
|
|
1418
1418
|
};
|
|
1419
1419
|
if (topK != null) {
|
|
@@ -1615,7 +1615,7 @@ import {
|
|
|
1615
1615
|
// src/embedding/openai-embedding-options.ts
|
|
1616
1616
|
import { lazySchema as lazySchema5, zodSchema as zodSchema5 } from "@ai-sdk/provider-utils";
|
|
1617
1617
|
import { z as z6 } from "zod/v4";
|
|
1618
|
-
var
|
|
1618
|
+
var openaiEmbeddingModelOptions = lazySchema5(
|
|
1619
1619
|
() => zodSchema5(
|
|
1620
1620
|
z6.object({
|
|
1621
1621
|
/**
|
|
@@ -1674,7 +1674,7 @@ var OpenAIEmbeddingModel = class {
|
|
|
1674
1674
|
const openaiOptions = (_a = await parseProviderOptions3({
|
|
1675
1675
|
provider: "openai",
|
|
1676
1676
|
providerOptions,
|
|
1677
|
-
schema:
|
|
1677
|
+
schema: openaiEmbeddingModelOptions
|
|
1678
1678
|
})) != null ? _a : {};
|
|
1679
1679
|
const {
|
|
1680
1680
|
responseHeaders,
|
|
@@ -1982,7 +1982,7 @@ var openaiTranscriptionResponseSchema = lazySchema8(
|
|
|
1982
1982
|
// src/transcription/openai-transcription-options.ts
|
|
1983
1983
|
import { lazySchema as lazySchema9, zodSchema as zodSchema9 } from "@ai-sdk/provider-utils";
|
|
1984
1984
|
import { z as z10 } from "zod/v4";
|
|
1985
|
-
var
|
|
1985
|
+
var openAITranscriptionModelOptions = lazySchema9(
|
|
1986
1986
|
() => zodSchema9(
|
|
1987
1987
|
z10.object({
|
|
1988
1988
|
/**
|
|
@@ -2089,7 +2089,7 @@ var OpenAITranscriptionModel = class {
|
|
|
2089
2089
|
const openAIOptions = await parseProviderOptions4({
|
|
2090
2090
|
provider: "openai",
|
|
2091
2091
|
providerOptions,
|
|
2092
|
-
schema:
|
|
2092
|
+
schema: openAITranscriptionModelOptions
|
|
2093
2093
|
});
|
|
2094
2094
|
const formData = new FormData();
|
|
2095
2095
|
const blob = audio instanceof Uint8Array ? new Blob([audio]) : new Blob([convertBase64ToUint8Array2(audio)]);
|
|
@@ -2189,7 +2189,7 @@ import {
|
|
|
2189
2189
|
// src/speech/openai-speech-options.ts
|
|
2190
2190
|
import { lazySchema as lazySchema10, zodSchema as zodSchema10 } from "@ai-sdk/provider-utils";
|
|
2191
2191
|
import { z as z11 } from "zod/v4";
|
|
2192
|
-
var
|
|
2192
|
+
var openaiSpeechModelOptionsSchema = lazySchema10(
|
|
2193
2193
|
() => zodSchema10(
|
|
2194
2194
|
z11.object({
|
|
2195
2195
|
instructions: z11.string().nullish(),
|
|
@@ -2221,7 +2221,7 @@ var OpenAISpeechModel = class {
|
|
|
2221
2221
|
const openAIOptions = await parseProviderOptions5({
|
|
2222
2222
|
provider: "openai",
|
|
2223
2223
|
providerOptions,
|
|
2224
|
-
schema:
|
|
2224
|
+
schema: openaiSpeechModelOptionsSchema
|
|
2225
2225
|
});
|
|
2226
2226
|
const requestBody = {
|
|
2227
2227
|
model: this.modelId,
|
|
@@ -3668,7 +3668,7 @@ var openaiResponsesModelIds = [
|
|
|
3668
3668
|
"gpt-5-chat-latest",
|
|
3669
3669
|
...openaiResponsesReasoningModelIds
|
|
3670
3670
|
];
|
|
3671
|
-
var
|
|
3671
|
+
var openaiLanguageModelResponsesOptionsSchema = lazySchema15(
|
|
3672
3672
|
() => zodSchema15(
|
|
3673
3673
|
z17.object({
|
|
3674
3674
|
/**
|
|
@@ -4389,13 +4389,13 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
4389
4389
|
let openaiOptions = await parseProviderOptions7({
|
|
4390
4390
|
provider: providerOptionsName,
|
|
4391
4391
|
providerOptions,
|
|
4392
|
-
schema:
|
|
4392
|
+
schema: openaiLanguageModelResponsesOptionsSchema
|
|
4393
4393
|
});
|
|
4394
4394
|
if (openaiOptions == null && providerOptionsName !== "openai") {
|
|
4395
4395
|
openaiOptions = await parseProviderOptions7({
|
|
4396
4396
|
provider: "openai",
|
|
4397
4397
|
providerOptions,
|
|
4398
|
-
schema:
|
|
4398
|
+
schema: openaiLanguageModelResponsesOptionsSchema
|
|
4399
4399
|
});
|
|
4400
4400
|
}
|
|
4401
4401
|
const isReasoningModel = (_a = openaiOptions == null ? void 0 : openaiOptions.forceReasoning) != null ? _a : modelCapabilities.isReasoningModel;
|
|
@@ -5788,11 +5788,11 @@ export {
|
|
|
5788
5788
|
imageGenerationArgsSchema,
|
|
5789
5789
|
imageGenerationOutputSchema,
|
|
5790
5790
|
modelMaxImagesPerCall,
|
|
5791
|
-
|
|
5792
|
-
|
|
5793
|
-
|
|
5794
|
-
|
|
5795
|
-
|
|
5791
|
+
openAITranscriptionModelOptions,
|
|
5792
|
+
openaiEmbeddingModelOptions,
|
|
5793
|
+
openaiLanguageModelChatOptions,
|
|
5794
|
+
openaiLanguageModelCompletionOptions,
|
|
5795
|
+
openaiSpeechModelOptionsSchema,
|
|
5796
5796
|
webSearchPreview,
|
|
5797
5797
|
webSearchPreviewArgsSchema,
|
|
5798
5798
|
webSearchPreviewInputSchema
|