@ai-sdk/openai 2.0.22 → 2.0.24
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +14 -0
- package/dist/index.js +21 -3
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +22 -3
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.js +21 -3
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +22 -3
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +2 -2
package/dist/internal/index.mjs
CHANGED
|
@@ -1817,6 +1817,7 @@ import {
|
|
|
1817
1817
|
combineHeaders as combineHeaders5,
|
|
1818
1818
|
convertBase64ToUint8Array,
|
|
1819
1819
|
createJsonResponseHandler as createJsonResponseHandler5,
|
|
1820
|
+
mediaTypeToExtension,
|
|
1820
1821
|
parseProviderOptions as parseProviderOptions4,
|
|
1821
1822
|
postFormDataToApi
|
|
1822
1823
|
} from "@ai-sdk/provider-utils";
|
|
@@ -1932,7 +1933,12 @@ var OpenAITranscriptionModel = class {
|
|
|
1932
1933
|
const formData = new FormData();
|
|
1933
1934
|
const blob = audio instanceof Uint8Array ? new Blob([audio]) : new Blob([convertBase64ToUint8Array(audio)]);
|
|
1934
1935
|
formData.append("model", this.modelId);
|
|
1935
|
-
|
|
1936
|
+
const fileExtension = mediaTypeToExtension(mediaType);
|
|
1937
|
+
formData.append(
|
|
1938
|
+
"file",
|
|
1939
|
+
new File([blob], "audio", { type: mediaType }),
|
|
1940
|
+
`audio.${fileExtension}`
|
|
1941
|
+
);
|
|
1936
1942
|
if (openAIOptions) {
|
|
1937
1943
|
const transcriptionModelOptions = {
|
|
1938
1944
|
include: openAIOptions.include,
|
|
@@ -2803,6 +2809,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2803
2809
|
})
|
|
2804
2810
|
])
|
|
2805
2811
|
),
|
|
2812
|
+
service_tier: z16.string().nullish(),
|
|
2806
2813
|
incomplete_details: z16.object({ reason: z16.string() }).nullable(),
|
|
2807
2814
|
usage: usageSchema2
|
|
2808
2815
|
})
|
|
@@ -2961,6 +2968,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2961
2968
|
if (logprobs.length > 0) {
|
|
2962
2969
|
providerMetadata.openai.logprobs = logprobs;
|
|
2963
2970
|
}
|
|
2971
|
+
if (typeof response.service_tier === "string") {
|
|
2972
|
+
providerMetadata.openai.serviceTier = response.service_tier;
|
|
2973
|
+
}
|
|
2964
2974
|
return {
|
|
2965
2975
|
content,
|
|
2966
2976
|
finishReason: mapOpenAIResponseFinishReason({
|
|
@@ -3017,6 +3027,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
3017
3027
|
const ongoingToolCalls = {};
|
|
3018
3028
|
let hasToolCalls = false;
|
|
3019
3029
|
const activeReasoning = {};
|
|
3030
|
+
let serviceTier;
|
|
3020
3031
|
return {
|
|
3021
3032
|
stream: response.pipeThrough(
|
|
3022
3033
|
new TransformStream({
|
|
@@ -3275,6 +3286,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
3275
3286
|
usage.totalTokens = value.response.usage.input_tokens + value.response.usage.output_tokens;
|
|
3276
3287
|
usage.reasoningTokens = (_j = (_i = value.response.usage.output_tokens_details) == null ? void 0 : _i.reasoning_tokens) != null ? _j : void 0;
|
|
3277
3288
|
usage.cachedInputTokens = (_l = (_k = value.response.usage.input_tokens_details) == null ? void 0 : _k.cached_tokens) != null ? _l : void 0;
|
|
3289
|
+
if (typeof value.response.service_tier === "string") {
|
|
3290
|
+
serviceTier = value.response.service_tier;
|
|
3291
|
+
}
|
|
3278
3292
|
} else if (isResponseAnnotationAddedChunk(value)) {
|
|
3279
3293
|
if (value.annotation.type === "url_citation") {
|
|
3280
3294
|
controller.enqueue({
|
|
@@ -3307,6 +3321,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
3307
3321
|
if (logprobs.length > 0) {
|
|
3308
3322
|
providerMetadata.openai.logprobs = logprobs;
|
|
3309
3323
|
}
|
|
3324
|
+
if (serviceTier !== void 0) {
|
|
3325
|
+
providerMetadata.openai.serviceTier = serviceTier;
|
|
3326
|
+
}
|
|
3310
3327
|
controller.enqueue({
|
|
3311
3328
|
type: "finish",
|
|
3312
3329
|
finishReason,
|
|
@@ -3344,7 +3361,8 @@ var responseFinishedChunkSchema = z16.object({
|
|
|
3344
3361
|
type: z16.enum(["response.completed", "response.incomplete"]),
|
|
3345
3362
|
response: z16.object({
|
|
3346
3363
|
incomplete_details: z16.object({ reason: z16.string() }).nullish(),
|
|
3347
|
-
usage: usageSchema2
|
|
3364
|
+
usage: usageSchema2,
|
|
3365
|
+
service_tier: z16.string().nullish()
|
|
3348
3366
|
})
|
|
3349
3367
|
});
|
|
3350
3368
|
var responseCreatedChunkSchema = z16.object({
|
|
@@ -3352,7 +3370,8 @@ var responseCreatedChunkSchema = z16.object({
|
|
|
3352
3370
|
response: z16.object({
|
|
3353
3371
|
id: z16.string(),
|
|
3354
3372
|
created_at: z16.number(),
|
|
3355
|
-
model: z16.string()
|
|
3373
|
+
model: z16.string(),
|
|
3374
|
+
service_tier: z16.string().nullish()
|
|
3356
3375
|
})
|
|
3357
3376
|
});
|
|
3358
3377
|
var responseOutputItemAddedSchema = z16.object({
|