@ai-sdk/openai 2.0.22 → 2.0.23
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +6 -0
- package/dist/index.js +15 -2
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +15 -2
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.js +15 -2
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +15 -2
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +1 -1
package/dist/index.mjs
CHANGED
|
@@ -2483,6 +2483,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2483
2483
|
})
|
|
2484
2484
|
])
|
|
2485
2485
|
),
|
|
2486
|
+
service_tier: z13.string().nullish(),
|
|
2486
2487
|
incomplete_details: z13.object({ reason: z13.string() }).nullable(),
|
|
2487
2488
|
usage: usageSchema2
|
|
2488
2489
|
})
|
|
@@ -2641,6 +2642,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2641
2642
|
if (logprobs.length > 0) {
|
|
2642
2643
|
providerMetadata.openai.logprobs = logprobs;
|
|
2643
2644
|
}
|
|
2645
|
+
if (typeof response.service_tier === "string") {
|
|
2646
|
+
providerMetadata.openai.serviceTier = response.service_tier;
|
|
2647
|
+
}
|
|
2644
2648
|
return {
|
|
2645
2649
|
content,
|
|
2646
2650
|
finishReason: mapOpenAIResponseFinishReason({
|
|
@@ -2697,6 +2701,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2697
2701
|
const ongoingToolCalls = {};
|
|
2698
2702
|
let hasToolCalls = false;
|
|
2699
2703
|
const activeReasoning = {};
|
|
2704
|
+
let serviceTier;
|
|
2700
2705
|
return {
|
|
2701
2706
|
stream: response.pipeThrough(
|
|
2702
2707
|
new TransformStream({
|
|
@@ -2955,6 +2960,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2955
2960
|
usage.totalTokens = value.response.usage.input_tokens + value.response.usage.output_tokens;
|
|
2956
2961
|
usage.reasoningTokens = (_j = (_i = value.response.usage.output_tokens_details) == null ? void 0 : _i.reasoning_tokens) != null ? _j : void 0;
|
|
2957
2962
|
usage.cachedInputTokens = (_l = (_k = value.response.usage.input_tokens_details) == null ? void 0 : _k.cached_tokens) != null ? _l : void 0;
|
|
2963
|
+
if (typeof value.response.service_tier === "string") {
|
|
2964
|
+
serviceTier = value.response.service_tier;
|
|
2965
|
+
}
|
|
2958
2966
|
} else if (isResponseAnnotationAddedChunk(value)) {
|
|
2959
2967
|
if (value.annotation.type === "url_citation") {
|
|
2960
2968
|
controller.enqueue({
|
|
@@ -2987,6 +2995,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2987
2995
|
if (logprobs.length > 0) {
|
|
2988
2996
|
providerMetadata.openai.logprobs = logprobs;
|
|
2989
2997
|
}
|
|
2998
|
+
if (serviceTier !== void 0) {
|
|
2999
|
+
providerMetadata.openai.serviceTier = serviceTier;
|
|
3000
|
+
}
|
|
2990
3001
|
controller.enqueue({
|
|
2991
3002
|
type: "finish",
|
|
2992
3003
|
finishReason,
|
|
@@ -3024,7 +3035,8 @@ var responseFinishedChunkSchema = z13.object({
|
|
|
3024
3035
|
type: z13.enum(["response.completed", "response.incomplete"]),
|
|
3025
3036
|
response: z13.object({
|
|
3026
3037
|
incomplete_details: z13.object({ reason: z13.string() }).nullish(),
|
|
3027
|
-
usage: usageSchema2
|
|
3038
|
+
usage: usageSchema2,
|
|
3039
|
+
service_tier: z13.string().nullish()
|
|
3028
3040
|
})
|
|
3029
3041
|
});
|
|
3030
3042
|
var responseCreatedChunkSchema = z13.object({
|
|
@@ -3032,7 +3044,8 @@ var responseCreatedChunkSchema = z13.object({
|
|
|
3032
3044
|
response: z13.object({
|
|
3033
3045
|
id: z13.string(),
|
|
3034
3046
|
created_at: z13.number(),
|
|
3035
|
-
model: z13.string()
|
|
3047
|
+
model: z13.string(),
|
|
3048
|
+
service_tier: z13.string().nullish()
|
|
3036
3049
|
})
|
|
3037
3050
|
});
|
|
3038
3051
|
var responseOutputItemAddedSchema = z13.object({
|