@ai-sdk/openai 2.0.22 → 2.0.23
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +6 -0
- package/dist/index.js +15 -2
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +15 -2
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.js +15 -2
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +15 -2
- package/dist/internal/index.mjs.map +1 -1
- package/package.json +1 -1
package/dist/internal/index.mjs
CHANGED
|
@@ -2803,6 +2803,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2803
2803
|
})
|
|
2804
2804
|
])
|
|
2805
2805
|
),
|
|
2806
|
+
service_tier: z16.string().nullish(),
|
|
2806
2807
|
incomplete_details: z16.object({ reason: z16.string() }).nullable(),
|
|
2807
2808
|
usage: usageSchema2
|
|
2808
2809
|
})
|
|
@@ -2961,6 +2962,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
2961
2962
|
if (logprobs.length > 0) {
|
|
2962
2963
|
providerMetadata.openai.logprobs = logprobs;
|
|
2963
2964
|
}
|
|
2965
|
+
if (typeof response.service_tier === "string") {
|
|
2966
|
+
providerMetadata.openai.serviceTier = response.service_tier;
|
|
2967
|
+
}
|
|
2964
2968
|
return {
|
|
2965
2969
|
content,
|
|
2966
2970
|
finishReason: mapOpenAIResponseFinishReason({
|
|
@@ -3017,6 +3021,7 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
3017
3021
|
const ongoingToolCalls = {};
|
|
3018
3022
|
let hasToolCalls = false;
|
|
3019
3023
|
const activeReasoning = {};
|
|
3024
|
+
let serviceTier;
|
|
3020
3025
|
return {
|
|
3021
3026
|
stream: response.pipeThrough(
|
|
3022
3027
|
new TransformStream({
|
|
@@ -3275,6 +3280,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
3275
3280
|
usage.totalTokens = value.response.usage.input_tokens + value.response.usage.output_tokens;
|
|
3276
3281
|
usage.reasoningTokens = (_j = (_i = value.response.usage.output_tokens_details) == null ? void 0 : _i.reasoning_tokens) != null ? _j : void 0;
|
|
3277
3282
|
usage.cachedInputTokens = (_l = (_k = value.response.usage.input_tokens_details) == null ? void 0 : _k.cached_tokens) != null ? _l : void 0;
|
|
3283
|
+
if (typeof value.response.service_tier === "string") {
|
|
3284
|
+
serviceTier = value.response.service_tier;
|
|
3285
|
+
}
|
|
3278
3286
|
} else if (isResponseAnnotationAddedChunk(value)) {
|
|
3279
3287
|
if (value.annotation.type === "url_citation") {
|
|
3280
3288
|
controller.enqueue({
|
|
@@ -3307,6 +3315,9 @@ var OpenAIResponsesLanguageModel = class {
|
|
|
3307
3315
|
if (logprobs.length > 0) {
|
|
3308
3316
|
providerMetadata.openai.logprobs = logprobs;
|
|
3309
3317
|
}
|
|
3318
|
+
if (serviceTier !== void 0) {
|
|
3319
|
+
providerMetadata.openai.serviceTier = serviceTier;
|
|
3320
|
+
}
|
|
3310
3321
|
controller.enqueue({
|
|
3311
3322
|
type: "finish",
|
|
3312
3323
|
finishReason,
|
|
@@ -3344,7 +3355,8 @@ var responseFinishedChunkSchema = z16.object({
|
|
|
3344
3355
|
type: z16.enum(["response.completed", "response.incomplete"]),
|
|
3345
3356
|
response: z16.object({
|
|
3346
3357
|
incomplete_details: z16.object({ reason: z16.string() }).nullish(),
|
|
3347
|
-
usage: usageSchema2
|
|
3358
|
+
usage: usageSchema2,
|
|
3359
|
+
service_tier: z16.string().nullish()
|
|
3348
3360
|
})
|
|
3349
3361
|
});
|
|
3350
3362
|
var responseCreatedChunkSchema = z16.object({
|
|
@@ -3352,7 +3364,8 @@ var responseCreatedChunkSchema = z16.object({
|
|
|
3352
3364
|
response: z16.object({
|
|
3353
3365
|
id: z16.string(),
|
|
3354
3366
|
created_at: z16.number(),
|
|
3355
|
-
model: z16.string()
|
|
3367
|
+
model: z16.string(),
|
|
3368
|
+
service_tier: z16.string().nullish()
|
|
3356
3369
|
})
|
|
3357
3370
|
});
|
|
3358
3371
|
var responseOutputItemAddedSchema = z16.object({
|