@ai-sdk/openai 2.0.22 → 2.0.23

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2803,6 +2803,7 @@ var OpenAIResponsesLanguageModel = class {
2803
2803
  })
2804
2804
  ])
2805
2805
  ),
2806
+ service_tier: z16.string().nullish(),
2806
2807
  incomplete_details: z16.object({ reason: z16.string() }).nullable(),
2807
2808
  usage: usageSchema2
2808
2809
  })
@@ -2961,6 +2962,9 @@ var OpenAIResponsesLanguageModel = class {
2961
2962
  if (logprobs.length > 0) {
2962
2963
  providerMetadata.openai.logprobs = logprobs;
2963
2964
  }
2965
+ if (typeof response.service_tier === "string") {
2966
+ providerMetadata.openai.serviceTier = response.service_tier;
2967
+ }
2964
2968
  return {
2965
2969
  content,
2966
2970
  finishReason: mapOpenAIResponseFinishReason({
@@ -3017,6 +3021,7 @@ var OpenAIResponsesLanguageModel = class {
3017
3021
  const ongoingToolCalls = {};
3018
3022
  let hasToolCalls = false;
3019
3023
  const activeReasoning = {};
3024
+ let serviceTier;
3020
3025
  return {
3021
3026
  stream: response.pipeThrough(
3022
3027
  new TransformStream({
@@ -3275,6 +3280,9 @@ var OpenAIResponsesLanguageModel = class {
3275
3280
  usage.totalTokens = value.response.usage.input_tokens + value.response.usage.output_tokens;
3276
3281
  usage.reasoningTokens = (_j = (_i = value.response.usage.output_tokens_details) == null ? void 0 : _i.reasoning_tokens) != null ? _j : void 0;
3277
3282
  usage.cachedInputTokens = (_l = (_k = value.response.usage.input_tokens_details) == null ? void 0 : _k.cached_tokens) != null ? _l : void 0;
3283
+ if (typeof value.response.service_tier === "string") {
3284
+ serviceTier = value.response.service_tier;
3285
+ }
3278
3286
  } else if (isResponseAnnotationAddedChunk(value)) {
3279
3287
  if (value.annotation.type === "url_citation") {
3280
3288
  controller.enqueue({
@@ -3307,6 +3315,9 @@ var OpenAIResponsesLanguageModel = class {
3307
3315
  if (logprobs.length > 0) {
3308
3316
  providerMetadata.openai.logprobs = logprobs;
3309
3317
  }
3318
+ if (serviceTier !== void 0) {
3319
+ providerMetadata.openai.serviceTier = serviceTier;
3320
+ }
3310
3321
  controller.enqueue({
3311
3322
  type: "finish",
3312
3323
  finishReason,
@@ -3344,7 +3355,8 @@ var responseFinishedChunkSchema = z16.object({
3344
3355
  type: z16.enum(["response.completed", "response.incomplete"]),
3345
3356
  response: z16.object({
3346
3357
  incomplete_details: z16.object({ reason: z16.string() }).nullish(),
3347
- usage: usageSchema2
3358
+ usage: usageSchema2,
3359
+ service_tier: z16.string().nullish()
3348
3360
  })
3349
3361
  });
3350
3362
  var responseCreatedChunkSchema = z16.object({
@@ -3352,7 +3364,8 @@ var responseCreatedChunkSchema = z16.object({
3352
3364
  response: z16.object({
3353
3365
  id: z16.string(),
3354
3366
  created_at: z16.number(),
3355
- model: z16.string()
3367
+ model: z16.string(),
3368
+ service_tier: z16.string().nullish()
3356
3369
  })
3357
3370
  });
3358
3371
  var responseOutputItemAddedSchema = z16.object({