@ai-sdk/openai 2.0.22 → 2.0.24

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1817,6 +1817,7 @@ import {
1817
1817
  combineHeaders as combineHeaders5,
1818
1818
  convertBase64ToUint8Array,
1819
1819
  createJsonResponseHandler as createJsonResponseHandler5,
1820
+ mediaTypeToExtension,
1820
1821
  parseProviderOptions as parseProviderOptions4,
1821
1822
  postFormDataToApi
1822
1823
  } from "@ai-sdk/provider-utils";
@@ -1932,7 +1933,12 @@ var OpenAITranscriptionModel = class {
1932
1933
  const formData = new FormData();
1933
1934
  const blob = audio instanceof Uint8Array ? new Blob([audio]) : new Blob([convertBase64ToUint8Array(audio)]);
1934
1935
  formData.append("model", this.modelId);
1935
- formData.append("file", new File([blob], "audio", { type: mediaType }));
1936
+ const fileExtension = mediaTypeToExtension(mediaType);
1937
+ formData.append(
1938
+ "file",
1939
+ new File([blob], "audio", { type: mediaType }),
1940
+ `audio.${fileExtension}`
1941
+ );
1936
1942
  if (openAIOptions) {
1937
1943
  const transcriptionModelOptions = {
1938
1944
  include: openAIOptions.include,
@@ -2803,6 +2809,7 @@ var OpenAIResponsesLanguageModel = class {
2803
2809
  })
2804
2810
  ])
2805
2811
  ),
2812
+ service_tier: z16.string().nullish(),
2806
2813
  incomplete_details: z16.object({ reason: z16.string() }).nullable(),
2807
2814
  usage: usageSchema2
2808
2815
  })
@@ -2961,6 +2968,9 @@ var OpenAIResponsesLanguageModel = class {
2961
2968
  if (logprobs.length > 0) {
2962
2969
  providerMetadata.openai.logprobs = logprobs;
2963
2970
  }
2971
+ if (typeof response.service_tier === "string") {
2972
+ providerMetadata.openai.serviceTier = response.service_tier;
2973
+ }
2964
2974
  return {
2965
2975
  content,
2966
2976
  finishReason: mapOpenAIResponseFinishReason({
@@ -3017,6 +3027,7 @@ var OpenAIResponsesLanguageModel = class {
3017
3027
  const ongoingToolCalls = {};
3018
3028
  let hasToolCalls = false;
3019
3029
  const activeReasoning = {};
3030
+ let serviceTier;
3020
3031
  return {
3021
3032
  stream: response.pipeThrough(
3022
3033
  new TransformStream({
@@ -3275,6 +3286,9 @@ var OpenAIResponsesLanguageModel = class {
3275
3286
  usage.totalTokens = value.response.usage.input_tokens + value.response.usage.output_tokens;
3276
3287
  usage.reasoningTokens = (_j = (_i = value.response.usage.output_tokens_details) == null ? void 0 : _i.reasoning_tokens) != null ? _j : void 0;
3277
3288
  usage.cachedInputTokens = (_l = (_k = value.response.usage.input_tokens_details) == null ? void 0 : _k.cached_tokens) != null ? _l : void 0;
3289
+ if (typeof value.response.service_tier === "string") {
3290
+ serviceTier = value.response.service_tier;
3291
+ }
3278
3292
  } else if (isResponseAnnotationAddedChunk(value)) {
3279
3293
  if (value.annotation.type === "url_citation") {
3280
3294
  controller.enqueue({
@@ -3307,6 +3321,9 @@ var OpenAIResponsesLanguageModel = class {
3307
3321
  if (logprobs.length > 0) {
3308
3322
  providerMetadata.openai.logprobs = logprobs;
3309
3323
  }
3324
+ if (serviceTier !== void 0) {
3325
+ providerMetadata.openai.serviceTier = serviceTier;
3326
+ }
3310
3327
  controller.enqueue({
3311
3328
  type: "finish",
3312
3329
  finishReason,
@@ -3344,7 +3361,8 @@ var responseFinishedChunkSchema = z16.object({
3344
3361
  type: z16.enum(["response.completed", "response.incomplete"]),
3345
3362
  response: z16.object({
3346
3363
  incomplete_details: z16.object({ reason: z16.string() }).nullish(),
3347
- usage: usageSchema2
3364
+ usage: usageSchema2,
3365
+ service_tier: z16.string().nullish()
3348
3366
  })
3349
3367
  });
3350
3368
  var responseCreatedChunkSchema = z16.object({
@@ -3352,7 +3370,8 @@ var responseCreatedChunkSchema = z16.object({
3352
3370
  response: z16.object({
3353
3371
  id: z16.string(),
3354
3372
  created_at: z16.number(),
3355
- model: z16.string()
3373
+ model: z16.string(),
3374
+ service_tier: z16.string().nullish()
3356
3375
  })
3357
3376
  });
3358
3377
  var responseOutputItemAddedSchema = z16.object({