@ai-sdk/openai 3.0.0-beta.60 → 3.0.0-beta.61

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2821,10 +2821,13 @@ var openaiResponsesChunkSchema = lazySchema12(
2821
2821
  }),
2822
2822
  z14.object({
2823
2823
  type: z14.literal("error"),
2824
- code: z14.string(),
2825
- message: z14.string(),
2826
- param: z14.string().nullish(),
2827
- sequence_number: z14.number()
2824
+ sequence_number: z14.number(),
2825
+ error: z14.object({
2826
+ type: z14.string(),
2827
+ code: z14.string(),
2828
+ message: z14.string(),
2829
+ param: z14.string().nullish()
2830
+ })
2828
2831
  }),
2829
2832
  z14.object({ type: z14.string() }).loose().transform((value) => ({
2830
2833
  type: "unknown_chunk",
@@ -2837,13 +2840,15 @@ var openaiResponsesChunkSchema = lazySchema12(
2837
2840
  var openaiResponsesResponseSchema = lazySchema12(
2838
2841
  () => zodSchema12(
2839
2842
  z14.object({
2840
- id: z14.string(),
2841
- created_at: z14.number(),
2843
+ id: z14.string().optional(),
2844
+ created_at: z14.number().optional(),
2842
2845
  error: z14.object({
2843
- code: z14.string(),
2844
- message: z14.string()
2846
+ message: z14.string(),
2847
+ type: z14.string(),
2848
+ param: z14.string().nullish(),
2849
+ code: z14.string()
2845
2850
  }).nullish(),
2846
- model: z14.string(),
2851
+ model: z14.string().optional(),
2847
2852
  output: z14.array(
2848
2853
  z14.discriminatedUnion("type", [
2849
2854
  z14.object({
@@ -3046,7 +3051,7 @@ var openaiResponsesResponseSchema = lazySchema12(
3046
3051
  approval_request_id: z14.string()
3047
3052
  })
3048
3053
  ])
3049
- ),
3054
+ ).optional(),
3050
3055
  service_tier: z14.string().nullish(),
3051
3056
  incomplete_details: z14.object({ reason: z14.string() }).nullish(),
3052
3057
  usage: z14.object({
@@ -3054,7 +3059,7 @@ var openaiResponsesResponseSchema = lazySchema12(
3054
3059
  input_tokens_details: z14.object({ cached_tokens: z14.number().nullish() }).nullish(),
3055
3060
  output_tokens: z14.number(),
3056
3061
  output_tokens_details: z14.object({ reasoning_tokens: z14.number().nullish() }).nullish()
3057
- })
3062
+ }).optional()
3058
3063
  })
3059
3064
  )
3060
3065
  );
@@ -4281,6 +4286,7 @@ var OpenAIResponsesLanguageModel = class {
4281
4286
  if (typeof response.service_tier === "string") {
4282
4287
  providerMetadata.openai.serviceTier = response.service_tier;
4283
4288
  }
4289
+ const usage = response.usage;
4284
4290
  return {
4285
4291
  content,
4286
4292
  finishReason: mapOpenAIResponseFinishReason({
@@ -4288,11 +4294,11 @@ var OpenAIResponsesLanguageModel = class {
4288
4294
  hasFunctionCall
4289
4295
  }),
4290
4296
  usage: {
4291
- inputTokens: response.usage.input_tokens,
4292
- outputTokens: response.usage.output_tokens,
4293
- totalTokens: response.usage.input_tokens + response.usage.output_tokens,
4294
- reasoningTokens: (_z = (_y = response.usage.output_tokens_details) == null ? void 0 : _y.reasoning_tokens) != null ? _z : void 0,
4295
- cachedInputTokens: (_B = (_A = response.usage.input_tokens_details) == null ? void 0 : _A.cached_tokens) != null ? _B : void 0
4297
+ inputTokens: usage.input_tokens,
4298
+ outputTokens: usage.output_tokens,
4299
+ totalTokens: usage.input_tokens + usage.output_tokens,
4300
+ reasoningTokens: (_z = (_y = usage.output_tokens_details) == null ? void 0 : _y.reasoning_tokens) != null ? _z : void 0,
4301
+ cachedInputTokens: (_B = (_A = usage.input_tokens_details) == null ? void 0 : _A.cached_tokens) != null ? _B : void 0
4296
4302
  },
4297
4303
  request: { body },
4298
4304
  response: {