@ai-sdk/openai 3.0.0-beta.59 → 3.0.0-beta.61

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.mjs CHANGED
@@ -2902,10 +2902,13 @@ var openaiResponsesChunkSchema = lazySchema15(
2902
2902
  }),
2903
2903
  z17.object({
2904
2904
  type: z17.literal("error"),
2905
- code: z17.string(),
2906
- message: z17.string(),
2907
- param: z17.string().nullish(),
2908
- sequence_number: z17.number()
2905
+ sequence_number: z17.number(),
2906
+ error: z17.object({
2907
+ type: z17.string(),
2908
+ code: z17.string(),
2909
+ message: z17.string(),
2910
+ param: z17.string().nullish()
2911
+ })
2909
2912
  }),
2910
2913
  z17.object({ type: z17.string() }).loose().transform((value) => ({
2911
2914
  type: "unknown_chunk",
@@ -2918,13 +2921,15 @@ var openaiResponsesChunkSchema = lazySchema15(
2918
2921
  var openaiResponsesResponseSchema = lazySchema15(
2919
2922
  () => zodSchema15(
2920
2923
  z17.object({
2921
- id: z17.string(),
2922
- created_at: z17.number(),
2924
+ id: z17.string().optional(),
2925
+ created_at: z17.number().optional(),
2923
2926
  error: z17.object({
2924
- code: z17.string(),
2925
- message: z17.string()
2927
+ message: z17.string(),
2928
+ type: z17.string(),
2929
+ param: z17.string().nullish(),
2930
+ code: z17.string()
2926
2931
  }).nullish(),
2927
- model: z17.string(),
2932
+ model: z17.string().optional(),
2928
2933
  output: z17.array(
2929
2934
  z17.discriminatedUnion("type", [
2930
2935
  z17.object({
@@ -3127,7 +3132,7 @@ var openaiResponsesResponseSchema = lazySchema15(
3127
3132
  approval_request_id: z17.string()
3128
3133
  })
3129
3134
  ])
3130
- ),
3135
+ ).optional(),
3131
3136
  service_tier: z17.string().nullish(),
3132
3137
  incomplete_details: z17.object({ reason: z17.string() }).nullish(),
3133
3138
  usage: z17.object({
@@ -3135,7 +3140,7 @@ var openaiResponsesResponseSchema = lazySchema15(
3135
3140
  input_tokens_details: z17.object({ cached_tokens: z17.number().nullish() }).nullish(),
3136
3141
  output_tokens: z17.number(),
3137
3142
  output_tokens_details: z17.object({ reasoning_tokens: z17.number().nullish() }).nullish()
3138
- })
3143
+ }).optional()
3139
3144
  })
3140
3145
  )
3141
3146
  );
@@ -3212,6 +3217,7 @@ var openaiResponsesModelIds = [
3212
3217
  var openaiResponsesProviderOptionsSchema = lazySchema16(
3213
3218
  () => zodSchema16(
3214
3219
  z18.object({
3220
+ conversation: z18.string().nullish(),
3215
3221
  include: z18.array(
3216
3222
  z18.enum([
3217
3223
  "reasoning.encrypted_content",
@@ -3497,6 +3503,13 @@ var OpenAIResponsesLanguageModel = class {
3497
3503
  providerOptions,
3498
3504
  schema: openaiResponsesProviderOptionsSchema
3499
3505
  });
3506
+ if ((openaiOptions == null ? void 0 : openaiOptions.conversation) && (openaiOptions == null ? void 0 : openaiOptions.previousResponseId)) {
3507
+ warnings.push({
3508
+ type: "unsupported-setting",
3509
+ setting: "conversation",
3510
+ details: "conversation and previousResponseId cannot be used together"
3511
+ });
3512
+ }
3500
3513
  const { input, warnings: inputWarnings } = await convertToOpenAIResponsesInput({
3501
3514
  prompt,
3502
3515
  systemMessageMode: modelConfig.systemMessageMode,
@@ -3559,6 +3572,7 @@ var OpenAIResponsesLanguageModel = class {
3559
3572
  }
3560
3573
  },
3561
3574
  // provider options:
3575
+ conversation: openaiOptions == null ? void 0 : openaiOptions.conversation,
3562
3576
  max_tool_calls: openaiOptions == null ? void 0 : openaiOptions.maxToolCalls,
3563
3577
  metadata: openaiOptions == null ? void 0 : openaiOptions.metadata,
3564
3578
  parallel_tool_calls: openaiOptions == null ? void 0 : openaiOptions.parallelToolCalls,
@@ -4008,6 +4022,7 @@ var OpenAIResponsesLanguageModel = class {
4008
4022
  if (typeof response.service_tier === "string") {
4009
4023
  providerMetadata.openai.serviceTier = response.service_tier;
4010
4024
  }
4025
+ const usage = response.usage;
4011
4026
  return {
4012
4027
  content,
4013
4028
  finishReason: mapOpenAIResponseFinishReason({
@@ -4015,11 +4030,11 @@ var OpenAIResponsesLanguageModel = class {
4015
4030
  hasFunctionCall
4016
4031
  }),
4017
4032
  usage: {
4018
- inputTokens: response.usage.input_tokens,
4019
- outputTokens: response.usage.output_tokens,
4020
- totalTokens: response.usage.input_tokens + response.usage.output_tokens,
4021
- reasoningTokens: (_z = (_y = response.usage.output_tokens_details) == null ? void 0 : _y.reasoning_tokens) != null ? _z : void 0,
4022
- cachedInputTokens: (_B = (_A = response.usage.input_tokens_details) == null ? void 0 : _A.cached_tokens) != null ? _B : void 0
4033
+ inputTokens: usage.input_tokens,
4034
+ outputTokens: usage.output_tokens,
4035
+ totalTokens: usage.input_tokens + usage.output_tokens,
4036
+ reasoningTokens: (_z = (_y = usage.output_tokens_details) == null ? void 0 : _y.reasoning_tokens) != null ? _z : void 0,
4037
+ cachedInputTokens: (_B = (_A = usage.input_tokens_details) == null ? void 0 : _A.cached_tokens) != null ? _B : void 0
4023
4038
  },
4024
4039
  request: { body },
4025
4040
  response: {
@@ -5030,7 +5045,7 @@ var OpenAITranscriptionModel = class {
5030
5045
  };
5031
5046
 
5032
5047
  // src/version.ts
5033
- var VERSION = true ? "3.0.0-beta.59" : "0.0.0-test";
5048
+ var VERSION = true ? "3.0.0-beta.61" : "0.0.0-test";
5034
5049
 
5035
5050
  // src/openai-provider.ts
5036
5051
  function createOpenAI(options = {}) {