@ai-sdk/xai 3.0.0-beta.38 → 3.0.0-beta.39

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.mjs CHANGED
@@ -451,7 +451,7 @@ var XaiChatLanguageModel = class {
451
451
  };
452
452
  }
453
453
  async doGenerate(options) {
454
- var _a, _b, _c;
454
+ var _a, _b, _c, _d, _e;
455
455
  const { args: body, warnings } = await this.getArgs(options);
456
456
  const {
457
457
  responseHeaders,
@@ -513,7 +513,8 @@ var XaiChatLanguageModel = class {
513
513
  inputTokens: response.usage.prompt_tokens,
514
514
  outputTokens: response.usage.completion_tokens,
515
515
  totalTokens: response.usage.total_tokens,
516
- reasoningTokens: (_c = (_b = response.usage.completion_tokens_details) == null ? void 0 : _b.reasoning_tokens) != null ? _c : void 0
516
+ reasoningTokens: (_c = (_b = response.usage.completion_tokens_details) == null ? void 0 : _b.reasoning_tokens) != null ? _c : void 0,
517
+ cachedInputTokens: (_e = (_d = response.usage.prompt_tokens_details) == null ? void 0 : _d.cached_tokens) != null ? _e : void 0
517
518
  },
518
519
  request: { body },
519
520
  response: {
@@ -547,7 +548,9 @@ var XaiChatLanguageModel = class {
547
548
  const usage = {
548
549
  inputTokens: void 0,
549
550
  outputTokens: void 0,
550
- totalTokens: void 0
551
+ totalTokens: void 0,
552
+ reasoningTokens: void 0,
553
+ cachedInputTokens: void 0
551
554
  };
552
555
  let isFirstChunk = true;
553
556
  const contentBlocks = {};
@@ -560,7 +563,7 @@ var XaiChatLanguageModel = class {
560
563
  controller.enqueue({ type: "stream-start", warnings });
561
564
  },
562
565
  transform(chunk, controller) {
563
- var _a2, _b;
566
+ var _a2, _b, _c, _d;
564
567
  if (options.includeRawChunks) {
565
568
  controller.enqueue({ type: "raw", rawValue: chunk.rawValue });
566
569
  }
@@ -591,6 +594,7 @@ var XaiChatLanguageModel = class {
591
594
  usage.outputTokens = value.usage.completion_tokens;
592
595
  usage.totalTokens = value.usage.total_tokens;
593
596
  usage.reasoningTokens = (_b = (_a2 = value.usage.completion_tokens_details) == null ? void 0 : _a2.reasoning_tokens) != null ? _b : void 0;
597
+ usage.cachedInputTokens = (_d = (_c = value.usage.prompt_tokens_details) == null ? void 0 : _c.cached_tokens) != null ? _d : void 0;
594
598
  }
595
599
  const choice = value.choices[0];
596
600
  if ((choice == null ? void 0 : choice.finish_reason) != null) {
@@ -686,8 +690,17 @@ var xaiUsageSchema = z3.object({
686
690
  prompt_tokens: z3.number(),
687
691
  completion_tokens: z3.number(),
688
692
  total_tokens: z3.number(),
693
+ prompt_tokens_details: z3.object({
694
+ text_tokens: z3.number().nullish(),
695
+ audio_tokens: z3.number().nullish(),
696
+ image_tokens: z3.number().nullish(),
697
+ cached_tokens: z3.number().nullish()
698
+ }).nullish(),
689
699
  completion_tokens_details: z3.object({
690
- reasoning_tokens: z3.number().nullish()
700
+ reasoning_tokens: z3.number().nullish(),
701
+ audio_tokens: z3.number().nullish(),
702
+ accepted_prediction_tokens: z3.number().nullish(),
703
+ rejected_prediction_tokens: z3.number().nullish()
691
704
  }).nullish()
692
705
  });
693
706
  var xaiChatResponseSchema = z3.object({
@@ -1834,7 +1847,7 @@ var xaiTools = {
1834
1847
  };
1835
1848
 
1836
1849
  // src/version.ts
1837
- var VERSION = true ? "3.0.0-beta.38" : "0.0.0-test";
1850
+ var VERSION = true ? "3.0.0-beta.39" : "0.0.0-test";
1838
1851
 
1839
1852
  // src/xai-provider.ts
1840
1853
  var xaiErrorStructure = {