ai 5.0.0-canary.17 → 5.0.0-canary.18

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,20 +1,7 @@
1
- import * as _ai_sdk_provider from '@ai-sdk/provider';
2
- import { SharedV2ProviderOptions, JSONValue as JSONValue$1, LanguageModelV2Source, LanguageModelV2Usage, LanguageModelV2FinishReason, JSONObject, LanguageModelV2FunctionTool, LanguageModelV2ProviderDefinedTool, LanguageModelV2ToolChoice, LanguageModelV2Prompt } from '@ai-sdk/provider';
1
+ import { ToolResultContent, ToolCall, ToolResult, Schema } from '@ai-sdk/provider-utils';
3
2
  import { z } from 'zod';
4
- import { ToolCall, ToolResult, Schema } from '@ai-sdk/provider-utils';
5
-
6
- type ToolResultContent = Array<{
7
- type: 'text';
8
- text: string;
9
- } | {
10
- type: 'image';
11
- data: string;
12
- mediaType?: string;
13
- /**
14
- * @deprecated Use `mediaType` instead.
15
- */
16
- mimeType?: string;
17
- }>;
3
+ import * as _ai_sdk_provider from '@ai-sdk/provider';
4
+ import { SharedV2ProviderOptions, JSONValue as JSONValue$1, LanguageModelV2Source, LanguageModelV2FinishReason, LanguageModelV2Usage, JSONObject, LanguageModelV2FunctionTool, LanguageModelV2ProviderDefinedTool, LanguageModelV2ToolChoice, LanguageModelV2Prompt } from '@ai-sdk/provider';
18
5
 
19
6
  /**
20
7
  Additional provider-specific options.
@@ -394,25 +381,6 @@ type StepStartUIPart = {
394
381
  type: 'step-start';
395
382
  };
396
383
 
397
- /**
398
- Represents the number of tokens used in a prompt and completion.
399
- */
400
- type LanguageModelUsage = {
401
- /**
402
- The number of tokens used in the prompt.
403
- */
404
- promptTokens: number;
405
- /**
406
- The number of tokens used in the completion.
407
- */
408
- completionTokens: number;
409
- /**
410
- The total number of tokens used (promptTokens + completionTokens).
411
- */
412
- totalTokens: number;
413
- };
414
- declare function calculateLanguageModelUsage({ inputTokens, outputTokens, }: LanguageModelV2Usage): LanguageModelUsage;
415
-
416
384
  type DataStreamString = `${(typeof DataStreamStringPrefixes)[keyof typeof DataStreamStringPrefixes]}:${string}\n`;
417
385
  interface DataStreamPart<CODE extends string, NAME extends string, TYPE> {
418
386
  code: CODE;
@@ -430,17 +398,11 @@ declare const dataStreamParts: readonly [DataStreamPart<"0", "text", string>, Da
430
398
  argsTextDelta: string;
431
399
  }>, DataStreamPart<"d", "finish_message", {
432
400
  finishReason: LanguageModelV2FinishReason;
433
- usage?: {
434
- promptTokens: number;
435
- completionTokens: number;
436
- };
401
+ usage?: LanguageModelV2Usage;
437
402
  }>, DataStreamPart<"e", "finish_step", {
438
403
  isContinued: boolean;
439
404
  finishReason: LanguageModelV2FinishReason;
440
- usage?: {
441
- promptTokens: number;
442
- completionTokens: number;
443
- };
405
+ usage?: LanguageModelV2Usage;
444
406
  }>, DataStreamPart<"f", "start_step", {
445
407
  messageId: string;
446
408
  }>, DataStreamPart<"g", "reasoning", {
@@ -825,4 +787,4 @@ declare function createCallbacksTransformer(callbacks?: StreamCallbacks | undefi
825
787
  */
826
788
  declare const HANGING_STREAM_WARNING_TIME_MS: number;
827
789
 
828
- export { DataStreamWriter, HANGING_STREAM_WARNING_TIME_MS, StreamCallbacks, StreamData, calculateLanguageModelUsage, convertToLanguageModelPrompt, createCallbacksTransformer, formatDataStreamPart, mergeStreams, prepareCallSettings, prepareResponseHeaders, prepareRetries, prepareToolsAndToolChoice, standardizePrompt };
790
+ export { DataStreamWriter, HANGING_STREAM_WARNING_TIME_MS, StreamCallbacks, StreamData, convertToLanguageModelPrompt, createCallbacksTransformer, formatDataStreamPart, mergeStreams, prepareCallSettings, prepareResponseHeaders, prepareRetries, prepareToolsAndToolChoice, standardizePrompt };
@@ -1,20 +1,7 @@
1
- import * as _ai_sdk_provider from '@ai-sdk/provider';
2
- import { SharedV2ProviderOptions, JSONValue as JSONValue$1, LanguageModelV2Source, LanguageModelV2Usage, LanguageModelV2FinishReason, JSONObject, LanguageModelV2FunctionTool, LanguageModelV2ProviderDefinedTool, LanguageModelV2ToolChoice, LanguageModelV2Prompt } from '@ai-sdk/provider';
1
+ import { ToolResultContent, ToolCall, ToolResult, Schema } from '@ai-sdk/provider-utils';
3
2
  import { z } from 'zod';
4
- import { ToolCall, ToolResult, Schema } from '@ai-sdk/provider-utils';
5
-
6
- type ToolResultContent = Array<{
7
- type: 'text';
8
- text: string;
9
- } | {
10
- type: 'image';
11
- data: string;
12
- mediaType?: string;
13
- /**
14
- * @deprecated Use `mediaType` instead.
15
- */
16
- mimeType?: string;
17
- }>;
3
+ import * as _ai_sdk_provider from '@ai-sdk/provider';
4
+ import { SharedV2ProviderOptions, JSONValue as JSONValue$1, LanguageModelV2Source, LanguageModelV2FinishReason, LanguageModelV2Usage, JSONObject, LanguageModelV2FunctionTool, LanguageModelV2ProviderDefinedTool, LanguageModelV2ToolChoice, LanguageModelV2Prompt } from '@ai-sdk/provider';
18
5
 
19
6
  /**
20
7
  Additional provider-specific options.
@@ -394,25 +381,6 @@ type StepStartUIPart = {
394
381
  type: 'step-start';
395
382
  };
396
383
 
397
- /**
398
- Represents the number of tokens used in a prompt and completion.
399
- */
400
- type LanguageModelUsage = {
401
- /**
402
- The number of tokens used in the prompt.
403
- */
404
- promptTokens: number;
405
- /**
406
- The number of tokens used in the completion.
407
- */
408
- completionTokens: number;
409
- /**
410
- The total number of tokens used (promptTokens + completionTokens).
411
- */
412
- totalTokens: number;
413
- };
414
- declare function calculateLanguageModelUsage({ inputTokens, outputTokens, }: LanguageModelV2Usage): LanguageModelUsage;
415
-
416
384
  type DataStreamString = `${(typeof DataStreamStringPrefixes)[keyof typeof DataStreamStringPrefixes]}:${string}\n`;
417
385
  interface DataStreamPart<CODE extends string, NAME extends string, TYPE> {
418
386
  code: CODE;
@@ -430,17 +398,11 @@ declare const dataStreamParts: readonly [DataStreamPart<"0", "text", string>, Da
430
398
  argsTextDelta: string;
431
399
  }>, DataStreamPart<"d", "finish_message", {
432
400
  finishReason: LanguageModelV2FinishReason;
433
- usage?: {
434
- promptTokens: number;
435
- completionTokens: number;
436
- };
401
+ usage?: LanguageModelV2Usage;
437
402
  }>, DataStreamPart<"e", "finish_step", {
438
403
  isContinued: boolean;
439
404
  finishReason: LanguageModelV2FinishReason;
440
- usage?: {
441
- promptTokens: number;
442
- completionTokens: number;
443
- };
405
+ usage?: LanguageModelV2Usage;
444
406
  }>, DataStreamPart<"f", "start_step", {
445
407
  messageId: string;
446
408
  }>, DataStreamPart<"g", "reasoning", {
@@ -825,4 +787,4 @@ declare function createCallbacksTransformer(callbacks?: StreamCallbacks | undefi
825
787
  */
826
788
  declare const HANGING_STREAM_WARNING_TIME_MS: number;
827
789
 
828
- export { DataStreamWriter, HANGING_STREAM_WARNING_TIME_MS, StreamCallbacks, StreamData, calculateLanguageModelUsage, convertToLanguageModelPrompt, createCallbacksTransformer, formatDataStreamPart, mergeStreams, prepareCallSettings, prepareResponseHeaders, prepareRetries, prepareToolsAndToolChoice, standardizePrompt };
790
+ export { DataStreamWriter, HANGING_STREAM_WARNING_TIME_MS, StreamCallbacks, StreamData, convertToLanguageModelPrompt, createCallbacksTransformer, formatDataStreamPart, mergeStreams, prepareCallSettings, prepareResponseHeaders, prepareRetries, prepareToolsAndToolChoice, standardizePrompt };
@@ -22,7 +22,6 @@ var internal_exports = {};
22
22
  __export(internal_exports, {
23
23
  HANGING_STREAM_WARNING_TIME_MS: () => HANGING_STREAM_WARNING_TIME_MS,
24
24
  StreamData: () => StreamData,
25
- calculateLanguageModelUsage: () => calculateLanguageModelUsage,
26
25
  convertToLanguageModelPrompt: () => convertToLanguageModelPrompt,
27
26
  createCallbacksTransformer: () => createCallbacksTransformer,
28
27
  formatDataStreamPart: () => formatDataStreamPart,
@@ -638,10 +637,13 @@ var finishMessageStreamPart = {
638
637
  const result = {
639
638
  finishReason: value.finishReason
640
639
  };
641
- if ("usage" in value && value.usage != null && typeof value.usage === "object" && "promptTokens" in value.usage && "completionTokens" in value.usage) {
640
+ if ("usage" in value && value.usage != null && typeof value.usage === "object") {
642
641
  result.usage = {
643
- promptTokens: typeof value.usage.promptTokens === "number" ? value.usage.promptTokens : Number.NaN,
644
- completionTokens: typeof value.usage.completionTokens === "number" ? value.usage.completionTokens : Number.NaN
642
+ inputTokens: "inputTokens" in value.usage && typeof value.usage.inputTokens === "number" ? value.usage.inputTokens : void 0,
643
+ outputTokens: "outputTokens" in value.usage && typeof value.usage.outputTokens === "number" ? value.usage.outputTokens : void 0,
644
+ totalTokens: "totalTokens" in value.usage && typeof value.usage.totalTokens === "number" ? value.usage.totalTokens : void 0,
645
+ reasoningTokens: "reasoningTokens" in value.usage && typeof value.usage.reasoningTokens === "number" ? value.usage.reasoningTokens : void 0,
646
+ cachedInputTokens: "cachedInputTokens" in value.usage && typeof value.usage.cachedInputTokens === "number" ? value.usage.cachedInputTokens : void 0
645
647
  };
646
648
  }
647
649
  return {
@@ -663,10 +665,13 @@ var finishStepStreamPart = {
663
665
  finishReason: value.finishReason,
664
666
  isContinued: false
665
667
  };
666
- if ("usage" in value && value.usage != null && typeof value.usage === "object" && "promptTokens" in value.usage && "completionTokens" in value.usage) {
668
+ if ("usage" in value && value.usage != null && typeof value.usage === "object") {
667
669
  result.usage = {
668
- promptTokens: typeof value.usage.promptTokens === "number" ? value.usage.promptTokens : Number.NaN,
669
- completionTokens: typeof value.usage.completionTokens === "number" ? value.usage.completionTokens : Number.NaN
670
+ inputTokens: "inputTokens" in value.usage && typeof value.usage.inputTokens === "number" ? value.usage.inputTokens : void 0,
671
+ outputTokens: "outputTokens" in value.usage && typeof value.usage.outputTokens === "number" ? value.usage.outputTokens : void 0,
672
+ totalTokens: "totalTokens" in value.usage && typeof value.usage.totalTokens === "number" ? value.usage.totalTokens : void 0,
673
+ reasoningTokens: "reasoningTokens" in value.usage && typeof value.usage.reasoningTokens === "number" ? value.usage.reasoningTokens : void 0,
674
+ cachedInputTokens: "cachedInputTokens" in value.usage && typeof value.usage.cachedInputTokens === "number" ? value.usage.cachedInputTokens : void 0
670
675
  };
671
676
  }
672
677
  if ("isContinued" in value && typeof value.isContinued === "boolean") {
@@ -1445,18 +1450,6 @@ function convertPartToLanguageModelPart(part, downloadedAssets) {
1445
1450
  }
1446
1451
  }
1447
1452
 
1448
- // core/types/usage.ts
1449
- function calculateLanguageModelUsage({
1450
- inputTokens,
1451
- outputTokens
1452
- }) {
1453
- return {
1454
- promptTokens: inputTokens != null ? inputTokens : NaN,
1455
- completionTokens: outputTokens != null ? outputTokens : NaN,
1456
- totalTokens: (inputTokens != null ? inputTokens : 0) + (outputTokens != null ? outputTokens : 0)
1457
- };
1458
- }
1459
-
1460
1453
  // core/util/prepare-response-headers.ts
1461
1454
  function prepareResponseHeaders(headers, {
1462
1455
  contentType,
@@ -1658,7 +1651,6 @@ var StreamData = class {
1658
1651
  0 && (module.exports = {
1659
1652
  HANGING_STREAM_WARNING_TIME_MS,
1660
1653
  StreamData,
1661
- calculateLanguageModelUsage,
1662
1654
  convertToLanguageModelPrompt,
1663
1655
  createCallbacksTransformer,
1664
1656
  formatDataStreamPart,