ai 5.0.0-canary.17 → 5.0.0-canary.18
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +19 -0
- package/README.md +4 -4
- package/dist/index.d.mts +39 -87
- package/dist/index.d.ts +39 -87
- package/dist/index.js +177 -166
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +177 -166
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +6 -44
- package/dist/internal/index.d.ts +6 -44
- package/dist/internal/index.js +12 -20
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +12 -19
- package/dist/internal/index.mjs.map +1 -1
- package/dist/mcp-stdio/index.d.mts +2 -2
- package/dist/mcp-stdio/index.d.ts +2 -2
- package/package.json +4 -4
@@ -1,20 +1,7 @@
|
|
1
|
-
import
|
2
|
-
import { SharedV2ProviderOptions, JSONValue as JSONValue$1, LanguageModelV2Source, LanguageModelV2Usage, LanguageModelV2FinishReason, JSONObject, LanguageModelV2FunctionTool, LanguageModelV2ProviderDefinedTool, LanguageModelV2ToolChoice, LanguageModelV2Prompt } from '@ai-sdk/provider';
|
1
|
+
import { ToolResultContent, ToolCall, ToolResult, Schema } from '@ai-sdk/provider-utils';
|
3
2
|
import { z } from 'zod';
|
4
|
-
import
|
5
|
-
|
6
|
-
type ToolResultContent = Array<{
|
7
|
-
type: 'text';
|
8
|
-
text: string;
|
9
|
-
} | {
|
10
|
-
type: 'image';
|
11
|
-
data: string;
|
12
|
-
mediaType?: string;
|
13
|
-
/**
|
14
|
-
* @deprecated Use `mediaType` instead.
|
15
|
-
*/
|
16
|
-
mimeType?: string;
|
17
|
-
}>;
|
3
|
+
import * as _ai_sdk_provider from '@ai-sdk/provider';
|
4
|
+
import { SharedV2ProviderOptions, JSONValue as JSONValue$1, LanguageModelV2Source, LanguageModelV2FinishReason, LanguageModelV2Usage, JSONObject, LanguageModelV2FunctionTool, LanguageModelV2ProviderDefinedTool, LanguageModelV2ToolChoice, LanguageModelV2Prompt } from '@ai-sdk/provider';
|
18
5
|
|
19
6
|
/**
|
20
7
|
Additional provider-specific options.
|
@@ -394,25 +381,6 @@ type StepStartUIPart = {
|
|
394
381
|
type: 'step-start';
|
395
382
|
};
|
396
383
|
|
397
|
-
/**
|
398
|
-
Represents the number of tokens used in a prompt and completion.
|
399
|
-
*/
|
400
|
-
type LanguageModelUsage = {
|
401
|
-
/**
|
402
|
-
The number of tokens used in the prompt.
|
403
|
-
*/
|
404
|
-
promptTokens: number;
|
405
|
-
/**
|
406
|
-
The number of tokens used in the completion.
|
407
|
-
*/
|
408
|
-
completionTokens: number;
|
409
|
-
/**
|
410
|
-
The total number of tokens used (promptTokens + completionTokens).
|
411
|
-
*/
|
412
|
-
totalTokens: number;
|
413
|
-
};
|
414
|
-
declare function calculateLanguageModelUsage({ inputTokens, outputTokens, }: LanguageModelV2Usage): LanguageModelUsage;
|
415
|
-
|
416
384
|
type DataStreamString = `${(typeof DataStreamStringPrefixes)[keyof typeof DataStreamStringPrefixes]}:${string}\n`;
|
417
385
|
interface DataStreamPart<CODE extends string, NAME extends string, TYPE> {
|
418
386
|
code: CODE;
|
@@ -430,17 +398,11 @@ declare const dataStreamParts: readonly [DataStreamPart<"0", "text", string>, Da
|
|
430
398
|
argsTextDelta: string;
|
431
399
|
}>, DataStreamPart<"d", "finish_message", {
|
432
400
|
finishReason: LanguageModelV2FinishReason;
|
433
|
-
usage?:
|
434
|
-
promptTokens: number;
|
435
|
-
completionTokens: number;
|
436
|
-
};
|
401
|
+
usage?: LanguageModelV2Usage;
|
437
402
|
}>, DataStreamPart<"e", "finish_step", {
|
438
403
|
isContinued: boolean;
|
439
404
|
finishReason: LanguageModelV2FinishReason;
|
440
|
-
usage?:
|
441
|
-
promptTokens: number;
|
442
|
-
completionTokens: number;
|
443
|
-
};
|
405
|
+
usage?: LanguageModelV2Usage;
|
444
406
|
}>, DataStreamPart<"f", "start_step", {
|
445
407
|
messageId: string;
|
446
408
|
}>, DataStreamPart<"g", "reasoning", {
|
@@ -825,4 +787,4 @@ declare function createCallbacksTransformer(callbacks?: StreamCallbacks | undefi
|
|
825
787
|
*/
|
826
788
|
declare const HANGING_STREAM_WARNING_TIME_MS: number;
|
827
789
|
|
828
|
-
export { DataStreamWriter, HANGING_STREAM_WARNING_TIME_MS, StreamCallbacks, StreamData,
|
790
|
+
export { DataStreamWriter, HANGING_STREAM_WARNING_TIME_MS, StreamCallbacks, StreamData, convertToLanguageModelPrompt, createCallbacksTransformer, formatDataStreamPart, mergeStreams, prepareCallSettings, prepareResponseHeaders, prepareRetries, prepareToolsAndToolChoice, standardizePrompt };
|
package/dist/internal/index.d.ts
CHANGED
@@ -1,20 +1,7 @@
|
|
1
|
-
import
|
2
|
-
import { SharedV2ProviderOptions, JSONValue as JSONValue$1, LanguageModelV2Source, LanguageModelV2Usage, LanguageModelV2FinishReason, JSONObject, LanguageModelV2FunctionTool, LanguageModelV2ProviderDefinedTool, LanguageModelV2ToolChoice, LanguageModelV2Prompt } from '@ai-sdk/provider';
|
1
|
+
import { ToolResultContent, ToolCall, ToolResult, Schema } from '@ai-sdk/provider-utils';
|
3
2
|
import { z } from 'zod';
|
4
|
-
import
|
5
|
-
|
6
|
-
type ToolResultContent = Array<{
|
7
|
-
type: 'text';
|
8
|
-
text: string;
|
9
|
-
} | {
|
10
|
-
type: 'image';
|
11
|
-
data: string;
|
12
|
-
mediaType?: string;
|
13
|
-
/**
|
14
|
-
* @deprecated Use `mediaType` instead.
|
15
|
-
*/
|
16
|
-
mimeType?: string;
|
17
|
-
}>;
|
3
|
+
import * as _ai_sdk_provider from '@ai-sdk/provider';
|
4
|
+
import { SharedV2ProviderOptions, JSONValue as JSONValue$1, LanguageModelV2Source, LanguageModelV2FinishReason, LanguageModelV2Usage, JSONObject, LanguageModelV2FunctionTool, LanguageModelV2ProviderDefinedTool, LanguageModelV2ToolChoice, LanguageModelV2Prompt } from '@ai-sdk/provider';
|
18
5
|
|
19
6
|
/**
|
20
7
|
Additional provider-specific options.
|
@@ -394,25 +381,6 @@ type StepStartUIPart = {
|
|
394
381
|
type: 'step-start';
|
395
382
|
};
|
396
383
|
|
397
|
-
/**
|
398
|
-
Represents the number of tokens used in a prompt and completion.
|
399
|
-
*/
|
400
|
-
type LanguageModelUsage = {
|
401
|
-
/**
|
402
|
-
The number of tokens used in the prompt.
|
403
|
-
*/
|
404
|
-
promptTokens: number;
|
405
|
-
/**
|
406
|
-
The number of tokens used in the completion.
|
407
|
-
*/
|
408
|
-
completionTokens: number;
|
409
|
-
/**
|
410
|
-
The total number of tokens used (promptTokens + completionTokens).
|
411
|
-
*/
|
412
|
-
totalTokens: number;
|
413
|
-
};
|
414
|
-
declare function calculateLanguageModelUsage({ inputTokens, outputTokens, }: LanguageModelV2Usage): LanguageModelUsage;
|
415
|
-
|
416
384
|
type DataStreamString = `${(typeof DataStreamStringPrefixes)[keyof typeof DataStreamStringPrefixes]}:${string}\n`;
|
417
385
|
interface DataStreamPart<CODE extends string, NAME extends string, TYPE> {
|
418
386
|
code: CODE;
|
@@ -430,17 +398,11 @@ declare const dataStreamParts: readonly [DataStreamPart<"0", "text", string>, Da
|
|
430
398
|
argsTextDelta: string;
|
431
399
|
}>, DataStreamPart<"d", "finish_message", {
|
432
400
|
finishReason: LanguageModelV2FinishReason;
|
433
|
-
usage?:
|
434
|
-
promptTokens: number;
|
435
|
-
completionTokens: number;
|
436
|
-
};
|
401
|
+
usage?: LanguageModelV2Usage;
|
437
402
|
}>, DataStreamPart<"e", "finish_step", {
|
438
403
|
isContinued: boolean;
|
439
404
|
finishReason: LanguageModelV2FinishReason;
|
440
|
-
usage?:
|
441
|
-
promptTokens: number;
|
442
|
-
completionTokens: number;
|
443
|
-
};
|
405
|
+
usage?: LanguageModelV2Usage;
|
444
406
|
}>, DataStreamPart<"f", "start_step", {
|
445
407
|
messageId: string;
|
446
408
|
}>, DataStreamPart<"g", "reasoning", {
|
@@ -825,4 +787,4 @@ declare function createCallbacksTransformer(callbacks?: StreamCallbacks | undefi
|
|
825
787
|
*/
|
826
788
|
declare const HANGING_STREAM_WARNING_TIME_MS: number;
|
827
789
|
|
828
|
-
export { DataStreamWriter, HANGING_STREAM_WARNING_TIME_MS, StreamCallbacks, StreamData,
|
790
|
+
export { DataStreamWriter, HANGING_STREAM_WARNING_TIME_MS, StreamCallbacks, StreamData, convertToLanguageModelPrompt, createCallbacksTransformer, formatDataStreamPart, mergeStreams, prepareCallSettings, prepareResponseHeaders, prepareRetries, prepareToolsAndToolChoice, standardizePrompt };
|
package/dist/internal/index.js
CHANGED
@@ -22,7 +22,6 @@ var internal_exports = {};
|
|
22
22
|
__export(internal_exports, {
|
23
23
|
HANGING_STREAM_WARNING_TIME_MS: () => HANGING_STREAM_WARNING_TIME_MS,
|
24
24
|
StreamData: () => StreamData,
|
25
|
-
calculateLanguageModelUsage: () => calculateLanguageModelUsage,
|
26
25
|
convertToLanguageModelPrompt: () => convertToLanguageModelPrompt,
|
27
26
|
createCallbacksTransformer: () => createCallbacksTransformer,
|
28
27
|
formatDataStreamPart: () => formatDataStreamPart,
|
@@ -638,10 +637,13 @@ var finishMessageStreamPart = {
|
|
638
637
|
const result = {
|
639
638
|
finishReason: value.finishReason
|
640
639
|
};
|
641
|
-
if ("usage" in value && value.usage != null && typeof value.usage === "object"
|
640
|
+
if ("usage" in value && value.usage != null && typeof value.usage === "object") {
|
642
641
|
result.usage = {
|
643
|
-
|
644
|
-
|
642
|
+
inputTokens: "inputTokens" in value.usage && typeof value.usage.inputTokens === "number" ? value.usage.inputTokens : void 0,
|
643
|
+
outputTokens: "outputTokens" in value.usage && typeof value.usage.outputTokens === "number" ? value.usage.outputTokens : void 0,
|
644
|
+
totalTokens: "totalTokens" in value.usage && typeof value.usage.totalTokens === "number" ? value.usage.totalTokens : void 0,
|
645
|
+
reasoningTokens: "reasoningTokens" in value.usage && typeof value.usage.reasoningTokens === "number" ? value.usage.reasoningTokens : void 0,
|
646
|
+
cachedInputTokens: "cachedInputTokens" in value.usage && typeof value.usage.cachedInputTokens === "number" ? value.usage.cachedInputTokens : void 0
|
645
647
|
};
|
646
648
|
}
|
647
649
|
return {
|
@@ -663,10 +665,13 @@ var finishStepStreamPart = {
|
|
663
665
|
finishReason: value.finishReason,
|
664
666
|
isContinued: false
|
665
667
|
};
|
666
|
-
if ("usage" in value && value.usage != null && typeof value.usage === "object"
|
668
|
+
if ("usage" in value && value.usage != null && typeof value.usage === "object") {
|
667
669
|
result.usage = {
|
668
|
-
|
669
|
-
|
670
|
+
inputTokens: "inputTokens" in value.usage && typeof value.usage.inputTokens === "number" ? value.usage.inputTokens : void 0,
|
671
|
+
outputTokens: "outputTokens" in value.usage && typeof value.usage.outputTokens === "number" ? value.usage.outputTokens : void 0,
|
672
|
+
totalTokens: "totalTokens" in value.usage && typeof value.usage.totalTokens === "number" ? value.usage.totalTokens : void 0,
|
673
|
+
reasoningTokens: "reasoningTokens" in value.usage && typeof value.usage.reasoningTokens === "number" ? value.usage.reasoningTokens : void 0,
|
674
|
+
cachedInputTokens: "cachedInputTokens" in value.usage && typeof value.usage.cachedInputTokens === "number" ? value.usage.cachedInputTokens : void 0
|
670
675
|
};
|
671
676
|
}
|
672
677
|
if ("isContinued" in value && typeof value.isContinued === "boolean") {
|
@@ -1445,18 +1450,6 @@ function convertPartToLanguageModelPart(part, downloadedAssets) {
|
|
1445
1450
|
}
|
1446
1451
|
}
|
1447
1452
|
|
1448
|
-
// core/types/usage.ts
|
1449
|
-
function calculateLanguageModelUsage({
|
1450
|
-
inputTokens,
|
1451
|
-
outputTokens
|
1452
|
-
}) {
|
1453
|
-
return {
|
1454
|
-
promptTokens: inputTokens != null ? inputTokens : NaN,
|
1455
|
-
completionTokens: outputTokens != null ? outputTokens : NaN,
|
1456
|
-
totalTokens: (inputTokens != null ? inputTokens : 0) + (outputTokens != null ? outputTokens : 0)
|
1457
|
-
};
|
1458
|
-
}
|
1459
|
-
|
1460
1453
|
// core/util/prepare-response-headers.ts
|
1461
1454
|
function prepareResponseHeaders(headers, {
|
1462
1455
|
contentType,
|
@@ -1658,7 +1651,6 @@ var StreamData = class {
|
|
1658
1651
|
0 && (module.exports = {
|
1659
1652
|
HANGING_STREAM_WARNING_TIME_MS,
|
1660
1653
|
StreamData,
|
1661
|
-
calculateLanguageModelUsage,
|
1662
1654
|
convertToLanguageModelPrompt,
|
1663
1655
|
createCallbacksTransformer,
|
1664
1656
|
formatDataStreamPart,
|