ai 5.0.0-canary.13 → 5.0.0-canary.15
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +30 -0
- package/dist/index.d.mts +133 -294
- package/dist/index.d.ts +133 -294
- package/dist/index.js +174 -242
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +121 -174
- package/dist/index.mjs.map +1 -1
- package/dist/internal/index.d.mts +24 -65
- package/dist/internal/index.d.ts +24 -65
- package/dist/internal/index.js +10 -69
- package/dist/internal/index.js.map +1 -1
- package/dist/internal/index.mjs +5 -50
- package/dist/internal/index.mjs.map +1 -1
- package/dist/test/index.d.mts +4 -3
- package/dist/test/index.d.ts +4 -3
- package/dist/test/index.js +5 -2
- package/dist/test/index.js.map +1 -1
- package/dist/test/index.mjs +5 -2
- package/dist/test/index.mjs.map +1 -1
- package/package.json +3 -4
package/dist/index.js
CHANGED
@@ -1,9 +1,7 @@
|
|
1
1
|
"use strict";
|
2
|
-
var __create = Object.create;
|
3
2
|
var __defProp = Object.defineProperty;
|
4
3
|
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
|
5
4
|
var __getOwnPropNames = Object.getOwnPropertyNames;
|
6
|
-
var __getProtoOf = Object.getPrototypeOf;
|
7
5
|
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
8
6
|
var __export = (target, all) => {
|
9
7
|
for (var name17 in all)
|
@@ -17,14 +15,6 @@ var __copyProps = (to, from, except, desc) => {
|
|
17
15
|
}
|
18
16
|
return to;
|
19
17
|
};
|
20
|
-
var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
|
21
|
-
// If the importer is in node compatibility mode or this is not an ESM
|
22
|
-
// file that has been converted to a CommonJS file using a Babel-
|
23
|
-
// compatible transform (i.e. "__esModule" has not been set), then set
|
24
|
-
// "default" to the CommonJS "module.exports" for node compatibility.
|
25
|
-
isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
|
26
|
-
mod
|
27
|
-
));
|
28
18
|
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
|
29
19
|
|
30
20
|
// index.ts
|
@@ -61,7 +51,7 @@ __export(ai_exports, {
|
|
61
51
|
UnsupportedFunctionalityError: () => import_provider21.UnsupportedFunctionalityError,
|
62
52
|
appendClientMessage: () => appendClientMessage,
|
63
53
|
appendResponseMessages: () => appendResponseMessages,
|
64
|
-
asSchema: () => asSchema,
|
54
|
+
asSchema: () => import_provider_utils4.asSchema,
|
65
55
|
callChatApi: () => callChatApi,
|
66
56
|
callCompletionApi: () => callCompletionApi,
|
67
57
|
convertToCoreMessages: () => convertToCoreMessages,
|
@@ -73,7 +63,7 @@ __export(ai_exports, {
|
|
73
63
|
cosineSimilarity: () => cosineSimilarity,
|
74
64
|
createDataStream: () => createDataStream,
|
75
65
|
createDataStreamResponse: () => createDataStreamResponse,
|
76
|
-
createIdGenerator: () =>
|
66
|
+
createIdGenerator: () => import_provider_utils21.createIdGenerator,
|
77
67
|
createProviderRegistry: () => createProviderRegistry,
|
78
68
|
customProvider: () => customProvider,
|
79
69
|
defaultSettingsMiddleware: () => defaultSettingsMiddleware,
|
@@ -89,14 +79,14 @@ __export(ai_exports, {
|
|
89
79
|
extractReasoningMiddleware: () => extractReasoningMiddleware,
|
90
80
|
fillMessageParts: () => fillMessageParts,
|
91
81
|
formatDataStreamPart: () => formatDataStreamPart,
|
92
|
-
generateId: () =>
|
82
|
+
generateId: () => import_provider_utils21.generateId,
|
93
83
|
generateObject: () => generateObject,
|
94
84
|
generateText: () => generateText,
|
95
85
|
getMessageParts: () => getMessageParts,
|
96
86
|
getTextFromDataUrl: () => getTextFromDataUrl,
|
97
87
|
isAssistantMessageWithCompletedToolCalls: () => isAssistantMessageWithCompletedToolCalls,
|
98
88
|
isDeepEqualData: () => isDeepEqualData,
|
99
|
-
jsonSchema: () => jsonSchema,
|
89
|
+
jsonSchema: () => import_provider_utils4.jsonSchema,
|
100
90
|
parseDataStreamPart: () => parseDataStreamPart,
|
101
91
|
parsePartialJson: () => parsePartialJson,
|
102
92
|
pipeDataStreamToResponse: () => pipeDataStreamToResponse,
|
@@ -111,16 +101,15 @@ __export(ai_exports, {
|
|
111
101
|
streamText: () => streamText,
|
112
102
|
tool: () => tool,
|
113
103
|
updateToolCallResult: () => updateToolCallResult,
|
114
|
-
wrapLanguageModel: () => wrapLanguageModel
|
115
|
-
zodSchema: () => zodSchema
|
104
|
+
wrapLanguageModel: () => wrapLanguageModel
|
116
105
|
});
|
117
106
|
module.exports = __toCommonJS(ai_exports);
|
118
107
|
|
119
108
|
// core/index.ts
|
120
|
-
var
|
109
|
+
var import_provider_utils21 = require("@ai-sdk/provider-utils");
|
121
110
|
|
122
111
|
// core/util/index.ts
|
123
|
-
var
|
112
|
+
var import_provider_utils4 = require("@ai-sdk/provider-utils");
|
124
113
|
|
125
114
|
// core/util/process-chat-response.ts
|
126
115
|
var import_provider_utils2 = require("@ai-sdk/provider-utils");
|
@@ -935,7 +924,6 @@ async function processChatResponse({
|
|
935
924
|
execUpdate();
|
936
925
|
},
|
937
926
|
onReasoningPart(value) {
|
938
|
-
var _a18;
|
939
927
|
if (currentReasoningPart == null) {
|
940
928
|
currentReasoningPart = {
|
941
929
|
type: "reasoning",
|
@@ -947,7 +935,6 @@ async function processChatResponse({
|
|
947
935
|
currentReasoningPart.reasoning += value.text;
|
948
936
|
currentReasoningPart.providerMetadata = value.providerMetadata;
|
949
937
|
}
|
950
|
-
message.reasoning = ((_a18 = message.reasoning) != null ? _a18 : "") + value.text;
|
951
938
|
execUpdate();
|
952
939
|
},
|
953
940
|
onReasoningPartFinish(value) {
|
@@ -1370,12 +1357,6 @@ function getMessageParts(message) {
|
|
1370
1357
|
type: "tool-invocation",
|
1371
1358
|
toolInvocation
|
1372
1359
|
})) : [],
|
1373
|
-
...message.reasoning ? [
|
1374
|
-
{
|
1375
|
-
type: "reasoning",
|
1376
|
-
reasoning: message.reasoning
|
1377
|
-
}
|
1378
|
-
] : [],
|
1379
1360
|
...message.content ? [{ type: "text", text: message.content }] : []
|
1380
1361
|
];
|
1381
1362
|
}
|
@@ -1455,51 +1436,29 @@ async function prepareAttachmentsForRequest(attachmentsFromOptions) {
|
|
1455
1436
|
throw new Error("Invalid attachments type");
|
1456
1437
|
}
|
1457
1438
|
|
1458
|
-
// core/util/
|
1459
|
-
|
1460
|
-
|
1461
|
-
|
1462
|
-
|
1463
|
-
|
1439
|
+
// core/util/update-tool-call-result.ts
|
1440
|
+
function updateToolCallResult({
|
1441
|
+
messages,
|
1442
|
+
toolCallId,
|
1443
|
+
toolResult: result
|
1444
|
+
}) {
|
1464
1445
|
var _a17;
|
1465
|
-
const
|
1466
|
-
|
1467
|
-
(
|
1468
|
-
$refStrategy: useReferences ? "root" : "none",
|
1469
|
-
target: "jsonSchema7"
|
1470
|
-
// note: openai mode breaks various gemini conversions
|
1471
|
-
}),
|
1472
|
-
{
|
1473
|
-
validate: (value) => {
|
1474
|
-
const result = zodSchema2.safeParse(value);
|
1475
|
-
return result.success ? { success: true, value: result.data } : { success: false, error: result.error };
|
1476
|
-
}
|
1477
|
-
}
|
1446
|
+
const lastMessage = messages[messages.length - 1];
|
1447
|
+
const invocationPart = lastMessage.parts.find(
|
1448
|
+
(part) => part.type === "tool-invocation" && part.toolInvocation.toolCallId === toolCallId
|
1478
1449
|
);
|
1479
|
-
|
1480
|
-
|
1481
|
-
|
1482
|
-
|
1483
|
-
|
1484
|
-
|
1485
|
-
|
1486
|
-
return {
|
1487
|
-
[schemaSymbol]: true,
|
1488
|
-
_type: void 0,
|
1489
|
-
// should never be used directly
|
1490
|
-
[import_provider_utils4.validatorSymbol]: true,
|
1491
|
-
jsonSchema: jsonSchema2,
|
1492
|
-
validate
|
1450
|
+
if (invocationPart == null) {
|
1451
|
+
return;
|
1452
|
+
}
|
1453
|
+
const toolResult = {
|
1454
|
+
...invocationPart.toolInvocation,
|
1455
|
+
state: "result",
|
1456
|
+
result
|
1493
1457
|
};
|
1494
|
-
|
1495
|
-
|
1496
|
-
|
1497
|
-
|
1498
|
-
function asSchema(schema) {
|
1499
|
-
return schema == null ? jsonSchema({
|
1500
|
-
properties: {},
|
1501
|
-
additionalProperties: false
|
1502
|
-
}) : isSchema(schema) ? schema : zodSchema(schema);
|
1458
|
+
invocationPart.toolInvocation = toolResult;
|
1459
|
+
lastMessage.toolInvocations = (_a17 = lastMessage.toolInvocations) == null ? void 0 : _a17.map(
|
1460
|
+
(toolInvocation) => toolInvocation.toolCallId === toolCallId ? toolResult : toolInvocation
|
1461
|
+
);
|
1503
1462
|
}
|
1504
1463
|
|
1505
1464
|
// core/util/should-resubmit-messages.ts
|
@@ -1531,31 +1490,6 @@ function isAssistantMessageWithCompletedToolCalls(message) {
|
|
1531
1490
|
return lastStepToolInvocations.length > 0 && lastStepToolInvocations.every((part) => "result" in part.toolInvocation);
|
1532
1491
|
}
|
1533
1492
|
|
1534
|
-
// core/util/update-tool-call-result.ts
|
1535
|
-
function updateToolCallResult({
|
1536
|
-
messages,
|
1537
|
-
toolCallId,
|
1538
|
-
toolResult: result
|
1539
|
-
}) {
|
1540
|
-
var _a17;
|
1541
|
-
const lastMessage = messages[messages.length - 1];
|
1542
|
-
const invocationPart = lastMessage.parts.find(
|
1543
|
-
(part) => part.type === "tool-invocation" && part.toolInvocation.toolCallId === toolCallId
|
1544
|
-
);
|
1545
|
-
if (invocationPart == null) {
|
1546
|
-
return;
|
1547
|
-
}
|
1548
|
-
const toolResult = {
|
1549
|
-
...invocationPart.toolInvocation,
|
1550
|
-
state: "result",
|
1551
|
-
result
|
1552
|
-
};
|
1553
|
-
invocationPart.toolInvocation = toolResult;
|
1554
|
-
lastMessage.toolInvocations = (_a17 = lastMessage.toolInvocations) == null ? void 0 : _a17.map(
|
1555
|
-
(toolInvocation) => toolInvocation.toolCallId === toolCallId ? toolResult : toolInvocation
|
1556
|
-
);
|
1557
|
-
}
|
1558
|
-
|
1559
1493
|
// core/data-stream/create-data-stream.ts
|
1560
1494
|
function createDataStream({
|
1561
1495
|
execute,
|
@@ -1764,7 +1698,7 @@ _a = symbol;
|
|
1764
1698
|
|
1765
1699
|
// util/retry-with-exponential-backoff.ts
|
1766
1700
|
var import_provider3 = require("@ai-sdk/provider");
|
1767
|
-
var
|
1701
|
+
var import_provider_utils5 = require("@ai-sdk/provider-utils");
|
1768
1702
|
|
1769
1703
|
// util/retry-error.ts
|
1770
1704
|
var import_provider2 = require("@ai-sdk/provider");
|
@@ -1808,13 +1742,13 @@ async function _retryWithExponentialBackoff(f, {
|
|
1808
1742
|
try {
|
1809
1743
|
return await f();
|
1810
1744
|
} catch (error) {
|
1811
|
-
if ((0,
|
1745
|
+
if ((0, import_provider_utils5.isAbortError)(error)) {
|
1812
1746
|
throw error;
|
1813
1747
|
}
|
1814
1748
|
if (maxRetries === 0) {
|
1815
1749
|
throw error;
|
1816
1750
|
}
|
1817
|
-
const errorMessage = (0,
|
1751
|
+
const errorMessage = (0, import_provider_utils5.getErrorMessage)(error);
|
1818
1752
|
const newErrors = [...errors, error];
|
1819
1753
|
const tryNumber = newErrors.length;
|
1820
1754
|
if (tryNumber > maxRetries) {
|
@@ -1825,7 +1759,7 @@ async function _retryWithExponentialBackoff(f, {
|
|
1825
1759
|
});
|
1826
1760
|
}
|
1827
1761
|
if (error instanceof Error && import_provider3.APICallError.isInstance(error) && error.isRetryable === true && tryNumber <= maxRetries) {
|
1828
|
-
await (0,
|
1762
|
+
await (0, import_provider_utils5.delay)(delayInMs);
|
1829
1763
|
return _retryWithExponentialBackoff(
|
1830
1764
|
f,
|
1831
1765
|
{ maxRetries, delayInMs: backoffFactor * delayInMs, backoffFactor },
|
@@ -2215,8 +2149,8 @@ async function embedMany({
|
|
2215
2149
|
}),
|
2216
2150
|
tracer,
|
2217
2151
|
fn: async (span) => {
|
2218
|
-
const maxEmbeddingsPerCall = model.maxEmbeddingsPerCall;
|
2219
|
-
if (maxEmbeddingsPerCall == null) {
|
2152
|
+
const maxEmbeddingsPerCall = await model.maxEmbeddingsPerCall;
|
2153
|
+
if (maxEmbeddingsPerCall == null || maxEmbeddingsPerCall === Infinity) {
|
2220
2154
|
const { embeddings: embeddings2, usage, response } = await retry(() => {
|
2221
2155
|
return recordSpan({
|
2222
2156
|
name: "ai.embedMany.doEmbed",
|
@@ -2394,7 +2328,7 @@ var NoImageGeneratedError = class extends import_provider4.AISDKError {
|
|
2394
2328
|
_a3 = symbol3;
|
2395
2329
|
|
2396
2330
|
// core/generate-text/generated-file.ts
|
2397
|
-
var
|
2331
|
+
var import_provider_utils6 = require("@ai-sdk/provider-utils");
|
2398
2332
|
var DefaultGeneratedFile = class {
|
2399
2333
|
constructor({
|
2400
2334
|
data,
|
@@ -2408,14 +2342,14 @@ var DefaultGeneratedFile = class {
|
|
2408
2342
|
// lazy conversion with caching to avoid unnecessary conversion overhead:
|
2409
2343
|
get base64() {
|
2410
2344
|
if (this.base64Data == null) {
|
2411
|
-
this.base64Data = (0,
|
2345
|
+
this.base64Data = (0, import_provider_utils6.convertUint8ArrayToBase64)(this.uint8ArrayData);
|
2412
2346
|
}
|
2413
2347
|
return this.base64Data;
|
2414
2348
|
}
|
2415
2349
|
// lazy conversion with caching to avoid unnecessary conversion overhead:
|
2416
2350
|
get uint8Array() {
|
2417
2351
|
if (this.uint8ArrayData == null) {
|
2418
|
-
this.uint8ArrayData = (0,
|
2352
|
+
this.uint8ArrayData = (0, import_provider_utils6.convertBase64ToUint8Array)(this.base64Data);
|
2419
2353
|
}
|
2420
2354
|
return this.uint8ArrayData;
|
2421
2355
|
}
|
@@ -2428,7 +2362,7 @@ var DefaultGeneratedFileWithType = class extends DefaultGeneratedFile {
|
|
2428
2362
|
};
|
2429
2363
|
|
2430
2364
|
// core/util/detect-media-type.ts
|
2431
|
-
var
|
2365
|
+
var import_provider_utils7 = require("@ai-sdk/provider-utils");
|
2432
2366
|
var imageMediaTypeSignatures = [
|
2433
2367
|
{
|
2434
2368
|
mediaType: "image/gif",
|
@@ -2535,7 +2469,7 @@ var audioMediaTypeSignatures = [
|
|
2535
2469
|
}
|
2536
2470
|
];
|
2537
2471
|
var stripID3 = (data) => {
|
2538
|
-
const bytes = typeof data === "string" ? (0,
|
2472
|
+
const bytes = typeof data === "string" ? (0, import_provider_utils7.convertBase64ToUint8Array)(data) : data;
|
2539
2473
|
const id3Size = (bytes[6] & 127) << 21 | (bytes[7] & 127) << 14 | (bytes[8] & 127) << 7 | bytes[9] & 127;
|
2540
2474
|
return bytes.slice(id3Size + 10);
|
2541
2475
|
};
|
@@ -2639,7 +2573,7 @@ var DefaultGenerateImageResult = class {
|
|
2639
2573
|
|
2640
2574
|
// core/generate-object/generate-object.ts
|
2641
2575
|
var import_provider13 = require("@ai-sdk/provider");
|
2642
|
-
var
|
2576
|
+
var import_provider_utils12 = require("@ai-sdk/provider-utils");
|
2643
2577
|
|
2644
2578
|
// errors/no-object-generated-error.ts
|
2645
2579
|
var import_provider5 = require("@ai-sdk/provider");
|
@@ -2733,7 +2667,7 @@ async function download({ url }) {
|
|
2733
2667
|
|
2734
2668
|
// core/prompt/data-content.ts
|
2735
2669
|
var import_provider8 = require("@ai-sdk/provider");
|
2736
|
-
var
|
2670
|
+
var import_provider_utils8 = require("@ai-sdk/provider-utils");
|
2737
2671
|
var import_zod = require("zod");
|
2738
2672
|
|
2739
2673
|
// core/prompt/invalid-data-content-error.ts
|
@@ -2820,9 +2754,9 @@ function convertDataContentToBase64String(content) {
|
|
2820
2754
|
return content;
|
2821
2755
|
}
|
2822
2756
|
if (content instanceof ArrayBuffer) {
|
2823
|
-
return (0,
|
2757
|
+
return (0, import_provider_utils8.convertUint8ArrayToBase64)(new Uint8Array(content));
|
2824
2758
|
}
|
2825
|
-
return (0,
|
2759
|
+
return (0, import_provider_utils8.convertUint8ArrayToBase64)(content);
|
2826
2760
|
}
|
2827
2761
|
function convertDataContentToUint8Array(content) {
|
2828
2762
|
if (content instanceof Uint8Array) {
|
@@ -2830,7 +2764,7 @@ function convertDataContentToUint8Array(content) {
|
|
2830
2764
|
}
|
2831
2765
|
if (typeof content === "string") {
|
2832
2766
|
try {
|
2833
|
-
return (0,
|
2767
|
+
return (0, import_provider_utils8.convertBase64ToUint8Array)(content);
|
2834
2768
|
} catch (error) {
|
2835
2769
|
throw new InvalidDataContentError({
|
2836
2770
|
message: "Invalid data content. Content string is not a base64-encoded media.",
|
@@ -2874,7 +2808,7 @@ var InvalidMessageRoleError = class extends import_provider9.AISDKError {
|
|
2874
2808
|
_a7 = symbol7;
|
2875
2809
|
|
2876
2810
|
// core/prompt/convert-to-language-model-prompt.ts
|
2877
|
-
var
|
2811
|
+
var import_provider_utils9 = require("@ai-sdk/provider-utils");
|
2878
2812
|
async function convertToLanguageModelPrompt({
|
2879
2813
|
prompt,
|
2880
2814
|
supportedUrls,
|
@@ -3011,7 +2945,7 @@ async function downloadAssets(messages, downloadImplementation, supportedUrls) {
|
|
3011
2945
|
}
|
3012
2946
|
return { mediaType, data };
|
3013
2947
|
}).filter(
|
3014
|
-
(part) => part.data instanceof URL && part.mediaType != null && !(0,
|
2948
|
+
(part) => part.data instanceof URL && part.mediaType != null && !(0, import_provider_utils9.isUrlSupported)({
|
3015
2949
|
url: part.data.toString(),
|
3016
2950
|
mediaType: part.mediaType,
|
3017
2951
|
supportedUrls
|
@@ -3182,7 +3116,7 @@ function prepareCallSettings({
|
|
3182
3116
|
|
3183
3117
|
// core/prompt/standardize-prompt.ts
|
3184
3118
|
var import_provider11 = require("@ai-sdk/provider");
|
3185
|
-
var
|
3119
|
+
var import_provider_utils10 = require("@ai-sdk/provider-utils");
|
3186
3120
|
var import_zod7 = require("zod");
|
3187
3121
|
|
3188
3122
|
// core/prompt/attachments-to-parts.ts
|
@@ -3705,7 +3639,6 @@ async function standardizePrompt({
|
|
3705
3639
|
});
|
3706
3640
|
}
|
3707
3641
|
return {
|
3708
|
-
type: "prompt",
|
3709
3642
|
system: prompt.system,
|
3710
3643
|
messages: [
|
3711
3644
|
{
|
@@ -3732,7 +3665,7 @@ async function standardizePrompt({
|
|
3732
3665
|
message: "messages must not be empty"
|
3733
3666
|
});
|
3734
3667
|
}
|
3735
|
-
const validationResult = await (0,
|
3668
|
+
const validationResult = await (0, import_provider_utils10.safeValidateTypes)({
|
3736
3669
|
value: messages,
|
3737
3670
|
schema: import_zod7.z.array(coreMessageSchema)
|
3738
3671
|
});
|
@@ -3744,7 +3677,6 @@ async function standardizePrompt({
|
|
3744
3677
|
});
|
3745
3678
|
}
|
3746
3679
|
return {
|
3747
|
-
type: "messages",
|
3748
3680
|
messages,
|
3749
3681
|
system: prompt.system
|
3750
3682
|
};
|
@@ -3773,7 +3705,7 @@ function addLanguageModelUsage(usage1, usage2) {
|
|
3773
3705
|
|
3774
3706
|
// core/generate-object/output-strategy.ts
|
3775
3707
|
var import_provider12 = require("@ai-sdk/provider");
|
3776
|
-
var
|
3708
|
+
var import_provider_utils11 = require("@ai-sdk/provider-utils");
|
3777
3709
|
|
3778
3710
|
// core/util/async-iterable-stream.ts
|
3779
3711
|
function createAsyncIterableStream(source) {
|
@@ -3829,7 +3761,7 @@ var objectOutputStrategy = (schema) => ({
|
|
3829
3761
|
};
|
3830
3762
|
},
|
3831
3763
|
async validateFinalResult(value) {
|
3832
|
-
return (0,
|
3764
|
+
return (0, import_provider_utils11.safeValidateTypes)({ value, schema });
|
3833
3765
|
},
|
3834
3766
|
createElementStream() {
|
3835
3767
|
throw new import_provider12.UnsupportedFunctionalityError({
|
@@ -3873,7 +3805,7 @@ var arrayOutputStrategy = (schema) => {
|
|
3873
3805
|
const resultArray = [];
|
3874
3806
|
for (let i = 0; i < inputArray.length; i++) {
|
3875
3807
|
const element = inputArray[i];
|
3876
|
-
const result = await (0,
|
3808
|
+
const result = await (0, import_provider_utils11.safeValidateTypes)({ value: element, schema });
|
3877
3809
|
if (i === inputArray.length - 1 && !isFinalDelta) {
|
3878
3810
|
continue;
|
3879
3811
|
}
|
@@ -3914,7 +3846,7 @@ var arrayOutputStrategy = (schema) => {
|
|
3914
3846
|
}
|
3915
3847
|
const inputArray = value.elements;
|
3916
3848
|
for (const element of inputArray) {
|
3917
|
-
const result = await (0,
|
3849
|
+
const result = await (0, import_provider_utils11.safeValidateTypes)({ value: element, schema });
|
3918
3850
|
if (!result.success) {
|
3919
3851
|
return result;
|
3920
3852
|
}
|
@@ -4006,9 +3938,9 @@ function getOutputStrategy({
|
|
4006
3938
|
}) {
|
4007
3939
|
switch (output) {
|
4008
3940
|
case "object":
|
4009
|
-
return objectOutputStrategy(asSchema(schema));
|
3941
|
+
return objectOutputStrategy((0, import_provider_utils4.asSchema)(schema));
|
4010
3942
|
case "array":
|
4011
|
-
return arrayOutputStrategy(asSchema(schema));
|
3943
|
+
return arrayOutputStrategy((0, import_provider_utils4.asSchema)(schema));
|
4012
3944
|
case "enum":
|
4013
3945
|
return enumOutputStrategy(enumValues);
|
4014
3946
|
case "no-schema":
|
@@ -4139,30 +4071,32 @@ function validateObjectGenerationInput({
|
|
4139
4071
|
}
|
4140
4072
|
|
4141
4073
|
// core/generate-object/generate-object.ts
|
4142
|
-
var originalGenerateId = (0,
|
4143
|
-
async function generateObject({
|
4144
|
-
|
4145
|
-
|
4146
|
-
|
4147
|
-
|
4148
|
-
|
4149
|
-
|
4150
|
-
|
4151
|
-
|
4152
|
-
|
4153
|
-
|
4154
|
-
|
4155
|
-
|
4156
|
-
|
4157
|
-
|
4158
|
-
|
4159
|
-
|
4160
|
-
|
4161
|
-
|
4162
|
-
|
4163
|
-
|
4164
|
-
|
4165
|
-
|
4074
|
+
var originalGenerateId = (0, import_provider_utils12.createIdGenerator)({ prefix: "aiobj", size: 24 });
|
4075
|
+
async function generateObject(options) {
|
4076
|
+
const {
|
4077
|
+
model,
|
4078
|
+
output = "object",
|
4079
|
+
system,
|
4080
|
+
prompt,
|
4081
|
+
messages,
|
4082
|
+
maxRetries: maxRetriesArg,
|
4083
|
+
abortSignal,
|
4084
|
+
headers,
|
4085
|
+
experimental_repairText: repairText,
|
4086
|
+
experimental_telemetry: telemetry,
|
4087
|
+
providerOptions,
|
4088
|
+
_internal: {
|
4089
|
+
generateId: generateId3 = originalGenerateId,
|
4090
|
+
currentDate = () => /* @__PURE__ */ new Date()
|
4091
|
+
} = {},
|
4092
|
+
...settings
|
4093
|
+
} = options;
|
4094
|
+
const enumValues = "enum" in options ? options.enum : void 0;
|
4095
|
+
const {
|
4096
|
+
schema: inputSchema,
|
4097
|
+
schemaDescription,
|
4098
|
+
schemaName
|
4099
|
+
} = "schema" in options ? options : {};
|
4166
4100
|
validateObjectGenerationInput({
|
4167
4101
|
output,
|
4168
4102
|
schema: inputSchema,
|
@@ -4220,7 +4154,7 @@ async function generateObject({
|
|
4220
4154
|
});
|
4221
4155
|
const promptMessages = await convertToLanguageModelPrompt({
|
4222
4156
|
prompt: standardizedPrompt,
|
4223
|
-
supportedUrls: await model.
|
4157
|
+
supportedUrls: await model.supportedUrls
|
4224
4158
|
});
|
4225
4159
|
const generateResult = await retry(
|
4226
4160
|
() => recordSpan({
|
@@ -4233,9 +4167,6 @@ async function generateObject({
|
|
4233
4167
|
telemetry
|
4234
4168
|
}),
|
4235
4169
|
...baseTelemetryAttributes,
|
4236
|
-
"ai.prompt.format": {
|
4237
|
-
input: () => standardizedPrompt.type
|
4238
|
-
},
|
4239
4170
|
"ai.prompt.messages": {
|
4240
4171
|
input: () => JSON.stringify(promptMessages)
|
4241
4172
|
},
|
@@ -4261,7 +4192,6 @@ async function generateObject({
|
|
4261
4192
|
description: schemaDescription
|
4262
4193
|
},
|
4263
4194
|
...prepareCallSettings(settings),
|
4264
|
-
inputFormat: standardizedPrompt.type,
|
4265
4195
|
prompt: promptMessages,
|
4266
4196
|
providerOptions,
|
4267
4197
|
abortSignal,
|
@@ -4316,7 +4246,7 @@ async function generateObject({
|
|
4316
4246
|
request = (_a17 = generateResult.request) != null ? _a17 : {};
|
4317
4247
|
response = generateResult.responseData;
|
4318
4248
|
async function processResult(result2) {
|
4319
|
-
const parseResult = await (0,
|
4249
|
+
const parseResult = await (0, import_provider_utils12.safeParseJSON)({ text: result2 });
|
4320
4250
|
if (!parseResult.success) {
|
4321
4251
|
throw new NoObjectGeneratedError({
|
4322
4252
|
message: "No object generated: could not parse the response.",
|
@@ -4412,7 +4342,7 @@ var DefaultGenerateObjectResult = class {
|
|
4412
4342
|
};
|
4413
4343
|
|
4414
4344
|
// core/generate-object/stream-object.ts
|
4415
|
-
var
|
4345
|
+
var import_provider_utils13 = require("@ai-sdk/provider-utils");
|
4416
4346
|
|
4417
4347
|
// util/delayed-promise.ts
|
4418
4348
|
var DelayedPromise = class {
|
@@ -4556,30 +4486,33 @@ function now() {
|
|
4556
4486
|
}
|
4557
4487
|
|
4558
4488
|
// core/generate-object/stream-object.ts
|
4559
|
-
var originalGenerateId2 = (0,
|
4560
|
-
function streamObject({
|
4561
|
-
|
4562
|
-
|
4563
|
-
|
4564
|
-
|
4565
|
-
|
4566
|
-
|
4567
|
-
|
4568
|
-
|
4569
|
-
|
4570
|
-
|
4571
|
-
|
4572
|
-
|
4573
|
-
|
4574
|
-
|
4575
|
-
|
4576
|
-
|
4577
|
-
|
4578
|
-
|
4579
|
-
|
4580
|
-
} =
|
4581
|
-
|
4582
|
-
|
4489
|
+
var originalGenerateId2 = (0, import_provider_utils13.createIdGenerator)({ prefix: "aiobj", size: 24 });
|
4490
|
+
function streamObject(options) {
|
4491
|
+
const {
|
4492
|
+
model,
|
4493
|
+
output = "object",
|
4494
|
+
system,
|
4495
|
+
prompt,
|
4496
|
+
messages,
|
4497
|
+
maxRetries,
|
4498
|
+
abortSignal,
|
4499
|
+
headers,
|
4500
|
+
experimental_telemetry: telemetry,
|
4501
|
+
providerOptions,
|
4502
|
+
onError,
|
4503
|
+
onFinish,
|
4504
|
+
_internal: {
|
4505
|
+
generateId: generateId3 = originalGenerateId2,
|
4506
|
+
currentDate = () => /* @__PURE__ */ new Date(),
|
4507
|
+
now: now2 = now
|
4508
|
+
} = {},
|
4509
|
+
...settings
|
4510
|
+
} = options;
|
4511
|
+
const {
|
4512
|
+
schema: inputSchema,
|
4513
|
+
schemaDescription,
|
4514
|
+
schemaName
|
4515
|
+
} = "schema" in options ? options : {};
|
4583
4516
|
validateObjectGenerationInput({
|
4584
4517
|
output,
|
4585
4518
|
schema: inputSchema,
|
@@ -4692,10 +4625,9 @@ var DefaultStreamObjectResult = class {
|
|
4692
4625
|
description: schemaDescription
|
4693
4626
|
},
|
4694
4627
|
...prepareCallSettings(settings),
|
4695
|
-
inputFormat: standardizedPrompt.type,
|
4696
4628
|
prompt: await convertToLanguageModelPrompt({
|
4697
4629
|
prompt: standardizedPrompt,
|
4698
|
-
supportedUrls: await model.
|
4630
|
+
supportedUrls: await model.supportedUrls
|
4699
4631
|
}),
|
4700
4632
|
providerOptions,
|
4701
4633
|
abortSignal,
|
@@ -4730,9 +4662,6 @@ var DefaultStreamObjectResult = class {
|
|
4730
4662
|
telemetry
|
4731
4663
|
}),
|
4732
4664
|
...baseTelemetryAttributes,
|
4733
|
-
"ai.prompt.format": {
|
4734
|
-
input: () => callOptions.inputFormat
|
4735
|
-
},
|
4736
4665
|
"ai.prompt.messages": {
|
4737
4666
|
input: () => JSON.stringify(callOptions.prompt)
|
4738
4667
|
},
|
@@ -5052,7 +4981,7 @@ var DefaultStreamObjectResult = class {
|
|
5052
4981
|
};
|
5053
4982
|
|
5054
4983
|
// core/generate-text/generate-text.ts
|
5055
|
-
var
|
4984
|
+
var import_provider_utils15 = require("@ai-sdk/provider-utils");
|
5056
4985
|
|
5057
4986
|
// errors/no-output-specified-error.ts
|
5058
4987
|
var import_provider14 = require("@ai-sdk/provider");
|
@@ -5128,7 +5057,7 @@ function prepareToolsAndToolChoice({
|
|
5128
5057
|
type: "function",
|
5129
5058
|
name: name17,
|
5130
5059
|
description: tool2.description,
|
5131
|
-
parameters: asSchema(tool2.parameters).jsonSchema
|
5060
|
+
parameters: (0, import_provider_utils4.asSchema)(tool2.parameters).jsonSchema
|
5132
5061
|
};
|
5133
5062
|
case "provider-defined":
|
5134
5063
|
return {
|
@@ -5161,7 +5090,7 @@ function removeTextAfterLastWhitespace(text2) {
|
|
5161
5090
|
}
|
5162
5091
|
|
5163
5092
|
// core/generate-text/parse-tool-call.ts
|
5164
|
-
var
|
5093
|
+
var import_provider_utils14 = require("@ai-sdk/provider-utils");
|
5165
5094
|
|
5166
5095
|
// errors/invalid-tool-arguments-error.ts
|
5167
5096
|
var import_provider16 = require("@ai-sdk/provider");
|
@@ -5258,7 +5187,7 @@ async function parseToolCall({
|
|
5258
5187
|
tools,
|
5259
5188
|
parameterSchema: ({ toolName }) => {
|
5260
5189
|
const { parameters } = tools[toolName];
|
5261
|
-
return asSchema(parameters).jsonSchema;
|
5190
|
+
return (0, import_provider_utils4.asSchema)(parameters).jsonSchema;
|
5262
5191
|
},
|
5263
5192
|
system,
|
5264
5193
|
messages,
|
@@ -5288,8 +5217,8 @@ async function doParseToolCall({
|
|
5288
5217
|
availableTools: Object.keys(tools)
|
5289
5218
|
});
|
5290
5219
|
}
|
5291
|
-
const schema = asSchema(tool2.parameters);
|
5292
|
-
const parseResult = toolCall.args.trim() === "" ? await (0,
|
5220
|
+
const schema = (0, import_provider_utils4.asSchema)(tool2.parameters);
|
5221
|
+
const parseResult = toolCall.args.trim() === "" ? await (0, import_provider_utils14.safeValidateTypes)({ value: {}, schema }) : await (0, import_provider_utils14.safeParseJSON)({ text: toolCall.args, schema });
|
5293
5222
|
if (parseResult.success === false) {
|
5294
5223
|
throw new InvalidToolArgumentsError({
|
5295
5224
|
toolName,
|
@@ -5383,11 +5312,11 @@ function toResponseMessages({
|
|
5383
5312
|
}
|
5384
5313
|
|
5385
5314
|
// core/generate-text/generate-text.ts
|
5386
|
-
var originalGenerateId3 = (0,
|
5315
|
+
var originalGenerateId3 = (0, import_provider_utils15.createIdGenerator)({
|
5387
5316
|
prefix: "aitxt",
|
5388
5317
|
size: 24
|
5389
5318
|
});
|
5390
|
-
var originalGenerateMessageId = (0,
|
5319
|
+
var originalGenerateMessageId = (0, import_provider_utils15.createIdGenerator)({
|
5391
5320
|
prefix: "msg",
|
5392
5321
|
size: 24
|
5393
5322
|
});
|
@@ -5408,6 +5337,7 @@ async function generateText({
|
|
5408
5337
|
experimental_telemetry: telemetry,
|
5409
5338
|
providerOptions,
|
5410
5339
|
experimental_activeTools: activeTools,
|
5340
|
+
experimental_prepareStep: prepareStep,
|
5411
5341
|
experimental_repairToolCall: repairToolCall,
|
5412
5342
|
_internal: {
|
5413
5343
|
generateId: generateId3 = originalGenerateId3,
|
@@ -5446,6 +5376,9 @@ async function generateText({
|
|
5446
5376
|
telemetry
|
5447
5377
|
}),
|
5448
5378
|
...baseTelemetryAttributes,
|
5379
|
+
// model:
|
5380
|
+
"ai.model.provider": model.provider,
|
5381
|
+
"ai.model.id": model.modelId,
|
5449
5382
|
// specific settings that only make sense on the outer level:
|
5450
5383
|
"ai.prompt": {
|
5451
5384
|
input: () => JSON.stringify({ system, prompt, messages })
|
@@ -5455,10 +5388,8 @@ async function generateText({
|
|
5455
5388
|
}),
|
5456
5389
|
tracer,
|
5457
5390
|
fn: async (span) => {
|
5458
|
-
var _a17, _b, _c;
|
5459
|
-
const
|
5460
|
-
...prepareToolsAndToolChoice({ tools, toolChoice, activeTools })
|
5461
|
-
};
|
5391
|
+
var _a17, _b, _c, _d, _e, _f;
|
5392
|
+
const callSettings2 = prepareCallSettings(settings);
|
5462
5393
|
let currentModelResponse;
|
5463
5394
|
let currentToolCalls = [];
|
5464
5395
|
let currentToolResults = [];
|
@@ -5475,18 +5406,28 @@ async function generateText({
|
|
5475
5406
|
};
|
5476
5407
|
let stepType = "initial";
|
5477
5408
|
do {
|
5478
|
-
const promptFormat = stepCount === 0 ? initialPrompt.type : "messages";
|
5479
5409
|
const stepInputMessages = [
|
5480
5410
|
...initialPrompt.messages,
|
5481
5411
|
...responseMessages
|
5482
5412
|
];
|
5413
|
+
const prepareStepResult = await (prepareStep == null ? void 0 : prepareStep({
|
5414
|
+
model,
|
5415
|
+
steps,
|
5416
|
+
maxSteps,
|
5417
|
+
stepNumber: stepCount
|
5418
|
+
}));
|
5483
5419
|
const promptMessages = await convertToLanguageModelPrompt({
|
5484
5420
|
prompt: {
|
5485
|
-
type: promptFormat,
|
5486
5421
|
system: initialPrompt.system,
|
5487
5422
|
messages: stepInputMessages
|
5488
5423
|
},
|
5489
|
-
supportedUrls: await model.
|
5424
|
+
supportedUrls: await model.supportedUrls
|
5425
|
+
});
|
5426
|
+
const stepModel = (_a17 = prepareStepResult == null ? void 0 : prepareStepResult.model) != null ? _a17 : model;
|
5427
|
+
const { toolChoice: stepToolChoice, tools: stepTools } = prepareToolsAndToolChoice({
|
5428
|
+
tools,
|
5429
|
+
toolChoice: (_b = prepareStepResult == null ? void 0 : prepareStepResult.toolChoice) != null ? _b : toolChoice,
|
5430
|
+
activeTools: (_c = prepareStepResult == null ? void 0 : prepareStepResult.experimental_activeTools) != null ? _c : activeTools
|
5490
5431
|
});
|
5491
5432
|
currentModelResponse = await retry(
|
5492
5433
|
() => {
|
@@ -5501,23 +5442,23 @@ async function generateText({
|
|
5501
5442
|
telemetry
|
5502
5443
|
}),
|
5503
5444
|
...baseTelemetryAttributes,
|
5504
|
-
|
5445
|
+
// model:
|
5446
|
+
"ai.model.provider": stepModel.provider,
|
5447
|
+
"ai.model.id": stepModel.modelId,
|
5448
|
+
// prompt:
|
5505
5449
|
"ai.prompt.messages": {
|
5506
5450
|
input: () => JSON.stringify(promptMessages)
|
5507
5451
|
},
|
5508
5452
|
"ai.prompt.tools": {
|
5509
5453
|
// convert the language model level tools:
|
5510
|
-
input: () =>
|
5511
|
-
var _a19;
|
5512
|
-
return (_a19 = toolsAndToolChoice.tools) == null ? void 0 : _a19.map((tool2) => JSON.stringify(tool2));
|
5513
|
-
}
|
5454
|
+
input: () => stepTools == null ? void 0 : stepTools.map((tool2) => JSON.stringify(tool2))
|
5514
5455
|
},
|
5515
5456
|
"ai.prompt.toolChoice": {
|
5516
|
-
input: () =>
|
5457
|
+
input: () => stepToolChoice != null ? JSON.stringify(stepToolChoice) : void 0
|
5517
5458
|
},
|
5518
5459
|
// standardized gen-ai llm span attributes:
|
5519
|
-
"gen_ai.system":
|
5520
|
-
"gen_ai.request.model":
|
5460
|
+
"gen_ai.system": stepModel.provider,
|
5461
|
+
"gen_ai.request.model": stepModel.modelId,
|
5521
5462
|
"gen_ai.request.frequency_penalty": settings.frequencyPenalty,
|
5522
5463
|
"gen_ai.request.max_tokens": settings.maxOutputTokens,
|
5523
5464
|
"gen_ai.request.presence_penalty": settings.presencePenalty,
|
@@ -5529,11 +5470,11 @@ async function generateText({
|
|
5529
5470
|
}),
|
5530
5471
|
tracer,
|
5531
5472
|
fn: async (span2) => {
|
5532
|
-
var _a19, _b2, _c2,
|
5533
|
-
const result = await
|
5534
|
-
...
|
5535
|
-
|
5536
|
-
|
5473
|
+
var _a19, _b2, _c2, _d2, _e2, _f2, _g, _h;
|
5474
|
+
const result = await stepModel.doGenerate({
|
5475
|
+
...callSettings2,
|
5476
|
+
tools: stepTools,
|
5477
|
+
toolChoice: stepToolChoice,
|
5537
5478
|
responseFormat: output == null ? void 0 : output.responseFormat,
|
5538
5479
|
prompt: promptMessages,
|
5539
5480
|
providerOptions,
|
@@ -5542,8 +5483,8 @@ async function generateText({
|
|
5542
5483
|
});
|
5543
5484
|
const responseData = {
|
5544
5485
|
id: (_b2 = (_a19 = result.response) == null ? void 0 : _a19.id) != null ? _b2 : generateId3(),
|
5545
|
-
timestamp: (
|
5546
|
-
modelId: (
|
5486
|
+
timestamp: (_d2 = (_c2 = result.response) == null ? void 0 : _c2.timestamp) != null ? _d2 : currentDate(),
|
5487
|
+
modelId: (_f2 = (_e2 = result.response) == null ? void 0 : _e2.modelId) != null ? _f2 : stepModel.modelId,
|
5547
5488
|
headers: (_g = result.response) == null ? void 0 : _g.headers,
|
5548
5489
|
body: (_h = result.response) == null ? void 0 : _h.body
|
5549
5490
|
};
|
@@ -5619,7 +5560,7 @@ async function generateText({
|
|
5619
5560
|
nextStepType = "tool-result";
|
5620
5561
|
}
|
5621
5562
|
}
|
5622
|
-
const originalText = (
|
5563
|
+
const originalText = (_d = extractContentText(currentModelResponse.content)) != null ? _d : "";
|
5623
5564
|
const stepTextLeadingWhitespaceTrimmed = stepType === "continue" && // only for continue steps
|
5624
5565
|
text2.trimEnd() !== text2 ? originalText.trimStart() : originalText;
|
5625
5566
|
const stepText = nextStepType === "continue" ? removeTextAfterLastWhitespace(stepTextLeadingWhitespaceTrimmed) : stepTextLeadingWhitespaceTrimmed;
|
@@ -5672,7 +5613,7 @@ async function generateText({
|
|
5672
5613
|
finishReason: currentModelResponse.finishReason,
|
5673
5614
|
usage: currentUsage,
|
5674
5615
|
warnings: currentModelResponse.warnings,
|
5675
|
-
request: (
|
5616
|
+
request: (_e = currentModelResponse.request) != null ? _e : {},
|
5676
5617
|
response: {
|
5677
5618
|
...currentModelResponse.response,
|
5678
5619
|
// deep clone msgs to avoid mutating past messages in multi-step:
|
@@ -5725,7 +5666,7 @@ async function generateText({
|
|
5725
5666
|
finishReason: currentModelResponse.finishReason,
|
5726
5667
|
usage,
|
5727
5668
|
warnings: currentModelResponse.warnings,
|
5728
|
-
request: (
|
5669
|
+
request: (_f = currentModelResponse.request) != null ? _f : {},
|
5729
5670
|
response: {
|
5730
5671
|
...currentModelResponse.response,
|
5731
5672
|
messages: responseMessages
|
@@ -5860,7 +5801,7 @@ __export(output_exports, {
|
|
5860
5801
|
object: () => object,
|
5861
5802
|
text: () => text
|
5862
5803
|
});
|
5863
|
-
var
|
5804
|
+
var import_provider_utils16 = require("@ai-sdk/provider-utils");
|
5864
5805
|
|
5865
5806
|
// errors/index.ts
|
5866
5807
|
var import_provider21 = require("@ai-sdk/provider");
|
@@ -5921,7 +5862,7 @@ var text = () => ({
|
|
5921
5862
|
var object = ({
|
5922
5863
|
schema: inputSchema
|
5923
5864
|
}) => {
|
5924
|
-
const schema = asSchema(inputSchema);
|
5865
|
+
const schema = (0, import_provider_utils4.asSchema)(inputSchema);
|
5925
5866
|
return {
|
5926
5867
|
type: "object",
|
5927
5868
|
responseFormat: {
|
@@ -5947,7 +5888,7 @@ var object = ({
|
|
5947
5888
|
}
|
5948
5889
|
},
|
5949
5890
|
async parseOutput({ text: text2 }, context) {
|
5950
|
-
const parseResult = await (0,
|
5891
|
+
const parseResult = await (0, import_provider_utils16.safeParseJSON)({ text: text2 });
|
5951
5892
|
if (!parseResult.success) {
|
5952
5893
|
throw new NoObjectGeneratedError({
|
5953
5894
|
message: "No object generated: could not parse the response.",
|
@@ -5958,7 +5899,7 @@ var object = ({
|
|
5958
5899
|
finishReason: context.finishReason
|
5959
5900
|
});
|
5960
5901
|
}
|
5961
|
-
const validationResult = await (0,
|
5902
|
+
const validationResult = await (0, import_provider_utils16.safeValidateTypes)({
|
5962
5903
|
value: parseResult.value,
|
5963
5904
|
schema
|
5964
5905
|
});
|
@@ -5978,7 +5919,7 @@ var object = ({
|
|
5978
5919
|
};
|
5979
5920
|
|
5980
5921
|
// core/generate-text/smooth-stream.ts
|
5981
|
-
var
|
5922
|
+
var import_provider_utils17 = require("@ai-sdk/provider-utils");
|
5982
5923
|
var import_provider22 = require("@ai-sdk/provider");
|
5983
5924
|
var CHUNKING_REGEXPS = {
|
5984
5925
|
word: /\S+\s+/m,
|
@@ -5987,7 +5928,7 @@ var CHUNKING_REGEXPS = {
|
|
5987
5928
|
function smoothStream({
|
5988
5929
|
delayInMs = 10,
|
5989
5930
|
chunking = "word",
|
5990
|
-
_internal: { delay: delay2 =
|
5931
|
+
_internal: { delay: delay2 = import_provider_utils17.delay } = {}
|
5991
5932
|
} = {}) {
|
5992
5933
|
let detectChunk;
|
5993
5934
|
if (typeof chunking === "function") {
|
@@ -6047,7 +5988,7 @@ function smoothStream({
|
|
6047
5988
|
}
|
6048
5989
|
|
6049
5990
|
// core/generate-text/stream-text.ts
|
6050
|
-
var
|
5991
|
+
var import_provider_utils18 = require("@ai-sdk/provider-utils");
|
6051
5992
|
|
6052
5993
|
// util/as-array.ts
|
6053
5994
|
function asArray(value) {
|
@@ -6246,7 +6187,7 @@ function runToolsTransformation({
|
|
6246
6187
|
controller.enqueue(toolCall);
|
6247
6188
|
const tool2 = tools[toolCall.toolName];
|
6248
6189
|
if (tool2.execute != null) {
|
6249
|
-
const toolExecutionId = (0,
|
6190
|
+
const toolExecutionId = (0, import_provider_utils4.generateId)();
|
6250
6191
|
outstandingToolResults.add(toolExecutionId);
|
6251
6192
|
recordSpan({
|
6252
6193
|
name: "ai.toolCall",
|
@@ -6364,11 +6305,11 @@ function runToolsTransformation({
|
|
6364
6305
|
}
|
6365
6306
|
|
6366
6307
|
// core/generate-text/stream-text.ts
|
6367
|
-
var originalGenerateId4 = (0,
|
6308
|
+
var originalGenerateId4 = (0, import_provider_utils18.createIdGenerator)({
|
6368
6309
|
prefix: "aitxt",
|
6369
6310
|
size: 24
|
6370
6311
|
});
|
6371
|
-
var originalGenerateMessageId2 = (0,
|
6312
|
+
var originalGenerateMessageId2 = (0, import_provider_utils18.createIdGenerator)({
|
6372
6313
|
prefix: "msg",
|
6373
6314
|
size: 24
|
6374
6315
|
});
|
@@ -6795,18 +6736,16 @@ var DefaultStreamTextResult = class {
|
|
6795
6736
|
prompt: { system, prompt, messages },
|
6796
6737
|
tools
|
6797
6738
|
});
|
6798
|
-
const promptFormat = responseMessages.length === 0 ? initialPrompt.type : "messages";
|
6799
6739
|
const stepInputMessages = [
|
6800
6740
|
...initialPrompt.messages,
|
6801
6741
|
...responseMessages
|
6802
6742
|
];
|
6803
6743
|
const promptMessages = await convertToLanguageModelPrompt({
|
6804
6744
|
prompt: {
|
6805
|
-
type: promptFormat,
|
6806
6745
|
system: initialPrompt.system,
|
6807
6746
|
messages: stepInputMessages
|
6808
6747
|
},
|
6809
|
-
supportedUrls: await model.
|
6748
|
+
supportedUrls: await model.supportedUrls
|
6810
6749
|
});
|
6811
6750
|
const toolsAndToolChoice = {
|
6812
6751
|
...prepareToolsAndToolChoice({ tools, toolChoice, activeTools })
|
@@ -6826,9 +6765,6 @@ var DefaultStreamTextResult = class {
|
|
6826
6765
|
telemetry
|
6827
6766
|
}),
|
6828
6767
|
...baseTelemetryAttributes,
|
6829
|
-
"ai.prompt.format": {
|
6830
|
-
input: () => promptFormat
|
6831
|
-
},
|
6832
6768
|
"ai.prompt.messages": {
|
6833
6769
|
input: () => JSON.stringify(promptMessages)
|
6834
6770
|
},
|
@@ -6866,7 +6802,6 @@ var DefaultStreamTextResult = class {
|
|
6866
6802
|
result: await model.doStream({
|
6867
6803
|
...callSettings,
|
6868
6804
|
...toolsAndToolChoice,
|
6869
|
-
inputFormat: promptFormat,
|
6870
6805
|
responseFormat: output == null ? void 0 : output.responseFormat,
|
6871
6806
|
prompt: promptMessages,
|
6872
6807
|
providerOptions,
|
@@ -7959,8 +7894,8 @@ var doWrap = ({
|
|
7959
7894
|
provider: providerId != null ? providerId : model.provider,
|
7960
7895
|
modelId: modelId != null ? modelId : model.modelId,
|
7961
7896
|
// TODO middleware should be able to modify the supported urls
|
7962
|
-
|
7963
|
-
return model.
|
7897
|
+
get supportedUrls() {
|
7898
|
+
return model.supportedUrls;
|
7964
7899
|
},
|
7965
7900
|
async doGenerate(params) {
|
7966
7901
|
const transformedParams = await doTransform({ params, type: "generate" });
|
@@ -8077,7 +8012,6 @@ function appendResponseMessages({
|
|
8077
8012
|
);
|
8078
8013
|
(_b = lastMessage.parts) != null ? _b : lastMessage.parts = [];
|
8079
8014
|
lastMessage.content = textContent;
|
8080
|
-
lastMessage.reasoning = reasoningTextContent;
|
8081
8015
|
lastMessage.parts.push(...parts);
|
8082
8016
|
lastMessage.toolInvocations = [
|
8083
8017
|
...(_c = lastMessage.toolInvocations) != null ? _c : [],
|
@@ -8096,7 +8030,6 @@ function appendResponseMessages({
|
|
8096
8030
|
createdAt: currentDate(),
|
8097
8031
|
// generate a createdAt date for the message, will be overridden by the client
|
8098
8032
|
content: textContent,
|
8099
|
-
reasoning: reasoningTextContent,
|
8100
8033
|
toolInvocations: getToolInvocations2(0),
|
8101
8034
|
parts: [
|
8102
8035
|
...parts,
|
@@ -8303,7 +8236,7 @@ function tool(tool2) {
|
|
8303
8236
|
}
|
8304
8237
|
|
8305
8238
|
// core/tool/mcp/mcp-sse-transport.ts
|
8306
|
-
var
|
8239
|
+
var import_provider_utils19 = require("@ai-sdk/provider-utils");
|
8307
8240
|
|
8308
8241
|
// core/tool/mcp/json-rpc-message.ts
|
8309
8242
|
var import_zod9 = require("zod");
|
@@ -8474,7 +8407,7 @@ var SseMCPTransport = class {
|
|
8474
8407
|
(_b = this.onerror) == null ? void 0 : _b.call(this, error);
|
8475
8408
|
return reject(error);
|
8476
8409
|
}
|
8477
|
-
const stream = response.body.pipeThrough(new TextDecoderStream()).pipeThrough((0,
|
8410
|
+
const stream = response.body.pipeThrough(new TextDecoderStream()).pipeThrough((0, import_provider_utils19.createEventSourceParserStream)());
|
8478
8411
|
const reader = stream.getReader();
|
8479
8412
|
const processEvents = async () => {
|
8480
8413
|
var _a18, _b2, _c2;
|
@@ -8789,7 +8722,7 @@ var MCPClient = class {
|
|
8789
8722
|
if (schemas !== "automatic" && !(name17 in schemas)) {
|
8790
8723
|
continue;
|
8791
8724
|
}
|
8792
|
-
const parameters = schemas === "automatic" ? jsonSchema({
|
8725
|
+
const parameters = schemas === "automatic" ? (0, import_provider_utils4.jsonSchema)({
|
8793
8726
|
...inputSchema,
|
8794
8727
|
properties: (_a17 = inputSchema.properties) != null ? _a17 : {},
|
8795
8728
|
additionalProperties: false
|
@@ -8879,7 +8812,7 @@ function cosineSimilarity(vector1, vector2) {
|
|
8879
8812
|
}
|
8880
8813
|
|
8881
8814
|
// core/util/simulate-readable-stream.ts
|
8882
|
-
var
|
8815
|
+
var import_provider_utils20 = require("@ai-sdk/provider-utils");
|
8883
8816
|
function simulateReadableStream({
|
8884
8817
|
chunks,
|
8885
8818
|
initialDelayInMs = 0,
|
@@ -8887,7 +8820,7 @@ function simulateReadableStream({
|
|
8887
8820
|
_internal
|
8888
8821
|
}) {
|
8889
8822
|
var _a17;
|
8890
|
-
const delay2 = (_a17 = _internal == null ? void 0 : _internal.delay) != null ? _a17 :
|
8823
|
+
const delay2 = (_a17 = _internal == null ? void 0 : _internal.delay) != null ? _a17 : import_provider_utils20.delay;
|
8891
8824
|
let index = 0;
|
8892
8825
|
return new ReadableStream({
|
8893
8826
|
async pull(controller) {
|
@@ -9049,7 +8982,6 @@ var StreamData = class {
|
|
9049
8982
|
streamText,
|
9050
8983
|
tool,
|
9051
8984
|
updateToolCallResult,
|
9052
|
-
wrapLanguageModel
|
9053
|
-
zodSchema
|
8985
|
+
wrapLanguageModel
|
9054
8986
|
});
|
9055
8987
|
//# sourceMappingURL=index.js.map
|