ai 4.0.37 → 4.0.39
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +18 -0
- package/dist/index.d.mts +69 -35
- package/dist/index.d.ts +69 -35
- package/dist/index.js +97 -46
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +110 -61
- package/dist/index.mjs.map +1 -1
- package/package.json +4 -4
- package/rsc/dist/rsc-server.mjs +1 -1
- package/rsc/dist/rsc-server.mjs.map +1 -1
- package/test/dist/index.d.mts +3 -1
- package/test/dist/index.d.ts +3 -1
- package/test/dist/index.js +4 -2
- package/test/dist/index.js.map +1 -1
- package/test/dist/index.mjs +4 -2
- package/test/dist/index.mjs.map +1 -1
package/dist/index.mjs
CHANGED
@@ -4,19 +4,17 @@ var __export = (target, all) => {
|
|
4
4
|
__defProp(target, name14, { get: all[name14], enumerable: true });
|
5
5
|
};
|
6
6
|
|
7
|
-
//
|
7
|
+
// core/index.ts
|
8
|
+
import { createIdGenerator as createIdGenerator5, generateId as generateId2 } from "@ai-sdk/provider-utils";
|
8
9
|
import {
|
9
|
-
formatAssistantStreamPart
|
10
|
-
formatDataStreamPart as
|
10
|
+
formatAssistantStreamPart,
|
11
|
+
formatDataStreamPart as formatDataStreamPart3,
|
12
|
+
jsonSchema,
|
11
13
|
parseAssistantStreamPart,
|
12
14
|
parseDataStreamPart,
|
13
15
|
processDataStream,
|
14
16
|
processTextStream
|
15
17
|
} from "@ai-sdk/ui-utils";
|
16
|
-
import { generateId as generateId3 } from "@ai-sdk/provider-utils";
|
17
|
-
|
18
|
-
// core/index.ts
|
19
|
-
import { jsonSchema } from "@ai-sdk/ui-utils";
|
20
18
|
|
21
19
|
// core/data-stream/create-data-stream.ts
|
22
20
|
import { formatDataStreamPart } from "@ai-sdk/ui-utils";
|
@@ -1657,7 +1655,7 @@ function convertToCoreMessages(messages, options) {
|
|
1657
1655
|
break;
|
1658
1656
|
}
|
1659
1657
|
case "assistant": {
|
1660
|
-
if (toolInvocations == null) {
|
1658
|
+
if (toolInvocations == null || toolInvocations.length === 0) {
|
1661
1659
|
coreMessages.push({ role: "assistant", content });
|
1662
1660
|
break;
|
1663
1661
|
}
|
@@ -2225,7 +2223,7 @@ async function generateObject({
|
|
2225
2223
|
experimental_telemetry: telemetry,
|
2226
2224
|
experimental_providerMetadata: providerMetadata,
|
2227
2225
|
_internal: {
|
2228
|
-
generateId:
|
2226
|
+
generateId: generateId3 = originalGenerateId,
|
2229
2227
|
currentDate = () => /* @__PURE__ */ new Date()
|
2230
2228
|
} = {},
|
2231
2229
|
...settings
|
@@ -2355,7 +2353,7 @@ async function generateObject({
|
|
2355
2353
|
headers
|
2356
2354
|
});
|
2357
2355
|
const responseData = {
|
2358
|
-
id: (_b2 = (_a15 = result2.response) == null ? void 0 : _a15.id) != null ? _b2 :
|
2356
|
+
id: (_b2 = (_a15 = result2.response) == null ? void 0 : _a15.id) != null ? _b2 : generateId3(),
|
2359
2357
|
timestamp: (_d = (_c = result2.response) == null ? void 0 : _c.timestamp) != null ? _d : currentDate(),
|
2360
2358
|
modelId: (_f = (_e = result2.response) == null ? void 0 : _e.modelId) != null ? _f : model.modelId
|
2361
2359
|
};
|
@@ -2463,7 +2461,7 @@ async function generateObject({
|
|
2463
2461
|
});
|
2464
2462
|
const objectText = (_b2 = (_a15 = result2.toolCalls) == null ? void 0 : _a15[0]) == null ? void 0 : _b2.args;
|
2465
2463
|
const responseData = {
|
2466
|
-
id: (_d = (_c = result2.response) == null ? void 0 : _c.id) != null ? _d :
|
2464
|
+
id: (_d = (_c = result2.response) == null ? void 0 : _c.id) != null ? _d : generateId3(),
|
2467
2465
|
timestamp: (_f = (_e = result2.response) == null ? void 0 : _e.timestamp) != null ? _f : currentDate(),
|
2468
2466
|
modelId: (_h = (_g = result2.response) == null ? void 0 : _g.modelId) != null ? _h : model.modelId
|
2469
2467
|
};
|
@@ -2764,7 +2762,7 @@ function streamObject({
|
|
2764
2762
|
experimental_providerMetadata: providerMetadata,
|
2765
2763
|
onFinish,
|
2766
2764
|
_internal: {
|
2767
|
-
generateId:
|
2765
|
+
generateId: generateId3 = originalGenerateId2,
|
2768
2766
|
currentDate = () => /* @__PURE__ */ new Date(),
|
2769
2767
|
now: now2 = now
|
2770
2768
|
} = {},
|
@@ -2797,7 +2795,7 @@ function streamObject({
|
|
2797
2795
|
inputProviderMetadata: providerMetadata,
|
2798
2796
|
mode,
|
2799
2797
|
onFinish,
|
2800
|
-
generateId:
|
2798
|
+
generateId: generateId3,
|
2801
2799
|
currentDate,
|
2802
2800
|
now: now2
|
2803
2801
|
});
|
@@ -2819,7 +2817,7 @@ var DefaultStreamObjectResult = class {
|
|
2819
2817
|
inputProviderMetadata,
|
2820
2818
|
mode,
|
2821
2819
|
onFinish,
|
2822
|
-
generateId:
|
2820
|
+
generateId: generateId3,
|
2823
2821
|
currentDate,
|
2824
2822
|
now: now2
|
2825
2823
|
}) {
|
@@ -3020,7 +3018,7 @@ var DefaultStreamObjectResult = class {
|
|
3020
3018
|
let accumulatedText = "";
|
3021
3019
|
let textDelta = "";
|
3022
3020
|
let response = {
|
3023
|
-
id:
|
3021
|
+
id: generateId3(),
|
3024
3022
|
timestamp: currentDate(),
|
3025
3023
|
modelId: model.modelId
|
3026
3024
|
};
|
@@ -3561,16 +3559,20 @@ function toResponseMessages({
|
|
3561
3559
|
text: text2 = "",
|
3562
3560
|
tools,
|
3563
3561
|
toolCalls,
|
3564
|
-
toolResults
|
3562
|
+
toolResults,
|
3563
|
+
messageId,
|
3564
|
+
generateMessageId
|
3565
3565
|
}) {
|
3566
3566
|
const responseMessages = [];
|
3567
3567
|
responseMessages.push({
|
3568
3568
|
role: "assistant",
|
3569
|
-
content: [{ type: "text", text: text2 }, ...toolCalls]
|
3569
|
+
content: [{ type: "text", text: text2 }, ...toolCalls],
|
3570
|
+
id: messageId
|
3570
3571
|
});
|
3571
3572
|
if (toolResults.length > 0) {
|
3572
3573
|
responseMessages.push({
|
3573
3574
|
role: "tool",
|
3575
|
+
id: generateMessageId(),
|
3574
3576
|
content: toolResults.map((toolResult) => {
|
3575
3577
|
const tool2 = tools[toolResult.toolName];
|
3576
3578
|
return (tool2 == null ? void 0 : tool2.experimental_toToolResultContent) != null ? {
|
@@ -3594,7 +3596,14 @@ function toResponseMessages({
|
|
3594
3596
|
}
|
3595
3597
|
|
3596
3598
|
// core/generate-text/generate-text.ts
|
3597
|
-
var originalGenerateId3 = createIdGenerator3({
|
3599
|
+
var originalGenerateId3 = createIdGenerator3({
|
3600
|
+
prefix: "aitxt",
|
3601
|
+
size: 24
|
3602
|
+
});
|
3603
|
+
var originalGenerateMessageId = createIdGenerator3({
|
3604
|
+
prefix: "msg",
|
3605
|
+
size: 24
|
3606
|
+
});
|
3598
3607
|
async function generateText({
|
3599
3608
|
model,
|
3600
3609
|
tools,
|
@@ -3606,6 +3615,7 @@ async function generateText({
|
|
3606
3615
|
abortSignal,
|
3607
3616
|
headers,
|
3608
3617
|
maxSteps = 1,
|
3618
|
+
experimental_generateMessageId: generateMessageId = originalGenerateMessageId,
|
3609
3619
|
experimental_output: output,
|
3610
3620
|
experimental_continueSteps: continueSteps = false,
|
3611
3621
|
experimental_telemetry: telemetry,
|
@@ -3613,7 +3623,7 @@ async function generateText({
|
|
3613
3623
|
experimental_activeTools: activeTools,
|
3614
3624
|
experimental_repairToolCall: repairToolCall,
|
3615
3625
|
_internal: {
|
3616
|
-
generateId:
|
3626
|
+
generateId: generateId3 = originalGenerateId3,
|
3617
3627
|
currentDate = () => /* @__PURE__ */ new Date()
|
3618
3628
|
} = {},
|
3619
3629
|
onStepFinish,
|
@@ -3747,7 +3757,7 @@ async function generateText({
|
|
3747
3757
|
headers
|
3748
3758
|
});
|
3749
3759
|
const responseData = {
|
3750
|
-
id: (_b2 = (_a16 = result.response) == null ? void 0 : _a16.id) != null ? _b2 :
|
3760
|
+
id: (_b2 = (_a16 = result.response) == null ? void 0 : _a16.id) != null ? _b2 : generateId3(),
|
3751
3761
|
timestamp: (_d2 = (_c2 = result.response) == null ? void 0 : _c2.timestamp) != null ? _d2 : currentDate(),
|
3752
3762
|
modelId: (_f2 = (_e2 = result.response) == null ? void 0 : _e2.modelId) != null ? _f2 : model.modelId
|
3753
3763
|
};
|
@@ -3837,7 +3847,9 @@ async function generateText({
|
|
3837
3847
|
text: text2,
|
3838
3848
|
tools: tools != null ? tools : {},
|
3839
3849
|
toolCalls: currentToolCalls,
|
3840
|
-
toolResults: currentToolResults
|
3850
|
+
toolResults: currentToolResults,
|
3851
|
+
messageId: generateMessageId(),
|
3852
|
+
generateMessageId
|
3841
3853
|
})
|
3842
3854
|
);
|
3843
3855
|
}
|
@@ -3855,7 +3867,7 @@ async function generateText({
|
|
3855
3867
|
...currentModelResponse.response,
|
3856
3868
|
headers: (_d = currentModelResponse.rawResponse) == null ? void 0 : _d.headers,
|
3857
3869
|
// deep clone msgs to avoid mutating past messages in multi-step:
|
3858
|
-
messages:
|
3870
|
+
messages: structuredClone(responseMessages)
|
3859
3871
|
},
|
3860
3872
|
experimental_providerMetadata: currentModelResponse.providerMetadata,
|
3861
3873
|
isContinued: nextStepType === "continue"
|
@@ -4159,6 +4171,11 @@ function smoothStream({
|
|
4159
4171
|
import { createIdGenerator as createIdGenerator4 } from "@ai-sdk/provider-utils";
|
4160
4172
|
import { formatDataStreamPart as formatDataStreamPart2 } from "@ai-sdk/ui-utils";
|
4161
4173
|
|
4174
|
+
// util/as-array.ts
|
4175
|
+
function asArray(value) {
|
4176
|
+
return value === void 0 ? [] : Array.isArray(value) ? value : [value];
|
4177
|
+
}
|
4178
|
+
|
4162
4179
|
// core/util/merge-streams.ts
|
4163
4180
|
function mergeStreams(stream1, stream2) {
|
4164
4181
|
const reader1 = stream1.getReader();
|
@@ -4438,7 +4455,14 @@ function runToolsTransformation({
|
|
4438
4455
|
}
|
4439
4456
|
|
4440
4457
|
// core/generate-text/stream-text.ts
|
4441
|
-
var originalGenerateId4 = createIdGenerator4({
|
4458
|
+
var originalGenerateId4 = createIdGenerator4({
|
4459
|
+
prefix: "aitxt",
|
4460
|
+
size: 24
|
4461
|
+
});
|
4462
|
+
var originalGenerateMessageId2 = createIdGenerator4({
|
4463
|
+
prefix: "msg",
|
4464
|
+
size: 24
|
4465
|
+
});
|
4442
4466
|
function streamText({
|
4443
4467
|
model,
|
4444
4468
|
tools,
|
@@ -4450,6 +4474,7 @@ function streamText({
|
|
4450
4474
|
abortSignal,
|
4451
4475
|
headers,
|
4452
4476
|
maxSteps = 1,
|
4477
|
+
experimental_generateMessageId: generateMessageId = originalGenerateMessageId2,
|
4453
4478
|
experimental_output: output,
|
4454
4479
|
experimental_continueSteps: continueSteps = false,
|
4455
4480
|
experimental_telemetry: telemetry,
|
@@ -4463,7 +4488,7 @@ function streamText({
|
|
4463
4488
|
onStepFinish,
|
4464
4489
|
_internal: {
|
4465
4490
|
now: now2 = now,
|
4466
|
-
generateId:
|
4491
|
+
generateId: generateId3 = originalGenerateId4,
|
4467
4492
|
currentDate = () => /* @__PURE__ */ new Date()
|
4468
4493
|
} = {},
|
4469
4494
|
...settings
|
@@ -4481,7 +4506,7 @@ function streamText({
|
|
4481
4506
|
tools,
|
4482
4507
|
toolChoice,
|
4483
4508
|
toolCallStreaming,
|
4484
|
-
transforms:
|
4509
|
+
transforms: asArray(transform),
|
4485
4510
|
activeTools,
|
4486
4511
|
repairToolCall,
|
4487
4512
|
maxSteps,
|
@@ -4493,7 +4518,8 @@ function streamText({
|
|
4493
4518
|
onStepFinish,
|
4494
4519
|
now: now2,
|
4495
4520
|
currentDate,
|
4496
|
-
generateId:
|
4521
|
+
generateId: generateId3,
|
4522
|
+
generateMessageId
|
4497
4523
|
});
|
4498
4524
|
}
|
4499
4525
|
function createOutputTransformStream(output) {
|
@@ -4573,7 +4599,8 @@ var DefaultStreamTextResult = class {
|
|
4573
4599
|
onStepFinish,
|
4574
4600
|
now: now2,
|
4575
4601
|
currentDate,
|
4576
|
-
generateId:
|
4602
|
+
generateId: generateId3,
|
4603
|
+
generateMessageId
|
4577
4604
|
}) {
|
4578
4605
|
this.warningsPromise = new DelayedPromise();
|
4579
4606
|
this.usagePromise = new DelayedPromise();
|
@@ -4598,7 +4625,7 @@ var DefaultStreamTextResult = class {
|
|
4598
4625
|
let recordedContinuationText = "";
|
4599
4626
|
let recordedFullText = "";
|
4600
4627
|
const recordedResponse = {
|
4601
|
-
id:
|
4628
|
+
id: generateId3(),
|
4602
4629
|
timestamp: currentDate(),
|
4603
4630
|
modelId: model.modelId,
|
4604
4631
|
messages: []
|
@@ -4633,7 +4660,9 @@ var DefaultStreamTextResult = class {
|
|
4633
4660
|
text: recordedContinuationText,
|
4634
4661
|
tools: tools != null ? tools : {},
|
4635
4662
|
toolCalls: recordedToolCalls,
|
4636
|
-
toolResults: recordedToolResults
|
4663
|
+
toolResults: recordedToolResults,
|
4664
|
+
messageId: part.messageId,
|
4665
|
+
generateMessageId
|
4637
4666
|
});
|
4638
4667
|
const currentStep = recordedSteps.length;
|
4639
4668
|
let nextStepType = "done";
|
@@ -4808,7 +4837,8 @@ var DefaultStreamTextResult = class {
|
|
4808
4837
|
usage,
|
4809
4838
|
stepType: stepType2,
|
4810
4839
|
previousStepText,
|
4811
|
-
hasLeadingWhitespace
|
4840
|
+
hasLeadingWhitespace,
|
4841
|
+
messageId
|
4812
4842
|
}) {
|
4813
4843
|
const promptFormat = responseMessages.length === 0 ? initialPrompt.type : "messages";
|
4814
4844
|
const stepInputMessages = [
|
@@ -4916,7 +4946,7 @@ var DefaultStreamTextResult = class {
|
|
4916
4946
|
let fullStepText = stepType2 === "continue" ? previousStepText : "";
|
4917
4947
|
let stepLogProbs;
|
4918
4948
|
let stepResponse = {
|
4919
|
-
id:
|
4949
|
+
id: generateId3(),
|
4920
4950
|
timestamp: currentDate(),
|
4921
4951
|
modelId: model.modelId
|
4922
4952
|
};
|
@@ -4948,6 +4978,12 @@ var DefaultStreamTextResult = class {
|
|
4948
4978
|
doStreamSpan.setAttributes({
|
4949
4979
|
"ai.response.msToFirstChunk": msToFirstChunk
|
4950
4980
|
});
|
4981
|
+
controller.enqueue({
|
4982
|
+
type: "step-start",
|
4983
|
+
messageId,
|
4984
|
+
request: stepRequest,
|
4985
|
+
warnings: warnings != null ? warnings : []
|
4986
|
+
});
|
4951
4987
|
}
|
4952
4988
|
if (chunk.type === "text-delta" && chunk.textDelta.length === 0) {
|
4953
4989
|
return;
|
@@ -5092,7 +5128,8 @@ var DefaultStreamTextResult = class {
|
|
5092
5128
|
headers: rawResponse == null ? void 0 : rawResponse.headers
|
5093
5129
|
},
|
5094
5130
|
warnings,
|
5095
|
-
isContinued: nextStepType === "continue"
|
5131
|
+
isContinued: nextStepType === "continue",
|
5132
|
+
messageId
|
5096
5133
|
});
|
5097
5134
|
const combinedUsage = addLanguageModelUsage(usage, stepUsage);
|
5098
5135
|
if (nextStepType === "done") {
|
@@ -5125,7 +5162,9 @@ var DefaultStreamTextResult = class {
|
|
5125
5162
|
text: stepText,
|
5126
5163
|
tools: tools != null ? tools : {},
|
5127
5164
|
toolCalls: stepToolCalls,
|
5128
|
-
toolResults: stepToolResults
|
5165
|
+
toolResults: stepToolResults,
|
5166
|
+
messageId,
|
5167
|
+
generateMessageId
|
5129
5168
|
})
|
5130
5169
|
);
|
5131
5170
|
}
|
@@ -5135,7 +5174,11 @@ var DefaultStreamTextResult = class {
|
|
5135
5174
|
usage: combinedUsage,
|
5136
5175
|
stepType: nextStepType,
|
5137
5176
|
previousStepText: fullStepText,
|
5138
|
-
hasLeadingWhitespace: hasWhitespaceSuffix
|
5177
|
+
hasLeadingWhitespace: hasWhitespaceSuffix,
|
5178
|
+
messageId: (
|
5179
|
+
// keep the same id when continuing a step:
|
5180
|
+
nextStepType === "continue" ? messageId : generateMessageId()
|
5181
|
+
)
|
5139
5182
|
});
|
5140
5183
|
}
|
5141
5184
|
}
|
@@ -5153,7 +5196,8 @@ var DefaultStreamTextResult = class {
|
|
5153
5196
|
},
|
5154
5197
|
previousStepText: "",
|
5155
5198
|
stepType: "initial",
|
5156
|
-
hasLeadingWhitespace: false
|
5199
|
+
hasLeadingWhitespace: false,
|
5200
|
+
messageId: generateMessageId()
|
5157
5201
|
});
|
5158
5202
|
}
|
5159
5203
|
}).catch((error) => {
|
@@ -5318,6 +5362,14 @@ var DefaultStreamTextResult = class {
|
|
5318
5362
|
);
|
5319
5363
|
break;
|
5320
5364
|
}
|
5365
|
+
case "step-start": {
|
5366
|
+
controller.enqueue(
|
5367
|
+
formatDataStreamPart2("start_step", {
|
5368
|
+
messageId: chunk.messageId
|
5369
|
+
})
|
5370
|
+
);
|
5371
|
+
break;
|
5372
|
+
}
|
5321
5373
|
case "step-finish": {
|
5322
5374
|
controller.enqueue(
|
5323
5375
|
formatDataStreamPart2("finish_step", {
|
@@ -5463,7 +5515,6 @@ var experimental_wrapLanguageModel = ({
|
|
5463
5515
|
};
|
5464
5516
|
|
5465
5517
|
// core/prompt/append-response-messages.ts
|
5466
|
-
import { generateId as generateId2 } from "@ai-sdk/provider-utils";
|
5467
5518
|
function appendResponseMessages({
|
5468
5519
|
messages,
|
5469
5520
|
responseMessages
|
@@ -5473,17 +5524,10 @@ function appendResponseMessages({
|
|
5473
5524
|
for (const message of responseMessages) {
|
5474
5525
|
const role = message.role;
|
5475
5526
|
switch (role) {
|
5476
|
-
case "system":
|
5477
|
-
case "user": {
|
5478
|
-
throw new Error(
|
5479
|
-
"AI response must not contain system or user messages: " + role
|
5480
|
-
);
|
5481
|
-
}
|
5482
5527
|
case "assistant": {
|
5483
5528
|
clonedMessages.push({
|
5484
5529
|
role: "assistant",
|
5485
|
-
id:
|
5486
|
-
// generate an id for the message, will be overridden by the client
|
5530
|
+
id: message.id,
|
5487
5531
|
createdAt: /* @__PURE__ */ new Date(),
|
5488
5532
|
// generate a createdAt date for the message, will be overridden by the client
|
5489
5533
|
// only include text in the content:
|
@@ -5517,6 +5561,10 @@ function appendResponseMessages({
|
|
5517
5561
|
}
|
5518
5562
|
break;
|
5519
5563
|
}
|
5564
|
+
default: {
|
5565
|
+
const _exhaustiveCheck = role;
|
5566
|
+
throw new Error(`Unsupported message role: ${_exhaustiveCheck}`);
|
5567
|
+
}
|
5520
5568
|
}
|
5521
5569
|
}
|
5522
5570
|
return clonedMessages;
|
@@ -5708,7 +5756,7 @@ function simulateReadableStream({
|
|
5708
5756
|
|
5709
5757
|
// streams/assistant-response.ts
|
5710
5758
|
import {
|
5711
|
-
formatAssistantStreamPart
|
5759
|
+
formatAssistantStreamPart as formatAssistantStreamPart2
|
5712
5760
|
} from "@ai-sdk/ui-utils";
|
5713
5761
|
function AssistantResponse({ threadId, messageId }, process2) {
|
5714
5762
|
const stream = new ReadableStream({
|
@@ -5718,20 +5766,20 @@ function AssistantResponse({ threadId, messageId }, process2) {
|
|
5718
5766
|
const sendMessage = (message) => {
|
5719
5767
|
controller.enqueue(
|
5720
5768
|
textEncoder.encode(
|
5721
|
-
|
5769
|
+
formatAssistantStreamPart2("assistant_message", message)
|
5722
5770
|
)
|
5723
5771
|
);
|
5724
5772
|
};
|
5725
5773
|
const sendDataMessage = (message) => {
|
5726
5774
|
controller.enqueue(
|
5727
5775
|
textEncoder.encode(
|
5728
|
-
|
5776
|
+
formatAssistantStreamPart2("data_message", message)
|
5729
5777
|
)
|
5730
5778
|
);
|
5731
5779
|
};
|
5732
5780
|
const sendError = (errorMessage) => {
|
5733
5781
|
controller.enqueue(
|
5734
|
-
textEncoder.encode(
|
5782
|
+
textEncoder.encode(formatAssistantStreamPart2("error", errorMessage))
|
5735
5783
|
);
|
5736
5784
|
};
|
5737
5785
|
const forwardStream = async (stream2) => {
|
@@ -5742,7 +5790,7 @@ function AssistantResponse({ threadId, messageId }, process2) {
|
|
5742
5790
|
case "thread.message.created": {
|
5743
5791
|
controller.enqueue(
|
5744
5792
|
textEncoder.encode(
|
5745
|
-
|
5793
|
+
formatAssistantStreamPart2("assistant_message", {
|
5746
5794
|
id: value.data.id,
|
5747
5795
|
role: "assistant",
|
5748
5796
|
content: [{ type: "text", text: { value: "" } }]
|
@@ -5756,7 +5804,7 @@ function AssistantResponse({ threadId, messageId }, process2) {
|
|
5756
5804
|
if ((content == null ? void 0 : content.type) === "text" && ((_b = content.text) == null ? void 0 : _b.value) != null) {
|
5757
5805
|
controller.enqueue(
|
5758
5806
|
textEncoder.encode(
|
5759
|
-
|
5807
|
+
formatAssistantStreamPart2("text", content.text.value)
|
5760
5808
|
)
|
5761
5809
|
);
|
5762
5810
|
}
|
@@ -5773,7 +5821,7 @@ function AssistantResponse({ threadId, messageId }, process2) {
|
|
5773
5821
|
};
|
5774
5822
|
controller.enqueue(
|
5775
5823
|
textEncoder.encode(
|
5776
|
-
|
5824
|
+
formatAssistantStreamPart2("assistant_control_data", {
|
5777
5825
|
threadId,
|
5778
5826
|
messageId
|
5779
5827
|
})
|
@@ -5811,7 +5859,7 @@ __export(langchain_adapter_exports, {
|
|
5811
5859
|
toDataStream: () => toDataStream,
|
5812
5860
|
toDataStreamResponse: () => toDataStreamResponse
|
5813
5861
|
});
|
5814
|
-
import { formatDataStreamPart as
|
5862
|
+
import { formatDataStreamPart as formatDataStreamPart4 } from "@ai-sdk/ui-utils";
|
5815
5863
|
|
5816
5864
|
// streams/stream-callbacks.ts
|
5817
5865
|
function createCallbacksTransformer(callbacks = {}) {
|
@@ -5867,7 +5915,7 @@ function toDataStreamInternal(stream, callbacks) {
|
|
5867
5915
|
).pipeThrough(createCallbacksTransformer(callbacks)).pipeThrough(new TextDecoderStream()).pipeThrough(
|
5868
5916
|
new TransformStream({
|
5869
5917
|
transform: async (chunk, controller) => {
|
5870
|
-
controller.enqueue(
|
5918
|
+
controller.enqueue(formatDataStreamPart4("text", chunk));
|
5871
5919
|
}
|
5872
5920
|
})
|
5873
5921
|
);
|
@@ -5919,7 +5967,7 @@ __export(llamaindex_adapter_exports, {
|
|
5919
5967
|
toDataStreamResponse: () => toDataStreamResponse2
|
5920
5968
|
});
|
5921
5969
|
import { convertAsyncIteratorToReadableStream } from "@ai-sdk/provider-utils";
|
5922
|
-
import { formatDataStreamPart as
|
5970
|
+
import { formatDataStreamPart as formatDataStreamPart5 } from "@ai-sdk/ui-utils";
|
5923
5971
|
function toDataStreamInternal2(stream, callbacks) {
|
5924
5972
|
const trimStart = trimStartOfStream();
|
5925
5973
|
return convertAsyncIteratorToReadableStream(stream[Symbol.asyncIterator]()).pipeThrough(
|
@@ -5931,7 +5979,7 @@ function toDataStreamInternal2(stream, callbacks) {
|
|
5931
5979
|
).pipeThrough(createCallbacksTransformer(callbacks)).pipeThrough(new TextDecoderStream()).pipeThrough(
|
5932
5980
|
new TransformStream({
|
5933
5981
|
transform: async (chunk, controller) => {
|
5934
|
-
controller.enqueue(
|
5982
|
+
controller.enqueue(formatDataStreamPart5("text", chunk));
|
5935
5983
|
}
|
5936
5984
|
})
|
5937
5985
|
);
|
@@ -5973,7 +6021,7 @@ function trimStartOfStream() {
|
|
5973
6021
|
}
|
5974
6022
|
|
5975
6023
|
// streams/stream-data.ts
|
5976
|
-
import { formatDataStreamPart as
|
6024
|
+
import { formatDataStreamPart as formatDataStreamPart6 } from "@ai-sdk/ui-utils";
|
5977
6025
|
|
5978
6026
|
// util/constants.ts
|
5979
6027
|
var HANGING_STREAM_WARNING_TIME_MS = 15 * 1e3;
|
@@ -6025,7 +6073,7 @@ var StreamData = class {
|
|
6025
6073
|
throw new Error("Stream controller is not initialized.");
|
6026
6074
|
}
|
6027
6075
|
this.controller.enqueue(
|
6028
|
-
this.encoder.encode(
|
6076
|
+
this.encoder.encode(formatDataStreamPart6("data", [value]))
|
6029
6077
|
);
|
6030
6078
|
}
|
6031
6079
|
appendMessageAnnotation(value) {
|
@@ -6036,7 +6084,7 @@ var StreamData = class {
|
|
6036
6084
|
throw new Error("Stream controller is not initialized.");
|
6037
6085
|
}
|
6038
6086
|
this.controller.enqueue(
|
6039
|
-
this.encoder.encode(
|
6087
|
+
this.encoder.encode(formatDataStreamPart6("message_annotations", [value]))
|
6040
6088
|
);
|
6041
6089
|
}
|
6042
6090
|
};
|
@@ -6075,15 +6123,16 @@ export {
|
|
6075
6123
|
cosineSimilarity,
|
6076
6124
|
createDataStream,
|
6077
6125
|
createDataStreamResponse,
|
6126
|
+
createIdGenerator5 as createIdGenerator,
|
6078
6127
|
embed,
|
6079
6128
|
embedMany,
|
6080
6129
|
experimental_createProviderRegistry,
|
6081
6130
|
experimental_customProvider,
|
6082
6131
|
generateImage as experimental_generateImage,
|
6083
6132
|
experimental_wrapLanguageModel,
|
6084
|
-
|
6085
|
-
|
6086
|
-
|
6133
|
+
formatAssistantStreamPart,
|
6134
|
+
formatDataStreamPart3 as formatDataStreamPart,
|
6135
|
+
generateId2 as generateId,
|
6087
6136
|
generateObject,
|
6088
6137
|
generateText,
|
6089
6138
|
jsonSchema,
|