@mastra/ai-sdk 0.0.0-feat-mcp-embedded-docs-tools-clean-20260105080340 → 0.0.0-feat-mcp-embedded-docs-tools-clean-20260108085651
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +98 -3
- package/dist/chat-route.d.ts +2 -0
- package/dist/chat-route.d.ts.map +1 -1
- package/dist/index.cjs +18 -14
- package/dist/index.cjs.map +1 -1
- package/dist/index.js +19 -15
- package/dist/index.js.map +1 -1
- package/dist/middleware.d.ts.map +1 -1
- package/package.json +8 -8
package/dist/index.js
CHANGED
|
@@ -3,7 +3,7 @@ import * as z4 from 'zod/v4';
|
|
|
3
3
|
import { z } from 'zod/v4';
|
|
4
4
|
import { ZodFirstPartyTypeKind } from 'zod/v3';
|
|
5
5
|
import { convertFullStreamChunkToMastra, DefaultGeneratedFile, DefaultGeneratedFileWithType } from '@mastra/core/stream';
|
|
6
|
-
import { TripWire, MessageList } from '@mastra/core/agent';
|
|
6
|
+
import { TripWire, MessageList, aiV5ModelMessageToV2PromptMessage } from '@mastra/core/agent';
|
|
7
7
|
import { RequestContext } from '@mastra/core/di';
|
|
8
8
|
import { WorkingMemory, MessageHistory, SemanticRecall } from '@mastra/core/processors';
|
|
9
9
|
|
|
@@ -218,7 +218,7 @@ function secureJsonParse(text2) {
|
|
|
218
218
|
Error.stackTraceLimit = stackTraceLimit;
|
|
219
219
|
}
|
|
220
220
|
}
|
|
221
|
-
var validatorSymbol = Symbol.for("vercel.ai.validator");
|
|
221
|
+
var validatorSymbol = /* @__PURE__ */ Symbol.for("vercel.ai.validator");
|
|
222
222
|
function validator(validate) {
|
|
223
223
|
return { [validatorSymbol]: true, validate };
|
|
224
224
|
}
|
|
@@ -301,7 +301,7 @@ var getRelativePath = (pathA, pathB) => {
|
|
|
301
301
|
}
|
|
302
302
|
return [(pathA.length - i).toString(), ...pathB.slice(i)].join("/");
|
|
303
303
|
};
|
|
304
|
-
var ignoreOverride = Symbol(
|
|
304
|
+
var ignoreOverride = /* @__PURE__ */ Symbol(
|
|
305
305
|
"Let zodToJsonSchema decide on which parser to use"
|
|
306
306
|
);
|
|
307
307
|
var defaultOptions = {
|
|
@@ -1421,7 +1421,7 @@ function zodSchema(zodSchema2, options) {
|
|
|
1421
1421
|
return zod3Schema(zodSchema2);
|
|
1422
1422
|
}
|
|
1423
1423
|
}
|
|
1424
|
-
var schemaSymbol = Symbol.for("vercel.ai.schema");
|
|
1424
|
+
var schemaSymbol = /* @__PURE__ */ Symbol.for("vercel.ai.schema");
|
|
1425
1425
|
function jsonSchema(jsonSchema2, {
|
|
1426
1426
|
validate
|
|
1427
1427
|
} = {}) {
|
|
@@ -3329,7 +3329,7 @@ function convertFullStreamChunkToUIMessageStream({
|
|
|
3329
3329
|
}
|
|
3330
3330
|
|
|
3331
3331
|
// src/transformers.ts
|
|
3332
|
-
var PRIMITIVE_CACHE_SYMBOL = Symbol("primitive-cache");
|
|
3332
|
+
var PRIMITIVE_CACHE_SYMBOL = /* @__PURE__ */ Symbol("primitive-cache");
|
|
3333
3333
|
function WorkflowStreamToAISDKTransformer({
|
|
3334
3334
|
includeTextStreamParts
|
|
3335
3335
|
} = {}) {
|
|
@@ -4101,7 +4101,7 @@ async function handleChatStream({
|
|
|
4101
4101
|
sendReasoning = false,
|
|
4102
4102
|
sendSources = false
|
|
4103
4103
|
}) {
|
|
4104
|
-
const { messages, resumeData, runId, requestContext, ...rest } = params;
|
|
4104
|
+
const { messages, resumeData, runId, requestContext, trigger, ...rest } = params;
|
|
4105
4105
|
if (resumeData && !runId) {
|
|
4106
4106
|
throw new Error("runId is required when resumeData is provided");
|
|
4107
4107
|
}
|
|
@@ -4112,20 +4112,24 @@ async function handleChatStream({
|
|
|
4112
4112
|
if (!Array.isArray(messages)) {
|
|
4113
4113
|
throw new Error("Messages must be an array of UIMessage objects");
|
|
4114
4114
|
}
|
|
4115
|
-
const mergedOptions = {
|
|
4116
|
-
...defaultOptions2,
|
|
4117
|
-
...rest,
|
|
4118
|
-
...runId && { runId },
|
|
4119
|
-
requestContext: requestContext || defaultOptions2?.requestContext
|
|
4120
|
-
};
|
|
4121
|
-
const result = resumeData ? await agentObj.resumeStream(resumeData, mergedOptions) : await agentObj.stream(messages, mergedOptions);
|
|
4122
4115
|
let lastMessageId;
|
|
4123
|
-
|
|
4116
|
+
let messagesToSend = messages;
|
|
4117
|
+
if (messages.length > 0) {
|
|
4124
4118
|
const lastMessage = messages[messages.length - 1];
|
|
4125
4119
|
if (lastMessage?.role === "assistant") {
|
|
4126
4120
|
lastMessageId = lastMessage.id;
|
|
4121
|
+
if (trigger === "regenerate-message") {
|
|
4122
|
+
messagesToSend = messages.slice(0, -1);
|
|
4123
|
+
}
|
|
4127
4124
|
}
|
|
4128
4125
|
}
|
|
4126
|
+
const mergedOptions = {
|
|
4127
|
+
...defaultOptions2,
|
|
4128
|
+
...rest,
|
|
4129
|
+
...runId && { runId },
|
|
4130
|
+
requestContext: requestContext || defaultOptions2?.requestContext
|
|
4131
|
+
};
|
|
4132
|
+
const result = resumeData ? await agentObj.resumeStream(resumeData, mergedOptions) : await agentObj.stream(messagesToSend, mergedOptions);
|
|
4129
4133
|
return createUIMessageStream({
|
|
4130
4134
|
originalMessages: messages,
|
|
4131
4135
|
execute: async ({ writer }) => {
|
|
@@ -4632,7 +4636,7 @@ function createProcessorMiddleware(options) {
|
|
|
4632
4636
|
}
|
|
4633
4637
|
}
|
|
4634
4638
|
}
|
|
4635
|
-
const newPrompt = messageList.get.all.aiV5.prompt().map(
|
|
4639
|
+
const newPrompt = messageList.get.all.aiV5.prompt().map(aiV5ModelMessageToV2PromptMessage);
|
|
4636
4640
|
return {
|
|
4637
4641
|
...params,
|
|
4638
4642
|
prompt: newPrompt
|