@mindstudio-ai/remy 0.1.23 → 0.1.25
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/headless.js +11 -5
- package/dist/index.js +11 -5
- package/dist/subagents/browserAutomation/prompt.md +1 -1
- package/package.json +1 -1
package/dist/headless.js
CHANGED
|
@@ -2048,7 +2048,7 @@ var screenshotTool = {
|
|
|
2048
2048
|
const { url } = await sidecarRequest(
|
|
2049
2049
|
"/screenshot",
|
|
2050
2050
|
{},
|
|
2051
|
-
{ timeout:
|
|
2051
|
+
{ timeout: 12e4 }
|
|
2052
2052
|
);
|
|
2053
2053
|
const analysisPrompt = input.prompt || DEFAULT_PROMPT;
|
|
2054
2054
|
const analysis = await runCli(
|
|
@@ -2526,7 +2526,7 @@ async function executeDesignExpertTool(name, input) {
|
|
|
2526
2526
|
const { url } = await sidecarRequest(
|
|
2527
2527
|
"/screenshot",
|
|
2528
2528
|
{},
|
|
2529
|
-
{ timeout:
|
|
2529
|
+
{ timeout: 12e4 }
|
|
2530
2530
|
);
|
|
2531
2531
|
const analysisPrompt = input.prompt || "Describe this app screenshot for a visual designer reviewing the current state. What is visible: layout, typography, colors, spacing, imagery. Note anything that looks broken or off. Be concise.";
|
|
2532
2532
|
const analysis = await runCli(
|
|
@@ -3643,7 +3643,9 @@ function startStatusWatcher(config) {
|
|
|
3643
3643
|
body: JSON.stringify({
|
|
3644
3644
|
assistantText: ctx.assistantText.slice(-500),
|
|
3645
3645
|
lastToolName: ctx.lastToolName,
|
|
3646
|
-
lastToolResult: ctx.lastToolResult?.slice(-200)
|
|
3646
|
+
lastToolResult: ctx.lastToolResult?.slice(-200),
|
|
3647
|
+
onboardingState: ctx.onboardingState,
|
|
3648
|
+
userMessage: ctx.userMessage?.slice(-200)
|
|
3647
3649
|
}),
|
|
3648
3650
|
signal
|
|
3649
3651
|
});
|
|
@@ -3873,7 +3875,9 @@ async function runTurn(params) {
|
|
|
3873
3875
|
getContext: () => ({
|
|
3874
3876
|
assistantText: getTextContent(contentBlocks).slice(-500),
|
|
3875
3877
|
lastToolName: getToolCalls(contentBlocks).filter((tc) => !STATUS_EXCLUDED_TOOLS.has(tc.name)).at(-1)?.name || lastCompletedTools || void 0,
|
|
3876
|
-
lastToolResult: lastCompletedResult || void 0
|
|
3878
|
+
lastToolResult: lastCompletedResult || void 0,
|
|
3879
|
+
onboardingState,
|
|
3880
|
+
userMessage
|
|
3877
3881
|
}),
|
|
3878
3882
|
onStatus: (label) => onEvent({ type: "status", message: label }),
|
|
3879
3883
|
signal
|
|
@@ -4048,7 +4052,9 @@ async function runTurn(params) {
|
|
|
4048
4052
|
getContext: () => ({
|
|
4049
4053
|
assistantText: subAgentText || getTextContent(contentBlocks).slice(-500),
|
|
4050
4054
|
lastToolName: toolCalls.filter((tc) => !STATUS_EXCLUDED_TOOLS.has(tc.name)).map((tc) => tc.name).join(", ") || void 0,
|
|
4051
|
-
lastToolResult: lastCompletedResult || void 0
|
|
4055
|
+
lastToolResult: lastCompletedResult || void 0,
|
|
4056
|
+
onboardingState,
|
|
4057
|
+
userMessage
|
|
4052
4058
|
}),
|
|
4053
4059
|
onStatus: (label) => origOnEvent({ type: "status", message: label }),
|
|
4054
4060
|
signal
|
package/dist/index.js
CHANGED
|
@@ -1993,7 +1993,7 @@ var init_screenshot = __esm({
|
|
|
1993
1993
|
const { url } = await sidecarRequest(
|
|
1994
1994
|
"/screenshot",
|
|
1995
1995
|
{},
|
|
1996
|
-
{ timeout:
|
|
1996
|
+
{ timeout: 12e4 }
|
|
1997
1997
|
);
|
|
1998
1998
|
const analysisPrompt = input.prompt || DEFAULT_PROMPT;
|
|
1999
1999
|
const analysis = await runCli(
|
|
@@ -2402,7 +2402,7 @@ async function executeDesignExpertTool(name, input) {
|
|
|
2402
2402
|
const { url } = await sidecarRequest(
|
|
2403
2403
|
"/screenshot",
|
|
2404
2404
|
{},
|
|
2405
|
-
{ timeout:
|
|
2405
|
+
{ timeout: 12e4 }
|
|
2406
2406
|
);
|
|
2407
2407
|
const analysisPrompt = input.prompt || "Describe this app screenshot for a visual designer reviewing the current state. What is visible: layout, typography, colors, spacing, imagery. Note anything that looks broken or off. Be concise.";
|
|
2408
2408
|
const analysis = await runCli(
|
|
@@ -3747,7 +3747,9 @@ function startStatusWatcher(config) {
|
|
|
3747
3747
|
body: JSON.stringify({
|
|
3748
3748
|
assistantText: ctx.assistantText.slice(-500),
|
|
3749
3749
|
lastToolName: ctx.lastToolName,
|
|
3750
|
-
lastToolResult: ctx.lastToolResult?.slice(-200)
|
|
3750
|
+
lastToolResult: ctx.lastToolResult?.slice(-200),
|
|
3751
|
+
onboardingState: ctx.onboardingState,
|
|
3752
|
+
userMessage: ctx.userMessage?.slice(-200)
|
|
3751
3753
|
}),
|
|
3752
3754
|
signal
|
|
3753
3755
|
});
|
|
@@ -3976,7 +3978,9 @@ async function runTurn(params) {
|
|
|
3976
3978
|
getContext: () => ({
|
|
3977
3979
|
assistantText: getTextContent(contentBlocks).slice(-500),
|
|
3978
3980
|
lastToolName: getToolCalls(contentBlocks).filter((tc) => !STATUS_EXCLUDED_TOOLS.has(tc.name)).at(-1)?.name || lastCompletedTools || void 0,
|
|
3979
|
-
lastToolResult: lastCompletedResult || void 0
|
|
3981
|
+
lastToolResult: lastCompletedResult || void 0,
|
|
3982
|
+
onboardingState,
|
|
3983
|
+
userMessage
|
|
3980
3984
|
}),
|
|
3981
3985
|
onStatus: (label) => onEvent({ type: "status", message: label }),
|
|
3982
3986
|
signal
|
|
@@ -4151,7 +4155,9 @@ async function runTurn(params) {
|
|
|
4151
4155
|
getContext: () => ({
|
|
4152
4156
|
assistantText: subAgentText || getTextContent(contentBlocks).slice(-500),
|
|
4153
4157
|
lastToolName: toolCalls.filter((tc) => !STATUS_EXCLUDED_TOOLS.has(tc.name)).map((tc) => tc.name).join(", ") || void 0,
|
|
4154
|
-
lastToolResult: lastCompletedResult || void 0
|
|
4158
|
+
lastToolResult: lastCompletedResult || void 0,
|
|
4159
|
+
onboardingState,
|
|
4160
|
+
userMessage
|
|
4155
4161
|
}),
|
|
4156
4162
|
onStatus: (label) => origOnEvent({ type: "status", message: label }),
|
|
4157
4163
|
signal
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
You are a browser smoke test agent. You verify that features work end to end by interacting with the live preview. Focus on outcomes: does the feature work? Did the expected content appear? Just do the thing and see if it worked.
|
|
2
2
|
|
|
3
|
-
The user is watching the automation happen on their screen in real-time. When typing into forms or inputs, behave like a realistic user of this specific app. Use the app context (if provided) to understand the audience and tone. Type the way that audience would actually type — not formal, not robotic.
|
|
3
|
+
The user is watching the automation happen on their screen in real-time. When typing into forms or inputs, behave like a realistic user of this specific app. Use the app context (if provided) to understand the audience and tone. Type the way that audience would actually type — not formal, not robotic. The coding agent's name is Remy, so use that and the email remy@mindstudio.ai for any testing that requires a name or email.
|
|
4
4
|
|
|
5
5
|
## Snapshot format
|
|
6
6
|
|