@mindstudio-ai/remy 0.1.23 → 0.1.25

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/headless.js CHANGED
@@ -2048,7 +2048,7 @@ var screenshotTool = {
2048
2048
  const { url } = await sidecarRequest(
2049
2049
  "/screenshot",
2050
2050
  {},
2051
- { timeout: 3e4 }
2051
+ { timeout: 12e4 }
2052
2052
  );
2053
2053
  const analysisPrompt = input.prompt || DEFAULT_PROMPT;
2054
2054
  const analysis = await runCli(
@@ -2526,7 +2526,7 @@ async function executeDesignExpertTool(name, input) {
2526
2526
  const { url } = await sidecarRequest(
2527
2527
  "/screenshot",
2528
2528
  {},
2529
- { timeout: 3e4 }
2529
+ { timeout: 12e4 }
2530
2530
  );
2531
2531
  const analysisPrompt = input.prompt || "Describe this app screenshot for a visual designer reviewing the current state. What is visible: layout, typography, colors, spacing, imagery. Note anything that looks broken or off. Be concise.";
2532
2532
  const analysis = await runCli(
@@ -3643,7 +3643,9 @@ function startStatusWatcher(config) {
3643
3643
  body: JSON.stringify({
3644
3644
  assistantText: ctx.assistantText.slice(-500),
3645
3645
  lastToolName: ctx.lastToolName,
3646
- lastToolResult: ctx.lastToolResult?.slice(-200)
3646
+ lastToolResult: ctx.lastToolResult?.slice(-200),
3647
+ onboardingState: ctx.onboardingState,
3648
+ userMessage: ctx.userMessage?.slice(-200)
3647
3649
  }),
3648
3650
  signal
3649
3651
  });
@@ -3873,7 +3875,9 @@ async function runTurn(params) {
3873
3875
  getContext: () => ({
3874
3876
  assistantText: getTextContent(contentBlocks).slice(-500),
3875
3877
  lastToolName: getToolCalls(contentBlocks).filter((tc) => !STATUS_EXCLUDED_TOOLS.has(tc.name)).at(-1)?.name || lastCompletedTools || void 0,
3876
- lastToolResult: lastCompletedResult || void 0
3878
+ lastToolResult: lastCompletedResult || void 0,
3879
+ onboardingState,
3880
+ userMessage
3877
3881
  }),
3878
3882
  onStatus: (label) => onEvent({ type: "status", message: label }),
3879
3883
  signal
@@ -4048,7 +4052,9 @@ async function runTurn(params) {
4048
4052
  getContext: () => ({
4049
4053
  assistantText: subAgentText || getTextContent(contentBlocks).slice(-500),
4050
4054
  lastToolName: toolCalls.filter((tc) => !STATUS_EXCLUDED_TOOLS.has(tc.name)).map((tc) => tc.name).join(", ") || void 0,
4051
- lastToolResult: lastCompletedResult || void 0
4055
+ lastToolResult: lastCompletedResult || void 0,
4056
+ onboardingState,
4057
+ userMessage
4052
4058
  }),
4053
4059
  onStatus: (label) => origOnEvent({ type: "status", message: label }),
4054
4060
  signal
package/dist/index.js CHANGED
@@ -1993,7 +1993,7 @@ var init_screenshot = __esm({
1993
1993
  const { url } = await sidecarRequest(
1994
1994
  "/screenshot",
1995
1995
  {},
1996
- { timeout: 3e4 }
1996
+ { timeout: 12e4 }
1997
1997
  );
1998
1998
  const analysisPrompt = input.prompt || DEFAULT_PROMPT;
1999
1999
  const analysis = await runCli(
@@ -2402,7 +2402,7 @@ async function executeDesignExpertTool(name, input) {
2402
2402
  const { url } = await sidecarRequest(
2403
2403
  "/screenshot",
2404
2404
  {},
2405
- { timeout: 3e4 }
2405
+ { timeout: 12e4 }
2406
2406
  );
2407
2407
  const analysisPrompt = input.prompt || "Describe this app screenshot for a visual designer reviewing the current state. What is visible: layout, typography, colors, spacing, imagery. Note anything that looks broken or off. Be concise.";
2408
2408
  const analysis = await runCli(
@@ -3747,7 +3747,9 @@ function startStatusWatcher(config) {
3747
3747
  body: JSON.stringify({
3748
3748
  assistantText: ctx.assistantText.slice(-500),
3749
3749
  lastToolName: ctx.lastToolName,
3750
- lastToolResult: ctx.lastToolResult?.slice(-200)
3750
+ lastToolResult: ctx.lastToolResult?.slice(-200),
3751
+ onboardingState: ctx.onboardingState,
3752
+ userMessage: ctx.userMessage?.slice(-200)
3751
3753
  }),
3752
3754
  signal
3753
3755
  });
@@ -3976,7 +3978,9 @@ async function runTurn(params) {
3976
3978
  getContext: () => ({
3977
3979
  assistantText: getTextContent(contentBlocks).slice(-500),
3978
3980
  lastToolName: getToolCalls(contentBlocks).filter((tc) => !STATUS_EXCLUDED_TOOLS.has(tc.name)).at(-1)?.name || lastCompletedTools || void 0,
3979
- lastToolResult: lastCompletedResult || void 0
3981
+ lastToolResult: lastCompletedResult || void 0,
3982
+ onboardingState,
3983
+ userMessage
3980
3984
  }),
3981
3985
  onStatus: (label) => onEvent({ type: "status", message: label }),
3982
3986
  signal
@@ -4151,7 +4155,9 @@ async function runTurn(params) {
4151
4155
  getContext: () => ({
4152
4156
  assistantText: subAgentText || getTextContent(contentBlocks).slice(-500),
4153
4157
  lastToolName: toolCalls.filter((tc) => !STATUS_EXCLUDED_TOOLS.has(tc.name)).map((tc) => tc.name).join(", ") || void 0,
4154
- lastToolResult: lastCompletedResult || void 0
4158
+ lastToolResult: lastCompletedResult || void 0,
4159
+ onboardingState,
4160
+ userMessage
4155
4161
  }),
4156
4162
  onStatus: (label) => origOnEvent({ type: "status", message: label }),
4157
4163
  signal
@@ -1,6 +1,6 @@
1
1
  You are a browser smoke test agent. You verify that features work end to end by interacting with the live preview. Focus on outcomes: does the feature work? Did the expected content appear? Just do the thing and see if it worked.
2
2
 
3
- The user is watching the automation happen on their screen in real-time. When typing into forms or inputs, behave like a realistic user of this specific app. Use the app context (if provided) to understand the audience and tone. Type the way that audience would actually type — not formal, not robotic.
3
+ The user is watching the automation happen on their screen in real-time. When typing into forms or inputs, behave like a realistic user of this specific app. Use the app context (if provided) to understand the audience and tone. Type the way that audience would actually type — not formal, not robotic. The coding agent's name is Remy, so use that and the email remy@mindstudio.ai for any testing that requires a name or email.
4
4
 
5
5
  ## Snapshot format
6
6
 
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@mindstudio-ai/remy",
3
- "version": "0.1.23",
3
+ "version": "0.1.25",
4
4
  "description": "MindStudio coding agent",
5
5
  "repository": {
6
6
  "type": "git",