@alwaysmeticulous/debug-workspace 2.261.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/LICENSE +15 -0
- package/dist/debug-constants.d.ts +3 -0
- package/dist/debug-constants.js +10 -0
- package/dist/debug-constants.js.map +1 -0
- package/dist/debug.types.d.ts +21 -0
- package/dist/debug.types.js +3 -0
- package/dist/debug.types.js.map +1 -0
- package/dist/download-debug-data.d.ts +10 -0
- package/dist/download-debug-data.js +141 -0
- package/dist/download-debug-data.js.map +1 -0
- package/dist/generate-debug-workspace.d.ts +31 -0
- package/dist/generate-debug-workspace.js +1302 -0
- package/dist/generate-debug-workspace.js.map +1 -0
- package/dist/index.d.ts +6 -0
- package/dist/index.js +16 -0
- package/dist/index.js.map +1 -0
- package/dist/pipeline.d.ts +21 -0
- package/dist/pipeline.js +63 -0
- package/dist/pipeline.js.map +1 -0
- package/dist/resolve-debug-context.d.ts +12 -0
- package/dist/resolve-debug-context.js +187 -0
- package/dist/resolve-debug-context.js.map +1 -0
- package/dist/templates/CLAUDE.md +206 -0
- package/dist/templates/agents/planner.md +65 -0
- package/dist/templates/agents/summarizer.md +75 -0
- package/dist/templates/hooks/check-file-size.sh +36 -0
- package/dist/templates/hooks/load-context.sh +20 -0
- package/dist/templates/rules/feedback.md +37 -0
- package/dist/templates/settings.json +82 -0
- package/dist/templates/skills/debugging-diffs/SKILL.md +57 -0
- package/dist/templates/skills/debugging-flakes/SKILL.md +52 -0
- package/dist/templates/skills/debugging-network/SKILL.md +45 -0
- package/dist/templates/skills/debugging-sessions/SKILL.md +47 -0
- package/dist/templates/skills/debugging-timelines/SKILL.md +51 -0
- package/dist/templates/skills/pr-analysis/SKILL.md +20 -0
- package/dist/templates/templates/CLAUDE.md +206 -0
- package/dist/templates/templates/agents/planner.md +65 -0
- package/dist/templates/templates/agents/summarizer.md +75 -0
- package/dist/templates/templates/hooks/check-file-size.sh +36 -0
- package/dist/templates/templates/hooks/load-context.sh +20 -0
- package/dist/templates/templates/rules/feedback.md +37 -0
- package/dist/templates/templates/settings.json +82 -0
- package/dist/templates/templates/skills/debugging-diffs/SKILL.md +57 -0
- package/dist/templates/templates/skills/debugging-flakes/SKILL.md +52 -0
- package/dist/templates/templates/skills/debugging-network/SKILL.md +45 -0
- package/dist/templates/templates/skills/debugging-sessions/SKILL.md +47 -0
- package/dist/templates/templates/skills/debugging-timelines/SKILL.md +51 -0
- package/dist/templates/templates/skills/pr-analysis/SKILL.md +20 -0
- package/package.json +49 -0
|
@@ -0,0 +1,82 @@
|
|
|
1
|
+
{
|
|
2
|
+
"permissions": {
|
|
3
|
+
"allow": [
|
|
4
|
+
"Read",
|
|
5
|
+
"Grep",
|
|
6
|
+
"Glob",
|
|
7
|
+
"Edit",
|
|
8
|
+
"Write",
|
|
9
|
+
"Task",
|
|
10
|
+
"Bash(find *)",
|
|
11
|
+
"Bash(grep *)",
|
|
12
|
+
"Bash(ls *)",
|
|
13
|
+
"Bash(git diff *)",
|
|
14
|
+
"Bash(git log *)",
|
|
15
|
+
"Bash(git show *)",
|
|
16
|
+
"Bash(git status)",
|
|
17
|
+
"Bash(git rev-parse *)",
|
|
18
|
+
"Bash(git blame *)",
|
|
19
|
+
"Bash(git grep *)",
|
|
20
|
+
"Bash(diff *)",
|
|
21
|
+
"Bash(wc *)",
|
|
22
|
+
"Bash(head *)",
|
|
23
|
+
"Bash(tail *)",
|
|
24
|
+
"Bash(sort *)",
|
|
25
|
+
"Bash(uniq *)",
|
|
26
|
+
"Bash(jq *)",
|
|
27
|
+
"Bash(cat *)"
|
|
28
|
+
],
|
|
29
|
+
"ask": ["WebFetch", "WebSearch"],
|
|
30
|
+
"deny": [
|
|
31
|
+
"Edit(debug-data/**)",
|
|
32
|
+
"Write(debug-data/**)",
|
|
33
|
+
"mcp__*",
|
|
34
|
+
"ToolSearch",
|
|
35
|
+
"ListMcpResourcesTool",
|
|
36
|
+
"ReadMcpResourceTool"
|
|
37
|
+
],
|
|
38
|
+
"defaultMode": "default"
|
|
39
|
+
},
|
|
40
|
+
"spinnerVerbs": {
|
|
41
|
+
"mode": "replace",
|
|
42
|
+
"verbs": [
|
|
43
|
+
"Analyzing",
|
|
44
|
+
"Investigating",
|
|
45
|
+
"Comparing",
|
|
46
|
+
"Searching",
|
|
47
|
+
"Processing",
|
|
48
|
+
"Squinting at screenshots",
|
|
49
|
+
"Traversing tangled timelines",
|
|
50
|
+
"Fighting fickle flakes",
|
|
51
|
+
"Diagnosing dubious diffs",
|
|
52
|
+
"Reconciling replays",
|
|
53
|
+
"Deciphering devious divergences",
|
|
54
|
+
"Debugging flakes",
|
|
55
|
+
"Patching network sessions",
|
|
56
|
+
"Determining nondeterminism"
|
|
57
|
+
]
|
|
58
|
+
},
|
|
59
|
+
"hooks": {
|
|
60
|
+
"SessionStart": [
|
|
61
|
+
{
|
|
62
|
+
"hooks": [
|
|
63
|
+
{
|
|
64
|
+
"type": "command",
|
|
65
|
+
"command": ".claude/hooks/load-context.sh"
|
|
66
|
+
}
|
|
67
|
+
]
|
|
68
|
+
}
|
|
69
|
+
],
|
|
70
|
+
"PreToolUse": [
|
|
71
|
+
{
|
|
72
|
+
"matcher": "Read",
|
|
73
|
+
"hooks": [
|
|
74
|
+
{
|
|
75
|
+
"type": "command",
|
|
76
|
+
"command": ".claude/hooks/check-file-size.sh"
|
|
77
|
+
}
|
|
78
|
+
]
|
|
79
|
+
}
|
|
80
|
+
]
|
|
81
|
+
}
|
|
82
|
+
}
|
|
@@ -0,0 +1,57 @@
|
|
|
1
|
+
---
|
|
2
|
+
name: debugging-diffs
|
|
3
|
+
description: Investigate unexpected visual differences between head and base replays. Use when screenshot diffs are flagged or visual regressions are reported.
|
|
4
|
+
---
|
|
5
|
+
|
|
6
|
+
# Debugging Screenshot Diffs
|
|
7
|
+
|
|
8
|
+
Use this guide when investigating unexpected visual differences between head and base replays.
|
|
9
|
+
|
|
10
|
+
## Investigation Steps
|
|
11
|
+
|
|
12
|
+
### 1. Understand the Diffs
|
|
13
|
+
|
|
14
|
+
- Read the replay diff JSON in `debug-data/diffs/<id>.json`.
|
|
15
|
+
- Check `screenshotDiffResults` for which screenshots differ.
|
|
16
|
+
- Note the diff percentage and pixel count for each screenshot.
|
|
17
|
+
|
|
18
|
+
### 2. Correlate with Code Changes
|
|
19
|
+
|
|
20
|
+
- Check `commitSha` in `context.json` to identify the code changes.
|
|
21
|
+
- If `project-repo/` is available, use `git log` and `git diff` to see what changed.
|
|
22
|
+
- Focus on CSS changes, component rendering logic, and layout modifications.
|
|
23
|
+
|
|
24
|
+
### 3. Compare Logs at Screenshot Time
|
|
25
|
+
|
|
26
|
+
- Find the screenshot timestamps in `timeline.json`.
|
|
27
|
+
- Compare what events occurred before each screenshot in head vs base.
|
|
28
|
+
- Look for missing or extra events that could cause visual differences.
|
|
29
|
+
|
|
30
|
+
### 4. Check for Expected vs Unexpected Diffs
|
|
31
|
+
|
|
32
|
+
- **Expected**: Code changes that intentionally modify the UI (new features, style updates).
|
|
33
|
+
- **Unexpected**: Same code producing different visual output, or unrelated areas changing.
|
|
34
|
+
- Check if the diff is in a dynamic content area (timestamps, counters, user-specific data).
|
|
35
|
+
|
|
36
|
+
### 5. Examine Snapshotted Assets
|
|
37
|
+
|
|
38
|
+
- If `debug-data/replays/{head,base}/<replayId>/snapshotted-assets/` exists, compare JS/CSS between head and base.
|
|
39
|
+
- Look for changes in CSS that could cause layout shifts.
|
|
40
|
+
- Check for new or modified JavaScript that affects rendering.
|
|
41
|
+
|
|
42
|
+
### 6. Review Screenshot Assertions Config
|
|
43
|
+
|
|
44
|
+
- Check `screenshotAssertionsOptions` in the diff JSON for threshold settings.
|
|
45
|
+
- Some diffs may be within acceptable tolerance but still flagged.
|
|
46
|
+
|
|
47
|
+
### 7. Known Meticulous Replay Behaviors (Safe to Approve)
|
|
48
|
+
|
|
49
|
+
The following patterns are caused by Meticulous replay engine behavior, not by your code changes. When these are the root cause of a diff, the diff is safe to approve.
|
|
50
|
+
|
|
51
|
+
**Animation flakes**: Continuous animations (Lottie, CSS `@keyframes`, `requestAnimationFrame` loops) can cause screenshots to land at slightly different animation states between base and head, even with no code changes. Meticulous's DOM stability detection triggers extra animation frames trying to reach a stable state, but continuous animations never fully stabilize. **How to confirm**: check `replayComparison` in `context.json` for differing `totalAnimationFrames`; check log diffs for different animation frame counts; search the codebase for `lottie`, `requestAnimationFrame`, CSS `animation`, `@keyframes`, `<canvas>`, `<video>`. If the only visual difference is an animation in a different frame, this diff is safe to approve.
|
|
52
|
+
|
|
53
|
+
**Network timing differences**: When multiple network responses complete at the same virtual time, slight ordering differences can cause minor rendering variations (e.g. a list rendering in a different order before settling). **How to confirm**: log diffs show network requests completing in a different order at the same virtual time, but the final rendered state is the same or nearly identical. If no code change caused the reordering, this diff is safe to approve.
|
|
54
|
+
|
|
55
|
+
**Extra DOM stability frames**: The head replay may show more timeline events before a screenshot than the base, caused by DOM stability detection running additional animation frames. If the extra events are all animation-frame-related and the visual diff is minor, this diff is safe to approve.
|
|
56
|
+
|
|
57
|
+
When any of these patterns is the root cause, tell the user: "This diff is caused by a known Meticulous replay behavior, not by your code changes. It is safe to approve."
|
|
@@ -0,0 +1,52 @@
|
|
|
1
|
+
---
|
|
2
|
+
name: debugging-flakes
|
|
3
|
+
description: Investigate flaky or non-deterministic replay behavior. Use when replays produce different results across runs of the same session against the same code.
|
|
4
|
+
---
|
|
5
|
+
|
|
6
|
+
# Debugging Flaky Replays
|
|
7
|
+
|
|
8
|
+
A flaky replay is one that produces different results (screenshots, logs, behavior) across
|
|
9
|
+
multiple runs of the same session against the same code. Use this guide when investigating
|
|
10
|
+
non-deterministic behavior.
|
|
11
|
+
|
|
12
|
+
## Investigation Steps
|
|
13
|
+
|
|
14
|
+
### 1. Compare Deterministic Logs
|
|
15
|
+
|
|
16
|
+
- Diff `logs.deterministic.txt` between the head and base replays.
|
|
17
|
+
- Look for the first point of divergence -- this is usually where the flake originates.
|
|
18
|
+
- Pay attention to event execution order differences.
|
|
19
|
+
|
|
20
|
+
### 2. Check Timeline for Timing Issues
|
|
21
|
+
|
|
22
|
+
- Read `timeline.json` and look for events with significantly different virtual timestamps.
|
|
23
|
+
- Look for race conditions: events that depend on network responses or animations completing.
|
|
24
|
+
- Check for `setTimeout`/`setInterval` patterns that may resolve differently.
|
|
25
|
+
|
|
26
|
+
### 3. Analyze Accuracy Data
|
|
27
|
+
|
|
28
|
+
- Read `accuracyData.json` for the replay's self-assessment of accuracy.
|
|
29
|
+
- Low accuracy scores often correlate with flakiness.
|
|
30
|
+
- Check which specific checks failed.
|
|
31
|
+
|
|
32
|
+
### 4. Look for Common Flake Patterns
|
|
33
|
+
|
|
34
|
+
- **Animation timing**: CSS transitions or JS animations that haven't completed when screenshot is taken.
|
|
35
|
+
- **Network race conditions**: API responses arriving in different order.
|
|
36
|
+
- **Date/time dependencies**: Code that uses `Date.now()` or renders timestamps.
|
|
37
|
+
- **Random values**: Code using `Math.random()` that wasn't properly seeded.
|
|
38
|
+
- **Lazy loading**: Content loaded on intersection observer timing.
|
|
39
|
+
- **Font loading**: Web fonts loading at different times causing layout shifts.
|
|
40
|
+
|
|
41
|
+
When any of these patterns is identified as the root cause and no user code change is involved, the resulting diff is caused by non-determinism in the replay environment, not the user's code. Advise the user that these diffs are **safe to approve**.
|
|
42
|
+
|
|
43
|
+
### 5. Check Replay Parameters
|
|
44
|
+
|
|
45
|
+
- Compare `launchBrowserAndReplayParams.json` between runs.
|
|
46
|
+
- Verify viewport size, user agent, and other environment settings match.
|
|
47
|
+
- Check if network stubbing configuration is consistent.
|
|
48
|
+
|
|
49
|
+
### 6. Examine Stack Traces
|
|
50
|
+
|
|
51
|
+
- Read `stackTraces.json` for any errors thrown during replay.
|
|
52
|
+
- Errors that occur in some runs but not others are strong flake indicators.
|
|
@@ -0,0 +1,45 @@
|
|
|
1
|
+
---
|
|
2
|
+
name: debugging-network
|
|
3
|
+
description: Investigate network-related replay failures and divergences. Use when replays fail due to network errors, stubbing issues, or request ordering problems.
|
|
4
|
+
---
|
|
5
|
+
|
|
6
|
+
# Debugging Network Issues
|
|
7
|
+
|
|
8
|
+
Use this guide when replays fail or diverge due to network request problems.
|
|
9
|
+
|
|
10
|
+
## Investigation Steps
|
|
11
|
+
|
|
12
|
+
### 1. Check Logs for Network Errors
|
|
13
|
+
|
|
14
|
+
- Search `logs.concise.txt` for "network", "fetch", "xhr", "request", "response", "timeout".
|
|
15
|
+
- Look for failed requests, unexpected status codes, or missing responses.
|
|
16
|
+
- Check for CORS errors or SSL issues.
|
|
17
|
+
|
|
18
|
+
### 2. Compare Network Activity in Timeline
|
|
19
|
+
|
|
20
|
+
- In `timeline.json`, look for network-related events.
|
|
21
|
+
- Compare the sequence and timing of network requests between head and base.
|
|
22
|
+
- Look for requests in one replay that are missing in the other.
|
|
23
|
+
|
|
24
|
+
### 3. Examine Session Data
|
|
25
|
+
|
|
26
|
+
- Read session data in `debug-data/sessions/<id>/data.json`.
|
|
27
|
+
- Check `recordedRequests` for the original HAR entries captured during recording.
|
|
28
|
+
- Compare recorded requests with what was replayed.
|
|
29
|
+
|
|
30
|
+
### 4. Look for Stubbing Issues
|
|
31
|
+
|
|
32
|
+
- Network requests are stubbed during replay using recorded data.
|
|
33
|
+
- Check if new API endpoints were added that don't have recorded responses.
|
|
34
|
+
- Look for requests with dynamic parameters (timestamps, tokens) that may not match stubs.
|
|
35
|
+
|
|
36
|
+
### 5. Check for Request Ordering Dependencies
|
|
37
|
+
|
|
38
|
+
- Some applications depend on requests completing in a specific order.
|
|
39
|
+
- Look for race conditions where parallel requests resolve differently.
|
|
40
|
+
- Check for waterfall dependencies (request B depends on response from request A).
|
|
41
|
+
|
|
42
|
+
### 6. Verify API Compatibility
|
|
43
|
+
|
|
44
|
+
- If the API was changed, recorded responses may no longer be valid.
|
|
45
|
+
- Check for schema changes, new required fields, or renamed endpoints.
|
|
@@ -0,0 +1,47 @@
|
|
|
1
|
+
---
|
|
2
|
+
name: debugging-sessions
|
|
3
|
+
description: Investigate problems with recorded session data. Use when session recordings appear incomplete, corrupted, or contain unexpected data.
|
|
4
|
+
---
|
|
5
|
+
|
|
6
|
+
# Debugging Session Data Issues
|
|
7
|
+
|
|
8
|
+
Use this guide when investigating problems with the recorded session data itself.
|
|
9
|
+
|
|
10
|
+
## Investigation Steps
|
|
11
|
+
|
|
12
|
+
### 1. Examine Session Structure
|
|
13
|
+
|
|
14
|
+
- Read `debug-data/sessions/<id>/data.json` (this can be very large, use grep/search).
|
|
15
|
+
- Key fields: `rrwebEvents`, `userEvents`, `recordedRequests`, `applicationStorage`, `webSockets`.
|
|
16
|
+
|
|
17
|
+
### 2. Check User Events
|
|
18
|
+
|
|
19
|
+
- `userEvents` contains the sequence of user interactions that will be replayed.
|
|
20
|
+
- Verify events are in chronological order.
|
|
21
|
+
- Check for truncated or incomplete event sequences.
|
|
22
|
+
- Look for unusually rapid event sequences that may indicate automated behavior.
|
|
23
|
+
|
|
24
|
+
### 3. Verify Network Recordings
|
|
25
|
+
|
|
26
|
+
- `recordedRequests` contains HAR-format entries of network activity.
|
|
27
|
+
- Check for missing responses (the request was recorded but the response wasn't).
|
|
28
|
+
- Look for very large responses that might have been truncated.
|
|
29
|
+
- Verify content types and encoding are preserved correctly.
|
|
30
|
+
|
|
31
|
+
### 4. Check Application Storage
|
|
32
|
+
|
|
33
|
+
- `applicationStorage` captures localStorage, sessionStorage, and cookies.
|
|
34
|
+
- Verify that authentication state is properly captured.
|
|
35
|
+
- Look for expired tokens or sessions that may cause different behavior during replay.
|
|
36
|
+
|
|
37
|
+
### 5. Look for Session Quality Issues
|
|
38
|
+
|
|
39
|
+
- Very short sessions (few events) may not provide meaningful coverage.
|
|
40
|
+
- Sessions with `abandoned: true` were not completed normally.
|
|
41
|
+
- Check `numberUserEvents` and `numberBytes` for unusually small or large values.
|
|
42
|
+
|
|
43
|
+
### 6. Verify Recording Environment
|
|
44
|
+
|
|
45
|
+
- Check session metadata for the recording environment (hostname, URL).
|
|
46
|
+
- Ensure the session was recorded against a compatible version of the application.
|
|
47
|
+
- Look for environment-specific behavior (staging vs production data).
|
|
@@ -0,0 +1,51 @@
|
|
|
1
|
+
---
|
|
2
|
+
name: debugging-timelines
|
|
3
|
+
description: Investigate timeline divergence between head and base replays. Use when event sequences, ordering, or timing differ unexpectedly.
|
|
4
|
+
---
|
|
5
|
+
|
|
6
|
+
# Debugging Timeline Divergence
|
|
7
|
+
|
|
8
|
+
Use this guide when replay timelines differ unexpectedly between head and base runs.
|
|
9
|
+
|
|
10
|
+
## Investigation Steps
|
|
11
|
+
|
|
12
|
+
### 1. Load and Compare Timelines
|
|
13
|
+
|
|
14
|
+
- Read `timeline.json` from both head and base replay directories.
|
|
15
|
+
- The timeline is an array of events with timestamps, types, and data.
|
|
16
|
+
- Look for the first event where the timelines diverge.
|
|
17
|
+
|
|
18
|
+
### 2. Understand Event Types
|
|
19
|
+
|
|
20
|
+
Key timeline event types:
|
|
21
|
+
|
|
22
|
+
- **user-event**: User interactions (click, type, scroll, hover).
|
|
23
|
+
- **network-request**: API calls and responses.
|
|
24
|
+
- **screenshot**: Screenshot capture points.
|
|
25
|
+
- **mutation**: DOM mutations observed during replay.
|
|
26
|
+
- **navigation**: Page navigation events.
|
|
27
|
+
- **error**: JavaScript errors.
|
|
28
|
+
- **console**: Console log messages.
|
|
29
|
+
|
|
30
|
+
### 3. Identify Divergence Patterns
|
|
31
|
+
|
|
32
|
+
- **Missing events**: Events in base that don't appear in head (or vice versa).
|
|
33
|
+
- **Reordered events**: Same events but in different sequence.
|
|
34
|
+
- **Timing shifts**: Events at significantly different virtual timestamps.
|
|
35
|
+
- **Extra events**: New events not present in the baseline.
|
|
36
|
+
|
|
37
|
+
### 4. Check Timeline Stats
|
|
38
|
+
|
|
39
|
+
- Read `timeline-stats.json` for aggregated statistics.
|
|
40
|
+
- Compare event counts, durations, and error counts between replays.
|
|
41
|
+
|
|
42
|
+
### 5. Trace Back to Root Cause
|
|
43
|
+
|
|
44
|
+
- Once you find the divergence point, look at what happened immediately before.
|
|
45
|
+
- Check if a user event triggered different behavior.
|
|
46
|
+
- Look for conditional logic in the application that might execute differently.
|
|
47
|
+
|
|
48
|
+
### 6. Cross-Reference with Logs
|
|
49
|
+
|
|
50
|
+
- Use timestamps from the timeline divergence to find corresponding log entries.
|
|
51
|
+
- Check `logs.deterministic.txt` at the same virtual time for additional context.
|
|
@@ -0,0 +1,20 @@
|
|
|
1
|
+
---
|
|
2
|
+
name: pr-analysis
|
|
3
|
+
description: Analyze PR source code changes and correlate with screenshot diffs. Use when pr-diff.txt is present and you need to understand which code changes caused visual differences.
|
|
4
|
+
---
|
|
5
|
+
|
|
6
|
+
# PR Analysis
|
|
7
|
+
|
|
8
|
+
When `debug-data/pr-diff.txt` is present in the workspace, analyze the source code changes and correlate
|
|
9
|
+
them with the screenshot diffs.
|
|
10
|
+
|
|
11
|
+
1. Read `debug-data/pr-diff.txt` to understand what code changed
|
|
12
|
+
2. Read the diff summaries in `debug-data/diffs/*.summary.json` to see which screenshots differ
|
|
13
|
+
3. For each screenshot that differs, identify which code changes are most likely responsible
|
|
14
|
+
|
|
15
|
+
Provide a structured analysis:
|
|
16
|
+
|
|
17
|
+
- Which files were modified and what the key changes are
|
|
18
|
+
- Which code changes are most likely to affect visual output (CSS, layout, component rendering)
|
|
19
|
+
- For each differing screenshot, the most likely code change that caused it
|
|
20
|
+
- Whether the visual changes appear intentional (matching the code intent) or unintentional
|
package/package.json
ADDED
|
@@ -0,0 +1,49 @@
|
|
|
1
|
+
{
|
|
2
|
+
"name": "@alwaysmeticulous/debug-workspace",
|
|
3
|
+
"version": "2.261.0",
|
|
4
|
+
"description": "Shared debug workspace pipeline for investigating Meticulous diffs and replays",
|
|
5
|
+
"license": "ISC",
|
|
6
|
+
"main": "dist/index.js",
|
|
7
|
+
"types": "dist/index.d.ts",
|
|
8
|
+
"files": [
|
|
9
|
+
"dist"
|
|
10
|
+
],
|
|
11
|
+
"dependencies": {
|
|
12
|
+
"chalk": "^4.1.2",
|
|
13
|
+
"@alwaysmeticulous/client": "2.261.0",
|
|
14
|
+
"@alwaysmeticulous/common": "2.260.2",
|
|
15
|
+
"@alwaysmeticulous/downloading-helpers": "2.261.0"
|
|
16
|
+
},
|
|
17
|
+
"devDependencies": {
|
|
18
|
+
"vitest": "^4.0.17"
|
|
19
|
+
},
|
|
20
|
+
"author": {
|
|
21
|
+
"name": "The Meticulous Team",
|
|
22
|
+
"email": "eng@meticulous.ai",
|
|
23
|
+
"url": "https://meticulous.ai"
|
|
24
|
+
},
|
|
25
|
+
"engines": {
|
|
26
|
+
"node": ">= 12"
|
|
27
|
+
},
|
|
28
|
+
"homepage": "https://github.com/alwaysmeticulous/meticulous-sdk",
|
|
29
|
+
"repository": {
|
|
30
|
+
"type": "git",
|
|
31
|
+
"url": "https://github.com/alwaysmeticulous/meticulous-sdk.git",
|
|
32
|
+
"directory": "packages/debug-workspace"
|
|
33
|
+
},
|
|
34
|
+
"bugs": {
|
|
35
|
+
"url": "https://github.com/alwaysmeticulous/meticulous-sdk/issues"
|
|
36
|
+
},
|
|
37
|
+
"scripts": {
|
|
38
|
+
"clean": "rimraf dist tsconfig.tsbuildinfo",
|
|
39
|
+
"build": "tsc --build tsconfig.json",
|
|
40
|
+
"postbuild": "cp -r src/templates dist/templates",
|
|
41
|
+
"dev": "tsc --build tsconfig.json --watch",
|
|
42
|
+
"format": "prettier --write src",
|
|
43
|
+
"lint": "eslint \"src/**/*.{js,ts,tsx}\" --cache",
|
|
44
|
+
"lint:commit": "eslint --cache $(git diff --relative --name-only --diff-filter=ACMRTUXB master | grep -E \"(.js$|.ts$|.tsx$)\")",
|
|
45
|
+
"lint:fix": "eslint \"src/**/*.{js,ts,tsx}\" --cache --fix",
|
|
46
|
+
"depcheck": "depcheck --ignore-patterns=dist",
|
|
47
|
+
"test": "vitest run --passWithNoTests"
|
|
48
|
+
}
|
|
49
|
+
}
|