superlab 0.1.0 → 0.1.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +23 -23
- package/README.zh-CN.md +22 -21
- package/lib/i18n.cjs +626 -23
- package/lib/install.cjs +31 -0
- package/package-assets/claude/commands/lab/spec.md +1 -1
- package/package-assets/claude/commands/lab/write.md +1 -1
- package/package-assets/codex/prompts/lab-spec.md +1 -1
- package/package-assets/codex/prompts/lab-write.md +1 -1
- package/package-assets/shared/changes/README.md +10 -0
- package/package-assets/shared/config/workflow.json +5 -0
- package/package-assets/shared/context/decisions.md +11 -0
- package/package-assets/shared/context/evidence-index.md +16 -0
- package/package-assets/shared/context/mission.md +27 -0
- package/package-assets/shared/context/open-questions.md +11 -0
- package/package-assets/shared/context/state.md +19 -0
- package/package-assets/shared/examples/minimal-uplift-workflow.md +4 -4
- package/package-assets/shared/skills/lab/SKILL.md +54 -9
- package/package-assets/shared/skills/lab/references/brainstorming-integration.md +21 -0
- package/package-assets/shared/skills/lab/references/paper-writing/abstract.md +102 -0
- package/package-assets/shared/skills/lab/references/paper-writing/conclusion.md +35 -0
- package/package-assets/shared/skills/lab/references/paper-writing/does-my-writing-flow-source.md +45 -0
- package/package-assets/shared/skills/lab/references/paper-writing/experiments.md +102 -0
- package/package-assets/shared/skills/lab/references/paper-writing/introduction.md +408 -0
- package/package-assets/shared/skills/lab/references/paper-writing/method.md +196 -0
- package/package-assets/shared/skills/lab/references/paper-writing/paper-review.md +86 -0
- package/package-assets/shared/skills/lab/references/paper-writing/related-work.md +41 -0
- package/package-assets/shared/skills/lab/references/paper-writing-integration.md +29 -28
- package/package-assets/shared/skills/lab/references/workflow.md +1 -1
- package/package-assets/shared/skills/lab/stages/idea.md +43 -7
- package/package-assets/shared/skills/lab/stages/iterate.md +32 -0
- package/package-assets/shared/skills/lab/stages/report.md +19 -0
- package/package-assets/shared/skills/lab/stages/review.md +30 -0
- package/package-assets/shared/skills/lab/stages/run.md +17 -0
- package/package-assets/shared/skills/lab/stages/spec.md +36 -4
- package/package-assets/shared/skills/lab/stages/write.md +47 -15
- package/package-assets/shared/templates/design.md +10 -0
- package/package-assets/shared/templates/idea.md +76 -8
- package/package-assets/shared/templates/iteration-report.md +4 -0
- package/package-assets/shared/templates/paper-plan.md +12 -0
- package/package-assets/shared/templates/paper-section.md +24 -6
- package/package-assets/shared/templates/paper-section.tex +10 -0
- package/package-assets/shared/templates/paper.tex +29 -0
- package/package-assets/shared/templates/proposal.md +10 -0
- package/package-assets/shared/templates/review-checklist.md +23 -0
- package/package-assets/shared/templates/spec.md +7 -2
- package/package-assets/shared/templates/tasks.md +3 -1
- package/package-assets/shared/templates/write-iteration.md +5 -0
- package/package.json +3 -3
- package/package-assets/shared/scripts/check_openspec.sh +0 -10
|
@@ -0,0 +1,86 @@
|
|
|
1
|
+
# Paper Review (Paper Rview)
|
|
2
|
+
|
|
3
|
+
## Goal
|
|
4
|
+
|
|
5
|
+
Use an adversarial, reviewer-style checklist to detect reject risks early and revise the paper before submission.
|
|
6
|
+
|
|
7
|
+
## Core Principle
|
|
8
|
+
|
|
9
|
+
Pursue perfectionism in paper quality: assume reviewers will probe every weak point and proactively fix them.
|
|
10
|
+
|
|
11
|
+
## Critical Rule (Do Not Violate)
|
|
12
|
+
|
|
13
|
+
Every major claim, especially in Abstract and Introduction, must be:
|
|
14
|
+
|
|
15
|
+
1. technically correct, and
|
|
16
|
+
2. explicitly supported by experimental evidence.
|
|
17
|
+
|
|
18
|
+
If a claim is not supported, either add evidence or weaken/remove the claim.
|
|
19
|
+
|
|
20
|
+
## What Usually Gets a Paper Accepted
|
|
21
|
+
|
|
22
|
+
1. Sufficient contribution (for example: novel task, novel pipeline, novel module, novel design choices, new experimental findings, or new insight).
|
|
23
|
+
2. Better empirical performance than prior methods under fair comparisons.
|
|
24
|
+
3. Sufficient comparison experiments and ablation studies.
|
|
25
|
+
|
|
26
|
+
## Common Rejection Dimensions
|
|
27
|
+
|
|
28
|
+
| Rejection Dimension | Typical Failure Signals |
|
|
29
|
+
| ---------------------------- | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ |
|
|
30
|
+
| 1. Insufficient contribution | 1.1 Targeted failure cases are too common.<br /> 1.2 Proposed technique is already well explored; expected gains are predictable/well-known. |
|
|
31
|
+
| 2. Unclear writing | 2.1 Missing technical details; work is not reproducible.<br />2.2 A method module lacks clear motivation. |
|
|
32
|
+
| 3. Weak empirical effect | 3.1 Improvement over prior methods is only marginal.<br /> 3.2 Even if better than previous methods, absolute performance is still not strong enough. |
|
|
33
|
+
| 4. Incomplete evaluation | 4.1 Missing ablation studies.<br />4.2 Missing important baselines or important evaluation metrics.<br /> 4.3 Datasets are too simple to prove the method truly works. |
|
|
34
|
+
| 5. Problematic method design | 5.1 Experimental setting is unrealistic.<br />5.2 Method has technical flaws and appears unreasonable.<br />5.3 Method is not robust and needs per-scenario hyperparameter tuning. <br /> 5.4 New design introduces stronger limitations than its benefits, leading to negative net value. |
|
|
35
|
+
|
|
36
|
+
## End-of-Paper Self-Review Question List
|
|
37
|
+
|
|
38
|
+
Add this checklist near the end of the draft while revising.
|
|
39
|
+
Use each question to trigger concrete edits before submission.
|
|
40
|
+
|
|
41
|
+
### 1. Contribution
|
|
42
|
+
|
|
43
|
+
1. What new knowledge does this paper give to readers?
|
|
44
|
+
2. Are we solving a truly meaningful failure case, not a trivial/common one?
|
|
45
|
+
3. Is the technical idea genuinely non-obvious beyond well-explored practice?
|
|
46
|
+
4. Is our gain surprising or insightful rather than a predictable improvement?
|
|
47
|
+
5. Is there at least one clear novelty type (task/pipeline/module/design finding/insight)?
|
|
48
|
+
|
|
49
|
+
### 2. Writing Clarity
|
|
50
|
+
|
|
51
|
+
1. Can a knowledgeable reader reproduce the method from the paper?
|
|
52
|
+
2. Did we provide enough technical detail for each key module?
|
|
53
|
+
3. Is the motivation of every module explicit and logically connected to a challenge?
|
|
54
|
+
4. Are terms and notation consistent across sections?
|
|
55
|
+
5. Does each paragraph carry one clear message with smooth transitions?
|
|
56
|
+
|
|
57
|
+
### 3. Experimental Strength
|
|
58
|
+
|
|
59
|
+
1. Are improvements over strong baselines meaningful, not just statistically tiny?
|
|
60
|
+
2. Is absolute performance competitive enough for the target venue?
|
|
61
|
+
3. Are gains consistent across multiple datasets/settings/metrics?
|
|
62
|
+
4. Do we report both strengths and failure cases honestly?
|
|
63
|
+
|
|
64
|
+
### 4. Evaluation Completeness
|
|
65
|
+
|
|
66
|
+
1. Do we include ablations for all key design choices?
|
|
67
|
+
2. Are all strong/recent baselines included under fair settings?
|
|
68
|
+
3. Are evaluation metrics standard and sufficient for this task?
|
|
69
|
+
4. Are datasets/scenarios challenging enough to validate real effectiveness?
|
|
70
|
+
5. Are comparison and ablation protocols clearly documented?
|
|
71
|
+
|
|
72
|
+
### 5. Method Design Soundness
|
|
73
|
+
|
|
74
|
+
1. Is the experimental setting realistic for practical use?
|
|
75
|
+
2. Does the method have hidden technical defects or unreasonable assumptions?
|
|
76
|
+
3. Is the method robust without heavy per-case hyperparameter retuning?
|
|
77
|
+
4. Do benefits outweigh added complexity and new limitations?
|
|
78
|
+
5. Could reviewers reasonably argue that the net benefit is negative?
|
|
79
|
+
|
|
80
|
+
## Adversarial Writing Workflow
|
|
81
|
+
|
|
82
|
+
1. Read the paper as a skeptical reviewer.
|
|
83
|
+
2. Answer every question above with explicit evidence from the paper.
|
|
84
|
+
3. Mark each item as `pass`, `needs revision`, or `needs new experiment`.
|
|
85
|
+
4. Revise claims, writing, experiments, or method scope accordingly.
|
|
86
|
+
5. Repeat until no major rejection risk remains.
|
|
@@ -0,0 +1,41 @@
|
|
|
1
|
+
# Related Work Writing Guide
|
|
2
|
+
|
|
3
|
+
## Goal
|
|
4
|
+
|
|
5
|
+
Position your work against the most relevant lines of research, and make your novelty easy to verify.
|
|
6
|
+
|
|
7
|
+
## Workflow
|
|
8
|
+
|
|
9
|
+
1. List directly competing and recent baseline papers first.
|
|
10
|
+
2. Group literature by technical topic (not by publication year alone).
|
|
11
|
+
3. For each topic: summarize common paradigm, then key limitation relevant to your challenge.
|
|
12
|
+
4. End each topic by clarifying your distinction.
|
|
13
|
+
|
|
14
|
+
## Topic Design
|
|
15
|
+
|
|
16
|
+
Use 2-4 focused topics, for example:
|
|
17
|
+
|
|
18
|
+
1. Task-specific mainstream methods.
|
|
19
|
+
2. Methods closest to your core idea.
|
|
20
|
+
3. Auxiliary techniques your method builds on.
|
|
21
|
+
|
|
22
|
+
## Paragraph Template
|
|
23
|
+
|
|
24
|
+
1. Topic sentence: define scope of this topic.
|
|
25
|
+
2. Representative methods: one compact summary.
|
|
26
|
+
3. Limitation tied to your target technical challenge.
|
|
27
|
+
4. Transition sentence that leads to your method.
|
|
28
|
+
|
|
29
|
+
## Do and Don't
|
|
30
|
+
|
|
31
|
+
1. Do compare mechanisms, assumptions, and failure modes.
|
|
32
|
+
2. Do emphasize the exact gap your method fills.
|
|
33
|
+
3. Do not make Related Work a citation dump.
|
|
34
|
+
4. Do not hide strongest baselines.
|
|
35
|
+
|
|
36
|
+
## Checklist
|
|
37
|
+
|
|
38
|
+
1. Are all strongest/recent competitors covered?
|
|
39
|
+
2. Is each topic connected to your problem setting?
|
|
40
|
+
3. Is your difference explained in technical terms, not marketing terms?
|
|
41
|
+
4. Is citation coverage complete for all core claims?
|
|
@@ -1,39 +1,40 @@
|
|
|
1
1
|
# Paper-Writing Integration
|
|
2
2
|
|
|
3
|
-
`/lab:write`
|
|
3
|
+
`/lab:write` vendors the paper-writing references directly into `skills/lab/references/paper-writing/`.
|
|
4
|
+
The goal is to keep the upstream writing discipline while removing brittle runtime dependence on a separately installed skill.
|
|
4
5
|
|
|
5
|
-
|
|
6
|
+
## Role Split
|
|
6
7
|
|
|
7
|
-
- `
|
|
8
|
-
|
|
9
|
-
|
|
8
|
+
- `lab` controls stage boundaries, evidence discipline, and durable artifacts.
|
|
9
|
+
- the vendored paper-writing references control section structure, paragraph logic, and reviewer-facing polish.
|
|
10
|
+
- `/lab:write` links evidence-backed research outputs to paper-ready text.
|
|
10
11
|
|
|
11
|
-
|
|
12
|
-
- paragraph flow
|
|
13
|
-
- claim-evidence alignment
|
|
14
|
-
- conservative academic tone
|
|
15
|
-
- reviewer-style self-checks
|
|
12
|
+
## Required Shared Constraints
|
|
16
13
|
|
|
17
|
-
|
|
14
|
+
- one section or one bounded subsection per round
|
|
15
|
+
- load only the active section guide, not every section guide
|
|
16
|
+
- mini-outline before prose
|
|
17
|
+
- explicit evidence links for major claims
|
|
18
|
+
- motivation, design, and technical advantage when applicable
|
|
19
|
+
- stable terminology across the whole paper
|
|
20
|
+
- unsupported claims must be weakened or removed
|
|
21
|
+
- section-level flow check and reviewer pass after each round
|
|
22
|
+
- five-dimension self-review before a round is accepted
|
|
23
|
+
- route back to `review` or `iterate` if the evidence is weak
|
|
18
24
|
|
|
19
|
-
|
|
20
|
-
- hide failed runs or limitations
|
|
21
|
-
- blur sourced evidence with generated interpretation
|
|
22
|
-
- rewrite the whole paper in one unconstrained pass
|
|
25
|
+
## Vendored References
|
|
23
26
|
|
|
24
|
-
|
|
27
|
+
- `paper-writing/abstract.md`
|
|
28
|
+
- `paper-writing/introduction.md`
|
|
29
|
+
- `paper-writing/related-work.md`
|
|
30
|
+
- `paper-writing/method.md`
|
|
31
|
+
- `paper-writing/experiments.md`
|
|
32
|
+
- `paper-writing/conclusion.md`
|
|
33
|
+
- `paper-writing/paper-review.md`
|
|
34
|
+
- `paper-writing/does-my-writing-flow-source.md`
|
|
25
35
|
|
|
26
|
-
|
|
27
|
-
- `research-paper-writing` references control how each section is written and revised
|
|
28
|
-
- evidence still comes from `report`, iteration reports, and normalized summaries
|
|
36
|
+
## Attribution
|
|
29
37
|
|
|
30
|
-
|
|
38
|
+
These references are adapted from:
|
|
31
39
|
|
|
32
|
-
- `
|
|
33
|
-
- `introduction.md`
|
|
34
|
-
- `related-work.md`
|
|
35
|
-
- `method.md`
|
|
36
|
-
- `experiments.md`
|
|
37
|
-
- `conclusion.md`
|
|
38
|
-
- `paper-review.md`
|
|
39
|
-
- `does-my-writing-flow-source.md`
|
|
40
|
+
- `https://github.com/Master-cai/Research-Paper-Writing-Skills/tree/main/research-paper-writing/references`
|
|
@@ -25,7 +25,7 @@ Escalate to a higher-level redesign instead of mutating the mission when the cur
|
|
|
25
25
|
## Required Artifacts
|
|
26
26
|
|
|
27
27
|
- `docs/lab/idea.md`
|
|
28
|
-
-
|
|
28
|
+
- one lab change directory under `.superlab/changes/<change-id>/`
|
|
29
29
|
- normalized JSON summary from `scripts/eval_report.py`
|
|
30
30
|
- per-round report in `docs/lab/iterations/`
|
|
31
31
|
- final report in `docs/lab/report.md`
|
|
@@ -2,25 +2,61 @@
|
|
|
2
2
|
|
|
3
3
|
## Required Output
|
|
4
4
|
|
|
5
|
+
- one-sentence problem statement
|
|
6
|
+
- failure case
|
|
7
|
+
- idea classification
|
|
8
|
+
- contribution category
|
|
9
|
+
- breakthrough level
|
|
10
|
+
- existing methods summary
|
|
11
|
+
- why the proposed idea is better than existing methods
|
|
12
|
+
- three meaningful points
|
|
5
13
|
- literature-backed framing
|
|
6
14
|
- sourced datasets and metrics
|
|
7
15
|
- credible baseline shortlist
|
|
16
|
+
- 2-3 candidate approaches with trade-offs
|
|
8
17
|
- generated innovation hypothesis
|
|
9
18
|
- critique before convergence
|
|
10
19
|
- minimum viable experiment
|
|
20
|
+
- explicit approval gate before `/lab:spec`
|
|
11
21
|
|
|
12
22
|
## Evidence Discipline
|
|
13
23
|
|
|
14
24
|
- Mark sourced facts clearly.
|
|
15
25
|
- Mark generated hypotheses clearly.
|
|
16
26
|
- Do not merge them into one undifferentiated summary.
|
|
27
|
+
- Ask one clarifying question at a time when a missing assumption would materially change the proposal.
|
|
28
|
+
- Before writing the full artifact, give the user a short summary with the one-sentence problem, why current methods fail, and the three meaningful points.
|
|
29
|
+
|
|
30
|
+
## Context Read Set
|
|
31
|
+
|
|
32
|
+
- `.superlab/context/mission.md`
|
|
33
|
+
- `.superlab/context/open-questions.md`
|
|
34
|
+
|
|
35
|
+
## Context Write Set
|
|
36
|
+
|
|
37
|
+
- `.superlab/context/mission.md`
|
|
38
|
+
- `.superlab/context/decisions.md`
|
|
39
|
+
- `.superlab/context/open-questions.md`
|
|
17
40
|
|
|
18
41
|
## Recommended Structure
|
|
19
42
|
|
|
20
|
-
1.
|
|
21
|
-
2.
|
|
22
|
-
3.
|
|
23
|
-
4.
|
|
24
|
-
5.
|
|
25
|
-
6.
|
|
26
|
-
7.
|
|
43
|
+
1. One-sentence problem
|
|
44
|
+
2. Failure of existing methods
|
|
45
|
+
3. Idea classification, contribution category, and breakthrough level
|
|
46
|
+
4. Existing methods and shared assumptions
|
|
47
|
+
5. Why the proposed idea is better
|
|
48
|
+
6. Three meaningful points
|
|
49
|
+
7. Candidate approaches and recommendation
|
|
50
|
+
8. Dataset and metric candidates
|
|
51
|
+
9. Falsifiable hypothesis
|
|
52
|
+
10. Expert critique
|
|
53
|
+
11. Revised proposal
|
|
54
|
+
12. Approval gate
|
|
55
|
+
13. Minimum viable experiment
|
|
56
|
+
|
|
57
|
+
## Writing Standard
|
|
58
|
+
|
|
59
|
+
- Keep the problem statement short, concrete, and easy to scan.
|
|
60
|
+
- State why the target problem matters before talking about the method.
|
|
61
|
+
- Compare against existing methods explicitly, not by vague novelty language.
|
|
62
|
+
- The three meaningful points should each fit in one direct sentence.
|
|
@@ -9,8 +9,23 @@ Declare and keep fixed:
|
|
|
9
9
|
- primary metric
|
|
10
10
|
- success threshold
|
|
11
11
|
- verification commands
|
|
12
|
+
- completion_promise
|
|
12
13
|
- maximum iteration count
|
|
13
14
|
|
|
15
|
+
## Context Read Set
|
|
16
|
+
|
|
17
|
+
- `.superlab/context/mission.md`
|
|
18
|
+
- `.superlab/context/state.md`
|
|
19
|
+
- `.superlab/context/decisions.md`
|
|
20
|
+
- `.superlab/context/evidence-index.md`
|
|
21
|
+
|
|
22
|
+
## Context Write Set
|
|
23
|
+
|
|
24
|
+
- `.superlab/context/state.md`
|
|
25
|
+
- `.superlab/context/decisions.md`
|
|
26
|
+
- `.superlab/context/evidence-index.md`
|
|
27
|
+
- `.superlab/context/open-questions.md`
|
|
28
|
+
|
|
14
29
|
## Per-Round Output
|
|
15
30
|
|
|
16
31
|
- round hypothesis
|
|
@@ -18,9 +33,26 @@ Declare and keep fixed:
|
|
|
18
33
|
- normalized evaluation summary
|
|
19
34
|
- written iteration report
|
|
20
35
|
- continue or stop decision
|
|
36
|
+
- diagnostic mode trigger when risk increases for two consecutive rounds
|
|
21
37
|
|
|
22
38
|
## Stop Conditions
|
|
23
39
|
|
|
24
40
|
- threshold reached
|
|
25
41
|
- iteration cap reached
|
|
26
42
|
- fatal methodological flaw discovered
|
|
43
|
+
|
|
44
|
+
## Failure Exit Contract
|
|
45
|
+
|
|
46
|
+
If the loop stops without success, record:
|
|
47
|
+
|
|
48
|
+
- what succeeded
|
|
49
|
+
- what failed
|
|
50
|
+
- top 3 blockers
|
|
51
|
+
- next best actions
|
|
52
|
+
|
|
53
|
+
## Interaction Contract
|
|
54
|
+
|
|
55
|
+
- Start each round with a concise summary of current status, the leading hypothesis, and the main round risk.
|
|
56
|
+
- If the next move depends on an unresolved assumption, ask one clarifying question at a time.
|
|
57
|
+
- If more than one next hypothesis is credible, present 2-3 approaches with trade-offs and recommend the next bounded experiment before changing the mission state.
|
|
58
|
+
- Keep an approval gate when a proposed change would alter the frozen mission instead of only changing the implementation hypothesis.
|
|
@@ -10,9 +10,28 @@
|
|
|
10
10
|
- limitations
|
|
11
11
|
- next steps
|
|
12
12
|
|
|
13
|
+
## Context Read Set
|
|
14
|
+
|
|
15
|
+
- `.superlab/context/mission.md`
|
|
16
|
+
- `.superlab/context/state.md`
|
|
17
|
+
- `.superlab/context/decisions.md`
|
|
18
|
+
- `.superlab/context/evidence-index.md`
|
|
19
|
+
|
|
20
|
+
## Context Write Set
|
|
21
|
+
|
|
22
|
+
- `.superlab/context/state.md`
|
|
23
|
+
- `.superlab/context/evidence-index.md`
|
|
24
|
+
|
|
13
25
|
## Evidence Rules
|
|
14
26
|
|
|
15
27
|
- Do not hide failed iterations.
|
|
16
28
|
- Tie every major claim to recorded summaries or iteration artifacts.
|
|
17
29
|
- Prefer conservative interpretation over marketing language.
|
|
18
30
|
- Leave a clear handoff path into `/lab:write` with evidence links that section drafts can cite.
|
|
31
|
+
|
|
32
|
+
## Interaction Contract
|
|
33
|
+
|
|
34
|
+
- Start with a concise summary of the campaign outcome, strongest supported claim, and biggest reporting risk.
|
|
35
|
+
- If a missing assumption would change report interpretation, ask one clarifying question at a time.
|
|
36
|
+
- If there are multiple defensible report framings, present 2-3 approaches with trade-offs and recommend the most evidence-faithful framing before writing.
|
|
37
|
+
- Keep an approval gate when the reporting frame would materially affect what the paper later claims.
|
|
@@ -1,5 +1,25 @@
|
|
|
1
1
|
# `/lab:review` Stage Guide
|
|
2
2
|
|
|
3
|
+
## Required Flow
|
|
4
|
+
|
|
5
|
+
1. Give a concise summary of the artifact or result under review.
|
|
6
|
+
2. State the top review question or risk focus.
|
|
7
|
+
3. Audit in reviewer mode.
|
|
8
|
+
4. Output fatal flaws first when present.
|
|
9
|
+
5. Rank the fix priority.
|
|
10
|
+
6. End with residual risks and a clear next action.
|
|
11
|
+
|
|
12
|
+
## Context Read Set
|
|
13
|
+
|
|
14
|
+
- `.superlab/context/mission.md`
|
|
15
|
+
- `.superlab/context/decisions.md`
|
|
16
|
+
- `.superlab/context/evidence-index.md`
|
|
17
|
+
|
|
18
|
+
## Context Write Set
|
|
19
|
+
|
|
20
|
+
- `.superlab/context/state.md`
|
|
21
|
+
- `.superlab/context/open-questions.md`
|
|
22
|
+
|
|
3
23
|
## Reviewer Priorities
|
|
4
24
|
|
|
5
25
|
- unfair or weak baselines
|
|
@@ -11,6 +31,16 @@
|
|
|
11
31
|
|
|
12
32
|
## Output Style
|
|
13
33
|
|
|
34
|
+
- concise summary first
|
|
14
35
|
- findings first
|
|
36
|
+
- fatal flaws called out explicitly
|
|
37
|
+
- fix priority stated clearly
|
|
15
38
|
- evidence-linked critique
|
|
16
39
|
- explicit residual risks
|
|
40
|
+
|
|
41
|
+
## Interaction Contract
|
|
42
|
+
|
|
43
|
+
- Start with a concise summary of the review target and the main reviewer question.
|
|
44
|
+
- Ask one clarifying question at a time only if review scope ambiguity would change findings or severity.
|
|
45
|
+
- If there are multiple legitimate review framings, present 2-3 approaches with trade-offs and recommend the strictest useful framing.
|
|
46
|
+
- Do not use brainstorming to soften critique; once scope is clear, stay in reviewer mode and deliver findings directly.
|
|
@@ -7,6 +7,16 @@
|
|
|
7
7
|
- normalized evaluation summary
|
|
8
8
|
- validation result for the normalized summary
|
|
9
9
|
|
|
10
|
+
## Context Read Set
|
|
11
|
+
|
|
12
|
+
- `.superlab/context/mission.md`
|
|
13
|
+
- `.superlab/context/state.md`
|
|
14
|
+
|
|
15
|
+
## Context Write Set
|
|
16
|
+
|
|
17
|
+
- `.superlab/context/state.md`
|
|
18
|
+
- `.superlab/context/evidence-index.md`
|
|
19
|
+
|
|
10
20
|
## Constraints
|
|
11
21
|
|
|
12
22
|
- Prefer the smallest experiment that exercises the full pipeline.
|
|
@@ -20,3 +30,10 @@
|
|
|
20
30
|
3. Execute the smallest useful experiment
|
|
21
31
|
4. Normalize raw metrics
|
|
22
32
|
5. Validate the normalized summary
|
|
33
|
+
|
|
34
|
+
## Interaction Contract
|
|
35
|
+
|
|
36
|
+
- Start with a concise summary of the run goal, smallest candidate experiment, and primary execution risk.
|
|
37
|
+
- If the next run depends on an unresolved assumption, ask one clarifying question at a time.
|
|
38
|
+
- If there are multiple defensible tiny-run options, present 2-3 approaches with trade-offs and recommend the cheapest informative run.
|
|
39
|
+
- Only ask for approval when choosing a run path would materially spend more time or compute than the default smallest experiment.
|
|
@@ -2,10 +2,33 @@
|
|
|
2
2
|
|
|
3
3
|
## Required Output
|
|
4
4
|
|
|
5
|
-
-
|
|
6
|
-
-
|
|
7
|
-
-
|
|
8
|
-
-
|
|
5
|
+
- a lab change id
|
|
6
|
+
- `.superlab/changes/<change-id>/proposal.md`
|
|
7
|
+
- `.superlab/changes/<change-id>/design.md`
|
|
8
|
+
- `.superlab/changes/<change-id>/spec.md`
|
|
9
|
+
- `.superlab/changes/<change-id>/tasks.md`
|
|
10
|
+
|
|
11
|
+
## Config Read Set
|
|
12
|
+
|
|
13
|
+
- `.superlab/config/workflow.json`
|
|
14
|
+
|
|
15
|
+
## Context Read Set
|
|
16
|
+
|
|
17
|
+
- `.superlab/context/mission.md`
|
|
18
|
+
- `.superlab/context/decisions.md`
|
|
19
|
+
- `.superlab/context/state.md`
|
|
20
|
+
|
|
21
|
+
## Context Write Set
|
|
22
|
+
|
|
23
|
+
- `.superlab/context/state.md`
|
|
24
|
+
- `.superlab/context/decisions.md`
|
|
25
|
+
|
|
26
|
+
## Required Change Layout
|
|
27
|
+
|
|
28
|
+
1. Identify or create a lab change id.
|
|
29
|
+
2. Create `.superlab/changes/<change-id>/`.
|
|
30
|
+
3. Draft `proposal.md`, `design.md`, `spec.md`, and `tasks.md` inside that directory.
|
|
31
|
+
4. Keep the change self-contained so downstream stages can read one directory and its linked context.
|
|
9
32
|
|
|
10
33
|
## Conversion Rules
|
|
11
34
|
|
|
@@ -13,9 +36,18 @@
|
|
|
13
36
|
- Preserve evaluation boundaries from the idea stage.
|
|
14
37
|
- Translate risks into concrete tasks when possible.
|
|
15
38
|
- Make task granularity small enough that `/lab:run` and `/lab:iterate` can execute predictably.
|
|
39
|
+
- Use one lab-native change directory per approved idea instead of scattering spec artifacts.
|
|
40
|
+
|
|
41
|
+
## Interaction Contract
|
|
42
|
+
|
|
43
|
+
- Start with a concise summary of the approved idea, target change, and the main spec risk.
|
|
44
|
+
- If change decomposition or validation strategy is materially ambiguous, ask one clarifying question at a time.
|
|
45
|
+
- If there are multiple viable ways to structure the lab change directory, present 2-3 approaches with trade-offs and a recommendation before freezing the change.
|
|
46
|
+
- Keep an approval gate before locking a change structure that will drive `/lab:run` and `/lab:iterate`.
|
|
16
47
|
|
|
17
48
|
## Minimum Task Coverage
|
|
18
49
|
|
|
50
|
+
- change setup
|
|
19
51
|
- artifact creation
|
|
20
52
|
- validation run
|
|
21
53
|
- evaluation normalization
|
|
@@ -6,40 +6,72 @@
|
|
|
6
6
|
- iteration reports
|
|
7
7
|
- normalized summaries
|
|
8
8
|
- reviewer notes when available
|
|
9
|
-
-
|
|
9
|
+
- vendored paper-writing references under `skills/lab/references/paper-writing/`
|
|
10
10
|
|
|
11
|
-
##
|
|
11
|
+
## Config Read Set
|
|
12
12
|
|
|
13
|
-
|
|
13
|
+
- `.superlab/config/workflow.json`
|
|
14
14
|
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
-
|
|
18
|
-
-
|
|
19
|
-
-
|
|
20
|
-
|
|
15
|
+
## Context Read Set
|
|
16
|
+
|
|
17
|
+
- `.superlab/context/mission.md`
|
|
18
|
+
- `.superlab/context/decisions.md`
|
|
19
|
+
- `.superlab/context/evidence-index.md`
|
|
20
|
+
|
|
21
|
+
## Context Write Set
|
|
22
|
+
|
|
23
|
+
- `.superlab/context/state.md`
|
|
24
|
+
- `.superlab/context/evidence-index.md`
|
|
25
|
+
|
|
26
|
+
## Required References
|
|
27
|
+
|
|
28
|
+
Load the exact vendored file that matches the current target:
|
|
29
|
+
|
|
30
|
+
- abstract -> `skills/lab/references/paper-writing/abstract.md`
|
|
31
|
+
- introduction -> `skills/lab/references/paper-writing/introduction.md`
|
|
32
|
+
- related work -> `skills/lab/references/paper-writing/related-work.md`
|
|
33
|
+
- method -> `skills/lab/references/paper-writing/method.md`
|
|
34
|
+
- experiments -> `skills/lab/references/paper-writing/experiments.md`
|
|
35
|
+
- conclusion -> `skills/lab/references/paper-writing/conclusion.md`
|
|
21
36
|
|
|
22
37
|
Run these on every round:
|
|
23
38
|
|
|
24
|
-
- section flow check -> `
|
|
25
|
-
- reviewer pass -> `
|
|
39
|
+
- section flow check -> `skills/lab/references/paper-writing/does-my-writing-flow-source.md`
|
|
40
|
+
- reviewer pass -> `skills/lab/references/paper-writing/paper-review.md`
|
|
26
41
|
|
|
27
42
|
## Small-Step Writing Rules
|
|
28
43
|
|
|
29
44
|
- Change one section or one clearly bounded subsection per round.
|
|
45
|
+
- LaTeX is the required manuscript output format.
|
|
46
|
+
- Load only the current section guide. Do not load every section guide at once.
|
|
47
|
+
- Build a compact mini-outline before prose.
|
|
48
|
+
- For each subsection, explicitly include motivation, design, and technical advantage when applicable.
|
|
49
|
+
- Avoid a writing style that reads like incremental patching of a naive baseline.
|
|
50
|
+
- Keep terminology stable across the full paper.
|
|
30
51
|
- Tie every new or revised claim to explicit evidence.
|
|
52
|
+
- If a claim cannot be supported by results, weaken or remove it.
|
|
31
53
|
- Record what changed and why in a write-iteration artifact.
|
|
54
|
+
- Return paragraph-level roles for the revised prose when drafting.
|
|
55
|
+
- Run the five-dimension self-review checklist before accepting a round.
|
|
32
56
|
- Run reviewer-style checks after every round.
|
|
33
|
-
- Stop instead of free-writing if the
|
|
57
|
+
- Stop instead of free-writing if the required references are missing.
|
|
34
58
|
|
|
35
59
|
## Required Artifacts
|
|
36
60
|
|
|
37
|
-
-
|
|
38
|
-
-
|
|
39
|
-
- `docs/
|
|
61
|
+
- `.superlab/write/plan.md`
|
|
62
|
+
- `.superlab/write/iterations/<n>.md`
|
|
63
|
+
- `docs/paper/paper.tex`
|
|
64
|
+
- `docs/paper/sections/<section>.tex`
|
|
40
65
|
|
|
41
66
|
## Stop Conditions
|
|
42
67
|
|
|
43
68
|
- the target section is accepted for this round
|
|
44
69
|
- evidence is missing and the workflow must return to `review` or `iterate`
|
|
45
70
|
- the writing loop reaches its declared round limit
|
|
71
|
+
|
|
72
|
+
## Interaction Contract
|
|
73
|
+
|
|
74
|
+
- Start with a concise summary of the target section, evidence base, and main writing risk for this round.
|
|
75
|
+
- If a missing assumption would change the section structure or claim strength, ask one clarifying question at a time.
|
|
76
|
+
- If there are multiple defensible narrative structures for the current section, present 2-3 approaches with trade-offs and recommend one before drafting.
|
|
77
|
+
- Keep an approval gate when the chosen narrative structure would materially affect downstream sections or paper-level framing.
|
|
@@ -1,5 +1,10 @@
|
|
|
1
1
|
# Design
|
|
2
2
|
|
|
3
|
+
## Lab Change
|
|
4
|
+
|
|
5
|
+
- Change id:
|
|
6
|
+
- Target path: `.superlab/changes/<change-id>/design.md`
|
|
7
|
+
|
|
3
8
|
## Research Approach
|
|
4
9
|
|
|
5
10
|
Describe the method and why it is plausible.
|
|
@@ -21,3 +26,8 @@ Describe the method and why it is plausible.
|
|
|
21
26
|
- Run registry path:
|
|
22
27
|
- Normalized summary path:
|
|
23
28
|
- Iteration report path:
|
|
29
|
+
|
|
30
|
+
## Change Links
|
|
31
|
+
|
|
32
|
+
- Spec path: `.superlab/changes/<change-id>/spec.md`
|
|
33
|
+
- Tasks path: `.superlab/changes/<change-id>/tasks.md`
|