superlab 0.1.0 → 0.1.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (49) hide show
  1. package/README.md +23 -23
  2. package/README.zh-CN.md +22 -21
  3. package/lib/i18n.cjs +626 -23
  4. package/lib/install.cjs +31 -0
  5. package/package-assets/claude/commands/lab/spec.md +1 -1
  6. package/package-assets/claude/commands/lab/write.md +1 -1
  7. package/package-assets/codex/prompts/lab-spec.md +1 -1
  8. package/package-assets/codex/prompts/lab-write.md +1 -1
  9. package/package-assets/shared/changes/README.md +10 -0
  10. package/package-assets/shared/config/workflow.json +5 -0
  11. package/package-assets/shared/context/decisions.md +11 -0
  12. package/package-assets/shared/context/evidence-index.md +16 -0
  13. package/package-assets/shared/context/mission.md +27 -0
  14. package/package-assets/shared/context/open-questions.md +11 -0
  15. package/package-assets/shared/context/state.md +19 -0
  16. package/package-assets/shared/examples/minimal-uplift-workflow.md +4 -4
  17. package/package-assets/shared/skills/lab/SKILL.md +54 -9
  18. package/package-assets/shared/skills/lab/references/brainstorming-integration.md +21 -0
  19. package/package-assets/shared/skills/lab/references/paper-writing/abstract.md +102 -0
  20. package/package-assets/shared/skills/lab/references/paper-writing/conclusion.md +35 -0
  21. package/package-assets/shared/skills/lab/references/paper-writing/does-my-writing-flow-source.md +45 -0
  22. package/package-assets/shared/skills/lab/references/paper-writing/experiments.md +102 -0
  23. package/package-assets/shared/skills/lab/references/paper-writing/introduction.md +408 -0
  24. package/package-assets/shared/skills/lab/references/paper-writing/method.md +196 -0
  25. package/package-assets/shared/skills/lab/references/paper-writing/paper-review.md +86 -0
  26. package/package-assets/shared/skills/lab/references/paper-writing/related-work.md +41 -0
  27. package/package-assets/shared/skills/lab/references/paper-writing-integration.md +29 -28
  28. package/package-assets/shared/skills/lab/references/workflow.md +1 -1
  29. package/package-assets/shared/skills/lab/stages/idea.md +43 -7
  30. package/package-assets/shared/skills/lab/stages/iterate.md +32 -0
  31. package/package-assets/shared/skills/lab/stages/report.md +19 -0
  32. package/package-assets/shared/skills/lab/stages/review.md +30 -0
  33. package/package-assets/shared/skills/lab/stages/run.md +17 -0
  34. package/package-assets/shared/skills/lab/stages/spec.md +36 -4
  35. package/package-assets/shared/skills/lab/stages/write.md +47 -15
  36. package/package-assets/shared/templates/design.md +10 -0
  37. package/package-assets/shared/templates/idea.md +76 -8
  38. package/package-assets/shared/templates/iteration-report.md +4 -0
  39. package/package-assets/shared/templates/paper-plan.md +12 -0
  40. package/package-assets/shared/templates/paper-section.md +24 -6
  41. package/package-assets/shared/templates/paper-section.tex +10 -0
  42. package/package-assets/shared/templates/paper.tex +29 -0
  43. package/package-assets/shared/templates/proposal.md +10 -0
  44. package/package-assets/shared/templates/review-checklist.md +23 -0
  45. package/package-assets/shared/templates/spec.md +7 -2
  46. package/package-assets/shared/templates/tasks.md +3 -1
  47. package/package-assets/shared/templates/write-iteration.md +5 -0
  48. package/package.json +3 -3
  49. package/package-assets/shared/scripts/check_openspec.sh +0 -10
@@ -0,0 +1,86 @@
1
+ # Paper Review (Paper Rview)
2
+
3
+ ## Goal
4
+
5
+ Use an adversarial, reviewer-style checklist to detect reject risks early and revise the paper before submission.
6
+
7
+ ## Core Principle
8
+
9
+ Pursue perfectionism in paper quality: assume reviewers will probe every weak point and proactively fix them.
10
+
11
+ ## Critical Rule (Do Not Violate)
12
+
13
+ Every major claim, especially in Abstract and Introduction, must be:
14
+
15
+ 1. technically correct, and
16
+ 2. explicitly supported by experimental evidence.
17
+
18
+ If a claim is not supported, either add evidence or weaken/remove the claim.
19
+
20
+ ## What Usually Gets a Paper Accepted
21
+
22
+ 1. Sufficient contribution (for example: novel task, novel pipeline, novel module, novel design choices, new experimental findings, or new insight).
23
+ 2. Better empirical performance than prior methods under fair comparisons.
24
+ 3. Sufficient comparison experiments and ablation studies.
25
+
26
+ ## Common Rejection Dimensions
27
+
28
+ | Rejection Dimension | Typical Failure Signals |
29
+ | ---------------------------- | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ |
30
+ | 1. Insufficient contribution | 1.1 Targeted failure cases are too common.<br /> 1.2 Proposed technique is already well explored; expected gains are predictable/well-known. |
31
+ | 2. Unclear writing | 2.1 Missing technical details; work is not reproducible.<br />2.2 A method module lacks clear motivation. |
32
+ | 3. Weak empirical effect | 3.1 Improvement over prior methods is only marginal.<br /> 3.2 Even if better than previous methods, absolute performance is still not strong enough. |
33
+ | 4. Incomplete evaluation | 4.1 Missing ablation studies.<br />4.2 Missing important baselines or important evaluation metrics.<br /> 4.3 Datasets are too simple to prove the method truly works. |
34
+ | 5. Problematic method design | 5.1 Experimental setting is unrealistic.<br />5.2 Method has technical flaws and appears unreasonable.<br />5.3 Method is not robust and needs per-scenario hyperparameter tuning. <br /> 5.4 New design introduces stronger limitations than its benefits, leading to negative net value. |
35
+
36
+ ## End-of-Paper Self-Review Question List
37
+
38
+ Add this checklist near the end of the draft while revising.
39
+ Use each question to trigger concrete edits before submission.
40
+
41
+ ### 1. Contribution
42
+
43
+ 1. What new knowledge does this paper give to readers?
44
+ 2. Are we solving a truly meaningful failure case, not a trivial/common one?
45
+ 3. Is the technical idea genuinely non-obvious beyond well-explored practice?
46
+ 4. Is our gain surprising or insightful rather than a predictable improvement?
47
+ 5. Is there at least one clear novelty type (task/pipeline/module/design finding/insight)?
48
+
49
+ ### 2. Writing Clarity
50
+
51
+ 1. Can a knowledgeable reader reproduce the method from the paper?
52
+ 2. Did we provide enough technical detail for each key module?
53
+ 3. Is the motivation of every module explicit and logically connected to a challenge?
54
+ 4. Are terms and notation consistent across sections?
55
+ 5. Does each paragraph carry one clear message with smooth transitions?
56
+
57
+ ### 3. Experimental Strength
58
+
59
+ 1. Are improvements over strong baselines meaningful, not just statistically tiny?
60
+ 2. Is absolute performance competitive enough for the target venue?
61
+ 3. Are gains consistent across multiple datasets/settings/metrics?
62
+ 4. Do we report both strengths and failure cases honestly?
63
+
64
+ ### 4. Evaluation Completeness
65
+
66
+ 1. Do we include ablations for all key design choices?
67
+ 2. Are all strong/recent baselines included under fair settings?
68
+ 3. Are evaluation metrics standard and sufficient for this task?
69
+ 4. Are datasets/scenarios challenging enough to validate real effectiveness?
70
+ 5. Are comparison and ablation protocols clearly documented?
71
+
72
+ ### 5. Method Design Soundness
73
+
74
+ 1. Is the experimental setting realistic for practical use?
75
+ 2. Does the method have hidden technical defects or unreasonable assumptions?
76
+ 3. Is the method robust without heavy per-case hyperparameter retuning?
77
+ 4. Do benefits outweigh added complexity and new limitations?
78
+ 5. Could reviewers reasonably argue that the net benefit is negative?
79
+
80
+ ## Adversarial Writing Workflow
81
+
82
+ 1. Read the paper as a skeptical reviewer.
83
+ 2. Answer every question above with explicit evidence from the paper.
84
+ 3. Mark each item as `pass`, `needs revision`, or `needs new experiment`.
85
+ 4. Revise claims, writing, experiments, or method scope accordingly.
86
+ 5. Repeat until no major rejection risk remains.
@@ -0,0 +1,41 @@
1
+ # Related Work Writing Guide
2
+
3
+ ## Goal
4
+
5
+ Position your work against the most relevant lines of research, and make your novelty easy to verify.
6
+
7
+ ## Workflow
8
+
9
+ 1. List directly competing and recent baseline papers first.
10
+ 2. Group literature by technical topic (not by publication year alone).
11
+ 3. For each topic: summarize common paradigm, then key limitation relevant to your challenge.
12
+ 4. End each topic by clarifying your distinction.
13
+
14
+ ## Topic Design
15
+
16
+ Use 2-4 focused topics, for example:
17
+
18
+ 1. Task-specific mainstream methods.
19
+ 2. Methods closest to your core idea.
20
+ 3. Auxiliary techniques your method builds on.
21
+
22
+ ## Paragraph Template
23
+
24
+ 1. Topic sentence: define scope of this topic.
25
+ 2. Representative methods: one compact summary.
26
+ 3. Limitation tied to your target technical challenge.
27
+ 4. Transition sentence that leads to your method.
28
+
29
+ ## Do and Don't
30
+
31
+ 1. Do compare mechanisms, assumptions, and failure modes.
32
+ 2. Do emphasize the exact gap your method fills.
33
+ 3. Do not make Related Work a citation dump.
34
+ 4. Do not hide strongest baselines.
35
+
36
+ ## Checklist
37
+
38
+ 1. Are all strongest/recent competitors covered?
39
+ 2. Is each topic connected to your problem setting?
40
+ 3. Is your difference explained in technical terms, not marketing terms?
41
+ 4. Is citation coverage complete for all core claims?
@@ -1,39 +1,40 @@
1
1
  # Paper-Writing Integration
2
2
 
3
- `/lab:write` should directly reuse the upstream `research-paper-writing` section references instead of replacing them with a home-grown summary.
3
+ `/lab:write` vendors the paper-writing references directly into `skills/lab/references/paper-writing/`.
4
+ The goal is to keep the upstream writing discipline while removing brittle runtime dependence on a separately installed skill.
4
5
 
5
- Upstream source:
6
+ ## Role Split
6
7
 
7
- - `https://github.com/Master-cai/Research-Paper-Writing-Skills/tree/main/research-paper-writing/references`
8
-
9
- ## Use paper-writing rules for
8
+ - `lab` controls stage boundaries, evidence discipline, and durable artifacts.
9
+ - the vendored paper-writing references control section structure, paragraph logic, and reviewer-facing polish.
10
+ - `/lab:write` links evidence-backed research outputs to paper-ready text.
10
11
 
11
- - section structure
12
- - paragraph flow
13
- - claim-evidence alignment
14
- - conservative academic tone
15
- - reviewer-style self-checks
12
+ ## Required Shared Constraints
16
13
 
17
- ## Do not use paper-writing rules to
14
+ - one section or one bounded subsection per round
15
+ - load only the active section guide, not every section guide
16
+ - mini-outline before prose
17
+ - explicit evidence links for major claims
18
+ - motivation, design, and technical advantage when applicable
19
+ - stable terminology across the whole paper
20
+ - unsupported claims must be weakened or removed
21
+ - section-level flow check and reviewer pass after each round
22
+ - five-dimension self-review before a round is accepted
23
+ - route back to `review` or `iterate` if the evidence is weak
18
24
 
19
- - invent unsupported results
20
- - hide failed runs or limitations
21
- - blur sourced evidence with generated interpretation
22
- - rewrite the whole paper in one unconstrained pass
25
+ ## Vendored References
23
26
 
24
- ## Operating Model
27
+ - `paper-writing/abstract.md`
28
+ - `paper-writing/introduction.md`
29
+ - `paper-writing/related-work.md`
30
+ - `paper-writing/method.md`
31
+ - `paper-writing/experiments.md`
32
+ - `paper-writing/conclusion.md`
33
+ - `paper-writing/paper-review.md`
34
+ - `paper-writing/does-my-writing-flow-source.md`
25
35
 
26
- - `lab` controls stage boundaries and artifacts
27
- - `research-paper-writing` references control how each section is written and revised
28
- - evidence still comes from `report`, iteration reports, and normalized summaries
36
+ ## Attribution
29
37
 
30
- ## Required Upstream Files
38
+ These references are adapted from:
31
39
 
32
- - `abstract.md`
33
- - `introduction.md`
34
- - `related-work.md`
35
- - `method.md`
36
- - `experiments.md`
37
- - `conclusion.md`
38
- - `paper-review.md`
39
- - `does-my-writing-flow-source.md`
40
+ - `https://github.com/Master-cai/Research-Paper-Writing-Skills/tree/main/research-paper-writing/references`
@@ -25,7 +25,7 @@ Escalate to a higher-level redesign instead of mutating the mission when the cur
25
25
  ## Required Artifacts
26
26
 
27
27
  - `docs/lab/idea.md`
28
- - OpenSpec `proposal`, `design`, `spec`, `tasks`
28
+ - one lab change directory under `.superlab/changes/<change-id>/`
29
29
  - normalized JSON summary from `scripts/eval_report.py`
30
30
  - per-round report in `docs/lab/iterations/`
31
31
  - final report in `docs/lab/report.md`
@@ -2,25 +2,61 @@
2
2
 
3
3
  ## Required Output
4
4
 
5
+ - one-sentence problem statement
6
+ - failure case
7
+ - idea classification
8
+ - contribution category
9
+ - breakthrough level
10
+ - existing methods summary
11
+ - why the proposed idea is better than existing methods
12
+ - three meaningful points
5
13
  - literature-backed framing
6
14
  - sourced datasets and metrics
7
15
  - credible baseline shortlist
16
+ - 2-3 candidate approaches with trade-offs
8
17
  - generated innovation hypothesis
9
18
  - critique before convergence
10
19
  - minimum viable experiment
20
+ - explicit approval gate before `/lab:spec`
11
21
 
12
22
  ## Evidence Discipline
13
23
 
14
24
  - Mark sourced facts clearly.
15
25
  - Mark generated hypotheses clearly.
16
26
  - Do not merge them into one undifferentiated summary.
27
+ - Ask one clarifying question at a time when a missing assumption would materially change the proposal.
28
+ - Before writing the full artifact, give the user a short summary with the one-sentence problem, why current methods fail, and the three meaningful points.
29
+
30
+ ## Context Read Set
31
+
32
+ - `.superlab/context/mission.md`
33
+ - `.superlab/context/open-questions.md`
34
+
35
+ ## Context Write Set
36
+
37
+ - `.superlab/context/mission.md`
38
+ - `.superlab/context/decisions.md`
39
+ - `.superlab/context/open-questions.md`
17
40
 
18
41
  ## Recommended Structure
19
42
 
20
- 1. Problem framing
21
- 2. Related work and baseline conventions
22
- 3. Dataset and metric candidates
23
- 4. Proposed innovation
24
- 5. Expert critique
25
- 6. Revised proposal
26
- 7. Minimum viable experiment
43
+ 1. One-sentence problem
44
+ 2. Failure of existing methods
45
+ 3. Idea classification, contribution category, and breakthrough level
46
+ 4. Existing methods and shared assumptions
47
+ 5. Why the proposed idea is better
48
+ 6. Three meaningful points
49
+ 7. Candidate approaches and recommendation
50
+ 8. Dataset and metric candidates
51
+ 9. Falsifiable hypothesis
52
+ 10. Expert critique
53
+ 11. Revised proposal
54
+ 12. Approval gate
55
+ 13. Minimum viable experiment
56
+
57
+ ## Writing Standard
58
+
59
+ - Keep the problem statement short, concrete, and easy to scan.
60
+ - State why the target problem matters before talking about the method.
61
+ - Compare against existing methods explicitly, not by vague novelty language.
62
+ - The three meaningful points should each fit in one direct sentence.
@@ -9,8 +9,23 @@ Declare and keep fixed:
9
9
  - primary metric
10
10
  - success threshold
11
11
  - verification commands
12
+ - completion_promise
12
13
  - maximum iteration count
13
14
 
15
+ ## Context Read Set
16
+
17
+ - `.superlab/context/mission.md`
18
+ - `.superlab/context/state.md`
19
+ - `.superlab/context/decisions.md`
20
+ - `.superlab/context/evidence-index.md`
21
+
22
+ ## Context Write Set
23
+
24
+ - `.superlab/context/state.md`
25
+ - `.superlab/context/decisions.md`
26
+ - `.superlab/context/evidence-index.md`
27
+ - `.superlab/context/open-questions.md`
28
+
14
29
  ## Per-Round Output
15
30
 
16
31
  - round hypothesis
@@ -18,9 +33,26 @@ Declare and keep fixed:
18
33
  - normalized evaluation summary
19
34
  - written iteration report
20
35
  - continue or stop decision
36
+ - diagnostic mode trigger when risk increases for two consecutive rounds
21
37
 
22
38
  ## Stop Conditions
23
39
 
24
40
  - threshold reached
25
41
  - iteration cap reached
26
42
  - fatal methodological flaw discovered
43
+
44
+ ## Failure Exit Contract
45
+
46
+ If the loop stops without success, record:
47
+
48
+ - what succeeded
49
+ - what failed
50
+ - top 3 blockers
51
+ - next best actions
52
+
53
+ ## Interaction Contract
54
+
55
+ - Start each round with a concise summary of current status, the leading hypothesis, and the main round risk.
56
+ - If the next move depends on an unresolved assumption, ask one clarifying question at a time.
57
+ - If more than one next hypothesis is credible, present 2-3 approaches with trade-offs and recommend the next bounded experiment before changing the mission state.
58
+ - Keep an approval gate when a proposed change would alter the frozen mission instead of only changing the implementation hypothesis.
@@ -10,9 +10,28 @@
10
10
  - limitations
11
11
  - next steps
12
12
 
13
+ ## Context Read Set
14
+
15
+ - `.superlab/context/mission.md`
16
+ - `.superlab/context/state.md`
17
+ - `.superlab/context/decisions.md`
18
+ - `.superlab/context/evidence-index.md`
19
+
20
+ ## Context Write Set
21
+
22
+ - `.superlab/context/state.md`
23
+ - `.superlab/context/evidence-index.md`
24
+
13
25
  ## Evidence Rules
14
26
 
15
27
  - Do not hide failed iterations.
16
28
  - Tie every major claim to recorded summaries or iteration artifacts.
17
29
  - Prefer conservative interpretation over marketing language.
18
30
  - Leave a clear handoff path into `/lab:write` with evidence links that section drafts can cite.
31
+
32
+ ## Interaction Contract
33
+
34
+ - Start with a concise summary of the campaign outcome, strongest supported claim, and biggest reporting risk.
35
+ - If a missing assumption would change report interpretation, ask one clarifying question at a time.
36
+ - If there are multiple defensible report framings, present 2-3 approaches with trade-offs and recommend the most evidence-faithful framing before writing.
37
+ - Keep an approval gate when the reporting frame would materially affect what the paper later claims.
@@ -1,5 +1,25 @@
1
1
  # `/lab:review` Stage Guide
2
2
 
3
+ ## Required Flow
4
+
5
+ 1. Give a concise summary of the artifact or result under review.
6
+ 2. State the top review question or risk focus.
7
+ 3. Audit in reviewer mode.
8
+ 4. Output fatal flaws first when present.
9
+ 5. Rank the fix priority.
10
+ 6. End with residual risks and a clear next action.
11
+
12
+ ## Context Read Set
13
+
14
+ - `.superlab/context/mission.md`
15
+ - `.superlab/context/decisions.md`
16
+ - `.superlab/context/evidence-index.md`
17
+
18
+ ## Context Write Set
19
+
20
+ - `.superlab/context/state.md`
21
+ - `.superlab/context/open-questions.md`
22
+
3
23
  ## Reviewer Priorities
4
24
 
5
25
  - unfair or weak baselines
@@ -11,6 +31,16 @@
11
31
 
12
32
  ## Output Style
13
33
 
34
+ - concise summary first
14
35
  - findings first
36
+ - fatal flaws called out explicitly
37
+ - fix priority stated clearly
15
38
  - evidence-linked critique
16
39
  - explicit residual risks
40
+
41
+ ## Interaction Contract
42
+
43
+ - Start with a concise summary of the review target and the main reviewer question.
44
+ - Ask one clarifying question at a time only if review scope ambiguity would change findings or severity.
45
+ - If there are multiple legitimate review framings, present 2-3 approaches with trade-offs and recommend the strictest useful framing.
46
+ - Do not use brainstorming to soften critique; once scope is clear, stay in reviewer mode and deliver findings directly.
@@ -7,6 +7,16 @@
7
7
  - normalized evaluation summary
8
8
  - validation result for the normalized summary
9
9
 
10
+ ## Context Read Set
11
+
12
+ - `.superlab/context/mission.md`
13
+ - `.superlab/context/state.md`
14
+
15
+ ## Context Write Set
16
+
17
+ - `.superlab/context/state.md`
18
+ - `.superlab/context/evidence-index.md`
19
+
10
20
  ## Constraints
11
21
 
12
22
  - Prefer the smallest experiment that exercises the full pipeline.
@@ -20,3 +30,10 @@
20
30
  3. Execute the smallest useful experiment
21
31
  4. Normalize raw metrics
22
32
  5. Validate the normalized summary
33
+
34
+ ## Interaction Contract
35
+
36
+ - Start with a concise summary of the run goal, smallest candidate experiment, and primary execution risk.
37
+ - If the next run depends on an unresolved assumption, ask one clarifying question at a time.
38
+ - If there are multiple defensible tiny-run options, present 2-3 approaches with trade-offs and recommend the cheapest informative run.
39
+ - Only ask for approval when choosing a run path would materially spend more time or compute than the default smallest experiment.
@@ -2,10 +2,33 @@
2
2
 
3
3
  ## Required Output
4
4
 
5
- - OpenSpec `proposal`
6
- - OpenSpec `design`
7
- - OpenSpec `spec`
8
- - OpenSpec `tasks`
5
+ - a lab change id
6
+ - `.superlab/changes/<change-id>/proposal.md`
7
+ - `.superlab/changes/<change-id>/design.md`
8
+ - `.superlab/changes/<change-id>/spec.md`
9
+ - `.superlab/changes/<change-id>/tasks.md`
10
+
11
+ ## Config Read Set
12
+
13
+ - `.superlab/config/workflow.json`
14
+
15
+ ## Context Read Set
16
+
17
+ - `.superlab/context/mission.md`
18
+ - `.superlab/context/decisions.md`
19
+ - `.superlab/context/state.md`
20
+
21
+ ## Context Write Set
22
+
23
+ - `.superlab/context/state.md`
24
+ - `.superlab/context/decisions.md`
25
+
26
+ ## Required Change Layout
27
+
28
+ 1. Identify or create a lab change id.
29
+ 2. Create `.superlab/changes/<change-id>/`.
30
+ 3. Draft `proposal.md`, `design.md`, `spec.md`, and `tasks.md` inside that directory.
31
+ 4. Keep the change self-contained so downstream stages can read one directory and its linked context.
9
32
 
10
33
  ## Conversion Rules
11
34
 
@@ -13,9 +36,18 @@
13
36
  - Preserve evaluation boundaries from the idea stage.
14
37
  - Translate risks into concrete tasks when possible.
15
38
  - Make task granularity small enough that `/lab:run` and `/lab:iterate` can execute predictably.
39
+ - Use one lab-native change directory per approved idea instead of scattering spec artifacts.
40
+
41
+ ## Interaction Contract
42
+
43
+ - Start with a concise summary of the approved idea, target change, and the main spec risk.
44
+ - If change decomposition or validation strategy is materially ambiguous, ask one clarifying question at a time.
45
+ - If there are multiple viable ways to structure the lab change directory, present 2-3 approaches with trade-offs and a recommendation before freezing the change.
46
+ - Keep an approval gate before locking a change structure that will drive `/lab:run` and `/lab:iterate`.
16
47
 
17
48
  ## Minimum Task Coverage
18
49
 
50
+ - change setup
19
51
  - artifact creation
20
52
  - validation run
21
53
  - evaluation normalization
@@ -6,40 +6,72 @@
6
6
  - iteration reports
7
7
  - normalized summaries
8
8
  - reviewer notes when available
9
- - companion `research-paper-writing` references for the active section
9
+ - vendored paper-writing references under `skills/lab/references/paper-writing/`
10
10
 
11
- ## Companion References
11
+ ## Config Read Set
12
12
 
13
- Load the exact companion file that matches the current target:
13
+ - `.superlab/config/workflow.json`
14
14
 
15
- - abstract -> `research-paper-writing/references/abstract.md`
16
- - introduction -> `research-paper-writing/references/introduction.md`
17
- - related work -> `research-paper-writing/references/related-work.md`
18
- - method -> `research-paper-writing/references/method.md`
19
- - experiments -> `research-paper-writing/references/experiments.md`
20
- - conclusion -> `research-paper-writing/references/conclusion.md`
15
+ ## Context Read Set
16
+
17
+ - `.superlab/context/mission.md`
18
+ - `.superlab/context/decisions.md`
19
+ - `.superlab/context/evidence-index.md`
20
+
21
+ ## Context Write Set
22
+
23
+ - `.superlab/context/state.md`
24
+ - `.superlab/context/evidence-index.md`
25
+
26
+ ## Required References
27
+
28
+ Load the exact vendored file that matches the current target:
29
+
30
+ - abstract -> `skills/lab/references/paper-writing/abstract.md`
31
+ - introduction -> `skills/lab/references/paper-writing/introduction.md`
32
+ - related work -> `skills/lab/references/paper-writing/related-work.md`
33
+ - method -> `skills/lab/references/paper-writing/method.md`
34
+ - experiments -> `skills/lab/references/paper-writing/experiments.md`
35
+ - conclusion -> `skills/lab/references/paper-writing/conclusion.md`
21
36
 
22
37
  Run these on every round:
23
38
 
24
- - section flow check -> `research-paper-writing/references/does-my-writing-flow-source.md`
25
- - reviewer pass -> `research-paper-writing/references/paper-review.md`
39
+ - section flow check -> `skills/lab/references/paper-writing/does-my-writing-flow-source.md`
40
+ - reviewer pass -> `skills/lab/references/paper-writing/paper-review.md`
26
41
 
27
42
  ## Small-Step Writing Rules
28
43
 
29
44
  - Change one section or one clearly bounded subsection per round.
45
+ - LaTeX is the required manuscript output format.
46
+ - Load only the current section guide. Do not load every section guide at once.
47
+ - Build a compact mini-outline before prose.
48
+ - For each subsection, explicitly include motivation, design, and technical advantage when applicable.
49
+ - Avoid a writing style that reads like incremental patching of a naive baseline.
50
+ - Keep terminology stable across the full paper.
30
51
  - Tie every new or revised claim to explicit evidence.
52
+ - If a claim cannot be supported by results, weaken or remove it.
31
53
  - Record what changed and why in a write-iteration artifact.
54
+ - Return paragraph-level roles for the revised prose when drafting.
55
+ - Run the five-dimension self-review checklist before accepting a round.
32
56
  - Run reviewer-style checks after every round.
33
- - Stop instead of free-writing if the companion references are missing.
57
+ - Stop instead of free-writing if the required references are missing.
34
58
 
35
59
  ## Required Artifacts
36
60
 
37
- - `docs/lab/paper/plan.md`
38
- - `docs/lab/paper/sections/<section>.md`
39
- - `docs/lab/paper/write-iterations/<n>.md`
61
+ - `.superlab/write/plan.md`
62
+ - `.superlab/write/iterations/<n>.md`
63
+ - `docs/paper/paper.tex`
64
+ - `docs/paper/sections/<section>.tex`
40
65
 
41
66
  ## Stop Conditions
42
67
 
43
68
  - the target section is accepted for this round
44
69
  - evidence is missing and the workflow must return to `review` or `iterate`
45
70
  - the writing loop reaches its declared round limit
71
+
72
+ ## Interaction Contract
73
+
74
+ - Start with a concise summary of the target section, evidence base, and main writing risk for this round.
75
+ - If a missing assumption would change the section structure or claim strength, ask one clarifying question at a time.
76
+ - If there are multiple defensible narrative structures for the current section, present 2-3 approaches with trade-offs and recommend one before drafting.
77
+ - Keep an approval gate when the chosen narrative structure would materially affect downstream sections or paper-level framing.
@@ -1,5 +1,10 @@
1
1
  # Design
2
2
 
3
+ ## Lab Change
4
+
5
+ - Change id:
6
+ - Target path: `.superlab/changes/<change-id>/design.md`
7
+
3
8
  ## Research Approach
4
9
 
5
10
  Describe the method and why it is plausible.
@@ -21,3 +26,8 @@ Describe the method and why it is plausible.
21
26
  - Run registry path:
22
27
  - Normalized summary path:
23
28
  - Iteration report path:
29
+
30
+ ## Change Links
31
+
32
+ - Spec path: `.superlab/changes/<change-id>/spec.md`
33
+ - Tasks path: `.superlab/changes/<change-id>/tasks.md`