superlab 0.1.27 → 0.1.28

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (50) hide show
  1. package/README.md +3 -0
  2. package/README.zh-CN.md +3 -0
  3. package/bin/superlab.cjs +11 -0
  4. package/lib/auto_contracts.cjs +1 -1
  5. package/lib/auto_runner.cjs +1 -1
  6. package/lib/context.cjs +30 -198
  7. package/lib/i18n.cjs +170 -18
  8. package/lib/lab_write_contract.json +8 -0
  9. package/package-assets/claude/commands/lab-idea.md +1 -1
  10. package/package-assets/claude/commands/lab-write.md +1 -1
  11. package/package-assets/claude/commands/lab.md +4 -4
  12. package/package-assets/codex/prompts/lab-idea.md +1 -1
  13. package/package-assets/codex/prompts/lab-write.md +1 -1
  14. package/package-assets/codex/prompts/lab.md +4 -4
  15. package/package-assets/shared/lab/.managed/scripts/validate_idea_artifact.py +147 -0
  16. package/package-assets/shared/lab/.managed/scripts/validate_manuscript_delivery.py +50 -4
  17. package/package-assets/shared/lab/.managed/scripts/validate_paper_claims.py +86 -0
  18. package/package-assets/shared/lab/.managed/scripts/validate_paper_plan.py +38 -3
  19. package/package-assets/shared/lab/.managed/scripts/validate_section_draft.py +181 -0
  20. package/package-assets/shared/lab/.managed/templates/idea.md +43 -0
  21. package/package-assets/shared/lab/.managed/templates/paper-plan.md +24 -0
  22. package/package-assets/shared/lab/config/workflow.json +2 -1
  23. package/package-assets/shared/lab/context/auto-mode.md +1 -1
  24. package/package-assets/shared/lab/context/next-action.md +4 -4
  25. package/package-assets/shared/lab/context/session-brief.md +8 -1
  26. package/package-assets/shared/lab/context/summary.md +14 -3
  27. package/package-assets/shared/skills/lab/SKILL.md +17 -16
  28. package/package-assets/shared/skills/lab/references/paper-writing/examples/abstract/template-b.md +2 -2
  29. package/package-assets/shared/skills/lab/references/paper-writing/examples/conclusion/conservative-claim-boundary.md +13 -13
  30. package/package-assets/shared/skills/lab/references/paper-writing/examples/experiments/main-results-and-ablation-latex.md +18 -17
  31. package/package-assets/shared/skills/lab/references/paper-writing/examples/experiments-examples.md +1 -1
  32. package/package-assets/shared/skills/lab/references/paper-writing/examples/index.md +1 -1
  33. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/pipeline-version-1-one-contribution-multi-advantages.md +3 -3
  34. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/pipeline-version-2-two-contributions.md +1 -1
  35. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/annotated-figure-to-text.md +66 -0
  36. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/example-of-the-three-elements.md +11 -11
  37. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/{module-design-instant-ngp.md → module-design-multiresolution-encoding.md} +1 -1
  38. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/{module-triad-neural-body.md → module-triad-anchored-representation.md} +4 -4
  39. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/overview-template.md +4 -4
  40. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/pre-writing-questions.md +4 -3
  41. package/package-assets/shared/skills/lab/references/paper-writing/examples/method-examples.md +4 -4
  42. package/package-assets/shared/skills/lab/references/paper-writing/examples/related-work/closest-prior-gap-template.md +12 -12
  43. package/package-assets/shared/skills/lab/references/paper-writing/examples/related-work/topic-comparison-template.md +2 -2
  44. package/package-assets/shared/skills/lab/stages/auto.md +6 -2
  45. package/package-assets/shared/skills/lab/stages/data.md +0 -1
  46. package/package-assets/shared/skills/lab/stages/framing.md +0 -1
  47. package/package-assets/shared/skills/lab/stages/idea.md +30 -13
  48. package/package-assets/shared/skills/lab/stages/write.md +24 -3
  49. package/package.json +1 -1
  50. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/neural-body-annotated-figure-text.md +0 -66
@@ -24,6 +24,13 @@
24
24
  - Limitation sources:
25
25
  - Claims that still need more evidence:
26
26
 
27
+ ## Asset Coverage Targets
28
+
29
+ - Core asset floor:
30
+ - Required coverage categories:
31
+ - Current planned core assets:
32
+ - Coverage risks or gaps:
33
+
27
34
  ## Table Plan
28
35
 
29
36
  - Main results table:
@@ -41,6 +48,12 @@
41
48
 
42
49
  ## Figure Plan
43
50
 
51
+ - Problem setting or teaser figure:
52
+ - Asset file:
53
+ - Section:
54
+ - Figure intent:
55
+ - Evidence:
56
+ - Status:
44
57
  - Method overview figure:
45
58
  - Asset file:
46
59
  - Section:
@@ -54,6 +67,16 @@
54
67
  - Evidence:
55
68
  - Status:
56
69
 
70
+ ## Analysis Asset Plan
71
+
72
+ - Analysis asset:
73
+ - Asset file:
74
+ - Asset type:
75
+ - Section:
76
+ - Asset intent:
77
+ - Evidence:
78
+ - Status:
79
+
57
80
  ## Citation Plan
58
81
 
59
82
  - Background anchor:
@@ -74,6 +97,7 @@
74
97
 
75
98
  ## Section-to-Asset Map
76
99
 
100
+ - Introduction:
77
101
  - Method:
78
102
  - Experiments:
79
103
  - Related Work:
@@ -7,5 +7,6 @@
7
7
  "deliverables_root": "docs/research",
8
8
  "paper_template_root": "",
9
9
  "paper_template_decision": "unconfirmed",
10
- "paper_template_final_reminder_acknowledged": false
10
+ "paper_template_final_reminder_acknowledged": false,
11
+ "paper_language_finalization_decision": "unconfirmed"
11
12
  }
@@ -68,4 +68,4 @@ If `eval-protocol.md` declares structured rung entries, auto mode follows those
68
68
 
69
69
  - Stop conditions:
70
70
  - Escalation conditions:
71
- - Canonical promotion writeback: update `.lab/context/data-decisions.md`, `.lab/context/decisions.md`, `.lab/context/state.md`, and `.lab/context/session-brief.md`.
71
+ - Canonical promotion writeback: update `.lab/context/data-decisions.md`, `.lab/context/decisions.md`, `.lab/context/state.md`, and `.lab/context/workflow-state.md`.
@@ -5,15 +5,15 @@
5
5
  - Action:
6
6
  - Success signal:
7
7
 
8
- ## If Success
8
+ ## After Completion
9
9
 
10
10
  - Next action:
11
11
 
12
- ## If Failure
12
+ ## If Blocked
13
13
 
14
14
  - Fallback action:
15
15
 
16
- ## Human Decision Needed
16
+ ## Escalation
17
17
 
18
18
  - Question:
19
- - Why it blocks progress:
19
+ - Escalate when:
@@ -3,6 +3,8 @@
3
3
  ## Active Stage
4
4
 
5
5
  - Stage:
6
+ - Current objective:
7
+ - Immediate next action:
6
8
 
7
9
  ## Mission
8
10
 
@@ -11,10 +13,15 @@ One sentence describing the active research mission.
11
13
  ## Best Current Path
12
14
 
13
15
  - Approved direction:
14
- - Why this is the active path:
16
+ - Strongest supported claim:
15
17
  - Auto mode:
16
18
  - Auto objective:
17
19
  - Auto decision:
20
+ - Collaborator report mode:
21
+ - Canonical context readiness:
22
+ - Method name:
23
+ - Primary metrics:
24
+ - Secondary metrics:
18
25
 
19
26
  ## Main Risk
20
27
 
@@ -1,12 +1,23 @@
1
1
  # Research Summary
2
2
 
3
3
  ## Current Direction
4
-
5
- Summarize the current approved research direction in 5-10 lines.
6
-
4
+ - Mission:
5
+ - Approved direction:
6
+ - Active stage:
7
+ - Current objective:
7
8
  - Auto mode:
8
9
  - Auto objective:
9
10
  - Auto decision:
11
+ - Collaborator report mode:
12
+ - Canonical context readiness:
13
+ - Method name:
14
+ - Contribution bullets:
15
+ - Eval objective:
16
+ - Primary metrics:
17
+ - Secondary metrics:
18
+ - Dataset package:
19
+ - Benchmark role:
20
+ - Comparison suite:
10
21
 
11
22
  ## Strongest Evidence
12
23
 
@@ -34,17 +34,22 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
34
34
  ### `/lab:idea`
35
35
 
36
36
  - Search relevant literature, baselines, datasets, and evaluation metrics before proposing a plan.
37
+ - Build a literature-scoping bundle before claiming novelty. The default target is 20 relevant sources unless the field is genuinely too narrow and that exception is written down.
37
38
  - Read `.lab/context/mission.md` and `.lab/context/open-questions.md` before drafting.
39
+ - Read `.lab/config/workflow.json` before drafting and follow its `workflow_language` for idea artifacts.
38
40
  - Ask one clarifying question at a time when critical ambiguity remains.
39
- - State the problem, the failure case, and why the problem matters before proposing solutions.
41
+ - State the scenario, the problem, the failure case, and why the problem matters before proposing solutions.
40
42
  - Classify the idea by contribution category and breakthrough level.
41
43
  - Compare against existing methods explicitly and state why the idea should be better.
44
+ - Include a closest-prior-work comparison and a plain-language description of how the proposed direction would work.
42
45
  - Distinguish sourced evidence from generated innovation claims.
43
46
  - End with three meaningful points that are clear, short, and easy to scan.
44
47
  - Produce 2-3 candidate approaches with trade-offs before recommending one.
45
48
  - Critique the idea before converging on it.
49
+ - Include a minimum viable experiment before approval.
46
50
  - Keep an explicit approval gate before `/lab:spec`.
47
51
  - Write idea artifacts with the template in `.lab/.managed/templates/idea.md`.
52
+ - Run `.lab/.managed/scripts/validate_idea_artifact.py --idea <idea-artifact> --workflow-config .lab/config/workflow.json` before treating the idea as converged.
48
53
  - Update `.lab/context/mission.md`, `.lab/context/decisions.md`, and `.lab/context/open-questions.md` after convergence.
49
54
  - Do not leave `.lab/context/mission.md` as a template shell once the problem statement and approved direction are known.
50
55
  - Do not implement code in this stage.
@@ -91,9 +96,9 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
91
96
  - Reuse `/lab:run`, `/lab:iterate`, `/lab:review`, `/lab:report`, and optional `/lab:write` instead of inventing a second workflow.
92
97
  - Do not automatically change the research mission, paper-facing framing, or core claims.
93
98
  - You may add exploratory datasets, benchmarks, and comparison methods inside the approved exploration envelope.
94
- - You may promote an exploratory addition to the primary package only after the promotion policy in `auto-mode.md` is satisfied and the promotion is written back into `.lab/context/data-decisions.md`, `.lab/context/decisions.md`, `.lab/context/state.md`, and `.lab/context/session-brief.md`.
99
+ - You may promote an exploratory addition to the primary package only after the promotion policy in `auto-mode.md` is satisfied and the promotion is written back into `.lab/context/data-decisions.md`, `.lab/context/decisions.md`, `.lab/context/state.md`, and `.lab/context/workflow-state.md`.
95
100
  - Poll long-running commands until they complete, time out, or hit a stop condition.
96
- - Update `.lab/context/auto-status.md`, `.lab/context/state.md`, `.lab/context/workflow-state.md`, `.lab/context/decisions.md`, `.lab/context/data-decisions.md`, `.lab/context/evidence-index.md`, and `.lab/context/session-brief.md` as the campaign advances.
101
+ - Update `.lab/context/auto-status.md`, `.lab/context/state.md`, `.lab/context/workflow-state.md`, `.lab/context/decisions.md`, `.lab/context/data-decisions.md`, and `.lab/context/evidence-index.md` as the campaign advances, then refresh the derived handoff files.
97
102
  - Keep an explicit approval gate when a proposed action would leave the frozen core defined by the auto-mode contract.
98
103
 
99
104
  ### `/lab:spec`
@@ -165,26 +170,21 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
165
170
 
166
171
  - Start only after `report` artifacts are stable enough to support paper claims.
167
172
  - Start only after an approved framing artifact exists at `.lab/writing/framing.md`.
168
- - Read `.lab/config/workflow.json` before drafting and enforce its `paper_language` and `paper_format`.
169
- - If `paper_template_root` is empty and `paper_template_decision` is `unconfirmed`, ask once whether to continue with the managed default scaffold or attach a template directory first; persist the answer before drafting `.tex`.
170
- - If the project is still on the default scaffold at a final export or final-draft boundary and `paper_template_final_reminder_acknowledged` is `false`, ask one final reminder question before finalizing.
173
+ - Read `.lab/config/workflow.json` before drafting and enforce its `workflow_language`, `paper_language`, and `paper_format`.
171
174
  - Read `.lab/context/mission.md`, `.lab/context/decisions.md`, `.lab/context/evidence-index.md`, and `.lab/context/data-decisions.md` before drafting.
172
175
  - Write one paper section or one explicit subproblem per round.
176
+ - Ordinary manuscript drafting rounds should follow `workflow_language`.
177
+ - If `workflow_language` and `paper_language` differ, the first final-draft or export round must ask once whether to keep the draft language or convert the final manuscript to `paper_language`, then persist that choice.
173
178
  - Bind each claim to evidence from `report`, iteration reports, or normalized summaries.
174
- - Write planning artifacts with `.lab/.managed/templates/paper-plan.md`, `.lab/.managed/templates/paper-section.md`, and `.lab/.managed/templates/write-iteration.md`.
175
- - Write final manuscript artifacts with `.lab/.managed/templates/paper.tex`, `.lab/.managed/templates/paper-section.tex`, `.lab/.managed/templates/paper-table.tex`, `.lab/.managed/templates/paper-figure.tex`, and `.lab/.managed/templates/paper-references.bib`.
176
- - Use the vendored paper-writing references under `skills/lab/references/paper-writing/`.
177
- - For any section with a bundled example bank, also use the vendored example-bank files under `skills/lab/references/paper-writing/examples/`.
178
- - Load only the current section guide, the matching examples index when one exists, 1-2 matching concrete example files, plus `paper-review.md` and `does-my-writing-flow-source.md`.
179
- - Build a compact mini-outline before prose.
180
- - Build the paper asset plan before prose when the section carries method or experiments claims.
179
+ - Use the write-stage contract in `.codex/skills/lab/stages/write.md` or `.claude/skills/lab/stages/write.md` as the single source of truth for template choice, paper-plan requirements, section-specific references, validator calls, asset coverage, and final manuscript gates.
180
+ - Use the vendored paper-writing references under `skills/lab/references/paper-writing/` and the matching example-bank files under `skills/lab/references/paper-writing/examples/`.
181
+ - Treat `.lab/writing/plan.md` as the write-time source of truth for tables, figures, citations, and asset coverage.
182
+ - Treat section-quality, claim-safety, and manuscript-delivery checks as the canonical acceptance gates for final-draft or export rounds.
181
183
  - For each subsection, explicitly cover motivation, design, and technical advantage when applicable.
182
184
  - Keep terminology stable across rounds and sections.
183
185
  - If a claim is not supported by evidence, weaken or remove it.
184
186
  - Treat tables, figures, citations, and bibliography as core manuscript content rather than optional polish.
185
187
  - Keep paper-facing LaTeX free of absolute local paths, rerun ids, shell transcripts, and internal workflow provenance.
186
- - Materialize real LaTeX tables and figure placeholders instead of leaving all evidence inside prose paragraphs.
187
- - Run `.lab/.managed/scripts/validate_manuscript_delivery.py --paper-dir <deliverables_root>/paper` before accepting a final-draft or export round.
188
188
  - Before finalizing a round, append and answer the five-dimension self-review checklist and revise unresolved items.
189
189
  - Apply paper-writing discipline without changing experimental truth.
190
190
  - If the evidence is insufficient, stop and route back to `review` or `iterate`.
@@ -199,7 +199,8 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
199
199
  - No unconstrained auto mode. Every `/lab:auto` campaign must declare allowed stages, stop conditions, and a promotion policy in `.lab/context/auto-mode.md`.
200
200
  - No auto start without an explicit autonomy level and `Approval status: approved`.
201
201
  - No final report without validated normalized results.
202
- - No paper-writing round without stable report artifacts, an approved framing artifact, evidence links, LaTeX manuscript output, and a passing manuscript-delivery validation for final-draft or export rounds.
202
+ - No paper-writing round without stable report artifacts, an approved framing artifact, evidence links, and LaTeX manuscript output.
203
+ - No final-draft or export round without passing section-quality, claim-safety, and manuscript-delivery validation.
203
204
 
204
205
  ## References
205
206
 
@@ -16,7 +16,7 @@
16
16
 
17
17
  % Introduce the technical contribution that implements the insight in one to two sentences (usually mention the technical term/name only, without describing every detailed step. The term should be easy to understand and should not create a jump in reading. This ability is very important for writing a good abstract.)
18
18
  %% Example 1: To do this, we first present a label-efficient depth estimation framework using the internal representations of diffusion models. At the sampling phase, we utilize two guidance techniques to self-condition the generated image using the estimated depth map, the first of which uses pseudo-labeling, and the subsequent one uses a depth-domain diffusion prior.
19
- %% Example 2: To this end, we propose Neural Body, a new human body representation which assumes that the learned neural representations at different frames share the same set of latent codes anchored to a deformable mesh
19
+ %% Example 2: To this end, we propose AnchorField, a structured representation in which different observations share the same set of latent codes anchored to a deformable support.
20
20
 
21
21
  % Introduce the benefits of technical novelty
22
22
  %% Example 2: so that the observations across frames can be naturally integrated. The deformable mesh also provides geometric guidance for the network to learn 3D representations more efficiently.
@@ -29,6 +29,6 @@
29
29
  1. `This paper addresses the challenge of novel view synthesis for a human performer from a very sparse set of camera views.`
30
30
  2. `... representation learning will be ill-posed if the views are highly sparse.`
31
31
  3. `To solve this ill-posed problem, our key idea is to integrate observations over video frames.`
32
- 4. `To this end, we propose Neural Body ...`
32
+ 4. `To this end, we propose AnchorField ...`
33
33
  5. `... observations across frames can be naturally integrated ... provides geometric guidance ...`
34
34
  6. `Experiments show [main result].`
@@ -6,22 +6,22 @@ boundary explicit.
6
6
  ```tex
7
7
  \section{Conclusion}
8
8
 
9
- This paper shows that adding a structured ranking backbone together with a
10
- post-hoc calibration stage improves uplift ranking under the frozen benchmark
11
- protocol. Across the three benchmark families used in this work, the full model
9
+ This paper shows that adding a structured intermediate module together with a
10
+ lightweight adjustment stage improves performance under a fixed evaluation
11
+ protocol. Across the benchmark families used in this work, the full model
12
12
  consistently matches or exceeds the strongest baselines and remains stronger
13
13
  than the key ablated variants. This makes the main claim narrower than a
14
- universal superiority claim but stronger than a single-dataset win.
14
+ universal superiority claim but stronger than a single-setting win.
15
15
 
16
- We do not claim that the current method solves uplift modeling in every domain
17
- or that every design choice helps equally on every benchmark. In particular, the
18
- calibration stage appears beneficial on some datasets and neutral on others,
19
- which means its value should be interpreted as setting-dependent rather than as
20
- a guaranteed gain. That boundary is consistent with recent benchmarking
21
- practice, which argues for claim discipline and protocol-specific interpretation
22
- rather than broad overgeneralization~\cite{carlini2019evaluating}.
16
+ We do not claim that the current method solves the broader problem in every
17
+ domain or that every design choice helps equally on every benchmark. In
18
+ particular, the adjustment stage appears beneficial in some settings and
19
+ neutral in others, which means its value should be interpreted as
20
+ setting-dependent rather than as a guaranteed gain. That boundary is consistent
21
+ with recent benchmarking practice, which argues for claim discipline and
22
+ protocol-specific interpretation rather than broad overgeneralization~\cite{carlini2019evaluating}.
23
23
 
24
24
  The most useful next step is to extend the evaluation to a broader set of
25
- benchmark slices and to test whether the same ranking-versus-calibration split
26
- remains useful when the label distribution shifts more aggressively.
25
+ benchmark slices and to test whether the same backbone-versus-adjustment split
26
+ remains useful when the data distribution shifts more aggressively.
27
27
  ```
@@ -1,6 +1,6 @@
1
1
  # Main Results and Ablation LaTeX Example
2
2
 
3
- This file is a complete paper-facing LaTeX example for the experiments section.
3
+ This file is a complete manuscript-ready LaTeX example for the experiments section.
4
4
  Reuse the structure, caption logic, and prose-to-table linkage. Replace the
5
5
  placeholder methods, metrics, and values with the current project's evidence.
6
6
 
@@ -14,13 +14,13 @@ Source inspiration:
14
14
 
15
15
  ```tex
16
16
  \begin{table}[t]
17
- \caption{Main benchmark results under the frozen evaluation protocol. Higher is better on all metrics.}
17
+ \caption{Main benchmark results under the fixed evaluation protocol. Higher is better on all metrics.}
18
18
  \label{tab:main-results}
19
19
  \centering
20
20
  \resizebox{0.92\linewidth}{!}{
21
21
  \begin{tabular}{lccc}
22
22
  \toprule
23
- Method & AUUC $\uparrow$ & Qini $\uparrow$ & Calibration Error $\downarrow$ \\
23
+ Method & Primary Metric $\uparrow$ & Secondary Metric $\uparrow$ & Error Metric $\downarrow$ \\
24
24
  \midrule
25
25
  Strongest baseline & 0.1421 & 0.0873 & 0.0612 \\
26
26
  Closest prior work & 0.1488 & 0.0915 & 0.0544 \\
@@ -32,7 +32,7 @@ Ours & \textbf{0.1564} & \textbf{0.0987} & \textbf{0.0418} \\
32
32
  ```
33
33
 
34
34
  Table message:
35
- - `Does the proposed method beat the strongest baselines under the frozen protocol?`
35
+ - `Does the proposed method beat the strongest baselines under the fixed evaluation protocol?`
36
36
 
37
37
  ## Ablation Table
38
38
 
@@ -47,7 +47,7 @@ Variant & AUUC $\uparrow$ \\
47
47
  \midrule
48
48
  Ours & \textbf{0.1564} \\
49
49
  w/o structure module & 0.1497 \\
50
- w/o calibration stage & 0.1510 \\
50
+ w/o final adjustment stage & 0.1510 \\
51
51
  w/ shuffled auxiliary signal & 0.1458 \\
52
52
  \bottomrule
53
53
  \end{tabular}
@@ -65,19 +65,20 @@ Table message:
65
65
 
66
66
  Table~\ref{tab:main-results} answers the main ranking question: whether the full
67
67
  method remains stronger than the closest prior work and the strongest practical
68
- baseline under the frozen protocol. Our method achieves the best AUUC and Qini
69
- while also reducing calibration error, which means the gain is not coming from
70
- trading ranking quality against stability.
68
+ baseline under the fixed evaluation protocol. Our method achieves the best
69
+ primary and secondary metrics while also reducing the error metric, which means
70
+ the gain is not coming from trading one objective against stability.
71
71
 
72
72
  Table~\ref{tab:ablations} then asks a narrower mechanism question. Removing the
73
- structure module causes the largest drop, so the main gain is tied to structured
74
- signal modeling rather than to a generic increase in capacity. Removing the
75
- calibration stage leads to a smaller but still visible drop, which supports the
76
- claim that calibration improves final ranking quality without being the sole
77
- driver of the result. The shuffled-signal variant acts as a negative control and
78
- shows that the gain does not survive when the auxiliary information is broken.
73
+ structure module causes the largest drop, so the main gain is tied to explicit
74
+ structure modeling rather than to a generic increase in capacity. Removing the
75
+ final adjustment stage leads to a smaller but still visible drop, which
76
+ supports the claim that the adjustment helps the exposed prediction without
77
+ being the sole driver of the result. The shuffled-signal variant acts as a
78
+ negative control and shows that the gain does not survive when the auxiliary
79
+ information is broken.
79
80
 
80
- One caveat is that the calibration gain may remain neutral on some benchmarks,
81
- so the paper should not overclaim that every component helps equally on every
82
- dataset.
81
+ One caveat is that the final adjustment gain may remain neutral in some
82
+ settings, so the paper should not overclaim that every component helps equally
83
+ on every dataset.
83
84
  ```
@@ -1,6 +1,6 @@
1
1
  # Experiments Example Patterns
2
2
 
3
- Use these examples when turning validated results into paper-facing LaTeX assets.
3
+ Use these examples when turning validated results into manuscript-ready LaTeX assets.
4
4
  The referenced files contain complete LaTeX environments and section-level prose
5
5
  glue, not just checklists.
6
6
 
@@ -12,7 +12,7 @@ locally organized cite targets.
12
12
  5. Introduction technical-challenge files: `references/examples/introduction/technical-challenge-version-1-existing-task.md`, `references/examples/introduction/technical-challenge-version-2-existing-task-insight-backed-by-traditional.md`, `references/examples/introduction/technical-challenge-version-3-novel-task.md`, `references/examples/introduction/novel-task-challenge-decomposition.md`
13
13
  6. Introduction pipeline files: `references/examples/introduction/pipeline-version-1-one-contribution-multi-advantages.md`, `references/examples/introduction/pipeline-version-2-two-contributions.md`, `references/examples/introduction/pipeline-version-3-new-module-on-existing-pipeline.md`, `references/examples/introduction/pipeline-version-4-observation-driven.md`, `references/examples/introduction/pipeline-not-recommended-abstract-only.md`
14
14
  7. Method examples index: `references/examples/method-examples.md`
15
- 8. Method detail files: `references/examples/method/pre-writing-questions.md`, `references/examples/method/module-triad-neural-body.md`, `references/examples/method/neural-body-annotated-figure-text.md`, `references/examples/method/module-design-instant-ngp.md`, `references/examples/method/module-motivation-patterns.md`, `references/examples/method/section-skeleton.md`, `references/examples/method/overview-template.md`, `references/examples/method/example-of-the-three-elements.md`, `references/examples/method/method-writing-common-issues-note.md`
15
+ 8. Method detail files: `references/examples/method/pre-writing-questions.md`, `references/examples/method/module-triad-anchored-representation.md`, `references/examples/method/annotated-figure-to-text.md`, `references/examples/method/module-design-multiresolution-encoding.md`, `references/examples/method/module-motivation-patterns.md`, `references/examples/method/section-skeleton.md`, `references/examples/method/overview-template.md`, `references/examples/method/example-of-the-three-elements.md`, `references/examples/method/method-writing-common-issues-note.md`
16
16
  9. Related work examples index: `references/examples/related-work-examples.md`
17
17
  10. Related work files: `references/examples/related-work/closest-prior-gap-template.md`, `references/examples/related-work/topic-comparison-template.md`
18
18
  11. Experiments examples index: `references/examples/experiments-examples.md`
@@ -5,7 +5,7 @@
5
5
 
6
6
  ```latex
7
7
  % In this paper, we propose a novel framework …
8
- %% Example: In this paper, we introduce a novel implicit neural representation for dynamic humans, named Neural Body, to solve the challenge of novel view synthesis from sparse views.
8
+ %% Example: In this paper, we introduce a structured anchor-based representation, named AnchorField, to solve the challenge of reconstructing a changing target from sparse observations.
9
9
  In this paper, we propose a novel framework/representation, named [method name] for [xxx task].
10
10
 
11
11
  % Teaser for basic idea
@@ -13,11 +13,11 @@ In this paper, we propose a novel framework/representation, named [method name]
13
13
  The basic idea is illustrated in [xxx Figure].
14
14
 
15
15
  % One-sentence key novelty/contribution (very important ability)
16
- %% Example: For the implicit fields at different frames, instead of learning them separately, Neural Body generates them from the same set of latent codes.
16
+ %% Example: Instead of learning each observation-specific field separately, AnchorField generates them from the same set of anchor codes.
17
17
  Our innovation is in [one sentence for key novelty].
18
18
 
19
19
  % Method details
20
- %% Example: Specifically, we anchor a set of latent codes to the vertices of a deformable human model (SMPL in this work), namely that their spatial locations vary with the human pose. To obtain the 3D representation at a frame, we first transform the code locations based on the human pose, which can be reliably estimated from sparse camera views. Then, a network is designed to regress the density and color for any 3D point based on these latent codes. Both the latent codes and the network are jointly learned from images of all video frames during the reconstruction.
20
+ %% Example: Specifically, we anchor a set of latent codes to a reference support structure so that their spatial locations follow the current observation. To obtain the target representation for one condition, we first transform the code locations using the available alignment signal. Then, a prediction network estimates the required outputs at each query point from these codes. Both the anchor codes and the prediction network are jointly learned from all observations.
21
21
  Specifically, [how it works in detail].
22
22
 
23
23
  % Advantage 1
@@ -5,7 +5,7 @@
5
5
 
6
6
  ```latex
7
7
  % In this paper, we propose a novel framework …
8
- %% Example: In this paper, we introduce a novel implicit neural representation for dynamic humans, named Neural Body, to solve the challenge of novel view synthesis from sparse views.
8
+ %% Example: In this paper, we introduce a structured anchor-based representation, named AnchorField, to solve the challenge of reconstructing a changing target from sparse observations.
9
9
  In this paper, we propose a novel framework/representation, named [method name] for [xxx task].
10
10
 
11
11
  % One-sentence key novelty
@@ -0,0 +1,66 @@
1
+ # Annotated Figure (Text Conversion)
2
+
3
+ This file converts an annotated method figure into reusable writing notes.
4
+
5
+ ## Purpose
6
+
7
+ Use this mapping to understand how one Method section can explicitly separate:
8
+
9
+ 1. Module motivation
10
+ 2. Module design (data structure)
11
+ 3. Module design (forward process)
12
+ 4. Technical advantages
13
+
14
+ ## Block-by-Block Mapping
15
+
16
+ ### Section 3.1: Structured Anchor Codes
17
+
18
+ 1. **Module design (data structure)**
19
+ - The paragraph defines structured codes anchored to a reference support structure.
20
+ - It explains what is constructed: latent codes, anchor positions, and the condition-dependent transformation that keeps them aligned with the current instance.
21
+
22
+ 2. **Technical advantages**
23
+ - The paragraph explains why this design works better: it preserves correspondence across observations and keeps the representation aligned with changing geometry.
24
+ - It highlights why anchoring codes to a structured support is beneficial.
25
+
26
+ ### Section 3.2: Code Diffusion
27
+
28
+ 1. **Motivation of this module**
29
+ - The paragraph states the remaining problem: direct interpolation of sparse anchor codes leaves many query locations weakly represented.
30
+ - This motivates propagation from anchor locations to nearby continuous space.
31
+
32
+ 2. **Module design (forward process)**
33
+ - The paragraph explains the execution pipeline: build sparse feature volumes, run sparse convolutions, interpolate codes at query points, and feed those codes to prediction networks.
34
+ - This is a canonical input -> steps -> output module description.
35
+
36
+ ### Section 3.3: Field Regression
37
+
38
+ 1. **Module design (forward process) for the first prediction head**
39
+ - The first paragraph defines how one target quantity is regressed from the latent code and condition inputs.
40
+
41
+ 2. **Module design (data structure) for the second prediction head**
42
+ - The next paragraph introduces required inputs or embeddings such as latent code, query direction, spatial location, or condition tokens.
43
+
44
+ 3. **Module design (forward process) for the second prediction head**
45
+ - The following paragraph describes how those inputs are encoded and passed into the second prediction network for the final estimate.
46
+
47
+ ### Section 3.4: Output Aggregation
48
+
49
+ 1. **Module design (forward process)**
50
+ - The paragraph describes the final aggregation process that converts pointwise predictions into the output seen by the reader or evaluator.
51
+
52
+ ## Reusable Writing Pattern from This Figure
53
+
54
+ For each module subsection, follow this order:
55
+
56
+ 1. `Motivation`: state unresolved challenge and technical reason.
57
+ 2. `Design-1`: define structure/representation/network.
58
+ 3. `Design-2`: describe forward process in execution order.
59
+ 4. `Advantage`: explain why this module improves over alternatives.
60
+
61
+ ## Suggested Paragraph Starters
62
+
63
+ 1. Motivation: `A remaining challenge is ...`
64
+ 2. Data structure design: `We represent ... with ...`
65
+ 3. Forward process: `Given [input], we first ... then ... finally ...`
66
+ 4. Technical advantage: `Compared with previous methods, this design ... because ...`
@@ -18,27 +18,27 @@ Each `% ...` annotation explains the paragraph(s) immediately below it.
18
18
  \subsection{3.1. Structured latent codes}
19
19
 
20
20
  % Module design: introduce the module's data structure
21
- To control the spatial locations of latent codes with the human pose, we anchor these latent codes to a deformable human body model (SMPL) [38]. SMPL is a skinned vertex-based model, which is defined as a function of shape parameters, pose parameters, and a rigid transformation relative to the SMPL coordinate system. The function outputs a posed 3D mesh with 6890 vertices. Specifically, we define a set of latent codes \( Z = \{z_1, z_2, ..., z_{6890}\} \) on vertices of the SMPL model. For the frame \( t \), SMPL parameters \( S_t \) are estimated from the multi-view images \( \{I_t^c \mid c = 1, ..., N_c\} \) using [26]. The spatial locations of the latent codes are then transformed based on the human pose \( S_t \) for the density and color regression. Figure 3 shows an example. The dimension of latent code \( z \) is set to 16 in our experiments.
21
+ To keep latent codes aligned with the current instance, we anchor them to a deformable reference structure [38]. This structure is parameterized by shape variables, condition variables, and a rigid transformation relative to its own coordinate system, and outputs a posed mesh or support geometry. Specifically, we define a set of latent codes \( Z = \{z_1, z_2, ..., z_M\} \) on the support points of that reference structure. For one observation \( t \), the condition parameters \( S_t \) are estimated from the available inputs \( \{I_t^c \mid c = 1, ..., N_c\} \) using an external alignment procedure [26]. The spatial locations of the latent codes are then transformed according to \( S_t \) before the downstream prediction heads consume them. Figure 3 shows an example. The latent-code dimension is set to 16 in this example.
22
22
 
23
23
  % Technical advantages of this module
24
- Similar to the local implicit representations [25, 5, 18], the latent codes are used with a neural network to represent the local geometry and appearance of a human. Anchoring these codes to a deformable model enables us to represent a dynamic human. With the dynamic human representation, we establish a latent variable model that maps the same set of latent codes to the implicit fields of density and color at different frames, which naturally integrates observations at different frames.
24
+ Similar to local implicit representations [25, 5, 18], the latent codes are used with a neural network to represent local structure and attributes. Anchoring these codes to a deformable support lets the model stay aligned with changing geometry instead of relearning a separate representation for every condition. This yields a latent-variable view in which the same anchor codes can explain multiple observations while sharing information across them.
25
25
 
26
26
  \subsection{3.2. Code diffusion}
27
27
 
28
28
  % Motivation of this module
29
- Figure 3(a) shows the process of code diffusion. The implicit fields assign the density and color to each point in the 3D space, which requires us to query the latent codes at continuous 3D locations. This can be achieved with the trilinear interpolation. However, since the structured latent codes are relatively sparse in the 3D space, directly interpolating the latent codes leads to zero vectors at most 3D points. To solve this problem, we diffuse the latent codes defined on the surface to nearby 3D space.
29
+ Figure 3(a) shows the process of code propagation. The target fields assign outputs to each point in the domain, which requires querying the latent codes at continuous locations. This can be achieved with trilinear interpolation. However, because the structured anchor codes are sparse, directly interpolating them leaves many query points weakly represented. To solve this problem, we propagate the codes defined on the support structure to nearby space.
30
30
 
31
31
  % Module design: introduce module design by describing the module forward process
32
- Inspired by [65, 56, 49], we choose the SparseConvNet [21] to efficiently process the structured latent codes, whose architecture is described in Table 1. Specifically, based on the SMPL parameters, we compute the 3D bounding box of the human and divide the box into small voxels with voxel size of \( 5mm \times 5mm \times 5mm \). The latent code of a non-empty voxel is the mean of latent codes of SMPL vertices inside this voxel. SparseConvNet utilizes 3D sparse convolutions to process the input volume and output latent code volumes with \( 2\times, 4\times, 8\times, 16\times \) downsampled sizes. With the convolution and downsampling, the input codes are diffused to nearby space. Following [56], for any point in 3D space, we interpolate the latent codes from multi-scale code volumes of network layers 5, 9, 13, 17, and concatenate them into the final latent code. Since the code diffusion should not be affected by the human position and orientation in the world coordinate system, we transform the code locations to the SMPL coordinate system.
32
+ Inspired by [65, 56, 49], we choose a sparse convolutional backbone [21] to efficiently process the structured latent codes, whose architecture is described in Table 1. Specifically, based on the reference-structure parameters, we compute a 3D bounding box around the active support and divide that box into small voxels with voxel size \( 5mm \times 5mm \times 5mm \). The latent code of a non-empty voxel is the mean of the anchor codes that fall inside it. The sparse backbone applies 3D sparse convolutions to this input volume and outputs latent code volumes at \( 2\times, 4\times, 8\times, 16\times \) downsampled scales. With convolution and downsampling, the input codes are propagated to nearby space. Following [56], for any point in the domain, we interpolate the latent codes from multi-scale code volumes of network layers 5, 9, 13, and 17, and concatenate them into the final latent code. Since propagation should not be affected by global position and orientation, we transform code locations to the reference coordinate system.
33
33
 
34
- For any point \( \mathbf{x} \) in 3D space, we query its latent code from the latent code volume. Specifically, the point \( \mathbf{x} \) is first transformed to the SMPL coordinate system, which aligns the point and the latent code volume in 3D space. Then, the latent code is computed using the trilinear interpolation. For the SMPL parameters \( S_t \), we denote the latent code at point \( \mathbf{x} \) as \( \psi(\mathbf{x}, Z, S_t) \). The code vector is passed into MLP networks to predict the density and color for point \( \mathbf{x} \).
34
+ For any query point \( \mathbf{x} \), we obtain its latent code from the propagated code volume. Specifically, the point \( \mathbf{x} \) is first transformed to the reference coordinate system so that the query and code volume are aligned. Then, the latent code is computed with trilinear interpolation. For condition parameters \( S_t \), we denote the latent code at point \( \mathbf{x} \) by \( \psi(\mathbf{x}, Z, S_t) \). The resulting code vector is passed to prediction networks that estimate the target outputs at \( \mathbf{x} \).
35
35
 
36
36
  \subsection{3.3. Density and color regression}
37
37
 
38
- Figure 3(b) overviews the regression of density and color for any point in 3D space. The density and color fields are represented by MLP networks. Details of network architectures are described in the supplementary material.
38
+ Figure 3(b) overviews the regression of the target outputs for any query point. The corresponding fields are represented by MLP networks. Details of the network architectures are described in the supplementary material.
39
39
 
40
40
  % Module design: introduce module design by describing the module forward process
41
- \textbf{Density model.} For the frame \( t \), the volume density at point \( \mathbf{x} \) is predicted as a function of only the latent code \( \psi(\mathbf{x}, Z, S_t) \), which is defined as:
41
+ \textbf{First prediction head.} For condition \( t \), the first target quantity at point \( \mathbf{x} \) is predicted as a function of only the latent code \( \psi(\mathbf{x}, Z, S_t) \), which is defined as:
42
42
 
43
43
  \[
44
44
  \sigma_t(\mathbf{x}) = M_{\sigma}(\psi(\mathbf{x}, Z, S_t)),
@@ -48,20 +48,20 @@ Figure 3(b) overviews the regression of density and color for any point in 3D sp
48
48
  where \( M_{\sigma} \) represents an MLP network with four layers.
49
49
 
50
50
  % Module design: introduce the module's data structure
51
- \textbf{Color model.} Similar to [37, 44], we take both the latent code \( \psi(\mathbf{x}, Z, S_t) \) and the viewing direction \( \mathbf{d} \) as input for the color regression. To model the location-dependent incident light, the color model also takes the spatial location \( \mathbf{x} \) as input. We observe that temporally-varying factors affect the human appearance, such as secondary lighting and self-shadowing. Inspired by the auto-decoder [48], we assign a latent embedding \( \ell_t \) for each video frame \( t \) to encode the temporally-varying factors.
51
+ \textbf{Second prediction head.} Similar to [37, 44], we take both the latent code \( \psi(\mathbf{x}, Z, S_t) \) and an observation-dependent direction \( \mathbf{d} \) as input for the second target. To model location-dependent effects, this head also takes the spatial location \( \mathbf{x} \) as input. We further observe that condition-specific factors can change the target output. Inspired by the auto-decoder [48], we assign a latent embedding \( \ell_t \) for each condition \( t \) to encode those varying factors.
52
52
 
53
53
  % Module design: introduce module design by describing the module forward process
54
- Specifically, for the frame \( t \), the color at \( \mathbf{x} \) is predicted as a function of the latent code \( \psi(\mathbf{x}, Z, S_t) \), the viewing direction \( \mathbf{d} \), the spatial location \( \mathbf{x} \), and the latent embedding \( \ell_t \). Following [51, 44], we apply the positional encoding to both the viewing direction \( \mathbf{d} \) and the spatial location \( \mathbf{x} \), which enables better learning of high frequency functions. The color model at frame \( t \) is defined as:
54
+ Specifically, for condition \( t \), the second target at \( \mathbf{x} \) is predicted as a function of the latent code \( \psi(\mathbf{x}, Z, S_t) \), the observation direction \( \mathbf{d} \), the spatial location \( \mathbf{x} \), and the latent embedding \( \ell_t \). Following [51, 44], we apply positional encoding to both \( \mathbf{d} \) and \( \mathbf{x} \), which enables better learning of high-frequency functions. The second prediction head at condition \( t \) is defined as:
55
55
 
56
56
  \[
57
57
  c_t(\mathbf{x}) = M_c(\psi(\mathbf{x}, Z, S_t), \gamma_d(\mathbf{d}), \gamma_x(\mathbf{x}), \ell_t),
58
58
  \tag{2}
59
59
  \]
60
60
 
61
- where \( M_c \) represents an MLP network with two layers, and \( \gamma_d \) and \( \gamma_x \) are positional encoding functions for viewing direction and spatial location, respectively. We set the dimension of \( \ell_t \) to 128 in experiments.
61
+ where \( M_c \) represents an MLP network with two layers, and \( \gamma_d \) and \( \gamma_x \) are positional encoding functions for direction and spatial location, respectively. We set the dimension of \( \ell_t \) to 128 in this example.
62
62
 
63
63
  \subsection{3.4. Volume rendering}
64
64
 
65
65
  % Module design: introduce module design by describing the module forward process
66
- Given a viewpoint, we utilize the classical volume rendering techniques to render the Neural Body into a 2D image. The pixel colors are estimated via the volume rendering integral equation [27] that accumulates volume densities and colors along the corresponding camera ray. In practice, the integral is approximated using numerical quadrature [41, 44]. Given a pixel, we first compute its camera ray \( \mathbf{r} \) using the camera parameters and sample \( N_k \) points \( \{\mathbf{x}_k\}_{k=1}^{N_k} \) along camera ray \( \mathbf{r} \) between near and far bounds. The scene bounds are estimated based on the SMPL model. Then, Neural Body predicts volume densities and colors at these points. For the video frame \( t \), the rendered color \( \hat{C}_t(\mathbf{r}) \) ...
66
+ Given an observation direction, we utilize a standard aggregation procedure to render the predicted field into a 2D output. Pixel values are estimated via a volume-integration equation [27] that accumulates the predicted quantities along the corresponding ray. In practice, the integral is approximated using numerical quadrature [41, 44]. Given a pixel, we first compute its ray \( \mathbf{r} \) using the camera parameters and sample \( N_k \) points \( \{\mathbf{x}_k\}_{k=1}^{N_k} \) along that ray between near and far bounds. The scene bounds are estimated from the reference structure. Then, the model predicts the target outputs at these points. For condition \( t \), the rendered value \( \hat{C}_t(\mathbf{r}) \) ...
67
67
  ```
@@ -1,4 +1,4 @@
1
- # Module Design Example
1
+ # Module Design Example (Multiresolution Encoding)
2
2
 
3
3
  This example uses `%` comments as annotations.
4
4
  Each `% ...` annotation explains the paragraph(s) immediately below it.
@@ -1,11 +1,11 @@
1
- # Module Triad Example (Neural Body)
1
+ # Module Triad Example (Anchored Representation)
2
2
 
3
- `Use Neural Body to understand the three elements of a module: design, motivation, and technical advantages.`
3
+ `Use this anchored-representation example to understand the three elements of a module: design, motivation, and technical advantages.`
4
4
 
5
5
  Local source references:
6
6
 
7
7
  1. Annotated figure showing motivation/design/advantages split.
8
- 3. Text-converted annotation notes: `references/examples/method/neural-body-annotated-figure-text.md`
8
+ 3. Text-converted annotation notes: `references/examples/method/annotated-figure-to-text.md`
9
9
 
10
10
  Triad mapping template:
11
11
 
@@ -15,5 +15,5 @@ Triad mapping template:
15
15
 
16
16
  Direct usage:
17
17
 
18
- 1. Read `neural-body-annotated-figure-text.md` to map each paragraph to one triad element.
18
+ 1. Read `annotated-figure-to-text.md` to map each paragraph to one triad element.
19
19
  2. Rebuild your own Method subsection with the same triad order.
@@ -18,12 +18,12 @@
18
18
  %% Example: The overview of the proposed model is illustrated in Figure 3.
19
19
 
20
20
  % Explain what Section 3.1 covers
21
- %% Example 1: Neural Body starts from a set of structured latent codes attached to the surface of a deformable human model (Section 3.1).
22
- %% Example 2: In this section, we first describe how to model 3D scenes with MLP maps (Section 3.1).
21
+ %% Example 1: The model starts from a set of structured anchor codes attached to a reference support structure (Section 3.1).
22
+ %% Example 2: In this section, we first describe how to model the target space with learnable field maps (Section 3.1).
23
23
 
24
24
  % Explain what Section 3.2 covers
25
- %% Example 1: The latent code at any location around the surface can be obtained with a code diffusion process (Section 3.2) and then decoded to density and color values by neural networks (Section 3.3).
26
- %% Example 2: Then, Section 3.2 discusses how to represent volumetric videos with dynamic MLP maps.
25
+ %% Example 1: The code at any nearby location is obtained with a propagation process (Section 3.2) and then decoded into the target outputs by prediction networks (Section 3.3).
26
+ %% Example 2: Then, Section 3.2 discusses how to represent changing scenes with dynamic field maps.
27
27
 
28
28
  % Explain what Section 3.3 covers
29
29
  %% Example 3: Finally, we introduce some strategies to speed up the rendering process (Section 3.3).