superlab 0.1.27 → 0.1.28
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +3 -0
- package/README.zh-CN.md +3 -0
- package/bin/superlab.cjs +11 -0
- package/lib/auto_contracts.cjs +1 -1
- package/lib/auto_runner.cjs +1 -1
- package/lib/context.cjs +30 -198
- package/lib/i18n.cjs +170 -18
- package/lib/lab_write_contract.json +8 -0
- package/package-assets/claude/commands/lab-idea.md +1 -1
- package/package-assets/claude/commands/lab-write.md +1 -1
- package/package-assets/claude/commands/lab.md +4 -4
- package/package-assets/codex/prompts/lab-idea.md +1 -1
- package/package-assets/codex/prompts/lab-write.md +1 -1
- package/package-assets/codex/prompts/lab.md +4 -4
- package/package-assets/shared/lab/.managed/scripts/validate_idea_artifact.py +147 -0
- package/package-assets/shared/lab/.managed/scripts/validate_manuscript_delivery.py +50 -4
- package/package-assets/shared/lab/.managed/scripts/validate_paper_claims.py +86 -0
- package/package-assets/shared/lab/.managed/scripts/validate_paper_plan.py +38 -3
- package/package-assets/shared/lab/.managed/scripts/validate_section_draft.py +181 -0
- package/package-assets/shared/lab/.managed/templates/idea.md +43 -0
- package/package-assets/shared/lab/.managed/templates/paper-plan.md +24 -0
- package/package-assets/shared/lab/config/workflow.json +2 -1
- package/package-assets/shared/lab/context/auto-mode.md +1 -1
- package/package-assets/shared/lab/context/next-action.md +4 -4
- package/package-assets/shared/lab/context/session-brief.md +8 -1
- package/package-assets/shared/lab/context/summary.md +14 -3
- package/package-assets/shared/skills/lab/SKILL.md +17 -16
- package/package-assets/shared/skills/lab/references/paper-writing/examples/abstract/template-b.md +2 -2
- package/package-assets/shared/skills/lab/references/paper-writing/examples/conclusion/conservative-claim-boundary.md +13 -13
- package/package-assets/shared/skills/lab/references/paper-writing/examples/experiments/main-results-and-ablation-latex.md +18 -17
- package/package-assets/shared/skills/lab/references/paper-writing/examples/experiments-examples.md +1 -1
- package/package-assets/shared/skills/lab/references/paper-writing/examples/index.md +1 -1
- package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/pipeline-version-1-one-contribution-multi-advantages.md +3 -3
- package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/pipeline-version-2-two-contributions.md +1 -1
- package/package-assets/shared/skills/lab/references/paper-writing/examples/method/annotated-figure-to-text.md +66 -0
- package/package-assets/shared/skills/lab/references/paper-writing/examples/method/example-of-the-three-elements.md +11 -11
- package/package-assets/shared/skills/lab/references/paper-writing/examples/method/{module-design-instant-ngp.md → module-design-multiresolution-encoding.md} +1 -1
- package/package-assets/shared/skills/lab/references/paper-writing/examples/method/{module-triad-neural-body.md → module-triad-anchored-representation.md} +4 -4
- package/package-assets/shared/skills/lab/references/paper-writing/examples/method/overview-template.md +4 -4
- package/package-assets/shared/skills/lab/references/paper-writing/examples/method/pre-writing-questions.md +4 -3
- package/package-assets/shared/skills/lab/references/paper-writing/examples/method-examples.md +4 -4
- package/package-assets/shared/skills/lab/references/paper-writing/examples/related-work/closest-prior-gap-template.md +12 -12
- package/package-assets/shared/skills/lab/references/paper-writing/examples/related-work/topic-comparison-template.md +2 -2
- package/package-assets/shared/skills/lab/stages/auto.md +6 -2
- package/package-assets/shared/skills/lab/stages/data.md +0 -1
- package/package-assets/shared/skills/lab/stages/framing.md +0 -1
- package/package-assets/shared/skills/lab/stages/idea.md +30 -13
- package/package-assets/shared/skills/lab/stages/write.md +24 -3
- package/package.json +1 -1
- package/package-assets/shared/skills/lab/references/paper-writing/examples/method/neural-body-annotated-figure-text.md +0 -66
|
@@ -24,6 +24,13 @@
|
|
|
24
24
|
- Limitation sources:
|
|
25
25
|
- Claims that still need more evidence:
|
|
26
26
|
|
|
27
|
+
## Asset Coverage Targets
|
|
28
|
+
|
|
29
|
+
- Core asset floor:
|
|
30
|
+
- Required coverage categories:
|
|
31
|
+
- Current planned core assets:
|
|
32
|
+
- Coverage risks or gaps:
|
|
33
|
+
|
|
27
34
|
## Table Plan
|
|
28
35
|
|
|
29
36
|
- Main results table:
|
|
@@ -41,6 +48,12 @@
|
|
|
41
48
|
|
|
42
49
|
## Figure Plan
|
|
43
50
|
|
|
51
|
+
- Problem setting or teaser figure:
|
|
52
|
+
- Asset file:
|
|
53
|
+
- Section:
|
|
54
|
+
- Figure intent:
|
|
55
|
+
- Evidence:
|
|
56
|
+
- Status:
|
|
44
57
|
- Method overview figure:
|
|
45
58
|
- Asset file:
|
|
46
59
|
- Section:
|
|
@@ -54,6 +67,16 @@
|
|
|
54
67
|
- Evidence:
|
|
55
68
|
- Status:
|
|
56
69
|
|
|
70
|
+
## Analysis Asset Plan
|
|
71
|
+
|
|
72
|
+
- Analysis asset:
|
|
73
|
+
- Asset file:
|
|
74
|
+
- Asset type:
|
|
75
|
+
- Section:
|
|
76
|
+
- Asset intent:
|
|
77
|
+
- Evidence:
|
|
78
|
+
- Status:
|
|
79
|
+
|
|
57
80
|
## Citation Plan
|
|
58
81
|
|
|
59
82
|
- Background anchor:
|
|
@@ -74,6 +97,7 @@
|
|
|
74
97
|
|
|
75
98
|
## Section-to-Asset Map
|
|
76
99
|
|
|
100
|
+
- Introduction:
|
|
77
101
|
- Method:
|
|
78
102
|
- Experiments:
|
|
79
103
|
- Related Work:
|
|
@@ -7,5 +7,6 @@
|
|
|
7
7
|
"deliverables_root": "docs/research",
|
|
8
8
|
"paper_template_root": "",
|
|
9
9
|
"paper_template_decision": "unconfirmed",
|
|
10
|
-
"paper_template_final_reminder_acknowledged": false
|
|
10
|
+
"paper_template_final_reminder_acknowledged": false,
|
|
11
|
+
"paper_language_finalization_decision": "unconfirmed"
|
|
11
12
|
}
|
|
@@ -68,4 +68,4 @@ If `eval-protocol.md` declares structured rung entries, auto mode follows those
|
|
|
68
68
|
|
|
69
69
|
- Stop conditions:
|
|
70
70
|
- Escalation conditions:
|
|
71
|
-
- Canonical promotion writeback: update `.lab/context/data-decisions.md`, `.lab/context/decisions.md`, `.lab/context/state.md`, and `.lab/context/
|
|
71
|
+
- Canonical promotion writeback: update `.lab/context/data-decisions.md`, `.lab/context/decisions.md`, `.lab/context/state.md`, and `.lab/context/workflow-state.md`.
|
|
@@ -5,15 +5,15 @@
|
|
|
5
5
|
- Action:
|
|
6
6
|
- Success signal:
|
|
7
7
|
|
|
8
|
-
##
|
|
8
|
+
## After Completion
|
|
9
9
|
|
|
10
10
|
- Next action:
|
|
11
11
|
|
|
12
|
-
## If
|
|
12
|
+
## If Blocked
|
|
13
13
|
|
|
14
14
|
- Fallback action:
|
|
15
15
|
|
|
16
|
-
##
|
|
16
|
+
## Escalation
|
|
17
17
|
|
|
18
18
|
- Question:
|
|
19
|
-
-
|
|
19
|
+
- Escalate when:
|
|
@@ -3,6 +3,8 @@
|
|
|
3
3
|
## Active Stage
|
|
4
4
|
|
|
5
5
|
- Stage:
|
|
6
|
+
- Current objective:
|
|
7
|
+
- Immediate next action:
|
|
6
8
|
|
|
7
9
|
## Mission
|
|
8
10
|
|
|
@@ -11,10 +13,15 @@ One sentence describing the active research mission.
|
|
|
11
13
|
## Best Current Path
|
|
12
14
|
|
|
13
15
|
- Approved direction:
|
|
14
|
-
-
|
|
16
|
+
- Strongest supported claim:
|
|
15
17
|
- Auto mode:
|
|
16
18
|
- Auto objective:
|
|
17
19
|
- Auto decision:
|
|
20
|
+
- Collaborator report mode:
|
|
21
|
+
- Canonical context readiness:
|
|
22
|
+
- Method name:
|
|
23
|
+
- Primary metrics:
|
|
24
|
+
- Secondary metrics:
|
|
18
25
|
|
|
19
26
|
## Main Risk
|
|
20
27
|
|
|
@@ -1,12 +1,23 @@
|
|
|
1
1
|
# Research Summary
|
|
2
2
|
|
|
3
3
|
## Current Direction
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
|
|
4
|
+
- Mission:
|
|
5
|
+
- Approved direction:
|
|
6
|
+
- Active stage:
|
|
7
|
+
- Current objective:
|
|
7
8
|
- Auto mode:
|
|
8
9
|
- Auto objective:
|
|
9
10
|
- Auto decision:
|
|
11
|
+
- Collaborator report mode:
|
|
12
|
+
- Canonical context readiness:
|
|
13
|
+
- Method name:
|
|
14
|
+
- Contribution bullets:
|
|
15
|
+
- Eval objective:
|
|
16
|
+
- Primary metrics:
|
|
17
|
+
- Secondary metrics:
|
|
18
|
+
- Dataset package:
|
|
19
|
+
- Benchmark role:
|
|
20
|
+
- Comparison suite:
|
|
10
21
|
|
|
11
22
|
## Strongest Evidence
|
|
12
23
|
|
|
@@ -34,17 +34,22 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
|
|
|
34
34
|
### `/lab:idea`
|
|
35
35
|
|
|
36
36
|
- Search relevant literature, baselines, datasets, and evaluation metrics before proposing a plan.
|
|
37
|
+
- Build a literature-scoping bundle before claiming novelty. The default target is 20 relevant sources unless the field is genuinely too narrow and that exception is written down.
|
|
37
38
|
- Read `.lab/context/mission.md` and `.lab/context/open-questions.md` before drafting.
|
|
39
|
+
- Read `.lab/config/workflow.json` before drafting and follow its `workflow_language` for idea artifacts.
|
|
38
40
|
- Ask one clarifying question at a time when critical ambiguity remains.
|
|
39
|
-
- State the problem, the failure case, and why the problem matters before proposing solutions.
|
|
41
|
+
- State the scenario, the problem, the failure case, and why the problem matters before proposing solutions.
|
|
40
42
|
- Classify the idea by contribution category and breakthrough level.
|
|
41
43
|
- Compare against existing methods explicitly and state why the idea should be better.
|
|
44
|
+
- Include a closest-prior-work comparison and a plain-language description of how the proposed direction would work.
|
|
42
45
|
- Distinguish sourced evidence from generated innovation claims.
|
|
43
46
|
- End with three meaningful points that are clear, short, and easy to scan.
|
|
44
47
|
- Produce 2-3 candidate approaches with trade-offs before recommending one.
|
|
45
48
|
- Critique the idea before converging on it.
|
|
49
|
+
- Include a minimum viable experiment before approval.
|
|
46
50
|
- Keep an explicit approval gate before `/lab:spec`.
|
|
47
51
|
- Write idea artifacts with the template in `.lab/.managed/templates/idea.md`.
|
|
52
|
+
- Run `.lab/.managed/scripts/validate_idea_artifact.py --idea <idea-artifact> --workflow-config .lab/config/workflow.json` before treating the idea as converged.
|
|
48
53
|
- Update `.lab/context/mission.md`, `.lab/context/decisions.md`, and `.lab/context/open-questions.md` after convergence.
|
|
49
54
|
- Do not leave `.lab/context/mission.md` as a template shell once the problem statement and approved direction are known.
|
|
50
55
|
- Do not implement code in this stage.
|
|
@@ -91,9 +96,9 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
|
|
|
91
96
|
- Reuse `/lab:run`, `/lab:iterate`, `/lab:review`, `/lab:report`, and optional `/lab:write` instead of inventing a second workflow.
|
|
92
97
|
- Do not automatically change the research mission, paper-facing framing, or core claims.
|
|
93
98
|
- You may add exploratory datasets, benchmarks, and comparison methods inside the approved exploration envelope.
|
|
94
|
-
- You may promote an exploratory addition to the primary package only after the promotion policy in `auto-mode.md` is satisfied and the promotion is written back into `.lab/context/data-decisions.md`, `.lab/context/decisions.md`, `.lab/context/state.md`, and `.lab/context/
|
|
99
|
+
- You may promote an exploratory addition to the primary package only after the promotion policy in `auto-mode.md` is satisfied and the promotion is written back into `.lab/context/data-decisions.md`, `.lab/context/decisions.md`, `.lab/context/state.md`, and `.lab/context/workflow-state.md`.
|
|
95
100
|
- Poll long-running commands until they complete, time out, or hit a stop condition.
|
|
96
|
-
- Update `.lab/context/auto-status.md`, `.lab/context/state.md`, `.lab/context/workflow-state.md`, `.lab/context/decisions.md`, `.lab/context/data-decisions.md`, `.lab/context/evidence-index.md
|
|
101
|
+
- Update `.lab/context/auto-status.md`, `.lab/context/state.md`, `.lab/context/workflow-state.md`, `.lab/context/decisions.md`, `.lab/context/data-decisions.md`, and `.lab/context/evidence-index.md` as the campaign advances, then refresh the derived handoff files.
|
|
97
102
|
- Keep an explicit approval gate when a proposed action would leave the frozen core defined by the auto-mode contract.
|
|
98
103
|
|
|
99
104
|
### `/lab:spec`
|
|
@@ -165,26 +170,21 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
|
|
|
165
170
|
|
|
166
171
|
- Start only after `report` artifacts are stable enough to support paper claims.
|
|
167
172
|
- Start only after an approved framing artifact exists at `.lab/writing/framing.md`.
|
|
168
|
-
- Read `.lab/config/workflow.json` before drafting and enforce its `paper_language
|
|
169
|
-
- If `paper_template_root` is empty and `paper_template_decision` is `unconfirmed`, ask once whether to continue with the managed default scaffold or attach a template directory first; persist the answer before drafting `.tex`.
|
|
170
|
-
- If the project is still on the default scaffold at a final export or final-draft boundary and `paper_template_final_reminder_acknowledged` is `false`, ask one final reminder question before finalizing.
|
|
173
|
+
- Read `.lab/config/workflow.json` before drafting and enforce its `workflow_language`, `paper_language`, and `paper_format`.
|
|
171
174
|
- Read `.lab/context/mission.md`, `.lab/context/decisions.md`, `.lab/context/evidence-index.md`, and `.lab/context/data-decisions.md` before drafting.
|
|
172
175
|
- Write one paper section or one explicit subproblem per round.
|
|
176
|
+
- Ordinary manuscript drafting rounds should follow `workflow_language`.
|
|
177
|
+
- If `workflow_language` and `paper_language` differ, the first final-draft or export round must ask once whether to keep the draft language or convert the final manuscript to `paper_language`, then persist that choice.
|
|
173
178
|
- Bind each claim to evidence from `report`, iteration reports, or normalized summaries.
|
|
174
|
-
-
|
|
175
|
-
-
|
|
176
|
-
-
|
|
177
|
-
-
|
|
178
|
-
- Load only the current section guide, the matching examples index when one exists, 1-2 matching concrete example files, plus `paper-review.md` and `does-my-writing-flow-source.md`.
|
|
179
|
-
- Build a compact mini-outline before prose.
|
|
180
|
-
- Build the paper asset plan before prose when the section carries method or experiments claims.
|
|
179
|
+
- Use the write-stage contract in `.codex/skills/lab/stages/write.md` or `.claude/skills/lab/stages/write.md` as the single source of truth for template choice, paper-plan requirements, section-specific references, validator calls, asset coverage, and final manuscript gates.
|
|
180
|
+
- Use the vendored paper-writing references under `skills/lab/references/paper-writing/` and the matching example-bank files under `skills/lab/references/paper-writing/examples/`.
|
|
181
|
+
- Treat `.lab/writing/plan.md` as the write-time source of truth for tables, figures, citations, and asset coverage.
|
|
182
|
+
- Treat section-quality, claim-safety, and manuscript-delivery checks as the canonical acceptance gates for final-draft or export rounds.
|
|
181
183
|
- For each subsection, explicitly cover motivation, design, and technical advantage when applicable.
|
|
182
184
|
- Keep terminology stable across rounds and sections.
|
|
183
185
|
- If a claim is not supported by evidence, weaken or remove it.
|
|
184
186
|
- Treat tables, figures, citations, and bibliography as core manuscript content rather than optional polish.
|
|
185
187
|
- Keep paper-facing LaTeX free of absolute local paths, rerun ids, shell transcripts, and internal workflow provenance.
|
|
186
|
-
- Materialize real LaTeX tables and figure placeholders instead of leaving all evidence inside prose paragraphs.
|
|
187
|
-
- Run `.lab/.managed/scripts/validate_manuscript_delivery.py --paper-dir <deliverables_root>/paper` before accepting a final-draft or export round.
|
|
188
188
|
- Before finalizing a round, append and answer the five-dimension self-review checklist and revise unresolved items.
|
|
189
189
|
- Apply paper-writing discipline without changing experimental truth.
|
|
190
190
|
- If the evidence is insufficient, stop and route back to `review` or `iterate`.
|
|
@@ -199,7 +199,8 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
|
|
|
199
199
|
- No unconstrained auto mode. Every `/lab:auto` campaign must declare allowed stages, stop conditions, and a promotion policy in `.lab/context/auto-mode.md`.
|
|
200
200
|
- No auto start without an explicit autonomy level and `Approval status: approved`.
|
|
201
201
|
- No final report without validated normalized results.
|
|
202
|
-
- No paper-writing round without stable report artifacts, an approved framing artifact, evidence links, LaTeX manuscript output
|
|
202
|
+
- No paper-writing round without stable report artifacts, an approved framing artifact, evidence links, and LaTeX manuscript output.
|
|
203
|
+
- No final-draft or export round without passing section-quality, claim-safety, and manuscript-delivery validation.
|
|
203
204
|
|
|
204
205
|
## References
|
|
205
206
|
|
package/package-assets/shared/skills/lab/references/paper-writing/examples/abstract/template-b.md
CHANGED
|
@@ -16,7 +16,7 @@
|
|
|
16
16
|
|
|
17
17
|
% Introduce the technical contribution that implements the insight in one to two sentences (usually mention the technical term/name only, without describing every detailed step. The term should be easy to understand and should not create a jump in reading. This ability is very important for writing a good abstract.)
|
|
18
18
|
%% Example 1: To do this, we first present a label-efficient depth estimation framework using the internal representations of diffusion models. At the sampling phase, we utilize two guidance techniques to self-condition the generated image using the estimated depth map, the first of which uses pseudo-labeling, and the subsequent one uses a depth-domain diffusion prior.
|
|
19
|
-
%% Example 2: To this end, we propose
|
|
19
|
+
%% Example 2: To this end, we propose AnchorField, a structured representation in which different observations share the same set of latent codes anchored to a deformable support.
|
|
20
20
|
|
|
21
21
|
% Introduce the benefits of technical novelty
|
|
22
22
|
%% Example 2: so that the observations across frames can be naturally integrated. The deformable mesh also provides geometric guidance for the network to learn 3D representations more efficiently.
|
|
@@ -29,6 +29,6 @@
|
|
|
29
29
|
1. `This paper addresses the challenge of novel view synthesis for a human performer from a very sparse set of camera views.`
|
|
30
30
|
2. `... representation learning will be ill-posed if the views are highly sparse.`
|
|
31
31
|
3. `To solve this ill-posed problem, our key idea is to integrate observations over video frames.`
|
|
32
|
-
4. `To this end, we propose
|
|
32
|
+
4. `To this end, we propose AnchorField ...`
|
|
33
33
|
5. `... observations across frames can be naturally integrated ... provides geometric guidance ...`
|
|
34
34
|
6. `Experiments show [main result].`
|
|
@@ -6,22 +6,22 @@ boundary explicit.
|
|
|
6
6
|
```tex
|
|
7
7
|
\section{Conclusion}
|
|
8
8
|
|
|
9
|
-
This paper shows that adding a structured
|
|
10
|
-
|
|
11
|
-
protocol. Across the
|
|
9
|
+
This paper shows that adding a structured intermediate module together with a
|
|
10
|
+
lightweight adjustment stage improves performance under a fixed evaluation
|
|
11
|
+
protocol. Across the benchmark families used in this work, the full model
|
|
12
12
|
consistently matches or exceeds the strongest baselines and remains stronger
|
|
13
13
|
than the key ablated variants. This makes the main claim narrower than a
|
|
14
|
-
universal superiority claim but stronger than a single-
|
|
14
|
+
universal superiority claim but stronger than a single-setting win.
|
|
15
15
|
|
|
16
|
-
We do not claim that the current method solves
|
|
17
|
-
or that every design choice helps equally on every benchmark. In
|
|
18
|
-
|
|
19
|
-
which means its value should be interpreted as
|
|
20
|
-
a guaranteed gain. That boundary is consistent
|
|
21
|
-
practice, which argues for claim discipline and
|
|
22
|
-
rather than broad overgeneralization~\cite{carlini2019evaluating}.
|
|
16
|
+
We do not claim that the current method solves the broader problem in every
|
|
17
|
+
domain or that every design choice helps equally on every benchmark. In
|
|
18
|
+
particular, the adjustment stage appears beneficial in some settings and
|
|
19
|
+
neutral in others, which means its value should be interpreted as
|
|
20
|
+
setting-dependent rather than as a guaranteed gain. That boundary is consistent
|
|
21
|
+
with recent benchmarking practice, which argues for claim discipline and
|
|
22
|
+
protocol-specific interpretation rather than broad overgeneralization~\cite{carlini2019evaluating}.
|
|
23
23
|
|
|
24
24
|
The most useful next step is to extend the evaluation to a broader set of
|
|
25
|
-
benchmark slices and to test whether the same
|
|
26
|
-
remains useful when the
|
|
25
|
+
benchmark slices and to test whether the same backbone-versus-adjustment split
|
|
26
|
+
remains useful when the data distribution shifts more aggressively.
|
|
27
27
|
```
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Main Results and Ablation LaTeX Example
|
|
2
2
|
|
|
3
|
-
This file is a complete
|
|
3
|
+
This file is a complete manuscript-ready LaTeX example for the experiments section.
|
|
4
4
|
Reuse the structure, caption logic, and prose-to-table linkage. Replace the
|
|
5
5
|
placeholder methods, metrics, and values with the current project's evidence.
|
|
6
6
|
|
|
@@ -14,13 +14,13 @@ Source inspiration:
|
|
|
14
14
|
|
|
15
15
|
```tex
|
|
16
16
|
\begin{table}[t]
|
|
17
|
-
\caption{Main benchmark results under the
|
|
17
|
+
\caption{Main benchmark results under the fixed evaluation protocol. Higher is better on all metrics.}
|
|
18
18
|
\label{tab:main-results}
|
|
19
19
|
\centering
|
|
20
20
|
\resizebox{0.92\linewidth}{!}{
|
|
21
21
|
\begin{tabular}{lccc}
|
|
22
22
|
\toprule
|
|
23
|
-
Method &
|
|
23
|
+
Method & Primary Metric $\uparrow$ & Secondary Metric $\uparrow$ & Error Metric $\downarrow$ \\
|
|
24
24
|
\midrule
|
|
25
25
|
Strongest baseline & 0.1421 & 0.0873 & 0.0612 \\
|
|
26
26
|
Closest prior work & 0.1488 & 0.0915 & 0.0544 \\
|
|
@@ -32,7 +32,7 @@ Ours & \textbf{0.1564} & \textbf{0.0987} & \textbf{0.0418} \\
|
|
|
32
32
|
```
|
|
33
33
|
|
|
34
34
|
Table message:
|
|
35
|
-
- `Does the proposed method beat the strongest baselines under the
|
|
35
|
+
- `Does the proposed method beat the strongest baselines under the fixed evaluation protocol?`
|
|
36
36
|
|
|
37
37
|
## Ablation Table
|
|
38
38
|
|
|
@@ -47,7 +47,7 @@ Variant & AUUC $\uparrow$ \\
|
|
|
47
47
|
\midrule
|
|
48
48
|
Ours & \textbf{0.1564} \\
|
|
49
49
|
w/o structure module & 0.1497 \\
|
|
50
|
-
w/o
|
|
50
|
+
w/o final adjustment stage & 0.1510 \\
|
|
51
51
|
w/ shuffled auxiliary signal & 0.1458 \\
|
|
52
52
|
\bottomrule
|
|
53
53
|
\end{tabular}
|
|
@@ -65,19 +65,20 @@ Table message:
|
|
|
65
65
|
|
|
66
66
|
Table~\ref{tab:main-results} answers the main ranking question: whether the full
|
|
67
67
|
method remains stronger than the closest prior work and the strongest practical
|
|
68
|
-
baseline under the
|
|
69
|
-
while also reducing
|
|
70
|
-
trading
|
|
68
|
+
baseline under the fixed evaluation protocol. Our method achieves the best
|
|
69
|
+
primary and secondary metrics while also reducing the error metric, which means
|
|
70
|
+
the gain is not coming from trading one objective against stability.
|
|
71
71
|
|
|
72
72
|
Table~\ref{tab:ablations} then asks a narrower mechanism question. Removing the
|
|
73
|
-
structure module causes the largest drop, so the main gain is tied to
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
claim that
|
|
77
|
-
driver of the result. The shuffled-signal variant acts as a
|
|
78
|
-
shows that the gain does not survive when the auxiliary
|
|
73
|
+
structure module causes the largest drop, so the main gain is tied to explicit
|
|
74
|
+
structure modeling rather than to a generic increase in capacity. Removing the
|
|
75
|
+
final adjustment stage leads to a smaller but still visible drop, which
|
|
76
|
+
supports the claim that the adjustment helps the exposed prediction without
|
|
77
|
+
being the sole driver of the result. The shuffled-signal variant acts as a
|
|
78
|
+
negative control and shows that the gain does not survive when the auxiliary
|
|
79
|
+
information is broken.
|
|
79
80
|
|
|
80
|
-
One caveat is that the
|
|
81
|
-
so the paper should not overclaim that every component helps equally
|
|
82
|
-
dataset.
|
|
81
|
+
One caveat is that the final adjustment gain may remain neutral in some
|
|
82
|
+
settings, so the paper should not overclaim that every component helps equally
|
|
83
|
+
on every dataset.
|
|
83
84
|
```
|
package/package-assets/shared/skills/lab/references/paper-writing/examples/experiments-examples.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Experiments Example Patterns
|
|
2
2
|
|
|
3
|
-
Use these examples when turning validated results into
|
|
3
|
+
Use these examples when turning validated results into manuscript-ready LaTeX assets.
|
|
4
4
|
The referenced files contain complete LaTeX environments and section-level prose
|
|
5
5
|
glue, not just checklists.
|
|
6
6
|
|
|
@@ -12,7 +12,7 @@ locally organized cite targets.
|
|
|
12
12
|
5. Introduction technical-challenge files: `references/examples/introduction/technical-challenge-version-1-existing-task.md`, `references/examples/introduction/technical-challenge-version-2-existing-task-insight-backed-by-traditional.md`, `references/examples/introduction/technical-challenge-version-3-novel-task.md`, `references/examples/introduction/novel-task-challenge-decomposition.md`
|
|
13
13
|
6. Introduction pipeline files: `references/examples/introduction/pipeline-version-1-one-contribution-multi-advantages.md`, `references/examples/introduction/pipeline-version-2-two-contributions.md`, `references/examples/introduction/pipeline-version-3-new-module-on-existing-pipeline.md`, `references/examples/introduction/pipeline-version-4-observation-driven.md`, `references/examples/introduction/pipeline-not-recommended-abstract-only.md`
|
|
14
14
|
7. Method examples index: `references/examples/method-examples.md`
|
|
15
|
-
8. Method detail files: `references/examples/method/pre-writing-questions.md`, `references/examples/method/module-triad-
|
|
15
|
+
8. Method detail files: `references/examples/method/pre-writing-questions.md`, `references/examples/method/module-triad-anchored-representation.md`, `references/examples/method/annotated-figure-to-text.md`, `references/examples/method/module-design-multiresolution-encoding.md`, `references/examples/method/module-motivation-patterns.md`, `references/examples/method/section-skeleton.md`, `references/examples/method/overview-template.md`, `references/examples/method/example-of-the-three-elements.md`, `references/examples/method/method-writing-common-issues-note.md`
|
|
16
16
|
9. Related work examples index: `references/examples/related-work-examples.md`
|
|
17
17
|
10. Related work files: `references/examples/related-work/closest-prior-gap-template.md`, `references/examples/related-work/topic-comparison-template.md`
|
|
18
18
|
11. Experiments examples index: `references/examples/experiments-examples.md`
|
|
@@ -5,7 +5,7 @@
|
|
|
5
5
|
|
|
6
6
|
```latex
|
|
7
7
|
% In this paper, we propose a novel framework …
|
|
8
|
-
%% Example: In this paper, we introduce a
|
|
8
|
+
%% Example: In this paper, we introduce a structured anchor-based representation, named AnchorField, to solve the challenge of reconstructing a changing target from sparse observations.
|
|
9
9
|
In this paper, we propose a novel framework/representation, named [method name] for [xxx task].
|
|
10
10
|
|
|
11
11
|
% Teaser for basic idea
|
|
@@ -13,11 +13,11 @@ In this paper, we propose a novel framework/representation, named [method name]
|
|
|
13
13
|
The basic idea is illustrated in [xxx Figure].
|
|
14
14
|
|
|
15
15
|
% One-sentence key novelty/contribution (very important ability)
|
|
16
|
-
%% Example:
|
|
16
|
+
%% Example: Instead of learning each observation-specific field separately, AnchorField generates them from the same set of anchor codes.
|
|
17
17
|
Our innovation is in [one sentence for key novelty].
|
|
18
18
|
|
|
19
19
|
% Method details
|
|
20
|
-
%% Example: Specifically, we anchor a set of latent codes to
|
|
20
|
+
%% Example: Specifically, we anchor a set of latent codes to a reference support structure so that their spatial locations follow the current observation. To obtain the target representation for one condition, we first transform the code locations using the available alignment signal. Then, a prediction network estimates the required outputs at each query point from these codes. Both the anchor codes and the prediction network are jointly learned from all observations.
|
|
21
21
|
Specifically, [how it works in detail].
|
|
22
22
|
|
|
23
23
|
% Advantage 1
|
|
@@ -5,7 +5,7 @@
|
|
|
5
5
|
|
|
6
6
|
```latex
|
|
7
7
|
% In this paper, we propose a novel framework …
|
|
8
|
-
%% Example: In this paper, we introduce a
|
|
8
|
+
%% Example: In this paper, we introduce a structured anchor-based representation, named AnchorField, to solve the challenge of reconstructing a changing target from sparse observations.
|
|
9
9
|
In this paper, we propose a novel framework/representation, named [method name] for [xxx task].
|
|
10
10
|
|
|
11
11
|
% One-sentence key novelty
|
|
@@ -0,0 +1,66 @@
|
|
|
1
|
+
# Annotated Figure (Text Conversion)
|
|
2
|
+
|
|
3
|
+
This file converts an annotated method figure into reusable writing notes.
|
|
4
|
+
|
|
5
|
+
## Purpose
|
|
6
|
+
|
|
7
|
+
Use this mapping to understand how one Method section can explicitly separate:
|
|
8
|
+
|
|
9
|
+
1. Module motivation
|
|
10
|
+
2. Module design (data structure)
|
|
11
|
+
3. Module design (forward process)
|
|
12
|
+
4. Technical advantages
|
|
13
|
+
|
|
14
|
+
## Block-by-Block Mapping
|
|
15
|
+
|
|
16
|
+
### Section 3.1: Structured Anchor Codes
|
|
17
|
+
|
|
18
|
+
1. **Module design (data structure)**
|
|
19
|
+
- The paragraph defines structured codes anchored to a reference support structure.
|
|
20
|
+
- It explains what is constructed: latent codes, anchor positions, and the condition-dependent transformation that keeps them aligned with the current instance.
|
|
21
|
+
|
|
22
|
+
2. **Technical advantages**
|
|
23
|
+
- The paragraph explains why this design works better: it preserves correspondence across observations and keeps the representation aligned with changing geometry.
|
|
24
|
+
- It highlights why anchoring codes to a structured support is beneficial.
|
|
25
|
+
|
|
26
|
+
### Section 3.2: Code Diffusion
|
|
27
|
+
|
|
28
|
+
1. **Motivation of this module**
|
|
29
|
+
- The paragraph states the remaining problem: direct interpolation of sparse anchor codes leaves many query locations weakly represented.
|
|
30
|
+
- This motivates propagation from anchor locations to nearby continuous space.
|
|
31
|
+
|
|
32
|
+
2. **Module design (forward process)**
|
|
33
|
+
- The paragraph explains the execution pipeline: build sparse feature volumes, run sparse convolutions, interpolate codes at query points, and feed those codes to prediction networks.
|
|
34
|
+
- This is a canonical input -> steps -> output module description.
|
|
35
|
+
|
|
36
|
+
### Section 3.3: Field Regression
|
|
37
|
+
|
|
38
|
+
1. **Module design (forward process) for the first prediction head**
|
|
39
|
+
- The first paragraph defines how one target quantity is regressed from the latent code and condition inputs.
|
|
40
|
+
|
|
41
|
+
2. **Module design (data structure) for the second prediction head**
|
|
42
|
+
- The next paragraph introduces required inputs or embeddings such as latent code, query direction, spatial location, or condition tokens.
|
|
43
|
+
|
|
44
|
+
3. **Module design (forward process) for the second prediction head**
|
|
45
|
+
- The following paragraph describes how those inputs are encoded and passed into the second prediction network for the final estimate.
|
|
46
|
+
|
|
47
|
+
### Section 3.4: Output Aggregation
|
|
48
|
+
|
|
49
|
+
1. **Module design (forward process)**
|
|
50
|
+
- The paragraph describes the final aggregation process that converts pointwise predictions into the output seen by the reader or evaluator.
|
|
51
|
+
|
|
52
|
+
## Reusable Writing Pattern from This Figure
|
|
53
|
+
|
|
54
|
+
For each module subsection, follow this order:
|
|
55
|
+
|
|
56
|
+
1. `Motivation`: state unresolved challenge and technical reason.
|
|
57
|
+
2. `Design-1`: define structure/representation/network.
|
|
58
|
+
3. `Design-2`: describe forward process in execution order.
|
|
59
|
+
4. `Advantage`: explain why this module improves over alternatives.
|
|
60
|
+
|
|
61
|
+
## Suggested Paragraph Starters
|
|
62
|
+
|
|
63
|
+
1. Motivation: `A remaining challenge is ...`
|
|
64
|
+
2. Data structure design: `We represent ... with ...`
|
|
65
|
+
3. Forward process: `Given [input], we first ... then ... finally ...`
|
|
66
|
+
4. Technical advantage: `Compared with previous methods, this design ... because ...`
|
|
@@ -18,27 +18,27 @@ Each `% ...` annotation explains the paragraph(s) immediately below it.
|
|
|
18
18
|
\subsection{3.1. Structured latent codes}
|
|
19
19
|
|
|
20
20
|
% Module design: introduce the module's data structure
|
|
21
|
-
To
|
|
21
|
+
To keep latent codes aligned with the current instance, we anchor them to a deformable reference structure [38]. This structure is parameterized by shape variables, condition variables, and a rigid transformation relative to its own coordinate system, and outputs a posed mesh or support geometry. Specifically, we define a set of latent codes \( Z = \{z_1, z_2, ..., z_M\} \) on the support points of that reference structure. For one observation \( t \), the condition parameters \( S_t \) are estimated from the available inputs \( \{I_t^c \mid c = 1, ..., N_c\} \) using an external alignment procedure [26]. The spatial locations of the latent codes are then transformed according to \( S_t \) before the downstream prediction heads consume them. Figure 3 shows an example. The latent-code dimension is set to 16 in this example.
|
|
22
22
|
|
|
23
23
|
% Technical advantages of this module
|
|
24
|
-
Similar to
|
|
24
|
+
Similar to local implicit representations [25, 5, 18], the latent codes are used with a neural network to represent local structure and attributes. Anchoring these codes to a deformable support lets the model stay aligned with changing geometry instead of relearning a separate representation for every condition. This yields a latent-variable view in which the same anchor codes can explain multiple observations while sharing information across them.
|
|
25
25
|
|
|
26
26
|
\subsection{3.2. Code diffusion}
|
|
27
27
|
|
|
28
28
|
% Motivation of this module
|
|
29
|
-
Figure 3(a) shows the process of code
|
|
29
|
+
Figure 3(a) shows the process of code propagation. The target fields assign outputs to each point in the domain, which requires querying the latent codes at continuous locations. This can be achieved with trilinear interpolation. However, because the structured anchor codes are sparse, directly interpolating them leaves many query points weakly represented. To solve this problem, we propagate the codes defined on the support structure to nearby space.
|
|
30
30
|
|
|
31
31
|
% Module design: introduce module design by describing the module forward process
|
|
32
|
-
Inspired by [65, 56, 49], we choose
|
|
32
|
+
Inspired by [65, 56, 49], we choose a sparse convolutional backbone [21] to efficiently process the structured latent codes, whose architecture is described in Table 1. Specifically, based on the reference-structure parameters, we compute a 3D bounding box around the active support and divide that box into small voxels with voxel size \( 5mm \times 5mm \times 5mm \). The latent code of a non-empty voxel is the mean of the anchor codes that fall inside it. The sparse backbone applies 3D sparse convolutions to this input volume and outputs latent code volumes at \( 2\times, 4\times, 8\times, 16\times \) downsampled scales. With convolution and downsampling, the input codes are propagated to nearby space. Following [56], for any point in the domain, we interpolate the latent codes from multi-scale code volumes of network layers 5, 9, 13, and 17, and concatenate them into the final latent code. Since propagation should not be affected by global position and orientation, we transform code locations to the reference coordinate system.
|
|
33
33
|
|
|
34
|
-
For any point \( \mathbf{x} \)
|
|
34
|
+
For any query point \( \mathbf{x} \), we obtain its latent code from the propagated code volume. Specifically, the point \( \mathbf{x} \) is first transformed to the reference coordinate system so that the query and code volume are aligned. Then, the latent code is computed with trilinear interpolation. For condition parameters \( S_t \), we denote the latent code at point \( \mathbf{x} \) by \( \psi(\mathbf{x}, Z, S_t) \). The resulting code vector is passed to prediction networks that estimate the target outputs at \( \mathbf{x} \).
|
|
35
35
|
|
|
36
36
|
\subsection{3.3. Density and color regression}
|
|
37
37
|
|
|
38
|
-
Figure 3(b) overviews the regression of
|
|
38
|
+
Figure 3(b) overviews the regression of the target outputs for any query point. The corresponding fields are represented by MLP networks. Details of the network architectures are described in the supplementary material.
|
|
39
39
|
|
|
40
40
|
% Module design: introduce module design by describing the module forward process
|
|
41
|
-
\textbf{
|
|
41
|
+
\textbf{First prediction head.} For condition \( t \), the first target quantity at point \( \mathbf{x} \) is predicted as a function of only the latent code \( \psi(\mathbf{x}, Z, S_t) \), which is defined as:
|
|
42
42
|
|
|
43
43
|
\[
|
|
44
44
|
\sigma_t(\mathbf{x}) = M_{\sigma}(\psi(\mathbf{x}, Z, S_t)),
|
|
@@ -48,20 +48,20 @@ Figure 3(b) overviews the regression of density and color for any point in 3D sp
|
|
|
48
48
|
where \( M_{\sigma} \) represents an MLP network with four layers.
|
|
49
49
|
|
|
50
50
|
% Module design: introduce the module's data structure
|
|
51
|
-
\textbf{
|
|
51
|
+
\textbf{Second prediction head.} Similar to [37, 44], we take both the latent code \( \psi(\mathbf{x}, Z, S_t) \) and an observation-dependent direction \( \mathbf{d} \) as input for the second target. To model location-dependent effects, this head also takes the spatial location \( \mathbf{x} \) as input. We further observe that condition-specific factors can change the target output. Inspired by the auto-decoder [48], we assign a latent embedding \( \ell_t \) for each condition \( t \) to encode those varying factors.
|
|
52
52
|
|
|
53
53
|
% Module design: introduce module design by describing the module forward process
|
|
54
|
-
Specifically, for
|
|
54
|
+
Specifically, for condition \( t \), the second target at \( \mathbf{x} \) is predicted as a function of the latent code \( \psi(\mathbf{x}, Z, S_t) \), the observation direction \( \mathbf{d} \), the spatial location \( \mathbf{x} \), and the latent embedding \( \ell_t \). Following [51, 44], we apply positional encoding to both \( \mathbf{d} \) and \( \mathbf{x} \), which enables better learning of high-frequency functions. The second prediction head at condition \( t \) is defined as:
|
|
55
55
|
|
|
56
56
|
\[
|
|
57
57
|
c_t(\mathbf{x}) = M_c(\psi(\mathbf{x}, Z, S_t), \gamma_d(\mathbf{d}), \gamma_x(\mathbf{x}), \ell_t),
|
|
58
58
|
\tag{2}
|
|
59
59
|
\]
|
|
60
60
|
|
|
61
|
-
where \( M_c \) represents an MLP network with two layers, and \( \gamma_d \) and \( \gamma_x \) are positional encoding functions for
|
|
61
|
+
where \( M_c \) represents an MLP network with two layers, and \( \gamma_d \) and \( \gamma_x \) are positional encoding functions for direction and spatial location, respectively. We set the dimension of \( \ell_t \) to 128 in this example.
|
|
62
62
|
|
|
63
63
|
\subsection{3.4. Volume rendering}
|
|
64
64
|
|
|
65
65
|
% Module design: introduce module design by describing the module forward process
|
|
66
|
-
Given
|
|
66
|
+
Given an observation direction, we utilize a standard aggregation procedure to render the predicted field into a 2D output. Pixel values are estimated via a volume-integration equation [27] that accumulates the predicted quantities along the corresponding ray. In practice, the integral is approximated using numerical quadrature [41, 44]. Given a pixel, we first compute its ray \( \mathbf{r} \) using the camera parameters and sample \( N_k \) points \( \{\mathbf{x}_k\}_{k=1}^{N_k} \) along that ray between near and far bounds. The scene bounds are estimated from the reference structure. Then, the model predicts the target outputs at these points. For condition \( t \), the rendered value \( \hat{C}_t(\mathbf{r}) \) ...
|
|
67
67
|
```
|
|
@@ -1,11 +1,11 @@
|
|
|
1
|
-
# Module Triad Example (
|
|
1
|
+
# Module Triad Example (Anchored Representation)
|
|
2
2
|
|
|
3
|
-
`Use
|
|
3
|
+
`Use this anchored-representation example to understand the three elements of a module: design, motivation, and technical advantages.`
|
|
4
4
|
|
|
5
5
|
Local source references:
|
|
6
6
|
|
|
7
7
|
1. Annotated figure showing motivation/design/advantages split.
|
|
8
|
-
3. Text-converted annotation notes: `references/examples/method/
|
|
8
|
+
3. Text-converted annotation notes: `references/examples/method/annotated-figure-to-text.md`
|
|
9
9
|
|
|
10
10
|
Triad mapping template:
|
|
11
11
|
|
|
@@ -15,5 +15,5 @@ Triad mapping template:
|
|
|
15
15
|
|
|
16
16
|
Direct usage:
|
|
17
17
|
|
|
18
|
-
1. Read `
|
|
18
|
+
1. Read `annotated-figure-to-text.md` to map each paragraph to one triad element.
|
|
19
19
|
2. Rebuild your own Method subsection with the same triad order.
|
|
@@ -18,12 +18,12 @@
|
|
|
18
18
|
%% Example: The overview of the proposed model is illustrated in Figure 3.
|
|
19
19
|
|
|
20
20
|
% Explain what Section 3.1 covers
|
|
21
|
-
%% Example 1:
|
|
22
|
-
%% Example 2: In this section, we first describe how to model
|
|
21
|
+
%% Example 1: The model starts from a set of structured anchor codes attached to a reference support structure (Section 3.1).
|
|
22
|
+
%% Example 2: In this section, we first describe how to model the target space with learnable field maps (Section 3.1).
|
|
23
23
|
|
|
24
24
|
% Explain what Section 3.2 covers
|
|
25
|
-
%% Example 1: The
|
|
26
|
-
%% Example 2: Then, Section 3.2 discusses how to represent
|
|
25
|
+
%% Example 1: The code at any nearby location is obtained with a propagation process (Section 3.2) and then decoded into the target outputs by prediction networks (Section 3.3).
|
|
26
|
+
%% Example 2: Then, Section 3.2 discusses how to represent changing scenes with dynamic field maps.
|
|
27
27
|
|
|
28
28
|
% Explain what Section 3.3 covers
|
|
29
29
|
%% Example 3: Finally, we introduce some strategies to speed up the rendering process (Section 3.3).
|