superlab 0.1.14 → 0.1.15

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (55) hide show
  1. package/README.md +7 -3
  2. package/README.zh-CN.md +7 -3
  3. package/bin/superlab.cjs +38 -0
  4. package/lib/auto_contracts.cjs +7 -3
  5. package/lib/auto_runner.cjs +33 -52
  6. package/lib/auto_state.cjs +27 -21
  7. package/lib/context.cjs +15 -0
  8. package/lib/i18n.cjs +106 -33
  9. package/lib/install.cjs +1 -0
  10. package/package-assets/claude/commands/lab/write.md +1 -1
  11. package/package-assets/claude/commands/lab.md +1 -0
  12. package/package-assets/codex/prompts/lab-write.md +1 -1
  13. package/package-assets/codex/prompts/lab.md +1 -0
  14. package/package-assets/shared/lab/.managed/templates/final-report.md +12 -0
  15. package/package-assets/shared/lab/.managed/templates/main-tables.md +37 -0
  16. package/package-assets/shared/lab/config/workflow.json +3 -1
  17. package/package-assets/shared/lab/context/auto-outcome.md +3 -0
  18. package/package-assets/shared/skills/lab/SKILL.md +6 -2
  19. package/package-assets/shared/skills/lab/references/paper-writing/abstract.md +7 -1
  20. package/package-assets/shared/skills/lab/references/paper-writing/examples/abstract/template-a.md +21 -0
  21. package/package-assets/shared/skills/lab/references/paper-writing/examples/abstract/template-b.md +34 -0
  22. package/package-assets/shared/skills/lab/references/paper-writing/examples/abstract/template-c.md +28 -0
  23. package/package-assets/shared/skills/lab/references/paper-writing/examples/abstract-examples.md +13 -0
  24. package/package-assets/shared/skills/lab/references/paper-writing/examples/index.md +21 -0
  25. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/novel-task-challenge-decomposition.md +18 -0
  26. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/pipeline-not-recommended-abstract-only.md +30 -0
  27. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/pipeline-version-1-one-contribution-multi-advantages.md +30 -0
  28. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/pipeline-version-2-two-contributions.md +34 -0
  29. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/pipeline-version-3-new-module-on-existing-pipeline.md +18 -0
  30. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/pipeline-version-4-observation-driven.md +16 -0
  31. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/technical-challenge-version-1-existing-task.md +32 -0
  32. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/technical-challenge-version-2-existing-task-insight-backed-by-traditional.md +33 -0
  33. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/technical-challenge-version-3-novel-task.md +21 -0
  34. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/version-1-task-then-application.md +14 -0
  35. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/version-2-application-first.md +10 -0
  36. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/version-3-general-to-specific-setting.md +14 -0
  37. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction/version-4-open-with-challenge.md +20 -0
  38. package/package-assets/shared/skills/lab/references/paper-writing/examples/introduction-examples.md +25 -0
  39. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/example-of-the-three-elements.md +67 -0
  40. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/method-writing-common-issues-note.md +10 -0
  41. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/module-design-instant-ngp.md +55 -0
  42. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/module-motivation-patterns.md +15 -0
  43. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/module-triad-neural-body.md +19 -0
  44. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/neural-body-annotated-figure-text.md +66 -0
  45. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/overview-template.md +30 -0
  46. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/pre-writing-questions.md +17 -0
  47. package/package-assets/shared/skills/lab/references/paper-writing/examples/method/section-skeleton.md +9 -0
  48. package/package-assets/shared/skills/lab/references/paper-writing/examples/method-examples.md +24 -0
  49. package/package-assets/shared/skills/lab/references/paper-writing/introduction.md +7 -1
  50. package/package-assets/shared/skills/lab/references/paper-writing/method.md +6 -2
  51. package/package-assets/shared/skills/lab/references/paper-writing-integration.md +26 -0
  52. package/package-assets/shared/skills/lab/stages/auto.md +9 -1
  53. package/package-assets/shared/skills/lab/stages/report.md +5 -1
  54. package/package-assets/shared/skills/lab/stages/write.md +16 -1
  55. package/package.json +1 -1
package/lib/install.cjs CHANGED
@@ -538,6 +538,7 @@ function localizeInstalledAssets(targetDir, lang, { newlyCreatedProjectOwnedPath
538
538
  path.join(".lab", ".managed", "templates", "iteration-report.md"),
539
539
  path.join(".lab", ".managed", "templates", "review-checklist.md"),
540
540
  path.join(".lab", ".managed", "templates", "final-report.md"),
541
+ path.join(".lab", ".managed", "templates", "main-tables.md"),
541
542
  path.join(".lab", ".managed", "templates", "paper-plan.md"),
542
543
  path.join(".lab", ".managed", "templates", "paper-section.md"),
543
544
  path.join(".lab", ".managed", "templates", "write-iteration.md"),
@@ -8,4 +8,4 @@ tags: [workflow, research, writing]
8
8
  Use the installed `lab` skill at `.claude/skills/lab/SKILL.md`.
9
9
 
10
10
  Execute the requested `/lab:write` stage against the user's argument now. Do not only recommend another lab stage. If a blocking prerequisite is missing, say exactly what is missing and ask at most one clarifying question.
11
- This command runs the `/lab:write` stage. It requires an approved framing artifact from `/lab:framing`, must read the matching section reference from `.claude/skills/lab/references/paper-writing/`, plus `paper-review.md` and `does-my-writing-flow-source.md`, build a mini-outline, and then revise only one section.
11
+ This command runs the `/lab:write` stage. It requires an approved framing artifact from `/lab:framing`, must read the matching section reference from `.claude/skills/lab/references/paper-writing/`, and for `abstract`, `introduction`, or `method` it must also read `.claude/skills/lab/references/paper-writing/examples/index.md` plus the matching examples index and 1-2 concrete example files. Then it should run `paper-review.md` and `does-my-writing-flow-source.md`, build a mini-outline, and revise only one section.
@@ -40,6 +40,7 @@ tags: [workflow, research, overview]
40
40
 
41
41
  - `/lab:write`
42
42
  Turn stable report artifacts into paper sections through small, evidence-bound writing rounds using the vendored paper-writing references under the installed `lab` skill.
43
+ On the first manuscript-writing round, if `paper_template_root` is empty, explicitly ask once whether to stay on the managed default LaTeX scaffold or attach a template directory first; persist the user's default-scaffold choice before continuing.
43
44
 
44
45
  ## Dispatch Rules
45
46
 
@@ -6,4 +6,4 @@ argument-hint: section or writing target
6
6
  Use the installed `lab` skill at `.codex/skills/lab/SKILL.md`.
7
7
 
8
8
  Execute the requested `/lab:write` stage against the user's argument now. Do not only recommend another lab stage. If a blocking prerequisite is missing, say exactly what is missing and ask at most one clarifying question.
9
- This command runs the `/lab:write` stage. It requires an approved framing artifact from `/lab:framing`, must read the matching section reference from `.codex/skills/lab/references/paper-writing/`, plus `paper-review.md` and `does-my-writing-flow-source.md`, build a mini-outline, and then revise only one section.
9
+ This command runs the `/lab:write` stage. It requires an approved framing artifact from `/lab:framing`, must read the matching section reference from `.codex/skills/lab/references/paper-writing/`, and for `abstract`, `introduction`, or `method` it must also read `.codex/skills/lab/references/paper-writing/examples/index.md` plus the matching examples index and 1-2 concrete example files. Then it should run `paper-review.md` and `does-my-writing-flow-source.md`, build a mini-outline, and revise only one section.
@@ -38,6 +38,7 @@ argument-hint: workflow question or stage choice
38
38
 
39
39
  - `/lab:write`
40
40
  Turn stable report artifacts into paper sections through small, evidence-bound writing rounds using the vendored paper-writing references under the installed `lab` skill.
41
+ On the first manuscript-writing round, if `paper_template_root` is empty, explicitly ask once whether to stay on the managed default LaTeX scaffold or attach a template directory first; persist the user's default-scaffold choice before continuing.
41
42
 
42
43
  ## Dispatch Rules
43
44
 
@@ -4,6 +4,12 @@
4
4
 
5
5
  Summarize the method and overall outcome.
6
6
 
7
+ ## Selected Metrics
8
+
9
+ - Primary metrics:
10
+ - Secondary metrics:
11
+ - Required terminal evidence:
12
+
7
13
  ## Experiment Setup
8
14
 
9
15
  - Datasets:
@@ -11,6 +17,12 @@ Summarize the method and overall outcome.
11
17
  - Baselines:
12
18
  - Metrics:
13
19
 
20
+ ## Main Tables
21
+
22
+ - Managed main tables artifact: `<deliverables_root>/main-tables.md`
23
+ - Final performance summary:
24
+ - Table coverage:
25
+
14
26
  ## Main Results
15
27
 
16
28
  Summarize validated iteration outcomes.
@@ -0,0 +1,37 @@
1
+ # Main Tables
2
+
3
+ ## Selected Metrics
4
+
5
+ - Primary metrics:
6
+ - Secondary metrics:
7
+ - Required terminal evidence:
8
+
9
+ ## Final Performance Summary
10
+
11
+ - Main result summary:
12
+ - Most important numbers:
13
+ - Reporting caveat:
14
+
15
+ ## Table 1
16
+
17
+ - Purpose:
18
+ - Metrics used:
19
+ - Strongest supported claim:
20
+
21
+ ## Table 2
22
+
23
+ - Purpose:
24
+ - Metrics used:
25
+ - Strongest supported claim:
26
+
27
+ ## Table 3
28
+
29
+ - Purpose:
30
+ - Metrics used:
31
+ - Strongest supported claim:
32
+
33
+ ## Table 4
34
+
35
+ - Purpose:
36
+ - Metrics used:
37
+ - Strongest supported claim:
@@ -5,5 +5,7 @@
5
5
  "results_root": "results",
6
6
  "figures_root": "figures",
7
7
  "deliverables_root": "docs/research",
8
- "paper_template_root": ""
8
+ "paper_template_root": "",
9
+ "paper_template_decision": "unconfirmed",
10
+ "paper_template_final_reminder_acknowledged": false
9
11
  }
@@ -3,6 +3,9 @@
3
3
  ## Goal
4
4
 
5
5
  - Objective:
6
+ - Primary metrics:
7
+ - Secondary metrics:
8
+ - Required terminal evidence:
6
9
  - Experiment ladder:
7
10
  - Metric glossary:
8
11
  - Metric source papers:
@@ -151,7 +151,7 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
151
151
  - Read `.lab/context/eval-protocol.md` before choosing tables, thresholds, or final result framing.
152
152
  - Keep metric definitions, comparison semantics, and implementation references anchored to the approved evaluation protocol instead of re-deriving them during reporting.
153
153
  - Aggregate them with `.lab/.managed/scripts/summarize_iterations.py`.
154
- - Write the final document with `.lab/.managed/templates/final-report.md`.
154
+ - Write the final document with `.lab/.managed/templates/final-report.md` and the managed table summary with `.lab/.managed/templates/main-tables.md`.
155
155
  - Keep failed attempts and limitations visible.
156
156
  - Update `.lab/context/state.md` and `.lab/context/evidence-index.md` with report-level handoff notes.
157
157
 
@@ -160,13 +160,16 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
160
160
  - Start only after `report` artifacts are stable enough to support paper claims.
161
161
  - Start only after an approved framing artifact exists at `.lab/writing/framing.md`.
162
162
  - Read `.lab/config/workflow.json` before drafting and enforce its `paper_language` and `paper_format`.
163
+ - If `paper_template_root` is empty and `paper_template_decision` is `unconfirmed`, ask once whether to continue with the managed default scaffold or attach a template directory first; persist the answer before drafting `.tex`.
164
+ - If the project is still on the default scaffold at a final export or final-draft boundary and `paper_template_final_reminder_acknowledged` is `false`, ask one final reminder question before finalizing.
163
165
  - Read `.lab/context/mission.md`, `.lab/context/decisions.md`, `.lab/context/evidence-index.md`, and `.lab/context/data-decisions.md` before drafting.
164
166
  - Write one paper section or one explicit subproblem per round.
165
167
  - Bind each claim to evidence from `report`, iteration reports, or normalized summaries.
166
168
  - Write planning artifacts with `.lab/.managed/templates/paper-plan.md`, `.lab/.managed/templates/paper-section.md`, and `.lab/.managed/templates/write-iteration.md`.
167
169
  - Write final manuscript artifacts with `.lab/.managed/templates/paper.tex` and `.lab/.managed/templates/paper-section.tex`.
168
170
  - Use the vendored paper-writing references under `skills/lab/references/paper-writing/`.
169
- - Load only the current section guide, plus `paper-review.md` and `does-my-writing-flow-source.md`.
171
+ - For `abstract`, `introduction`, and `method`, also use the vendored example-bank files under `skills/lab/references/paper-writing/examples/`.
172
+ - Load only the current section guide, the matching examples index when one exists, 1-2 matching concrete example files, plus `paper-review.md` and `does-my-writing-flow-source.md`.
170
173
  - Build a compact mini-outline before prose.
171
174
  - For each subsection, explicitly cover motivation, design, and technical advantage when applicable.
172
175
  - Keep terminology stable across rounds and sections.
@@ -203,6 +206,7 @@ Use this skill when the user invokes `/lab:*` or asks for the structured researc
203
206
  - Write stage guide: `.codex/skills/lab/stages/write.md` or `.claude/skills/lab/stages/write.md`
204
207
  - Paper-writing integration: `.codex/skills/lab/references/paper-writing-integration.md` or `.claude/skills/lab/references/paper-writing-integration.md`
205
208
  - Vendored paper-writing references: `.codex/skills/lab/references/paper-writing/{abstract,introduction,related-work,method,experiments,conclusion,paper-review,does-my-writing-flow-source}.md` or `.claude/skills/lab/references/paper-writing/{abstract,introduction,related-work,method,experiments,conclusion,paper-review,does-my-writing-flow-source}.md`
209
+ - Vendored paper-writing example bank: `.codex/skills/lab/references/paper-writing/examples/{index,abstract-examples,introduction-examples,method-examples}.md` or `.claude/skills/lab/references/paper-writing/examples/{index,abstract-examples,introduction-examples,method-examples}.md`, plus the matching section subdirectories
206
210
  - Command adapters: the installed `/lab:*` command assets
207
211
  - Shared workflow config: `.lab/config/workflow.json`
208
212
  - Shared project context: `.lab/context/{mission,state,decisions,evidence-index,open-questions,data-decisions,eval-protocol,auto-mode,auto-status}.md`
@@ -77,7 +77,13 @@ Version 3: When there are multiple technical contributions, describe each contri
77
77
 
78
78
  ## Usage Note
79
79
 
80
- This vendored guide is intentionally self-contained. The original example-bank files are not bundled in `superlab`, so use the structures and checklists in this document directly rather than chasing external example paths.
80
+ This vendored guide should be paired with the local abstract example bank:
81
+
82
+ - `references/examples/index.md`
83
+ - `references/examples/abstract-examples.md`
84
+ - one matching file under `references/examples/abstract/`
85
+
86
+ Use the structures and checklists here first, then use the example files to choose a concrete abstract shape and sentence logic. Reuse structure, not wording.
81
87
 
82
88
  ## Abstract Quality Checklist
83
89
 
@@ -0,0 +1,21 @@
1
+ # Abstract Template A Examples (Challenge -> Contribution)
2
+
3
+ Source scope: your original notes, "Version 1".
4
+
5
+ ```latex
6
+ \section{Abstract}
7
+ % Task
8
+ % Technical challenge for previous methods (discuss around the technical challenge that we solved)
9
+ % Introduce the technical contribution for solving the challenge in one to two sentences (usually mention the technical term/name only, without describing every detailed step. The term should be easy to understand and should not create a jump in reading. This ability is very important for writing a good abstract.)
10
+ % Introduce the benefits of the technical contribution
11
+ % Experiment
12
+ ```
13
+
14
+ ## Reusable skeleton
15
+
16
+ 1. `[Task sentence]`
17
+ 2. `However, previous methods suffer from [technical challenge].`
18
+ 3. `To solve this challenge, we propose [technical contribution name].`
19
+ 4. `[One more contribution sentence if needed].`
20
+ 5. `This contribution brings [technical benefit].`
21
+ 6. `Experiments show [main result].`
@@ -0,0 +1,34 @@
1
+ # Abstract Template B Examples (Challenge -> Insight -> Contribution)
2
+
3
+ ```latex
4
+ \section{Abstract}
5
+ % Task
6
+ %% Example 1: In recent years, generative models have undergone significant advancement due to the success of diffusion models.
7
+ %% Example 2: This paper addresses the challenge of novel view synthesis for a human performer from a very sparse set of camera views.
8
+
9
+ % Technical challenge for previous methods (discuss around the technical challenge that we solved)
10
+ %% Example 1: The success of these models is often attributed to their use of guidance techniques, such as classifier and classifier-free methods, which provides effective mechanisms to tradeoff between fidelity and diversity. However, these methods are not capable of guiding a generated image to be aware of its geometric configuration, e.g., depth, which hinders the application of diffusion models to areas that require a certain level of depth awareness.
11
+ %% Example 2: Some recent works have shown that learning implicit neural representations of 3D scenes achieves remarkable view synthesis quality given dense input views. However, the representation learning will be ill-posed if the views are highly sparse.
12
+
13
+ % Introduce the insight for solving the challenge in one sentence
14
+ %% Example 1: To address this limitation, we propose a novel guidance approach for diffusion models that uses estimated depth information derived from the rich intermediate representations of diffusion models.
15
+ %% Example 2: To solve this ill-posed problem, our key idea is to integrate observations over video frames.
16
+
17
+ % Introduce the technical contribution that implements the insight in one to two sentences (usually mention the technical term/name only, without describing every detailed step. The term should be easy to understand and should not create a jump in reading. This ability is very important for writing a good abstract.)
18
+ %% Example 1: To do this, we first present a label-efficient depth estimation framework using the internal representations of diffusion models. At the sampling phase, we utilize two guidance techniques to self-condition the generated image using the estimated depth map, the first of which uses pseudo-labeling, and the subsequent one uses a depth-domain diffusion prior.
19
+ %% Example 2: To this end, we propose Neural Body, a new human body representation which assumes that the learned neural representations at different frames share the same set of latent codes anchored to a deformable mesh
20
+
21
+ % Introduce the benefits of technical novelty
22
+ %% Example 2: so that the observations across frames can be naturally integrated. The deformable mesh also provides geometric guidance for the network to learn 3D representations more efficiently.
23
+
24
+ % Experiment
25
+ ```
26
+
27
+ ## Given example pattern 2
28
+
29
+ 1. `This paper addresses the challenge of novel view synthesis for a human performer from a very sparse set of camera views.`
30
+ 2. `... representation learning will be ill-posed if the views are highly sparse.`
31
+ 3. `To solve this ill-posed problem, our key idea is to integrate observations over video frames.`
32
+ 4. `To this end, we propose Neural Body ...`
33
+ 5. `... observations across frames can be naturally integrated ... provides geometric guidance ...`
34
+ 6. `Experiments show [main result].`
@@ -0,0 +1,28 @@
1
+ # Abstract Template C Examples (Multiple Contributions)
2
+
3
+ ```latex
4
+ % Task
5
+ %% This paper introduces a novel contour-based approach named deep snake for real-time instance segmentation.
6
+
7
+ %% Unlike some recent methods that directly regress the coordinates of the object boundary points from an image
8
+
9
+ % Introduce technical contribution and technical advantage in one sentence (this ability is very important for writing a good abstract.)
10
+ %% deep snake uses a neural network to iteratively deform an initial contour to match the object boundary, which implements the classic idea of snake algorithms with a learning-based approach.
11
+
12
+ % Introduce technical contribution and technical advantage in one sentence
13
+ %% For structured feature learning on the contour, we propose to use circular convolution in deep snake, which better exploits the cycle-graph structure of a contour compared against generic graph convolution.
14
+
15
+ % Introduce technical contribution and technical advantage in one sentence
16
+ %% Based on deep snake, we develop a two-stage pipeline for instance segmentation: initial contour proposal and contour deformation, which can handle errors in object localization.
17
+
18
+ % Experiment
19
+ ```
20
+
21
+ ## Given example pattern (Deep Snake style from your text)
22
+
23
+ 1. `This paper introduces a novel contour-based approach named deep snake for real-time instance segmentation.`
24
+ 2. `Unlike some recent methods that directly regress the coordinates of the object boundary points from an image ...`
25
+ 3. `deep snake uses a neural network to iteratively deform an initial contour ...`
26
+ 4. `For structured feature learning on the contour, we propose circular convolution ...`
27
+ 5. `Based on deep snake, we develop a two-stage pipeline ...`
28
+ 6. `Experiments show [main result].`
@@ -0,0 +1,13 @@
1
+ # Abstract Examples Index
2
+
3
+ All abstract example cites should point to the local files below.
4
+
5
+ 1. Version 1 (Challenge -> Contribution)
6
+ `Version 1: Introduce the technical challenge, then use one to two sentences to present the technical contribution that solves the challenge.`
7
+ `references/examples/abstract/template-a.md`
8
+ 2. Version 2 (Challenge -> Insight -> Contribution)
9
+ `Version 2: Introduce the technical challenge, then use one to two sentences to present the insight for solving the challenge, and then one sentence to present the technical contribution that implements this insight. (Personally recommended.)`
10
+ `references/examples/abstract/template-b.md`
11
+ 3. Version 3 (Multiple Contributions)
12
+ `Version 3: When there are multiple technical contributions, describe each contribution together with its technical advantage.`
13
+ `references/examples/abstract/template-c.md`
@@ -0,0 +1,21 @@
1
+ # Example Bank Index
2
+
3
+ Use this folder for concrete writing patterns and locally organized cite targets.
4
+
5
+ ## Files
6
+
7
+ 1. Abstract examples index: `references/examples/abstract-examples.md`
8
+ 2. Introduction examples index: `references/examples/introduction-examples.md`
9
+ 3. Abstract template files: `references/examples/abstract/template-a.md`, `references/examples/abstract/template-b.md`, `references/examples/abstract/template-c.md`
10
+ 4. Introduction task/application files: `references/examples/introduction/version-1-task-then-application.md`, `references/examples/introduction/version-2-application-first.md`, `references/examples/introduction/version-3-general-to-specific-setting.md`, `references/examples/introduction/version-4-open-with-challenge.md`
11
+ 5. Introduction technical-challenge files: `references/examples/introduction/technical-challenge-version-1-existing-task.md`, `references/examples/introduction/technical-challenge-version-2-existing-task-insight-backed-by-traditional.md`, `references/examples/introduction/technical-challenge-version-3-novel-task.md`, `references/examples/introduction/novel-task-challenge-decomposition.md`
12
+ 6. Introduction pipeline files: `references/examples/introduction/pipeline-version-1-one-contribution-multi-advantages.md`, `references/examples/introduction/pipeline-version-2-two-contributions.md`, `references/examples/introduction/pipeline-version-3-new-module-on-existing-pipeline.md`, `references/examples/introduction/pipeline-version-4-observation-driven.md`, `references/examples/introduction/pipeline-not-recommended-abstract-only.md`
13
+ 7. Method examples index: `references/examples/method-examples.md`
14
+ 8. Method detail files: `references/examples/method/pre-writing-questions.md`, `references/examples/method/module-triad-neural-body.md`, `references/examples/method/neural-body-annotated-figure-text.md`, `references/examples/method/module-design-instant-ngp.md`, `references/examples/method/module-motivation-patterns.md`, `references/examples/method/section-skeleton.md`, `references/examples/method/overview-template.md`, `references/examples/method/example-of-the-three-elements.md`, `references/examples/method/method-writing-common-issues-note.md`
15
+
16
+ ## Usage
17
+
18
+ 1. Pick one template from a section guide.
19
+ 2. Open the matching examples file.
20
+ 3. Reuse the sentence logic, not exact wording.
21
+ 4. Keep citation links in your notes for traceability.
@@ -0,0 +1,18 @@
1
+ # Introduction Novel-Task Challenge Decomposition
2
+
3
+
4
+ `For novel tasks without direct methods, decompose the challenge into clear requirement/challenge points.`
5
+
6
+ ```latex
7
+ % To achieve xx goal, several requirements must be satisfied (or several challenges must be handled).
8
+ %% Example: In this work, our goal is to build a model that captures such object intrinsics from a single image. This problem is challenging for three reasons.
9
+
10
+ % Describe point 1
11
+ %% Example: First, we only have a single image. This makes our work fundamentally different from existing works on 3D-aware image generation models [8, 9, 27, 28], which typically require a large dataset of thousands of instances for training. In comparison, the single image contains at most a few dozen instances, making the inference problem highly under-constrained.
12
+
13
+ % Describe point 2
14
+ %% Example: Second, these already limited instances may vary significantly in pixel values. This is because they have different poses and illumination conditions, but neither of these factors are annotated or known. We also cannot resort to existing tools for pose estimation based on structure from motion, such as COLMAP [35], because the appearance variations violate the assumptions of epipolar geometry.
15
+
16
+ % Describe point 3
17
+ %% Example: Finally, the object intrinsics we aim to infer are probabilistic, not deterministic: no two roses in the natural world are identical, and we want to capture a distribution of their geometry, texture, and material to exploit the underlying multi-view information.
18
+ ```
@@ -0,0 +1,30 @@
1
+ # Not Recommended: Abstract-Only Method Description in Introduction
2
+
3
+
4
+ `Not recommended: If the method is simple, do not avoid concrete method details in Introduction and only discuss abstract insight to make it look novel.`
5
+
6
+ Expert note (faithful translation):
7
+
8
+ 1. The craft of this writing template is how to make a simple pipeline look novel.
9
+ 2. Note: this is not about making the insight look novel, but about making the pipeline steps look novel.
10
+ 3. In most cases this is not recommended.
11
+ 4. The better target is to clearly explain how the core contribution is implemented in Introduction.
12
+
13
+ ```latex
14
+ % To tackle this problem, we propose a novel 3D GAN training method to generate photo-realistic images irrespective of the viewing angle.
15
+
16
+ % Introduce key idea
17
+ % Our key idea is as follows. To ease the challenging problem of learning photorealistic and multi-view consistent image synthesis, we cast the problem into two subproblems, each of which can be solved more easily.
18
+
19
+ % Explain why the key idea works, but without concretely discussing the full pipeline (or only discuss abstract benefit)
20
+ %% Example: Specifically, we formulate the problem as a combination of two simple discrimination problems, one of which learns to discriminate whether a synthesized image looks real or not, and the other learns to discriminate whether a synthesized image agrees with the camera pose. Unlike the formulations of the previous methods, which try to learn the real image distribution for each pose, or to learn pose estimation, our subproblems are much easier as each of them is analogous to a basic binary classification problem.
21
+
22
+ % Introduce pipeline modules with new terms but without clearly explaining the full pipeline (or skip concrete pipeline details)
23
+ %% Example: Based on this key idea, we propose a dual-branched discriminator, which has two branches for learning photorealism and pose consistency, respectively. As these branches are supervised explicitly for their respective purposes, high-quality images with pose consistency can be produced at each viewing angle, and consequently, the generator creates high-quality images and shapes. (This paragraph does not clearly explain how the pipeline works.)
24
+
25
+ % Introduce another contribution
26
+ %% Example: In addition, we propose a pose-matching loss to give supervision to the discriminator for the pose consistency, by considering a positive pose (i.e., rendering pose or ground truth pose) and a negative pose (i.e., irrelevant pose) for a given image. (This paragraph does not clearly explain how the pipeline works.)
27
+
28
+ % Explain expected benefit over prior methods
29
+ %% Example: For example, the frontal viewpoint is one of the irrelevant poses for a side-view image. As reported in the experiments, this loss helps improve image and shape quality. This can be interpreted as a simplification of a classification problem from a large number of classes into binary, which is composed of positive and negative pairs.
30
+ ```
@@ -0,0 +1,30 @@
1
+ # Pipeline Version 1 (One Contribution, Multiple Advantages)
2
+
3
+
4
+ `Version 1: One contribution with multiple advantages, and one teaser figure to present the basic idea.`
5
+
6
+ ```latex
7
+ % In this paper, we propose a novel framework …
8
+ %% Example: In this paper, we introduce a novel implicit neural representation for dynamic humans, named Neural Body, to solve the challenge of novel view synthesis from sparse views.
9
+ In this paper, we propose a novel framework/representation, named [method name] for [xxx task].
10
+
11
+ % Teaser for basic idea
12
+ %% Example: The basic idea is illustrated in Figure 2.
13
+ The basic idea is illustrated in [xxx Figure].
14
+
15
+ % One-sentence key novelty/contribution (very important ability)
16
+ %% Example: For the implicit fields at different frames, instead of learning them separately, Neural Body generates them from the same set of latent codes.
17
+ Our innovation is in [one sentence for key novelty].
18
+
19
+ % Method details
20
+ %% Example: Specifically, we anchor a set of latent codes to the vertices of a deformable human model (SMPL in this work), namely that their spatial locations vary with the human pose. To obtain the 3D representation at a frame, we first transform the code locations based on the human pose, which can be reliably estimated from sparse camera views. Then, a network is designed to regress the density and color for any 3D point based on these latent codes. Both the latent codes and the network are jointly learned from images of all video frames during the reconstruction.
21
+ Specifically, [how it works in detail].
22
+
23
+ % Advantage 1
24
+ %% Example: This model is inspired by the latent variable model in statistics, which enables us to effectively integrate observations at different frames.
25
+ In contrast to previous methods, [our advantage].
26
+
27
+ % Advantage 2
28
+ %% Example: Another advantage of the proposed method is that the deformable model provides a geometric prior (rough surface location) to enable more efficient learning of implicit fields.
29
+ Another advantage of the proposed method is that [another advantage].
30
+ ```
@@ -0,0 +1,34 @@
1
+ # Pipeline Version 2 (Two Contributions)
2
+
3
+
4
+ `Version 2: Two contributions, and one teaser figure to present the basic idea.`
5
+
6
+ ```latex
7
+ % In this paper, we propose a novel framework …
8
+ %% Example: In this paper, we introduce a novel implicit neural representation for dynamic humans, named Neural Body, to solve the challenge of novel view synthesis from sparse views.
9
+ In this paper, we propose a novel framework/representation, named [method name] for [xxx task].
10
+
11
+ % One-sentence key novelty
12
+ %% Example: To that end, we propose techniques to represent a given subject with rare token identifiers and fine-tune a pre-trained, diffusion-based text-to-image framework that operates in two steps; generating a low-resolution image from text and subsequently applying super-resolution (SR) diffusion models.
13
+ Our innovation is in [one sentence for key novelty].
14
+
15
+ % Teaser
16
+ %% Example: The basic idea is illustrated in Figure 2.
17
+ The basic idea is illustrated in [xxx Figure].
18
+
19
+ % Contribution 1 details
20
+ %% Example: We first fine-tune the low-resolution text-to-image model with the input images and text prompts containing a unique identifier followed by the class name of the subject (e.g., “A [V] dog”).
21
+ Specifically, [how contribution 1 works].
22
+
23
+ % Advantage of contribution 1
24
+ %% Example: This model is inspired by the latent variable model in statistics, which enables us to effectively integrate observations at different frames.
25
+ In contrast to previous methods, [advantage of contribution 1].
26
+
27
+ % Challenge motivating contribution 2
28
+ %% Example: In order to prevent overfitting and language drift [35, 40] that cause the model to associate the class name (e.g., “dog”) with the specific instance
29
+ However, [another technical challenge].
30
+
31
+ % Contribution 2 details
32
+ %% Example: we propose an autogenous, class-specific prior preservation loss, which leverages the semantic prior on the class that is embedded in the model, and encourages it to generate diverse instances of the same class as our subject.
33
+ Specifically, [how contribution 2 works].
34
+ ```
@@ -0,0 +1,18 @@
1
+ # Pipeline Version 3 (New Module on Existing Pipeline)
2
+
3
+
4
+ `Version 3: Build on a prior pipeline and introduce one new module, with a teaser figure for the basic idea.`
5
+
6
+ ```latex
7
+ % In this paper, we propose a learning-based snake algorithm, named deep snake, for real-time instance segmentation.
8
+
9
+ % Inspired by previous methods [21, 25], deep snake takes an initial contour as input and deforms it by regressing vertex-wise offsets.
10
+
11
+ % Our innovation is introducing the circular convolution for efficient feature learning on a contour, as illustrated in Figure 1.
12
+
13
+ % We observe that the contour is a cycle graph that consists of a sequence of vertices connected in a closed cycle. Since every vertex has the same degree equal to two, we can apply the standard 1D convolution on the vertex features.
14
+
15
+ % Considering that the contour is periodic, deep snake introduces the circular convolution, which indicates that an aperiodic function (1D kernel) is convolved in the standard way with a periodic function (features defined on the contour).
16
+
17
+ % The kernel of circular convolution encodes not only the feature of each vertex but also the relationship among neighboring vertices. In contrast, the generic GCN performs pooling to aggregate information from neighboring vertices. The kernel function in our circular convolution amounts to a learnable aggregation function, which is more expressive and results in better performance than using a generic GCN, as demonstrated by our experimental results in Section 5.2.
18
+ ```
@@ -0,0 +1,16 @@
1
+ # Pipeline Version 4 (Observation-Driven Contribution)
2
+
3
+
4
+ `Version 4: Contribution comes from one important observation. Introduce key innovation first, then intuitive observation as motivation, then method details, then benefits.`
5
+
6
+ ```latex
7
+ % In this paper, we propose a learning-based snake algorithm, named deep snake, for real-time instance segmentation.
8
+
9
+ % Our innovation is introducing the circular convolution for efficient feature learning on a contour, as illustrated in Figure 1.
10
+
11
+ % We observe that the contour is a cycle graph that consists of a sequence of vertices connected in a closed cycle. Since every vertex has the same degree equal to two, we can apply the standard 1D convolution on the vertex features.
12
+
13
+ % Considering that the contour is periodic, deep snake introduces the circular convolution, which indicates that an aperiodic function (1D kernel) is convolved in the standard way with a periodic function (features defined on the contour).
14
+
15
+ % The kernel of circular convolution encodes not only the feature of each vertex but also the relationship among neighboring vertices. In contrast, the generic GCN performs pooling to aggregate information from neighboring vertices. The kernel function in our circular convolution amounts to a learnable aggregation function, which is more expressive and results in better performance than using a generic GCN, as demonstrated by our experimental results in Section 5.2.
16
+ ```
@@ -0,0 +1,32 @@
1
+ # Technical Challenge Version 1 (Existing Task, Existing Methods)
2
+
3
+
4
+ `Version 1: For existing tasks with existing methods, discuss the challenge chain from traditional methods to recent methods and finally to the challenge we solve.`
5
+
6
+ ```latex
7
+ % Discuss general technical challenges of this task (to lead into recent methods)
8
+ %% Example 1: This problem is quite challenging from many perspectives, including object detection under severe occlusions, variations in lighting and appearance, and cluttered background objects.
9
+ %% Example 2: This problem is particularly challenging due to the inherent ambiguity on acquiring human geometry, materials and motions from images.
10
+ This problem is particularly challenging due to several factors, including [xxx reason], [xxx reason], and [xxx reason].
11
+
12
+ % Briefly introduce one class of traditional methods, then discuss their technical challenge
13
+ %% Example: Traditional methods have shown that pose estimation can be achieved by establishing the correspondences between an object image and the object model.
14
+ To overcome these challenges, traditional methods [how they work], [what they achieve].
15
+
16
+ %% Example: They rely on hand-crafted features, which are not robust to image variations and background clutters.
17
+ However, they [technical challenge they face].
18
+
19
+ % Briefly introduce one class of recent methods 1 (optional), then discuss their challenge
20
+ %% Example: Deep learning based methods train end-to-end neural networks that take an image as input and output its corresponding pose.
21
+ Recently, [xxx methods] [how they work], [what they achieve].
22
+
23
+ %% Example: However, generalization remains as an issue, as it is unclear that such end-to-end methods learn sufficient feature representations for pose estimation.
24
+ However, they [limitation], because [xxx technical reason].
25
+
26
+ % Briefly introduce one class of recent methods 2, then discuss their challenge (must lead to our solved challenge)
27
+ %% Example: Some recent methods use CNNs to first regress 2D keypoints and then compute 6D pose parameters using the Perspective-n-Point (PnP) algorithm. In other words, the detected keypoints serve as an intermediate representation for pose estimation. Such two-stage approaches achieve state-of-the-art performance, thanks to robust detection of keypoints.
28
+ To overcome this challenge, [xxx methods] [how they work], [what they achieve].
29
+
30
+ %% Example: However, these methods have difficulty in tackling occluded and truncated objects, since part of their keypoints are invisible. Although CNNs may predict these unseen keypoints by memorizing similar patterns, generalization remains difficult.
31
+ However, they [limitation], because [xxx technical reason].
32
+ ```
@@ -0,0 +1,33 @@
1
+ # Technical Challenge Version 2 (Existing Task, Insight Backed by Traditional Methods)
2
+
3
+
4
+ `Version 2: For existing tasks, if our technical insight was used in traditional methods, discuss that line to provide conceptual backing.`
5
+
6
+ ```latex
7
+ % Introduce one class of traditional/recent methods and discuss their technical challenge (to lead to our insight)
8
+ %% Example (Deep Snake): Most of the state-of-the-art instance segmentation methods perform pixel-wise segmentation within a bounding box given by an object detector.
9
+ %% Example (ManhattanSDF): Given input images, traditional methods generally estimate the depth map for each image based on the multi-view stereo (MVS) algorithms and then fuse estimated depth maps into 3D models.
10
+ Traditional/recent methods [how they work], [what they achieve].
11
+
12
+ %% Example (Deep Snake): They may be sensitive to the inaccurate bounding box. Moreover, representing an object shape as dense binary pixels generally results in costly post-processing.
13
+ %% Example (ManhattanSDF): Although these methods achieve successful reconstruction in most cases, they have difficulty in handling low-textured regions, e.g., floors and walls of indoor scenes, due to the unreliable stereo matching in these regions.
14
+ However, they [limitation], because [xxx technical reason].
15
+
16
+ % Discuss traditional methods that used an insight similar to ours (implicitly backing our idea)
17
+ %% Example (Deep Snake): An alternative shape representation is the object contour, which is a set of vertices along the object silhouette. In contrast to pixel-based representation, a contour is not limited within a bounding box and has fewer parameters. Such a contour-based representation has long been used in image segmentation since the seminal work by Kass et al., which is well known as snakes or active contours.
18
+ %% Example (ManhattanSDF): To improve the reconstruction of low-textured regions, a typical approach is leveraging the planar prior of manmade scenes, which has long been explored in literature. A renowned example is the Manhattanworld assumption, i.e., the surfaces of man-made scenes should be aligned with three dominant directions.
19
+ To overcome this problem, a typical approach is [xxx insight], which has long been explored in literature.
20
+
21
+ These methods [how they work].
22
+
23
+ %% Example (Deep Snake): While many variants have been developed in literature, these methods are prone to local optima as the objective functions are handcrafted and typically nonconvex.
24
+ %% Example (ManhattanSDF): However, all of them focus on optimizing per-view depth maps instead of the full scene models in 3D space. As a result, depth estimation and plane segmentation could still be inconsistent among views, yielding suboptimal reconstruction quality as demonstrated by our experimental results in Section 5.3.
25
+ However, they [limitation], because [xxx technical reason].
26
+
27
+ % Then discuss newer methods and their remaining challenge (must lead to our solved challenge)
28
+ %% Example: There is a recent trend to represent 3D scenes as implicit neural representations and learn the representations from images with differentiable renderers. In particular, [49, 54, 55] use a signed distance field (SDF) to represent the scene and render it into images based on the sphere tracing or volume rendering. Thanks to the well-defined surfaces of SDFs, they recover high-quality 3D geometries from images.
29
+ To overcome this challenge, [xxx methods] [how they work], [what they achieve].
30
+
31
+ %% Example: However, these methods essentially rely on the multi-view photometric consistency to learn the SDFs. So they still suffer from poor performance in low-textured planar regions, as shown in Figure 1, as many plausible solutions may satisfy the photometric constraint in low-textured planar regions.
32
+ However, they [limitation], because [xxx technical reason].
33
+ ```
@@ -0,0 +1,21 @@
1
+ # Technical Challenge Version 3 (Novel Task)
2
+
3
+
4
+ `Version 3: For novel tasks without direct methods, define the challenge directly and decompose it by requirement/challenge points.`
5
+
6
+ ```latex
7
+ % To achieve xx goal, several requirements/challenges must be satisfied.
8
+ %% Example: In this work, our goal is to build a model that captures such object intrinsics from a single image. This problem is challenging for three reasons.
9
+
10
+ % Describe point 1
11
+ %% Example: First, we only have a single image. This makes our work fundamentally different from existing works on 3D-aware image generation models [8, 9, 27, 28], which typically require a large dataset of thousands of instances for training. In comparison, the single image contains at most a few dozen instances, making the inference problem highly under-constrained.
12
+
13
+ % Describe point 2
14
+ %% Example: Second, these already limited instances may vary significantly in pixel values. This is because they have different poses and illumination conditions, but neither of these factors are annotated or known. We also cannot resort to existing tools for pose estimation based on structure from motion, such as COLMAP [35], because the appearance variations violate the assumptions of epipolar geometry.
15
+
16
+ % Describe point 3
17
+ %% Example: Finally, the object intrinsics we aim to infer are probabilistic, not deterministic: no two roses in the natural world are identical, and we want to capture a distribution of their geometry, texture, and material to exploit the underlying multi-view information.
18
+ ```
19
+
20
+ See also:
21
+ 1. `references/examples/introduction/novel-task-challenge-decomposition.md`
@@ -0,0 +1,14 @@
1
+ # Introduction Version 1: Task First, Then Application
2
+
3
+
4
+ `Version 1: If the task is relatively niche, introduce the task first, then introduce applications.`
5
+
6
+ ```latex
7
+ % Introduce Task (if the task is very familiar, this part can be skipped)
8
+ %% Example: Object pose estimation aims to estimate object's orientation and translation relative to a canonical frame from a single image.
9
+ [xxx task] targets at recovering/reconstructing/estimating [xxx output] from [xxx input].
10
+
11
+ % Introduce Application
12
+ %% Example: Accurate pose estimation is essential for a variety of applications such as augmented reality, autonomous driving and robotic manipulation.
13
+ [xxx task] has a variety of applications such as [xxx], [xxx], and [xxx].
14
+ ```
@@ -0,0 +1,10 @@
1
+ # Introduction Version 2: Application First
2
+
3
+
4
+ `Version 2: If the task is already familiar to most readers, introduce applications directly.`
5
+
6
+ ```latex
7
+ % Introduce Application
8
+ %% Example: Accurate pose estimation is essential for a variety of applications such as augmented reality, autonomous driving and robotic manipulation.
9
+ [xxx task] has a variety of applications such as [xxx], [xxx], and [xxx].
10
+ ```
@@ -0,0 +1,14 @@
1
+ # Introduction Version 3: General Application -> Specific Setting
2
+
3
+
4
+ `Version 3: Introduce applications of the general task first, then introduce the specific task setting. (Personally recommended when the setting is relatively new.)`
5
+
6
+ ```latex
7
+ % Introduce applications of the general task
8
+ %% Example: Accurate pose estimation is essential for a variety of applications such as augmented reality, autonomous driving and robotic manipulation.
9
+ [xxx task] has a variety of applications such as [xxx], [xxx], and [xxx].
10
+
11
+ % Introduce the specific task setting
12
+ %% Example: This paper focuses on the specific setting of recovering the 6DoF pose of an object, i.e., rotation and translation in 3D, from a single RGB image of that object.
13
+ This paper focuses on the specific setting of recovering/reconstructing/estimating [xxx output] from [xxx input].
14
+ ```