arscontexta 0.6.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (418) hide show
  1. package/.claude-plugin/marketplace.json +11 -0
  2. package/.claude-plugin/plugin.json +22 -0
  3. package/README.md +683 -0
  4. package/agents/knowledge-guide.md +49 -0
  5. package/bin/cli.mjs +66 -0
  6. package/generators/agents-md.md +240 -0
  7. package/generators/claude-md.md +379 -0
  8. package/generators/features/atomic-notes.md +124 -0
  9. package/generators/features/ethical-guardrails.md +58 -0
  10. package/generators/features/graph-analysis.md +188 -0
  11. package/generators/features/helper-functions.md +92 -0
  12. package/generators/features/maintenance.md +164 -0
  13. package/generators/features/methodology-knowledge.md +70 -0
  14. package/generators/features/mocs.md +144 -0
  15. package/generators/features/multi-domain.md +61 -0
  16. package/generators/features/personality.md +71 -0
  17. package/generators/features/processing-pipeline.md +428 -0
  18. package/generators/features/schema.md +149 -0
  19. package/generators/features/self-evolution.md +229 -0
  20. package/generators/features/self-space.md +78 -0
  21. package/generators/features/semantic-search.md +99 -0
  22. package/generators/features/session-rhythm.md +85 -0
  23. package/generators/features/templates.md +85 -0
  24. package/generators/features/wiki-links.md +88 -0
  25. package/generators/soul-md.md +121 -0
  26. package/hooks/hooks.json +45 -0
  27. package/hooks/scripts/auto-commit.sh +44 -0
  28. package/hooks/scripts/session-capture.sh +35 -0
  29. package/hooks/scripts/session-orient.sh +86 -0
  30. package/hooks/scripts/write-validate.sh +42 -0
  31. package/methodology/AI shifts knowledge systems from externalizing memory to externalizing attention.md +59 -0
  32. package/methodology/BM25 retrieval fails on full-length descriptions because query term dilution reduces match scores.md +39 -0
  33. package/methodology/IBIS framework maps claim-based architecture to structured argumentation.md +58 -0
  34. package/methodology/LLM attention degrades as context fills.md +49 -0
  35. package/methodology/MOC construction forces synthesis that automated generation from metadata cannot replicate.md +49 -0
  36. package/methodology/MOC maintenance investment compounds because orientation savings multiply across every future session.md +41 -0
  37. package/methodology/MOCs are attention management devices not just organizational tools.md +51 -0
  38. package/methodology/PKM failure follows a predictable cycle.md +50 -0
  39. package/methodology/ThreadMode to DocumentMode transformation is the core value creation step.md +52 -0
  40. package/methodology/WIP limits force processing over accumulation.md +53 -0
  41. package/methodology/Zeigarnik effect validates capture-first philosophy because open loops drain attention.md +42 -0
  42. package/methodology/academic research uses structured extraction with cross-source synthesis.md +566 -0
  43. package/methodology/adapt the four-phase processing pipeline to domain-specific throughput needs.md +197 -0
  44. package/methodology/agent notes externalize navigation intuition that search cannot discover and traversal cannot reconstruct.md +48 -0
  45. package/methodology/agent self-memory should be architecturally separate from user knowledge systems.md +48 -0
  46. package/methodology/agent session boundaries create natural automation checkpoints that human-operated systems lack.md +56 -0
  47. package/methodology/agent-cognition.md +107 -0
  48. package/methodology/agents are simultaneously methodology executors and subjects creating a unique trust asymmetry.md +66 -0
  49. package/methodology/aspect-oriented programming solved the same cross-cutting concern problem that hooks solve.md +39 -0
  50. package/methodology/associative ontologies beat hierarchical taxonomies because heterarchy adapts while hierarchy brittles.md +53 -0
  51. package/methodology/attention residue may have a minimum granularity that cannot be subdivided.md +46 -0
  52. package/methodology/auto-commit hooks eliminate prospective memory failures by converting remember-to-act into guaranteed execution.md +47 -0
  53. package/methodology/automated detection is always safe because it only reads state while automated remediation risks content corruption.md +42 -0
  54. package/methodology/automation should be retired when its false positive rate exceeds its true positive rate or it catches zero issues.md +56 -0
  55. package/methodology/backlinks implicitly define notes by revealing usage context.md +35 -0
  56. package/methodology/backward maintenance asks what would be different if written today.md +62 -0
  57. package/methodology/balance onboarding enforcement and questions to prevent premature complexity.md +229 -0
  58. package/methodology/basic level categorization determines optimal MOC granularity.md +51 -0
  59. package/methodology/batching by context similarity reduces switching costs in agent processing.md +43 -0
  60. package/methodology/behavioral anti-patterns matter more than tool selection.md +42 -0
  61. package/methodology/betweenness centrality identifies bridge notes connecting disparate knowledge domains.md +57 -0
  62. package/methodology/blueprints that teach construction outperform downloads that provide pre-built code for platform-dependent modules.md +42 -0
  63. package/methodology/bootstrapping principle enables self-improving systems.md +62 -0
  64. package/methodology/build automatic memory through cognitive offloading and session handoffs.md +285 -0
  65. package/methodology/capture the reaction to content not just the content itself.md +41 -0
  66. package/methodology/claims must be specific enough to be wrong.md +36 -0
  67. package/methodology/closure rituals create clean breaks that prevent attention residue bleed.md +44 -0
  68. package/methodology/cognitive offloading is the architectural foundation for vault design.md +46 -0
  69. package/methodology/cognitive outsourcing risk in agent-operated systems.md +55 -0
  70. package/methodology/coherence maintains consistency despite inconsistent inputs.md +96 -0
  71. package/methodology/coherent architecture emerges from wiki links spreading activation and small-world topology.md +48 -0
  72. package/methodology/community detection algorithms can inform when MOCs should split or merge.md +52 -0
  73. package/methodology/complete navigation requires four complementary types that no single mechanism provides.md +43 -0
  74. package/methodology/complex systems evolve from simple working systems.md +59 -0
  75. package/methodology/composable knowledge architecture builds systems from independent toggleable modules not monolithic templates.md +61 -0
  76. package/methodology/compose multi-domain systems through separate templates and shared graph.md +372 -0
  77. package/methodology/concept-orientation beats source-orientation for cross-domain connections.md +51 -0
  78. package/methodology/confidence thresholds gate automated action between the mechanical and judgment zones.md +50 -0
  79. package/methodology/configuration dimensions interact so choices in one create pressure on others.md +58 -0
  80. package/methodology/configuration paralysis emerges when derivation surfaces too many decisions.md +44 -0
  81. package/methodology/context files function as agent operating systems through self-referential self-extension.md +46 -0
  82. package/methodology/context phrase clarity determines how deep a navigation hierarchy can scale.md +46 -0
  83. package/methodology/continuous small-batch processing eliminates review dread.md +48 -0
  84. package/methodology/controlled disorder engineers serendipity through semantic rather than topical linking.md +51 -0
  85. package/methodology/creative writing uses worldbuilding consistency with character tracking.md +672 -0
  86. package/methodology/cross-links between MOC territories indicate creative leaps and integration depth.md +43 -0
  87. package/methodology/dangling links reveal which notes want to exist.md +62 -0
  88. package/methodology/data exit velocity measures how quickly content escapes vendor lock-in.md +74 -0
  89. package/methodology/decontextualization risk means atomicity may strip meaning that cannot be recovered.md +48 -0
  90. package/methodology/dense interlinked research claims enable derivation while sparse references only enable templating.md +47 -0
  91. package/methodology/dependency resolution through topological sort makes module composition transparent and verifiable.md +56 -0
  92. package/methodology/derivation generates knowledge systems from composable research claims not template customization.md +63 -0
  93. package/methodology/derivation-engine.md +27 -0
  94. package/methodology/derived systems follow a seed-evolve-reseed lifecycle.md +56 -0
  95. package/methodology/description quality for humans diverges from description quality for keyword search.md +73 -0
  96. package/methodology/descriptions are retrieval filters not summaries.md +112 -0
  97. package/methodology/design MOCs as attention management devices with lifecycle governance.md +318 -0
  98. package/methodology/design-dimensions.md +66 -0
  99. package/methodology/digital mutability enables note evolution that physical permanence forbids.md +54 -0
  100. package/methodology/discovery-retrieval.md +48 -0
  101. package/methodology/distinctiveness scoring treats description quality as measurable.md +69 -0
  102. package/methodology/does agent processing recover what fast capture loses.md +43 -0
  103. package/methodology/domain-compositions.md +37 -0
  104. package/methodology/dual-coding with visual elements could enhance agent traversal.md +55 -0
  105. package/methodology/each module must be describable in one sentence under 200 characters or it does too many things.md +45 -0
  106. package/methodology/each new note compounds value by creating traversal paths.md +55 -0
  107. package/methodology/eight configuration dimensions parameterize the space of possible knowledge systems.md +56 -0
  108. package/methodology/elaborative encoding is the quality gate for new notes.md +55 -0
  109. package/methodology/enforce schema with graduated strictness across capture processing and query zones.md +221 -0
  110. package/methodology/enforcing atomicity can create paralysis when ideas resist decomposition.md +43 -0
  111. package/methodology/engineering uses technical decision tracking with architectural memory.md +766 -0
  112. package/methodology/every knowledge domain shares a four-phase processing skeleton that diverges only in the process step.md +53 -0
  113. package/methodology/evolution observations provide actionable signals for system adaptation.md +67 -0
  114. package/methodology/external memory shapes cognition more than base model.md +60 -0
  115. package/methodology/faceted classification treats notes as multi-dimensional objects rather than folder contents.md +65 -0
  116. package/methodology/failure-modes.md +27 -0
  117. package/methodology/false universalism applies same processing logic regardless of domain.md +49 -0
  118. package/methodology/federated wiki pattern enables multi-agent divergence as feature not bug.md +59 -0
  119. package/methodology/flat files break at retrieval scale.md +75 -0
  120. package/methodology/forced engagement produces weak connections.md +48 -0
  121. package/methodology/four abstraction layers separate platform-agnostic from platform-dependent knowledge system features.md +47 -0
  122. package/methodology/fresh context per task preserves quality better than chaining phases.md +44 -0
  123. package/methodology/friction reveals architecture.md +63 -0
  124. package/methodology/friction-driven module adoption prevents configuration debt by adding complexity only at pain points.md +48 -0
  125. package/methodology/gardening cycle implements tend prune fertilize operations.md +41 -0
  126. package/methodology/generation effect gate blocks processing without transformation.md +40 -0
  127. package/methodology/goal-driven memory orchestration enables autonomous domain learning through directed compute allocation.md +41 -0
  128. package/methodology/good descriptions layer heuristic then mechanism then implication.md +57 -0
  129. package/methodology/graph-structure.md +65 -0
  130. package/methodology/guided notes might outperform post-hoc structuring for high-volume capture.md +37 -0
  131. package/methodology/health wellness uses symptom-trigger correlation with multi-dimensional tracking.md +819 -0
  132. package/methodology/hook composition creates emergent methodology from independent single-concern components.md +47 -0
  133. package/methodology/hook enforcement guarantees quality while instruction enforcement merely suggests it.md +51 -0
  134. package/methodology/hook-driven learning loops create self-improving methodology through observation accumulation.md +62 -0
  135. package/methodology/hooks are the agent habit system that replaces the missing basal ganglia.md +40 -0
  136. package/methodology/hooks cannot replace genuine cognitive engagement yet more automation is always tempting.md +87 -0
  137. package/methodology/hooks enable context window efficiency by delegating deterministic checks to external processes.md +47 -0
  138. package/methodology/idempotent maintenance operations are safe to automate because running them twice produces the same result as running them once.md +44 -0
  139. package/methodology/implement condition-based maintenance triggers for derived systems.md +255 -0
  140. package/methodology/implicit dependencies create distributed monoliths that fail silently across configurations.md +58 -0
  141. package/methodology/implicit knowledge emerges from traversal.md +55 -0
  142. package/methodology/incremental formalization happens through repeated touching of old notes.md +60 -0
  143. package/methodology/incremental reading enables cross-source connection finding.md +39 -0
  144. package/methodology/index.md +32 -0
  145. package/methodology/inline links carry richer relationship data than metadata fields.md +91 -0
  146. package/methodology/insight accretion differs from productivity in knowledge systems.md +41 -0
  147. package/methodology/intermediate packets enable assembly over creation.md +52 -0
  148. package/methodology/intermediate representation pattern enables reliable vault operations beyond regex.md +62 -0
  149. package/methodology/justification chains enable forward backward and evolution reasoning about configuration decisions.md +46 -0
  150. package/methodology/knowledge system architecture is parameterized by platform capabilities not fixed by methodology.md +51 -0
  151. package/methodology/knowledge systems become communication partners through complexity and memory humans cannot sustain.md +47 -0
  152. package/methodology/knowledge systems share universal operations and structural components across all methodology traditions.md +46 -0
  153. package/methodology/legal case management uses precedent chains with regulatory change propagation.md +892 -0
  154. package/methodology/live index via periodic regeneration keeps discovery current.md +58 -0
  155. package/methodology/local-first file formats are inherently agent-native.md +69 -0
  156. package/methodology/logic column pattern separates reasoning from procedure.md +35 -0
  157. package/methodology/maintenance operations are more universal than creative pipelines because structural health is domain-invariant.md +47 -0
  158. package/methodology/maintenance scheduling frequency should match consequence speed not detection capability.md +50 -0
  159. package/methodology/maintenance targeting should prioritize mechanism and theory notes.md +26 -0
  160. package/methodology/maintenance-patterns.md +72 -0
  161. package/methodology/markdown plus YAML plus ripgrep implements a queryable graph database without infrastructure.md +55 -0
  162. package/methodology/maturity field enables agent context prioritization.md +33 -0
  163. package/methodology/memory-architecture.md +27 -0
  164. package/methodology/metacognitive confidence can diverge from retrieval capability.md +42 -0
  165. package/methodology/metadata reduces entropy enabling precision over recall.md +91 -0
  166. package/methodology/methodology development should follow the trajectory from documentation to skill to hook as understanding hardens.md +80 -0
  167. package/methodology/methodology traditions are named points in a shared configuration space not competing paradigms.md +64 -0
  168. package/methodology/mnemonic medium embeds verification into navigation.md +46 -0
  169. package/methodology/module communication through shared YAML fields creates loose coupling without direct dependencies.md +44 -0
  170. package/methodology/module deactivation must account for structural artifacts that survive the toggle.md +49 -0
  171. package/methodology/multi-domain systems compose through separate templates and shared graph.md +61 -0
  172. package/methodology/multi-domain-composition.md +27 -0
  173. package/methodology/narrow folksonomy optimizes for single-operator retrieval unlike broad consensus tagging.md +53 -0
  174. package/methodology/navigation infrastructure passes through distinct scaling regimes that require qualitative strategy shifts.md +48 -0
  175. package/methodology/navigational vertigo emerges in pure association systems without local hierarchy.md +54 -0
  176. package/methodology/note titles should function as APIs enabling sentence transclusion.md +51 -0
  177. package/methodology/note-design.md +57 -0
  178. package/methodology/notes are skills /342/200/224 curated knowledge injected when relevant.md" +62 -0
  179. package/methodology/notes function as cognitive anchors that stabilize attention during complex tasks.md +41 -0
  180. package/methodology/novel domains derive by mapping knowledge type to closest reference domain then adapting.md +50 -0
  181. package/methodology/nudge theory explains graduated hook enforcement as choice architecture for agents.md +59 -0
  182. package/methodology/observation and tension logs function as dead-letter queues for failed automation.md +51 -0
  183. package/methodology/operational memory and knowledge memory serve different functions in agent architecture.md +48 -0
  184. package/methodology/operational wisdom requires contextual observation.md +52 -0
  185. package/methodology/orchestrated vault creation transforms arscontexta from tool to autonomous knowledge factory.md +40 -0
  186. package/methodology/organic emergence versus active curation creates a fundamental vault governance tension.md +68 -0
  187. package/methodology/orphan notes are seeds not failures.md +38 -0
  188. package/methodology/over-automation corrupts quality when hooks encode judgment rather than verification.md +62 -0
  189. package/methodology/people relationships uses Dunbar-layered graphs with interaction tracking.md +659 -0
  190. package/methodology/personal assistant uses life area management with review automation.md +610 -0
  191. package/methodology/platform adapter translation is semantic not mechanical because hook event meanings differ.md +40 -0
  192. package/methodology/platform capability tiers determine which knowledge system features can be implemented.md +48 -0
  193. package/methodology/platform fragmentation means identical conceptual operations require different implementations across agent environments.md +44 -0
  194. package/methodology/premature complexity is the most common derivation failure mode.md +45 -0
  195. package/methodology/prevent domain-specific failure modes through the vulnerability matrix.md +336 -0
  196. package/methodology/processing effort should follow retrieval demand.md +57 -0
  197. package/methodology/processing-workflows.md +75 -0
  198. package/methodology/product management uses feedback pipelines with experiment tracking.md +789 -0
  199. package/methodology/productivity porn risk in meta-system building.md +30 -0
  200. package/methodology/programmable notes could enable property-triggered workflows.md +64 -0
  201. package/methodology/progressive disclosure means reading right not reading less.md +69 -0
  202. package/methodology/progressive schema validates only what active modules require not the full system schema.md +49 -0
  203. package/methodology/project management uses decision tracking with stakeholder context.md +776 -0
  204. package/methodology/propositional link semantics transform wiki links from associative to reasoned.md +87 -0
  205. package/methodology/prospective memory requires externalization.md +53 -0
  206. package/methodology/provenance tracks where beliefs come from.md +62 -0
  207. package/methodology/queries evolve during search so agents should checkpoint.md +35 -0
  208. package/methodology/question-answer metadata enables inverted search patterns.md +39 -0
  209. package/methodology/random note resurfacing prevents write-only memory.md +33 -0
  210. package/methodology/reconciliation loops that compare desired state to actual state enable drift correction without continuous monitoring.md +59 -0
  211. package/methodology/reflection synthesizes existing notes into new insight.md +100 -0
  212. package/methodology/retrieval utility should drive design over capture completeness.md +69 -0
  213. package/methodology/retrieval verification loop tests description quality at scale.md +81 -0
  214. package/methodology/role field makes graph structure explicit.md +94 -0
  215. package/methodology/scaffolding enables divergence that fine-tuning cannot.md +67 -0
  216. package/methodology/schema enforcement via validation agents enables soft consistency.md +60 -0
  217. package/methodology/schema evolution follows observe-then-formalize not design-then-enforce.md +65 -0
  218. package/methodology/schema field names are the only domain specific element in the universal note pattern.md +46 -0
  219. package/methodology/schema fields should use domain-native vocabulary not abstract terminology.md +47 -0
  220. package/methodology/schema templates reduce cognitive overhead at capture time.md +55 -0
  221. package/methodology/schema validation hooks externalize inhibitory control that degrades under cognitive load.md +48 -0
  222. package/methodology/schema-enforcement.md +27 -0
  223. package/methodology/self-extension requires context files to contain platform operations knowledge not just methodology.md +47 -0
  224. package/methodology/sense-making vs storage does compression lose essential nuance.md +73 -0
  225. package/methodology/session boundary hooks implement cognitive bookends for orientation and reflection.md +60 -0
  226. package/methodology/session handoff creates continuity without persistent memory.md +43 -0
  227. package/methodology/session outputs are packets for future selves.md +43 -0
  228. package/methodology/session transcript mining enables experiential validation that structural tests cannot provide.md +38 -0
  229. package/methodology/skill context budgets constrain knowledge system complexity on agent platforms.md +52 -0
  230. package/methodology/skills encode methodology so manual execution bypasses quality gates.md +50 -0
  231. package/methodology/small-world topology requires hubs and dense local links.md +99 -0
  232. package/methodology/source attribution enables tracing claims to foundations.md +38 -0
  233. package/methodology/spaced repetition scheduling could optimize vault maintenance.md +44 -0
  234. package/methodology/spreading activation models how agents should traverse.md +79 -0
  235. package/methodology/stale navigation actively misleads because agents trust curated maps completely.md +43 -0
  236. package/methodology/stigmergy coordinates agents through environmental traces without direct communication.md +62 -0
  237. package/methodology/storage versus thinking distinction determines which tool patterns apply.md +56 -0
  238. package/methodology/structure enables navigation without reading everything.md +52 -0
  239. package/methodology/structure without processing provides no value.md +56 -0
  240. package/methodology/student learning uses prerequisite graphs with spaced retrieval.md +770 -0
  241. package/methodology/summary coherence tests composability before filing.md +37 -0
  242. package/methodology/tag rot applies to wiki links because titles serve as both identifier and display text.md +50 -0
  243. package/methodology/temporal media must convert to spatial text for agent traversal.md +43 -0
  244. package/methodology/temporal processing priority creates age-based inbox urgency.md +45 -0
  245. package/methodology/temporal separation of capture and processing preserves context freshness.md +39 -0
  246. package/methodology/ten universal primitives form the kernel of every viable agent knowledge system.md +162 -0
  247. package/methodology/testing effect could enable agent knowledge verification.md +38 -0
  248. package/methodology/the AgentSkills standard embodies progressive disclosure at the skill level.md +40 -0
  249. package/methodology/the derivation engine improves recursively as deployed systems generate observations.md +49 -0
  250. package/methodology/the determinism boundary separates hook methodology from skill methodology.md +46 -0
  251. package/methodology/the fix-versus-report decision depends on determinism reversibility and accumulated trust.md +45 -0
  252. package/methodology/the generation effect requires active transformation not just storage.md +57 -0
  253. package/methodology/the no wrong patches guarantee ensures any valid module combination produces a valid system.md +58 -0
  254. package/methodology/the system is the argument.md +46 -0
  255. package/methodology/the vault constitutes identity for agents.md +86 -0
  256. package/methodology/the vault methodology transfers because it encodes cognitive science not domain specifics.md +47 -0
  257. package/methodology/therapy journal uses warm personality with pattern detection for emotional processing.md +584 -0
  258. package/methodology/three capture schools converge through agent-mediated synthesis.md +55 -0
  259. package/methodology/three concurrent maintenance loops operate at different timescales to catch different classes of problems.md +56 -0
  260. package/methodology/throughput matters more than accumulation.md +58 -0
  261. package/methodology/title as claim enables traversal as reasoning.md +50 -0
  262. package/methodology/topological organization beats temporal for knowledge work.md +52 -0
  263. package/methodology/trading uses conviction tracking with thesis-outcome correlation.md +699 -0
  264. package/methodology/trails transform ephemeral navigation into persistent artifacts.md +39 -0
  265. package/methodology/transform universal vocabulary to domain-native language through six levels.md +259 -0
  266. package/methodology/type field enables structured queries without folder hierarchies.md +53 -0
  267. package/methodology/use-case presets dissolve the tension between composability and simplicity.md +44 -0
  268. package/methodology/vault conventions may impose hidden rigidity on thinking.md +44 -0
  269. package/methodology/verbatim risk applies to agents too.md +31 -0
  270. package/methodology/vibe notetaking is the emerging industry consensus for AI-native self-organization.md +56 -0
  271. package/methodology/vivid memories need verification.md +45 -0
  272. package/methodology/vocabulary-transformation.md +27 -0
  273. package/methodology/voice capture is the highest-bandwidth channel for agent-delegated knowledge systems.md +45 -0
  274. package/methodology/wiki links are the digital evolution of analog indexing.md +73 -0
  275. package/methodology/wiki links as social contract transforms agents into stewards of incomplete references.md +52 -0
  276. package/methodology/wiki links create navigation paths that shape retrieval.md +63 -0
  277. package/methodology/wiki links implement GraphRAG without the infrastructure.md +101 -0
  278. package/methodology/writing for audience blocks authentic creation.md +22 -0
  279. package/methodology/you operate a system that takes notes.md +79 -0
  280. package/openclaw/SKILL.md +110 -0
  281. package/package.json +45 -0
  282. package/platforms/README.md +51 -0
  283. package/platforms/claude-code/generator.md +61 -0
  284. package/platforms/claude-code/hooks/README.md +186 -0
  285. package/platforms/claude-code/hooks/auto-commit.sh.template +38 -0
  286. package/platforms/claude-code/hooks/session-capture.sh.template +72 -0
  287. package/platforms/claude-code/hooks/session-orient.sh.template +189 -0
  288. package/platforms/claude-code/hooks/write-validate.sh.template +106 -0
  289. package/platforms/openclaw/generator.md +82 -0
  290. package/platforms/openclaw/hooks/README.md +89 -0
  291. package/platforms/openclaw/hooks/bootstrap.ts.template +224 -0
  292. package/platforms/openclaw/hooks/command-new.ts.template +165 -0
  293. package/platforms/openclaw/hooks/heartbeat.ts.template +214 -0
  294. package/platforms/shared/features/README.md +70 -0
  295. package/platforms/shared/skill-blocks/graph.md +145 -0
  296. package/platforms/shared/skill-blocks/learn.md +119 -0
  297. package/platforms/shared/skill-blocks/next.md +131 -0
  298. package/platforms/shared/skill-blocks/pipeline.md +326 -0
  299. package/platforms/shared/skill-blocks/ralph.md +616 -0
  300. package/platforms/shared/skill-blocks/reduce.md +1142 -0
  301. package/platforms/shared/skill-blocks/refactor.md +129 -0
  302. package/platforms/shared/skill-blocks/reflect.md +780 -0
  303. package/platforms/shared/skill-blocks/remember.md +524 -0
  304. package/platforms/shared/skill-blocks/rethink.md +574 -0
  305. package/platforms/shared/skill-blocks/reweave.md +680 -0
  306. package/platforms/shared/skill-blocks/seed.md +320 -0
  307. package/platforms/shared/skill-blocks/stats.md +145 -0
  308. package/platforms/shared/skill-blocks/tasks.md +171 -0
  309. package/platforms/shared/skill-blocks/validate.md +323 -0
  310. package/platforms/shared/skill-blocks/verify.md +562 -0
  311. package/platforms/shared/templates/README.md +35 -0
  312. package/presets/experimental/categories.yaml +1 -0
  313. package/presets/experimental/preset.yaml +38 -0
  314. package/presets/experimental/starter/README.md +7 -0
  315. package/presets/experimental/vocabulary.yaml +7 -0
  316. package/presets/personal/categories.yaml +7 -0
  317. package/presets/personal/preset.yaml +41 -0
  318. package/presets/personal/starter/goals.md +21 -0
  319. package/presets/personal/starter/index.md +17 -0
  320. package/presets/personal/starter/life-areas.md +21 -0
  321. package/presets/personal/starter/people.md +21 -0
  322. package/presets/personal/vocabulary.yaml +32 -0
  323. package/presets/research/categories.yaml +8 -0
  324. package/presets/research/preset.yaml +41 -0
  325. package/presets/research/starter/index.md +17 -0
  326. package/presets/research/starter/methods.md +21 -0
  327. package/presets/research/starter/open-questions.md +21 -0
  328. package/presets/research/vocabulary.yaml +33 -0
  329. package/reference/AUDIT-REPORT.md +238 -0
  330. package/reference/claim-map.md +172 -0
  331. package/reference/components.md +327 -0
  332. package/reference/conversation-patterns.md +542 -0
  333. package/reference/derivation-validation.md +649 -0
  334. package/reference/dimension-claim-map.md +134 -0
  335. package/reference/evolution-lifecycle.md +297 -0
  336. package/reference/failure-modes.md +235 -0
  337. package/reference/interaction-constraints.md +204 -0
  338. package/reference/kernel.yaml +242 -0
  339. package/reference/methodology.md +283 -0
  340. package/reference/open-questions.md +279 -0
  341. package/reference/personality-layer.md +302 -0
  342. package/reference/self-space.md +299 -0
  343. package/reference/semantic-vs-keyword.md +288 -0
  344. package/reference/session-lifecycle.md +298 -0
  345. package/reference/templates/base-note.md +16 -0
  346. package/reference/templates/companion-note.md +70 -0
  347. package/reference/templates/creative-note.md +16 -0
  348. package/reference/templates/learning-note.md +16 -0
  349. package/reference/templates/life-note.md +16 -0
  350. package/reference/templates/moc.md +26 -0
  351. package/reference/templates/relationship-note.md +17 -0
  352. package/reference/templates/research-note.md +19 -0
  353. package/reference/templates/session-log.md +24 -0
  354. package/reference/templates/therapy-note.md +16 -0
  355. package/reference/test-fixtures/edge-case-constraints.md +148 -0
  356. package/reference/test-fixtures/multi-domain.md +164 -0
  357. package/reference/test-fixtures/novel-domain-gaming.md +138 -0
  358. package/reference/test-fixtures/research-minimal.md +102 -0
  359. package/reference/test-fixtures/therapy-full.md +155 -0
  360. package/reference/testing-milestones.md +1087 -0
  361. package/reference/three-spaces.md +363 -0
  362. package/reference/tradition-presets.md +203 -0
  363. package/reference/use-case-presets.md +341 -0
  364. package/reference/validate-kernel.sh +432 -0
  365. package/reference/vocabulary-transforms.md +85 -0
  366. package/scripts/sync-thinking.sh +147 -0
  367. package/skill-sources/graph/SKILL.md +567 -0
  368. package/skill-sources/graph/skill.json +17 -0
  369. package/skill-sources/learn/SKILL.md +254 -0
  370. package/skill-sources/learn/skill.json +17 -0
  371. package/skill-sources/next/SKILL.md +407 -0
  372. package/skill-sources/next/skill.json +17 -0
  373. package/skill-sources/pipeline/SKILL.md +314 -0
  374. package/skill-sources/pipeline/skill.json +17 -0
  375. package/skill-sources/ralph/SKILL.md +604 -0
  376. package/skill-sources/ralph/skill.json +17 -0
  377. package/skill-sources/reduce/SKILL.md +1113 -0
  378. package/skill-sources/reduce/skill.json +17 -0
  379. package/skill-sources/refactor/SKILL.md +448 -0
  380. package/skill-sources/refactor/skill.json +17 -0
  381. package/skill-sources/reflect/SKILL.md +747 -0
  382. package/skill-sources/reflect/skill.json +17 -0
  383. package/skill-sources/remember/SKILL.md +534 -0
  384. package/skill-sources/remember/skill.json +17 -0
  385. package/skill-sources/rethink/SKILL.md +658 -0
  386. package/skill-sources/rethink/skill.json +17 -0
  387. package/skill-sources/reweave/SKILL.md +657 -0
  388. package/skill-sources/reweave/skill.json +17 -0
  389. package/skill-sources/seed/SKILL.md +303 -0
  390. package/skill-sources/seed/skill.json +17 -0
  391. package/skill-sources/stats/SKILL.md +371 -0
  392. package/skill-sources/stats/skill.json +17 -0
  393. package/skill-sources/tasks/SKILL.md +402 -0
  394. package/skill-sources/tasks/skill.json +17 -0
  395. package/skill-sources/validate/SKILL.md +310 -0
  396. package/skill-sources/validate/skill.json +17 -0
  397. package/skill-sources/verify/SKILL.md +532 -0
  398. package/skill-sources/verify/skill.json +17 -0
  399. package/skills/add-domain/SKILL.md +441 -0
  400. package/skills/add-domain/skill.json +17 -0
  401. package/skills/architect/SKILL.md +568 -0
  402. package/skills/architect/skill.json +17 -0
  403. package/skills/ask/SKILL.md +388 -0
  404. package/skills/ask/skill.json +17 -0
  405. package/skills/health/SKILL.md +760 -0
  406. package/skills/health/skill.json +17 -0
  407. package/skills/help/SKILL.md +348 -0
  408. package/skills/help/skill.json +17 -0
  409. package/skills/recommend/SKILL.md +553 -0
  410. package/skills/recommend/skill.json +17 -0
  411. package/skills/reseed/SKILL.md +385 -0
  412. package/skills/reseed/skill.json +17 -0
  413. package/skills/setup/SKILL.md +1688 -0
  414. package/skills/setup/skill.json +17 -0
  415. package/skills/tutorial/SKILL.md +496 -0
  416. package/skills/tutorial/skill.json +17 -0
  417. package/skills/upgrade/SKILL.md +395 -0
  418. package/skills/upgrade/skill.json +17 -0
@@ -0,0 +1,56 @@
1
+ ---
2
+ description: Granularity, organization, linking, processing intensity, navigation depth, maintenance cadence, schema density, and automation level — each a spectrum with poles, together defining what knobs a
3
+ kind: research
4
+ topics: ["[[design-dimensions]]"]
5
+ methodology: ["Original"]
6
+ source: [[knowledge-system-derivation-blueprint]]
7
+ ---
8
+
9
+ # eight configuration dimensions parameterize the space of possible knowledge systems
10
+
11
+ The claim that knowledge systems are parameterized rather than fixed needs teeth. Saying "it depends on the use case" is vacuously true. What makes the design space tractable is identifying the specific dimensions along which systems vary and understanding what each dimension's poles actually trade off. Eight dimensions emerge from the research, and together they define the space a knowledge system generator navigates.
12
+
13
+ **Note granularity** ranges from atomic (one claim per note, as this vault practices) to coarse (one note per source, preserving original structure). Atomic granularity maximizes composability — since [[enforcing atomicity can create paralysis when ideas resist decomposition]], the cost is real, but atomic notes can be linked, rearranged, and synthesized independently in ways that compound notes cannot. Coarse granularity reduces capture friction and preserves source coherence, which matters when the primary operation is reference rather than synthesis. The right position depends on whether the system prioritizes retrieval of original context or generation of new connections.
14
+
15
+ **Organization** ranges from flat (all files in one directory, structure via wiki links) to hierarchical (nested folders with path-based navigation). Since [[associative ontologies beat hierarchical taxonomies because heterarchy adapts while hierarchy brittles]], flat organization accommodates evolving understanding because notes are not locked into categories at creation time. But hierarchical systems provide predictable locations for team collaboration and reduce navigational vertigo in large collections. Wiki link resolution constrains this choice — flat structures enable filename-only linking while hierarchical structures require path disambiguation.
16
+
17
+ **Linking philosophy** ranges from explicit only (every connection is manually created with articulated reasoning) to explicit plus implicit discovery (semantic search, embedding similarity, and algorithmic suggestions augment manual linking). Since [[propositional link semantics transform wiki links from associative to reasoned]], explicit-only linking produces higher-quality connections but scales poorly. Adding implicit discovery compensates for human or agent attention limits but introduces noise — because [[spreading activation models how agents should traverse]], the traversal mechanism must handle both curated and suggested edges without conflating their reliability.
18
+
19
+ **Processing intensity** ranges from heavy (extract, reflect, reweave, verify — the full pipeline this vault implements) to light (capture, route, review — minimal transformation). Since [[fresh context per task preserves quality better than chaining phases]], heavy processing produces higher-quality notes but demands more infrastructure and compute. Since [[every knowledge domain shares a four-phase processing skeleton that diverges only in the process step]], the intensity dimension specifically governs how much transformation happens in the process phase — capture, connect, and verify remain structural constants regardless of intensity level. Light processing suits domains where content arrives near-final (curated feeds, expert-written sources) or where the volume-to-value ratio is low enough that deep processing per item isn't justified.
20
+
21
+ **Navigation depth** ranges from two levels (hub linking directly to topic notes) to four levels (hub to domain to topic to sub-topic). Deeper navigation serves larger collections by providing intermediate orientation points, but each level adds maintenance burden and cognitive overhead during traversal. The right depth correlates with expected note count — a 50-note vault needs two levels, a 5000-note vault needs four — but also with the conceptual complexity of the domain. A domain with natural sub-structure (like "processing" splitting into capture, transformation, and retrieval) benefits from depth even at modest scale.
22
+
23
+ **Maintenance cadence** ranges from daily (a trading journal where yesterday's analysis becomes stale) to quarterly (a stable reference vault where core concepts change slowly). This dimension is often invisible in system design because maintenance happens — or fails to happen — as a social practice rather than an architectural feature. But cadence constrains other dimensions: daily maintenance requires lightweight review processes, while quarterly maintenance can afford deep reweaving passes. Domains with high temporal dynamics (markets, active research) need faster cadence than domains with slow conceptual evolution (mathematical foundations, historical analysis). Cadence also interacts with the governance rhythm: since [[organic emergence versus active curation creates a fundamental vault governance tension]], the cadence setting determines how frequently curation interventions occur, and setting cadence too high suppresses the emergence that generates organic structure while setting it too low lets structural debt accumulate past easy resolution.
24
+
25
+ **Schema density** ranges from minimal (description and topics only, roughly what any note needs to be findable) to dense (eight or more fields including methodology, classification, provenance, status, and custom domain fields). Dense schemas make notes queryable like a graph database — since [[faceted classification treats notes as multi-dimensional objects rather than folder contents]], each YAML field is an independent classification facet that composes multiplicatively for retrieval precision. But since [[storage versus thinking distinction determines which tool patterns apply]], the trade-off is real: each required field adds capture friction and maintenance burden. And since [[schema fields should use domain-native vocabulary not abstract terminology]], the density choice interacts with vocabulary: dense schemas amplify the cost of vocabulary mismatch because every abstractly-named field forces a translation step at capture time, so higher density demands more careful domain-native naming to remain usable. The right density depends on whether the primary access pattern is structured query (favoring dense) or free-form traversal (favoring minimal). Critically, since [[schema evolution follows observe-then-formalize not design-then-enforce]], the position on this spectrum should not be chosen upfront — systems should start minimal and move toward density only when usage evidence (manual field additions, patterned free text) justifies the cost of maintaining additional fields.
26
+
27
+ **Automation level** ranges from full (hooks enforce schemas on every write, skills encode methodology as executable workflows, pipelines coordinate multi-phase processing) to manual (all quality standards exist as written instructions that the agent follows by convention). Since [[four abstraction layers separate platform-agnostic from platform-dependent knowledge system features]], automation level is partly determined by platform capability — you cannot automate what the platform does not support — and partly a design choice about system maturity. New systems benefit from starting manual (convention layer) and automating as understanding hardens, following the trajectory from documentation to skill to hook.
28
+
29
+ What makes these eight dimensions a useful parameterization rather than an arbitrary list is that they are largely independent in definition but interact in practice. A system positioned at "atomic granularity" does not logically require "heavy processing," but in practice, atomic notes need more connection work than coarse notes because the relationships lost by decomposing sources must be explicitly recreated. Since [[configuration dimensions interact so choices in one create pressure on others]], the space of viable configurations is much smaller than the combinatorial product of eight independent spectrums. This interaction structure also explains why [[methodology traditions are named points in a shared configuration space not competing paradigms]] — Zettelkasten, PARA, Cornell, and Evergreen each represent a coherent region where dimension interactions have been resolved through practice, not competing answers to the same question. The dimensions together define what a knowledge system generator decides when [[derivation generates knowledge systems from composable research claims not template customization]], while the universal components (notes, prose-as-title, wiki links, YAML base fields, graph mechanics, progressive disclosure) define what it never varies. But since [[dense interlinked research claims enable derivation while sparse references only enable templating]], the dimensions are only navigable when the claim graph supporting each dimension is dense enough — sparse research per dimension forces the derivation agent to guess rather than compose, degrading derivation to templating at the per-dimension level even if the overall graph appears large. And when a domain has no direct reference model, since [[novel domains derive by mapping knowledge type to closest reference domain then adapting]], knowledge type classification identifies which reference domain provides initial settings for these eight dimensions, with adaptation adjusting from that starting configuration rather than choosing each dimension from scratch.
30
+
31
+ ---
32
+ ---
33
+
34
+ Relevant Notes:
35
+ - [[knowledge system architecture is parameterized by platform capabilities not fixed by methodology]] — the upstream principle this concretizes: parameterization as concept needs specific dimensions to become actionable
36
+ - [[four abstraction layers separate platform-agnostic from platform-dependent knowledge system features]] — layers describe WHERE features live while dimensions describe HOW features vary; the two are orthogonal decompositions of the same design space
37
+ - [[storage versus thinking distinction determines which tool patterns apply]] — the upstream classification before dimensions apply: a storage system and a thinking system occupy different regions of the eight-dimensional space
38
+ - [[enforcing atomicity can create paralysis when ideas resist decomposition]] — the failure mode at the atomic pole of the granularity dimension, grounding it in real trade-offs rather than abstract spectrum
39
+ - [[associative ontologies beat hierarchical taxonomies because heterarchy adapts while hierarchy brittles]] — the research case for one pole of the organization dimension
40
+ - [[propositional link semantics transform wiki links from associative to reasoned]] — moves along the linking philosophy dimension from implicit association toward explicit typed connections
41
+ - [[fresh context per task preserves quality better than chaining phases]] — grounds the heavy end of the processing intensity dimension in attention science
42
+ - [[configuration dimensions interact so choices in one create pressure on others]] — develops the interaction thesis: dimensions are independent in definition but coupled in practice, reducing the viable configuration space
43
+ - [[methodology traditions are named points in a shared configuration space not competing paradigms]] — extends: maps five major traditions as specific points in this eight-dimensional space, showing what instantiated dimension choices look like
44
+ - [[derivation generates knowledge systems from composable research claims not template customization]] — enables: derivation is the process that traverses these dimensions to compose justified configurations rather than selecting from templates
45
+ - [[every knowledge domain shares a four-phase processing skeleton that diverges only in the process step]] — constrains the processing intensity dimension by showing that capture, connect, and verify are structural constants while only the process step varies by domain
46
+ - [[faceted classification treats notes as multi-dimensional objects rather than folder contents]] — grounds the schema density dimension in Ranganathan's formal classification theory: each YAML field is an independent facet enabling multiplicative retrieval precision
47
+ - [[novel domains derive by mapping knowledge type to closest reference domain then adapting]] — operationalizes these dimensions for unfamiliar domains: knowledge type classification identifies which reference domain provides initial dimension settings, with adaptation adjusting from that starting configuration
48
+ - [[schema evolution follows observe-then-formalize not design-then-enforce]] — the temporal mechanism for the schema density dimension: rather than choosing density upfront, start minimal and let the five-signal quarterly review protocol move density in response to observed usage patterns
49
+ - [[schema fields should use domain-native vocabulary not abstract terminology]] — the naming constraint on the schema density dimension: dense schemas amplify vocabulary mismatch costs because every abstractly-named field forces a translation step, so higher density demands more careful domain-native naming
50
+ - [[organic emergence versus active curation creates a fundamental vault governance tension]] — the governance meta-dimension spanning maintenance cadence and automation level: cadence governs how frequently curation interventions occur, and automation level governs how much curation is delegated to infrastructure, with the governance rhythm determining the effective position on both spectrums
51
+ - [[configuration paralysis emerges when derivation surfaces too many decisions]] — the UX consequence of surfacing all eight dimensions: presenting every dimension as a question produces analysis paralysis; sensible defaults and inference from dimension coupling reduce the decision surface to genuine choice points
52
+ - [[premature complexity is the most common derivation failure mode]] — the deployment consequence of the combinatorial space: a derivation engine can justify choices along all eight dimensions, but the composed system exceeds user capacity, requiring a complexity budget that constrains initial deployment to a subset of these dimensions
53
+ - [[dense interlinked research claims enable derivation while sparse references only enable templating]] — the research quality prerequisite: dimensions are only navigable when each has dense, interlinked claims with methodology provenance; sparse coverage per dimension forces derivation to guess rather than compose, degrading to templating at the per-dimension level
54
+
55
+ Topics:
56
+ - [[design-dimensions]]
@@ -0,0 +1,55 @@
1
+ ---
2
+ description: Zettelkasten works because connecting new information to existing knowledge — not just filing it — creates encoding depth, and explicit relationship articulation in wiki links is how the vault
3
+ kind: research
4
+ topics: ["[[note-design]]", "[[graph-structure]]"]
5
+ methodology: ["Zettelkasten", "Cognitive Science"]
6
+ source: [[tft-research-part3]]
7
+ ---
8
+
9
+ # elaborative encoding is the quality gate for new notes
10
+
11
+ Elaborative encoding is the cognitive mechanism that distinguishes productive note-taking from sophisticated filing. The principle is specific: information is better retained and more usable when the learner connects it to what they already know, rather than storing it in isolation. Craik and Lockhart's levels of processing framework established the hierarchy — shallow processing (structural features) produces weaker memory traces than deep processing (semantic meaning) — and elaborative rehearsal is the deepest form because it requires actively relating new material to existing mental structures.
12
+
13
+ Zettelkasten operationalizes this through a simple rule: every new note must link to at least one existing note with an explicit statement of why. Not "related to" — that is structural processing, the same level as organizing by topic. The relationship must be articulated: "extends this by adding the temporal dimension," "contradicts this because agents lack motor encoding," "provides the mechanism that this note only names." Since [[inline links carry richer relationship data than metadata fields]], the prose context around a wiki link IS the elaborative encoding — it forces the author to process how the new claim relates to the existing knowledge structure, which is the cognitive act that creates durable understanding.
14
+
15
+ This is narrower and more specific than [[the generation effect requires active transformation not just storage]]. The generation effect covers all forms of active production — writing descriptions, synthesizing sources, articulating claims. Elaborative encoding specifies the particular form of generation that builds network coherence: connecting to what already exists. You can generate a perfectly good description for a note without connecting it to anything, and that description demonstrates the generation effect. But without elaborative encoding — without the "this extends X because Y" articulation — the note floats as an isolated node. The generation created local value (a good description), but missed the network-level value (traversal paths to existing knowledge).
16
+
17
+ This is why the vault's quality standard requires context phrases on wiki links rather than bare link lists. A MOC entry that reads `- [[note title]]` has failed the elaborative encoding test. It says a connection exists without saying what the connection means. Compare `- [[note title]] — contradicts the assumption that volume drives value` — this forces the author to process the relationship, and that processing is what transforms a link from a filing operation into a cognitive operation. Since [[propositional link semantics transform wiki links from associative to reasoned]], the articulation vocabulary (extends, contradicts, enables, specifies) provides scaffolding for elaboration. Each relationship type demands a different form of processing: "extends" requires understanding what the original note argues and how the new note adds to it; "contradicts" requires understanding the tension and why both positions have force.
18
+
19
+ The connection to capture is immediate. Since [[capture the reaction to content not just the content itself]], reactions are elaborative encoding happening at capture time — relating the new content to existing understanding before the context decays. Without a reaction, the agent must reconstruct the elaboration later, and since [[structure without processing provides no value]], reconstruction that only achieves structural processing (filing, tagging, organizing) produces nothing. The vault's processing pipeline encodes elaborative encoding as a structural requirement: the reflect phase exists specifically to find and articulate connections, and the quality gates on that phase check whether the articulation is genuine or perfunctory.
20
+
21
+ But there is a shadow side. Mandatory elaboration can become a ritual that produces form without substance, and since [[verbatim risk applies to agents too]], an agent that writes "extends [[note]] by adding detail" on every connection has satisfied the articulation requirement without achieving genuine elaborative processing. The articulation vocabulary helps here because different relationship types demand different cognitive work, but the deeper safeguard is specificity. Since [[claims must be specific enough to be wrong]], vague elaboration fails the same test that vague claims fail — if the relationship articulation could apply to any pair of notes, it hasn't actually processed the specific connection between these two. "Extends by adding the temporal dimension" can only describe one relationship. "Extends by adding detail" could describe anything. The elaboration must be as specific as the claims it connects.
22
+
23
+ This grounds the vault's approach to note quality in cognitive science rather than aesthetic preference. The requirement for explicit relationship articulation is not a formatting convention. It is the mechanism through which new knowledge integrates with existing knowledge, creating the encoding depth that makes content retrievable and the network density that makes traversal productive. Notes without elaborated connections are structurally complete but cognitively orphaned — they exist in the graph without participating in the reasoning the graph enables. Elaborative encoding also serves as the calibration mechanism for [[controlled disorder engineers serendipity through semantic rather than topical linking]] — the quality gate that keeps disorder "controlled" rather than random. Every semantic cross-link must pass the articulation test ("WHY do these connect?"), which means the network-level unpredictability that generates serendipity is composed entirely of judgment-tested individual connections. Without this gate, semantic linking degenerates into noise; with it, each cross-topical edge carries genuine reasoning that makes the unexpected adjacency productive rather than arbitrary.
24
+
25
+ The encoding depth is testable. Since [[testing effect could enable agent knowledge verification]], an agent that reads only title and description and then predicts note content is testing whether the elaborative encoding produced retrievable hooks. Notes whose connections were genuinely elaborated create distinctive enough descriptions that prediction succeeds. Notes whose connections were structurally filed produce descriptions that merely paraphrase, and prediction fails. The testing effect becomes the verification mechanism for elaborative encoding quality.
26
+
27
+ Elaborative encoding also has a temporal dimension. The initial encoding at creation time — connecting a new note to existing knowledge — is only the first pass. Since [[incremental formalization happens through repeated touching of old notes]], each subsequent encounter with a note triggers fresh elaboration against current understanding. A link added six months ago said "extends by adding the temporal dimension" based on six-month-old understanding. Today, with new notes and deeper comprehension, that same connection might need richer articulation or might reveal itself as superficial. This is why [[backward maintenance asks what would be different if written today]] — the reweaving question forces re-elaboration, preventing connections from fossilizing at whatever depth they had when first written.
28
+
29
+ There is a deeper question about WHO does the elaboration. Since [[cognitive outsourcing risk in agent-operated systems]] identifies the delegation shadow, the encoding depth that elaboration creates accrues to whoever does the processing. When agents handle all connection-finding and articulation, the vault becomes richly elaborated — dense with context phrases, typed relationships, and genuine connection reasoning — but the human's understanding may remain at the structural processing level of someone who reads but does not generate. The elaborative encoding gate ensures vault quality, but it does not ensure human understanding unless the human participates in at least some of the elaboration.
30
+
31
+ The vault's claim-based architecture maps onto a formal framework that names what elaboration produces. Since [[IBIS framework maps claim-based architecture to structured argumentation]], each context phrase on a wiki link functions as an Argument in IBIS terms — evidence connecting one Position (claim note) to another. A bare link without elaboration is a non-argument: it asserts a connection exists without providing any argumentative force. The elaborative encoding test and the IBIS argument test converge: both require that connections carry reasoning, not just reference.
32
+ ---
33
+
34
+ Relevant Notes:
35
+ - [[the generation effect requires active transformation not just storage]] — foundation: generation is the broader principle; elaborative encoding specifies the mechanism — connecting to existing knowledge is the particular form of generation that creates durable understanding
36
+ - [[inline links carry richer relationship data than metadata fields]] — implements elaborative encoding at the link level: 'since X, therefore Y' is elaboration, 'related_to: X' is filing
37
+ - [[propositional link semantics transform wiki links from associative to reasoned]] — extends the articulation requirement: typed relationships (causes, enables, contradicts) are deeper elaboration than untyped association
38
+ - [[structure without processing provides no value]] — the Lazy Cornell anti-pattern is the failure mode when elaboration is skipped: structure without connecting-to-existing-knowledge produces no benefit
39
+ - [[generation effect gate blocks processing without transformation]] — operationalizes the gate this note describes: requires a generated artifact before inbox exit, but elaborative encoding specifies what KIND of generation matters most
40
+ - [[capture the reaction to content not just the content itself]] — reactions ARE elaborative encoding at capture time: relating new content to existing understanding is the reaction that creates synthesis seeds
41
+ - [[claims must be specific enough to be wrong]] — specificity enables elaboration: vague claims cannot be connected to existing knowledge because they lack the precision needed for relationship articulation
42
+ - [[testing effect could enable agent knowledge verification]] — verifies whether elaboration succeeded: if an agent cannot predict note content from title and description, the elaborative encoding that should have created retrievable connections was shallow
43
+ - [[verbatim risk applies to agents too]] — the failure mode when elaboration becomes ritual: agents can produce context phrases that satisfy the articulation requirement structurally while achieving no genuine elaborative processing
44
+ - [[incremental formalization happens through repeated touching of old notes]] — extends elaborative encoding from a single-pass mechanism to a longitudinal one: each re-encounter with an old note triggers fresh elaboration against current understanding, deepening connections over time
45
+ - [[backward maintenance asks what would be different if written today]] — re-elaboration as maintenance: the 'what would be different' question forces fresh elaborative encoding of existing content against new knowledge, preventing connections from calcifying at their original depth
46
+ - [[cognitive outsourcing risk in agent-operated systems]] — the delegation shadow: when agents perform all elaborative encoding, the vault benefits from deep connections but the human loses the encoding depth that elaboration provides to whoever does the processing
47
+ - [[IBIS framework maps claim-based architecture to structured argumentation]] — provides formal vocabulary: elaborated context phrases on wiki links are Arguments in IBIS terms, and bare links without articulation are non-arguments that fail both the IBIS and elaborative encoding tests
48
+ - [[wiki links as social contract transforms agents into stewards of incomplete references]] — extends elaborative encoding from a quality gate to a stewardship standard: the social contract says agents must fulfill link promises, and elaborative encoding specifies what fulfillment requires — not just creating the target note but connecting it with articulated relationships
49
+ - [[vibe notetaking is the emerging industry consensus for AI-native self-organization]] — industry-scale example of the elaboration gap: most vibe notetaking tools use embedding-based linking that skips elaboration entirely, while agent-curated wiki links with context phrases implement elaborative encoding at system scale
50
+ - [[controlled disorder engineers serendipity through semantic rather than topical linking]] — elaborative encoding is the calibration mechanism: the requirement that every link articulate WHY it connects is what keeps Luhmann's controlled disorder productive rather than random; without the quality gate, semantic cross-linking degenerates into noise
51
+ - [[MOC construction forces synthesis that automated generation from metadata cannot replicate]] — navigation-scale application: writing context phrases on MOC entries IS elaborative encoding applied to navigation; the Dump-Lump-Jump pattern reveals that the Lump phase (context phrase creation) is precisely where elaboration happens, and automated MOC generation skips this encoding step entirely, producing lists without the relationship articulation that makes MOCs navigable
52
+
53
+ Topics:
54
+ - [[note-design]]
55
+ - [[graph-structure]]
@@ -0,0 +1,221 @@
1
+ ---
2
+ description: Why schema enforcement is non-negotiable for agent-operated knowledge systems and how to implement it across domains — soft enforcement at capture, hard enforcement at pipeline boundaries, and the
3
+ kind: guidance
4
+ status: active
5
+ topics: ["[[schema-enforcement]]"]
6
+ ---
7
+
8
+ # enforce schema with graduated strictness across capture processing and query zones
9
+
10
+ Schema enforcement is the data integrity layer that makes a knowledge vault queryable rather than browsable. Without it, the vault degrades from a structured graph database into an unstructured pile of files that happens to have YAML at the top. Since [[markdown plus YAML plus ripgrep implements a queryable graph database without infrastructure]], schema enforcement is what keeps that database functional — the YAML fields are the columns, and enforcement is what ensures the columns are populated.
11
+
12
+ This doc tells the plugin WHEN to enforce, HOW strictly, and WHAT to enforce for each domain. It is consulted by /ask when users ask about schema compliance, by /architect when designing new note types, and by /recommend when diagnosing retrieval failures.
13
+
14
+ ## Why Non-Negotiable
15
+
16
+ Every domain composition the plugin generates depends on queryable metadata. Therapy systems query `mood` and `trigger` fields to detect emotional patterns across entries. Project management queries `status` and `stakeholders` to prepare meeting briefings. Research queries `methodology` and `classification` to find methodologically comparable claims. Trading queries `ticker` and `thesis` to check conviction against outcome. Without schema enforcement, these queries return incomplete or incorrect results — the system works for a week, then silently degrades as notes accumulate without proper metadata.
17
+
18
+ The degradation is invisible. A therapy vault with 50 reflections looks healthy. But if 12 of them lack `mood` fields, the pattern detection algorithm sees only 38 data points and the mood-trigger correlations it surfaces are skewed toward entries that happened to be well-formatted. The user never knows the analysis is incomplete because missing data does not announce itself — it simply disappears from query results. This is "silent data loss," and it is the primary failure mode that schema enforcement prevents.
19
+
20
+ Since [[schema enforcement via validation agents enables soft consistency]], the enforcement approach matters as much as the enforcement itself. The research is clear: hard enforcement at capture time creates friction that causes system abandonment, while no enforcement at all produces metadata drift that corrupts retrieval. The solution is graduated enforcement — strict where it matters, gentle where friction would cost more than compliance.
21
+
22
+ ## The Cognitive Science of Externalized Enforcement
23
+
24
+ Why can't we simply instruct agents to fill in all the fields? Because since [[schema validation hooks externalize inhibitory control that degrades under cognitive load]], instruction-based enforcement degrades precisely when it matters most. Inhibitory control — the capacity to suppress automatic behavior and follow rules — is the first executive function to degrade under cognitive load. An agent deep in a complex synthesis task, with its context window filling past the smart zone, stops attending to procedural instructions like "validate frontmatter against the template schema." The agent does not decide to skip validation. It simply stops attending to the instruction, the same way a tired surgeon proceeds without the checklist.
25
+
26
+ Externalized enforcement through hooks solves this. A validation hook fires after every file write regardless of the agent's cognitive state — whether it is in the first 10% of context or the last 5%. The check happens outside the context window, immune to the attention degradation that compromises everything inside it. Since [[hooks enable context window efficiency by delegating deterministic checks to external processes]], externalized enforcement provides a double benefit: it both guarantees compliance and frees context tokens for substantive work.
27
+
28
+ This is the theoretical foundation for the entire enforcement architecture. Enforcement that lives in instructions degrades. Enforcement that lives in infrastructure persists.
29
+
30
+ ## The Enforcement Gradient
31
+
32
+ The plugin implements a graduated enforcement model across three zones:
33
+
34
+ | Zone | Strictness | When | Why |
35
+ |------|-----------|------|-----|
36
+ | **Capture** | Soft (warn, don't block) | Creating new notes | Since [[temporal separation of capture and processing preserves context freshness]], blocking capture for schema issues trades context preservation for format purity — a bad trade |
37
+ | **Processing** | Medium (require core fields) | Pipeline reduce/reflect/verify phases | Processing phases can ensure quality because they are not time-sensitive and run with dedicated context |
38
+ | **Query** | Hard (missing fields = invisible) | Graph queries, pattern detection | If a therapy note lacks `mood`, it does not appear in mood analysis — enforcement by consequence |
39
+
40
+ The gradient maps to Thaler and Sunstein's insight from nudge theory: intervention strength should match violation severity. Since [[nudge theory explains graduated hook enforcement as choice architecture for agents]], blocking a write (exit code 2) is a mandate appropriate for structural failures. Warning via context injection is a nudge appropriate for qualitative issues. The choice between them is architectural, not arbitrary.
41
+
42
+ ### What to Enforce at Each Level
43
+
44
+ **Always enforce (hard, at all levels):**
45
+ - `description` field exists and is non-empty — since [[descriptions are retrieval filters not summaries]], every note needs one for discovery
46
+ - `topics` field exists with at least one MOC reference — orphan notes are lost notes
47
+ - Valid wiki links (targets exist) — since [[stale navigation actively misleads because agents trust curated maps completely]], dangling links do not just clutter; they actively misdirect
48
+
49
+ **Enforce at processing (medium):**
50
+ - Domain-specific required fields populated (e.g., `mood` for therapy entries, `ticker` for trade journals, `jurisdiction` for legal notes)
51
+ - Enum values within valid ranges (e.g., `status: preliminary | open | dissolved`, `conviction_level: low | moderate | high | extreme`)
52
+ - Description adds information beyond title — since [[good descriptions layer heuristic then mechanism then implication]], descriptions that merely restate the title waste the retrieval layer
53
+ - `relevant_notes` entries include context phrases explaining the relationship
54
+
55
+ **Warn at capture (soft):**
56
+ - Missing optional fields
57
+ - Description quality issues (too short, restates title)
58
+ - Missing relevant_notes context phrases
59
+ - Non-standard enum values that might be typos
60
+
61
+ ## Domain-Specific Enforcement
62
+
63
+ Each domain has different critical fields. The plugin derives enforcement rules from domain schemas, following the principle that fields the domain QUERIES against are critical:
64
+
65
+ | Domain | Critical Fields (Hard) | Important Fields (Medium) | Nice-to-Have (Soft) |
66
+ |--------|----------------------|--------------------------|---------------------|
67
+ | Research | description, methodology, topics | classification, source, confidence | adapted_from, evidence_type |
68
+ | Therapy | description, mood, date | triggers, coping_used, mood_context | physical_sensations, sleep_quality |
69
+ | PM | description, project, status | stakeholders, rationale | follow_ups, blocked_by |
70
+ | Creative | description, canon_status | characters_involved, timeline_position | draft_number, revision_notes |
71
+ | Learning | description, domain, prerequisites | mastery_level, knowledge_type | last_reviewed, practice_count |
72
+ | Trading | description, ticker, thesis | conviction_level, position_size | emotions_during, market_regime |
73
+ | Health | description, date, type | severity, triggers | correlation_notes, duration |
74
+ | Legal | description, jurisdiction, matter | precedent_status, holding | opposing_arguments, appeal_risk |
75
+
76
+ The principle: fields the domain QUERIES against are critical (hard enforcement). Fields that enrich context but are not machine-processed are important (medium). Fields that provide optional color are nice-to-have (soft). A research vault that never queries `adapted_from` does not need to enforce it — but a research vault that queries `methodology` to compare evidence across traditions must enforce it, because missing values corrupt the comparison.
77
+
78
+ Since [[schema fields should use domain-native vocabulary not abstract terminology]], enforcement must respect the language of each domain. A therapy system validates that `trigger` uses the person's own words, not clinical codes. A trading system validates that `thesis` contains an actual investment rationale, not a ticker repeat. Field names are the only domain-specific element in the universal note pattern, and they carry the entire burden of making the system feel native rather than imposed.
79
+
80
+ ## The Observe-Then-Formalize Lifecycle
81
+
82
+ Since [[schema evolution follows observe-then-formalize not design-then-enforce]], the plugin never designs a 15-field schema on day one. Every field starts its life as an observation, graduates to a convention, and is formalized as schema.
83
+
84
+ ### The Lifecycle Stages
85
+
86
+ **Stage 1: Minimal seed.** The plugin generates a starting schema with 2-3 required fields (always `description` and `topics`, plus one domain-critical field). Everything else is optional or absent. The schema is deliberately incomplete — it captures just enough to make the vault queryable, not enough to be burdensome.
87
+
88
+ **Stage 2: Observation.** As the user works, the agent notices patterns: fields being added manually, fields being stuffed with "N/A," free-text fields developing internal structure. These observations accumulate as evidence for schema change. Since [[hook-driven learning loops create self-improving methodology through observation accumulation]], the observation mechanism already exists — schema evolution piggybacks on the general learning loop.
89
+
90
+ **Stage 3: Convention.** When an observation recurs 5+ times, it becomes a convention — documented in the context file as guidance but not yet enforced. "We tend to track triggers on therapy reflections" is a convention. It guides behavior without blocking it.
91
+
92
+ **Stage 4: Formalization.** When a convention has been followed consistently and the field is being queried, formalize it: add to the template, set required/optional status, define valid values if applicable, add validation. The field has earned its place through demonstrated use.
93
+
94
+ This lifecycle mirrors since [[methodology development should follow the trajectory from documentation to skill to hook as understanding hardens]] — schema hardens as evidence accumulates, not as speculation compounds.
95
+
96
+ ## The Five Evolution Signals
97
+
98
+ Since [[evolution observations provide actionable signals for system adaptation]], the quarterly review protocol uses five diagnostic signals to identify schema that needs to change. These signals are the concrete mechanism for the observe-then-formalize lifecycle:
99
+
100
+ ### 1. Manual Additions
101
+
102
+ The agent or user consistently adds a field manually that is not in the template. When the same ad-hoc field appears across five or more notes, it has demonstrated its value through use. **Action:** Formalize it — add to the template, decide required vs optional, define valid values. **Example:** A therapy system's agent starts adding `trigger` to reflections even though it was not in the initial template. After five reflections with manually added triggers, the field earns formalization.
103
+
104
+ ### 2. Placeholder Stuffing
105
+
106
+ A required field is consistently filled with "N/A," "none," or empty placeholder text. This means the field does not fit every note in the type. **Action:** Demote from required to optional. **Example:** A research system requires `adapted_from` on every claim, but original claims are stuffed with "null." The field should be optional with guidance to fill it only when genuinely adapting a human pattern.
107
+
108
+ ### 3. Dead Enums
109
+
110
+ An enum value defined in the template is never selected across 50+ notes. **Action:** Remove it — it is noise in the selection space. **Example:** A therapy system defines "ecstatic" as a mood enum but no reflection ever uses it. Drop it. Dead enums are a sign of upfront speculation that did not match reality.
111
+
112
+ ### 4. Patterned Free Text
113
+
114
+ A free-text field develops consistent internal structure across multiple notes. **Action:** If the structure is consistent enough, decompose into separate fields. If it is a convention, document it in the template guidance but keep the field free-text. **Example:** A free-text "context" field in a therapy system consistently contains "trigger: X, reaction: Y, insight: Z." That is structure waiting to be formalized into three separate fields. But a description field that consistently starts with a verb phrase is a stylistic convention — document it, do not decompose it.
115
+
116
+ ### 5. Oversized MOCs
117
+
118
+ A MOC consistently exceeds the healthy threshold (40+ notes for agent-operated systems, 30+ for human-operated). **Action:** Create sub-MOCs along the natural fault lines. **Signal:** If the split follows a pattern across multiple MOCs (e.g., every topic MOC splits into "theory" and "practice"), that pattern itself is a schema dimension worth formalizing. This signal is architectural rather than field-level, but it belongs in the quarterly review because it indicates the current taxonomy has outgrown its design.
119
+
120
+ The five signals compose into a quarterly review protocol: scan for manual additions, check for placeholder stuffing, audit enum usage, inspect free-text structure, review MOC sizes. Each check produces either "no action" or a specific schema proposal with evidence.
121
+
122
+ ## Progressive Validation
123
+
124
+ Since [[progressive schema validates only what active modules require not the full system schema]], the plugin never enforces fields from inactive modules. A user with just basic wiki-links and yaml-schema enabled sees no warnings about `methodology` or `relevant_notes` until they activate the modules that use those fields.
125
+
126
+ This is critical for onboarding. New users encounter minimal schema requirements. As they activate more features, enforcement grows with them. The progression:
127
+
128
+ 1. **Day 1:** Only `description` and `topics` required. The vault is immediately queryable for basic navigation.
129
+ 2. **Week 2:** User activates processing pipeline. Domain-specific fields become required at the processing boundary (not at capture).
130
+ 3. **Month 2:** User activates pattern detection. The fields that pattern detection queries become critical — now the system depends on them.
131
+ 4. **Ongoing:** The quarterly review protocol adds fields that have earned formalization and removes fields that have proven unnecessary.
132
+
133
+ Since [[premature complexity is the most common derivation failure mode]], progressive validation prevents the single most common cause of system abandonment: overwhelming new users with a schema they do not yet understand the purpose of.
134
+
135
+ ## Hook-Based Enforcement for Agents
136
+
137
+ For platforms that support hooks (tier 1), the plugin generates enforcement as automated checks:
138
+
139
+ **PostToolUse on Write:** Fires after every file creation or modification. Checks required fields for the note's type. Returns exit code 2 (blocking) for hard enforcement violations, exit code 0 with warning in additionalContext for soft violations. The agent sees warnings and can choose to address them; it cannot bypass blocks.
140
+
141
+ **Session-start health check:** Fires at session orientation. Counts notes per type missing critical fields. Reports as "schema compliance: 94% (3 therapy reflections missing mood field)." The number makes the gap visible without blocking work.
142
+
143
+ **Pipeline boundary gate:** Since [[generation effect gate blocks processing without transformation]], the pipeline boundary enforces a similar principle for schema: content cannot advance from one pipeline phase to the next if critical fields are missing. This is where processing-level enforcement lives — the gate is structural, not attentional.
144
+
145
+ For platforms that do not support hooks (tier 2-3), the plugin generates equivalent guidance in the context file: explicit checklists for note creation, processing phase instructions that include field validation, and periodic manual review prompts. Since [[platform capability tiers determine which knowledge system features can be implemented]], the enforcement mechanism adapts to platform capability while the enforcement principle remains constant.
146
+
147
+ ## Cross-Domain Enforcement
148
+
149
+ When users have multiple domains (e.g., research + personal life + therapy), each domain maintains its own schema scope. Since [[multi-domain systems compose through separate templates and shared graph]], hard enforcement that blocks on a therapy field violation must never stall research processing.
150
+
151
+ **Shared fields** (`description`, `topics`) have universal enforcement — every note in every domain must have them. **Domain-specific fields** (`mood`, `ticker`, `precedent`) have domain-scoped enforcement — they only apply to notes of the relevant type. The graph is shared; the schemas are separate.
152
+
153
+ The practical implication: a validation hook must know which template applies to which note. The plugin generates this routing from the template registry — each template declares which folder or note type it governs, and the hook dispatches accordingly.
154
+
155
+ ## Implementation Pattern
156
+
157
+ ### For /setup (Onboarding)
158
+
159
+ When generating a new vault, the plugin:
160
+ 1. Identifies domain-specific note types from the domain composition
161
+ 2. Generates template files with `_schema` blocks defining required/optional/enum fields
162
+ 3. Creates validation hooks appropriate to the user's platform tier
163
+ 4. Explains the enforcement gradient: "Your notes will be checked for [critical fields]. Missing fields get flagged during processing, not blocked at capture."
164
+
165
+ ### For /architect (Extension)
166
+
167
+ When a user adds a new note type:
168
+ 1. The plugin proposes a schema based on similar domains (since [[novel domains derive by mapping knowledge type to closest reference domain then adapting]])
169
+ 2. Critical fields are identified by asking: "What will you query this note type for?"
170
+ 3. The schema starts minimal and grows — since [[schema evolution follows observe-then-formalize not design-then-enforce]]
171
+
172
+ ### For /recommend (Maintenance)
173
+
174
+ When checking vault health:
175
+ 1. Count notes per type missing critical fields — report as "schema compliance"
176
+ 2. Detect fields being manually added that are not in the schema — suggest schema evolution
177
+ 3. Identify queries that return incomplete results due to missing metadata — flag as "silent data loss"
178
+
179
+ ## Anti-Patterns
180
+
181
+ | Anti-Pattern | Why It Fails | Better Approach |
182
+ |-------------|-------------|-----------------|
183
+ | Block all writes on any schema violation | Capture friction causes system abandonment | Soft enforcement at capture, hard at query |
184
+ | No enforcement at all | Metadata drift corrupts query results within weeks | At minimum, enforce description + topics universally |
185
+ | Enforce full schema from day one | Overwhelms users before they understand what fields are for | Progressive schema: minimal start, grow with observed use |
186
+ | Same enforcement across domains | Cross-domain blocking (therapy validation stalls research) | Domain-scoped validation via template routing |
187
+ | Enforce fields nobody queries | Wasted effort and unnecessary friction | Tie enforcement to actual query patterns |
188
+ | Design schema upfront without observation | Speculation produces dead fields and missing useful ones | Observe-then-formalize: fields earn their place through use |
189
+
190
+ ## Domain Examples
191
+
192
+ These domain compositions demonstrate schema enforcement in practice:
193
+
194
+ - [[academic research uses structured extraction with cross-source synthesis]] — Hard enforcement on `methodology` and `classification` enables evidence-quality queries and cross-methodology synthesis. The dense schema serves a purpose: every field powers a specific academic query.
195
+ - [[therapy journal uses warm personality with pattern detection for emotional processing]] — `mood` and `trigger` fields are critical for pattern detection, but soft enforcement at journaling time prevents friction during vulnerable moments. The ethical constraint governs: never let schema demands interfere with emotional capture.
196
+ - [[trading uses conviction tracking with thesis-outcome correlation]] — `ticker`, `thesis`, and `conviction_level` fields power strategy compliance checking. Enforcement by consequence when missing fields exclude trades from performance analysis.
197
+ - [[legal case management uses precedent chains with regulatory change propagation]] — `jurisdiction` and `precedent_status` require hard enforcement because regulatory compliance depends on complete metadata. This is the domain where enforcement is most naturally strict.
198
+ - [[health wellness uses symptom-trigger correlation with multi-dimensional tracking]] — Accumulation-based domain where schema enforcement gates pattern detection. Missing `date` or `type` fields make correlation analysis impossible, but capture must stay as frictionless as a health log requires.
199
+
200
+ ## Grounding
201
+
202
+ This guidance is grounded in:
203
+ - [[schema enforcement via validation agents enables soft consistency]] — the foundational research on soft vs hard enforcement
204
+ - [[schema validation hooks externalize inhibitory control that degrades under cognitive load]] — why enforcement must be externalized, not instructed
205
+ - [[schema fields should use domain-native vocabulary not abstract terminology]] — why enforcement must respect domain language
206
+ - [[progressive schema validates only what active modules require not the full system schema]] — module-aware enforcement
207
+ - [[schema evolution follows observe-then-formalize not design-then-enforce]] — enforcement generates the observation data for evolution
208
+ - [[nudge theory explains graduated hook enforcement as choice architecture for agents]] — theoretical basis for graduated enforcement
209
+ - [[schema templates reduce cognitive overhead at capture time]] — templates as the capture-side of enforcement
210
+ - [[evolution observations provide actionable signals for system adaptation]] — the five signals that drive schema evolution
211
+ - [[generation effect gate blocks processing without transformation]] — pipeline boundary enforcement as a quality gate
212
+
213
+ ---
214
+
215
+ Topics:
216
+ - [[index]]
217
+ - [[index]]
218
+ ---
219
+
220
+ Topics:
221
+ - [[schema-enforcement]]
@@ -0,0 +1,43 @@
1
+ ---
2
+ description: The cognitive effort of splitting complex arguments into single-concept notes can exceed productive friction, becoming system management overhead that crowds out actual thinking
3
+ kind: research
4
+ topics: ["[[note-design]]"]
5
+ confidence: speculative
6
+ methodology: ["Zettelkasten", "Evergreen"]
7
+ source: [[tft-research-part2]]
8
+ ---
9
+
10
+ # enforcing atomicity can create paralysis when ideas resist decomposition
11
+
12
+ Critics of strict atomicity in note-taking systems report that the requirement to decompose every insight into single-concept notes can be paralyzing. Breaking complex arguments into atomic units requires significant cognitive effort — identifying the seams, deciding where one concept ends and another begins, writing connective tissue to link the fragments. Some practitioners describe this as "note gymnastics: spending more time managing the system than actually thinking." Since [[insight accretion differs from productivity in knowledge systems]], this describes effort spent on productivity (structural splitting, fragment management) rather than accretion (deepening understanding) — the atomicity requirement may channel work toward system management at the expense of actual thinking.
13
+
14
+ This creates a diagnostic problem for ars contexta. Andy Matuschak's response to atomicity paralysis is instructive: "If it is hard to atomize the idea, it means the idea is not yet clearly understood." On this view, atomization difficulty is a signal, not an obstacle. The friction reveals that the insight hasn't resolved into clean components yet — which means forcing the decomposition is desirable difficulty that prompts deeper understanding. Since [[the generation effect requires active transformation not just storage]], the struggle to atomize might be exactly the generative work that creates cognitive hooks.
15
+
16
+ But there's an alternative interpretation: some ideas are genuinely relational and resist decomposition not because they're fuzzy but because they're structural. A claim about how three concepts interact may be one insight about their interaction, not three insights awkwardly stitched together. Forcing atomization of inherently relational thinking might destroy the insight rather than clarify it — like insisting that a molecule is really just atoms, when the molecular structure is precisely what matters. This parallels the concern that [[sense-making vs storage does compression lose essential nuance]] raises for descriptions: some features that make an idea valuable may be exactly the features that resist compression or decomposition.
17
+
18
+ Since [[configuration dimensions interact so choices in one create pressure on others]], the paralysis is not just a granularity problem in isolation — choosing atomic notes forces explicit linking, deep navigation, and heavy processing simultaneously, and the total burden can exceed the system's processing capacity. The vault's position — composability over atomicity — suggests a middle path, and since [[methodology traditions are named points in a shared configuration space not competing paradigms]], the mapping confirms that granularity is a spectrum with successfully operating traditions at each pole: Zettelkasten and Evergreen at atomic, Cornell at medium, PARA at coarse. The test is composability, not atomicity: can this note be linked from elsewhere without dragging unwanted context? If yes, the note is fine regardless of how many concepts it touches. If no, split. But this reframe doesn't eliminate the diagnostic problem: when linking drags context, is that because the note bundles separate claims, or because the claims are genuinely entangled?
19
+
20
+ For agent-operated systems, this creates a design tension. Since [[summary coherence tests composability before filing]] uses summary failure as a bundling diagnostic, agents might flag notes as "incoherent" when they're actually expressing relational insights that resist propositional summary. The agent's atomicity pressure could fragment ideas that humans understood as wholes. Since [[vault conventions may impose hidden rigidity on thinking]], atomicity enforcement might be one of those conventions that channels thinking into patterns favoring certain styles — sequential, decomposable, propositional — over others.
21
+
22
+ The research suggests friction is sometimes "desirable difficulty" and sometimes paralysis. Distinguishing these cases is non-trivial. One heuristic: if the struggle to split reveals new understanding (you realize what you thought was one idea is actually two), that's desirable difficulty. If the struggle feels like paperwork (you know exactly what you mean but can't make it fit the format), that's paralysis. But this heuristic relies on felt sense, which agents don't have. This is an operationalization problem shared with [[claims must be specific enough to be wrong]], which also asks when forced reformulation helps (incomplete thinking resolving into clarity) vs. constrains (valid non-linear insight being distorted).
23
+
24
+ There is a third structural option beyond "split" or "leave as-is" that partially dissolves this question. Since [[federated wiki pattern enables multi-agent divergence as feature not bug]], when an idea resists decomposition because its relational character IS the insight, parallel interpretive frames can coexist as federated notes. Instead of forcing a molecular insight into atomic fragments, two agents can each write a version that preserves the relational structure from its own angle. The test shifts from "can this be split?" to "do these perspectives illuminate different facets?" This does not fully operationalize the productive/paralysis distinction, but it removes one class of paralysis: the case where the thinker knows the idea is coherent but cannot make it fit the atomic format. Federation acknowledges that the coherence is real and offers coexistence instead of forced resolution.
25
+
26
+ The open question: can we operationalize the distinction between productive atomization friction and destructive atomization paralysis? Or does the vault's composability-over-atomicity framing dissolve the question by making atomicity a means rather than an end?
27
+ ---
28
+
29
+ Relevant Notes:
30
+ - [[vault conventions may impose hidden rigidity on thinking]] — this note names one specific form of that rigidity: atomicity enforcement as a convention that may constrain thinking
31
+ - [[the generation effect requires active transformation not just storage]] — the counterargument: if atomization difficulty signals incomplete understanding, the friction is desirable difficulty that prompts genuine transformation
32
+ - [[summary coherence tests composability before filing]] — operationalizes atomicity as a diagnostic: summary failure reveals bundling, but raises the question of whether all summary failures indicate bundling or whether some insights resist propositional compression
33
+ - [[claims must be specific enough to be wrong]] — sibling convention constraint: specificity requirement and atomicity requirement both demand that thinking resolve into particular forms, and both may channel thinking into patterns favoring certain styles
34
+ - [[sense-making vs storage does compression lose essential nuance]] — parallel tension at the description layer: that note asks whether ~150-character descriptions lose essential nuance, this note asks whether decomposition into atomic units loses essential structure
35
+ - [[insight accretion differs from productivity in knowledge systems]] — the note gymnastics critique names exactly this pattern: effort spent managing atomicity is productivity (filing, splitting, connecting fragments) without accretion (deepening understanding)
36
+ - [[writing for audience blocks authentic creation]] — sibling pattern of cognitive split: atomicity paralysis divides effort between decomposition and synthesis, audience awareness divides effort between presentation and synthesis; both identify how format requirements compete with genuine thinking
37
+ - [[federated wiki pattern enables multi-agent divergence as feature not bug]] — offers a structural alternative to forced decomposition: when ideas resist splitting because the relational structure IS the insight, parallel interpretive frames can coexist as federated notes without either losing the molecular character
38
+ - [[configuration dimensions interact so choices in one create pressure on others]] — generalizes this paralysis as a predicted outcome of the granularity cascade: fine granularity creates pressure toward heavy processing, and when processing burden exceeds capacity the system collapses into exactly this paralysis
39
+ - [[methodology traditions are named points in a shared configuration space not competing paradigms]] — maps the granularity spectrum to real traditions: Zettelkasten at atomic, Cornell at medium, PARA at coarse — showing that multiple successful operating points exist, making atomicity a choice not a requirement
40
+ - [[configuration paralysis emerges when derivation surfaces too many decisions]] — sibling paralysis pattern at the system level: atomicity paralysis comes from methodology requirements during note creation, configuration paralysis comes from methodology requirements during system derivation; both are cases where surfacing too many decisions overwhelms the decision-maker, but configuration paralysis has a cleaner solution because dimension coupling makes inference tractable
41
+
42
+ Topics:
43
+ - [[note-design]]