arscontexta 0.6.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.claude-plugin/marketplace.json +11 -0
- package/.claude-plugin/plugin.json +22 -0
- package/README.md +683 -0
- package/agents/knowledge-guide.md +49 -0
- package/bin/cli.mjs +66 -0
- package/generators/agents-md.md +240 -0
- package/generators/claude-md.md +379 -0
- package/generators/features/atomic-notes.md +124 -0
- package/generators/features/ethical-guardrails.md +58 -0
- package/generators/features/graph-analysis.md +188 -0
- package/generators/features/helper-functions.md +92 -0
- package/generators/features/maintenance.md +164 -0
- package/generators/features/methodology-knowledge.md +70 -0
- package/generators/features/mocs.md +144 -0
- package/generators/features/multi-domain.md +61 -0
- package/generators/features/personality.md +71 -0
- package/generators/features/processing-pipeline.md +428 -0
- package/generators/features/schema.md +149 -0
- package/generators/features/self-evolution.md +229 -0
- package/generators/features/self-space.md +78 -0
- package/generators/features/semantic-search.md +99 -0
- package/generators/features/session-rhythm.md +85 -0
- package/generators/features/templates.md +85 -0
- package/generators/features/wiki-links.md +88 -0
- package/generators/soul-md.md +121 -0
- package/hooks/hooks.json +45 -0
- package/hooks/scripts/auto-commit.sh +44 -0
- package/hooks/scripts/session-capture.sh +35 -0
- package/hooks/scripts/session-orient.sh +86 -0
- package/hooks/scripts/write-validate.sh +42 -0
- package/methodology/AI shifts knowledge systems from externalizing memory to externalizing attention.md +59 -0
- package/methodology/BM25 retrieval fails on full-length descriptions because query term dilution reduces match scores.md +39 -0
- package/methodology/IBIS framework maps claim-based architecture to structured argumentation.md +58 -0
- package/methodology/LLM attention degrades as context fills.md +49 -0
- package/methodology/MOC construction forces synthesis that automated generation from metadata cannot replicate.md +49 -0
- package/methodology/MOC maintenance investment compounds because orientation savings multiply across every future session.md +41 -0
- package/methodology/MOCs are attention management devices not just organizational tools.md +51 -0
- package/methodology/PKM failure follows a predictable cycle.md +50 -0
- package/methodology/ThreadMode to DocumentMode transformation is the core value creation step.md +52 -0
- package/methodology/WIP limits force processing over accumulation.md +53 -0
- package/methodology/Zeigarnik effect validates capture-first philosophy because open loops drain attention.md +42 -0
- package/methodology/academic research uses structured extraction with cross-source synthesis.md +566 -0
- package/methodology/adapt the four-phase processing pipeline to domain-specific throughput needs.md +197 -0
- package/methodology/agent notes externalize navigation intuition that search cannot discover and traversal cannot reconstruct.md +48 -0
- package/methodology/agent self-memory should be architecturally separate from user knowledge systems.md +48 -0
- package/methodology/agent session boundaries create natural automation checkpoints that human-operated systems lack.md +56 -0
- package/methodology/agent-cognition.md +107 -0
- package/methodology/agents are simultaneously methodology executors and subjects creating a unique trust asymmetry.md +66 -0
- package/methodology/aspect-oriented programming solved the same cross-cutting concern problem that hooks solve.md +39 -0
- package/methodology/associative ontologies beat hierarchical taxonomies because heterarchy adapts while hierarchy brittles.md +53 -0
- package/methodology/attention residue may have a minimum granularity that cannot be subdivided.md +46 -0
- package/methodology/auto-commit hooks eliminate prospective memory failures by converting remember-to-act into guaranteed execution.md +47 -0
- package/methodology/automated detection is always safe because it only reads state while automated remediation risks content corruption.md +42 -0
- package/methodology/automation should be retired when its false positive rate exceeds its true positive rate or it catches zero issues.md +56 -0
- package/methodology/backlinks implicitly define notes by revealing usage context.md +35 -0
- package/methodology/backward maintenance asks what would be different if written today.md +62 -0
- package/methodology/balance onboarding enforcement and questions to prevent premature complexity.md +229 -0
- package/methodology/basic level categorization determines optimal MOC granularity.md +51 -0
- package/methodology/batching by context similarity reduces switching costs in agent processing.md +43 -0
- package/methodology/behavioral anti-patterns matter more than tool selection.md +42 -0
- package/methodology/betweenness centrality identifies bridge notes connecting disparate knowledge domains.md +57 -0
- package/methodology/blueprints that teach construction outperform downloads that provide pre-built code for platform-dependent modules.md +42 -0
- package/methodology/bootstrapping principle enables self-improving systems.md +62 -0
- package/methodology/build automatic memory through cognitive offloading and session handoffs.md +285 -0
- package/methodology/capture the reaction to content not just the content itself.md +41 -0
- package/methodology/claims must be specific enough to be wrong.md +36 -0
- package/methodology/closure rituals create clean breaks that prevent attention residue bleed.md +44 -0
- package/methodology/cognitive offloading is the architectural foundation for vault design.md +46 -0
- package/methodology/cognitive outsourcing risk in agent-operated systems.md +55 -0
- package/methodology/coherence maintains consistency despite inconsistent inputs.md +96 -0
- package/methodology/coherent architecture emerges from wiki links spreading activation and small-world topology.md +48 -0
- package/methodology/community detection algorithms can inform when MOCs should split or merge.md +52 -0
- package/methodology/complete navigation requires four complementary types that no single mechanism provides.md +43 -0
- package/methodology/complex systems evolve from simple working systems.md +59 -0
- package/methodology/composable knowledge architecture builds systems from independent toggleable modules not monolithic templates.md +61 -0
- package/methodology/compose multi-domain systems through separate templates and shared graph.md +372 -0
- package/methodology/concept-orientation beats source-orientation for cross-domain connections.md +51 -0
- package/methodology/confidence thresholds gate automated action between the mechanical and judgment zones.md +50 -0
- package/methodology/configuration dimensions interact so choices in one create pressure on others.md +58 -0
- package/methodology/configuration paralysis emerges when derivation surfaces too many decisions.md +44 -0
- package/methodology/context files function as agent operating systems through self-referential self-extension.md +46 -0
- package/methodology/context phrase clarity determines how deep a navigation hierarchy can scale.md +46 -0
- package/methodology/continuous small-batch processing eliminates review dread.md +48 -0
- package/methodology/controlled disorder engineers serendipity through semantic rather than topical linking.md +51 -0
- package/methodology/creative writing uses worldbuilding consistency with character tracking.md +672 -0
- package/methodology/cross-links between MOC territories indicate creative leaps and integration depth.md +43 -0
- package/methodology/dangling links reveal which notes want to exist.md +62 -0
- package/methodology/data exit velocity measures how quickly content escapes vendor lock-in.md +74 -0
- package/methodology/decontextualization risk means atomicity may strip meaning that cannot be recovered.md +48 -0
- package/methodology/dense interlinked research claims enable derivation while sparse references only enable templating.md +47 -0
- package/methodology/dependency resolution through topological sort makes module composition transparent and verifiable.md +56 -0
- package/methodology/derivation generates knowledge systems from composable research claims not template customization.md +63 -0
- package/methodology/derivation-engine.md +27 -0
- package/methodology/derived systems follow a seed-evolve-reseed lifecycle.md +56 -0
- package/methodology/description quality for humans diverges from description quality for keyword search.md +73 -0
- package/methodology/descriptions are retrieval filters not summaries.md +112 -0
- package/methodology/design MOCs as attention management devices with lifecycle governance.md +318 -0
- package/methodology/design-dimensions.md +66 -0
- package/methodology/digital mutability enables note evolution that physical permanence forbids.md +54 -0
- package/methodology/discovery-retrieval.md +48 -0
- package/methodology/distinctiveness scoring treats description quality as measurable.md +69 -0
- package/methodology/does agent processing recover what fast capture loses.md +43 -0
- package/methodology/domain-compositions.md +37 -0
- package/methodology/dual-coding with visual elements could enhance agent traversal.md +55 -0
- package/methodology/each module must be describable in one sentence under 200 characters or it does too many things.md +45 -0
- package/methodology/each new note compounds value by creating traversal paths.md +55 -0
- package/methodology/eight configuration dimensions parameterize the space of possible knowledge systems.md +56 -0
- package/methodology/elaborative encoding is the quality gate for new notes.md +55 -0
- package/methodology/enforce schema with graduated strictness across capture processing and query zones.md +221 -0
- package/methodology/enforcing atomicity can create paralysis when ideas resist decomposition.md +43 -0
- package/methodology/engineering uses technical decision tracking with architectural memory.md +766 -0
- package/methodology/every knowledge domain shares a four-phase processing skeleton that diverges only in the process step.md +53 -0
- package/methodology/evolution observations provide actionable signals for system adaptation.md +67 -0
- package/methodology/external memory shapes cognition more than base model.md +60 -0
- package/methodology/faceted classification treats notes as multi-dimensional objects rather than folder contents.md +65 -0
- package/methodology/failure-modes.md +27 -0
- package/methodology/false universalism applies same processing logic regardless of domain.md +49 -0
- package/methodology/federated wiki pattern enables multi-agent divergence as feature not bug.md +59 -0
- package/methodology/flat files break at retrieval scale.md +75 -0
- package/methodology/forced engagement produces weak connections.md +48 -0
- package/methodology/four abstraction layers separate platform-agnostic from platform-dependent knowledge system features.md +47 -0
- package/methodology/fresh context per task preserves quality better than chaining phases.md +44 -0
- package/methodology/friction reveals architecture.md +63 -0
- package/methodology/friction-driven module adoption prevents configuration debt by adding complexity only at pain points.md +48 -0
- package/methodology/gardening cycle implements tend prune fertilize operations.md +41 -0
- package/methodology/generation effect gate blocks processing without transformation.md +40 -0
- package/methodology/goal-driven memory orchestration enables autonomous domain learning through directed compute allocation.md +41 -0
- package/methodology/good descriptions layer heuristic then mechanism then implication.md +57 -0
- package/methodology/graph-structure.md +65 -0
- package/methodology/guided notes might outperform post-hoc structuring for high-volume capture.md +37 -0
- package/methodology/health wellness uses symptom-trigger correlation with multi-dimensional tracking.md +819 -0
- package/methodology/hook composition creates emergent methodology from independent single-concern components.md +47 -0
- package/methodology/hook enforcement guarantees quality while instruction enforcement merely suggests it.md +51 -0
- package/methodology/hook-driven learning loops create self-improving methodology through observation accumulation.md +62 -0
- package/methodology/hooks are the agent habit system that replaces the missing basal ganglia.md +40 -0
- package/methodology/hooks cannot replace genuine cognitive engagement yet more automation is always tempting.md +87 -0
- package/methodology/hooks enable context window efficiency by delegating deterministic checks to external processes.md +47 -0
- package/methodology/idempotent maintenance operations are safe to automate because running them twice produces the same result as running them once.md +44 -0
- package/methodology/implement condition-based maintenance triggers for derived systems.md +255 -0
- package/methodology/implicit dependencies create distributed monoliths that fail silently across configurations.md +58 -0
- package/methodology/implicit knowledge emerges from traversal.md +55 -0
- package/methodology/incremental formalization happens through repeated touching of old notes.md +60 -0
- package/methodology/incremental reading enables cross-source connection finding.md +39 -0
- package/methodology/index.md +32 -0
- package/methodology/inline links carry richer relationship data than metadata fields.md +91 -0
- package/methodology/insight accretion differs from productivity in knowledge systems.md +41 -0
- package/methodology/intermediate packets enable assembly over creation.md +52 -0
- package/methodology/intermediate representation pattern enables reliable vault operations beyond regex.md +62 -0
- package/methodology/justification chains enable forward backward and evolution reasoning about configuration decisions.md +46 -0
- package/methodology/knowledge system architecture is parameterized by platform capabilities not fixed by methodology.md +51 -0
- package/methodology/knowledge systems become communication partners through complexity and memory humans cannot sustain.md +47 -0
- package/methodology/knowledge systems share universal operations and structural components across all methodology traditions.md +46 -0
- package/methodology/legal case management uses precedent chains with regulatory change propagation.md +892 -0
- package/methodology/live index via periodic regeneration keeps discovery current.md +58 -0
- package/methodology/local-first file formats are inherently agent-native.md +69 -0
- package/methodology/logic column pattern separates reasoning from procedure.md +35 -0
- package/methodology/maintenance operations are more universal than creative pipelines because structural health is domain-invariant.md +47 -0
- package/methodology/maintenance scheduling frequency should match consequence speed not detection capability.md +50 -0
- package/methodology/maintenance targeting should prioritize mechanism and theory notes.md +26 -0
- package/methodology/maintenance-patterns.md +72 -0
- package/methodology/markdown plus YAML plus ripgrep implements a queryable graph database without infrastructure.md +55 -0
- package/methodology/maturity field enables agent context prioritization.md +33 -0
- package/methodology/memory-architecture.md +27 -0
- package/methodology/metacognitive confidence can diverge from retrieval capability.md +42 -0
- package/methodology/metadata reduces entropy enabling precision over recall.md +91 -0
- package/methodology/methodology development should follow the trajectory from documentation to skill to hook as understanding hardens.md +80 -0
- package/methodology/methodology traditions are named points in a shared configuration space not competing paradigms.md +64 -0
- package/methodology/mnemonic medium embeds verification into navigation.md +46 -0
- package/methodology/module communication through shared YAML fields creates loose coupling without direct dependencies.md +44 -0
- package/methodology/module deactivation must account for structural artifacts that survive the toggle.md +49 -0
- package/methodology/multi-domain systems compose through separate templates and shared graph.md +61 -0
- package/methodology/multi-domain-composition.md +27 -0
- package/methodology/narrow folksonomy optimizes for single-operator retrieval unlike broad consensus tagging.md +53 -0
- package/methodology/navigation infrastructure passes through distinct scaling regimes that require qualitative strategy shifts.md +48 -0
- package/methodology/navigational vertigo emerges in pure association systems without local hierarchy.md +54 -0
- package/methodology/note titles should function as APIs enabling sentence transclusion.md +51 -0
- package/methodology/note-design.md +57 -0
- package/methodology/notes are skills /342/200/224 curated knowledge injected when relevant.md" +62 -0
- package/methodology/notes function as cognitive anchors that stabilize attention during complex tasks.md +41 -0
- package/methodology/novel domains derive by mapping knowledge type to closest reference domain then adapting.md +50 -0
- package/methodology/nudge theory explains graduated hook enforcement as choice architecture for agents.md +59 -0
- package/methodology/observation and tension logs function as dead-letter queues for failed automation.md +51 -0
- package/methodology/operational memory and knowledge memory serve different functions in agent architecture.md +48 -0
- package/methodology/operational wisdom requires contextual observation.md +52 -0
- package/methodology/orchestrated vault creation transforms arscontexta from tool to autonomous knowledge factory.md +40 -0
- package/methodology/organic emergence versus active curation creates a fundamental vault governance tension.md +68 -0
- package/methodology/orphan notes are seeds not failures.md +38 -0
- package/methodology/over-automation corrupts quality when hooks encode judgment rather than verification.md +62 -0
- package/methodology/people relationships uses Dunbar-layered graphs with interaction tracking.md +659 -0
- package/methodology/personal assistant uses life area management with review automation.md +610 -0
- package/methodology/platform adapter translation is semantic not mechanical because hook event meanings differ.md +40 -0
- package/methodology/platform capability tiers determine which knowledge system features can be implemented.md +48 -0
- package/methodology/platform fragmentation means identical conceptual operations require different implementations across agent environments.md +44 -0
- package/methodology/premature complexity is the most common derivation failure mode.md +45 -0
- package/methodology/prevent domain-specific failure modes through the vulnerability matrix.md +336 -0
- package/methodology/processing effort should follow retrieval demand.md +57 -0
- package/methodology/processing-workflows.md +75 -0
- package/methodology/product management uses feedback pipelines with experiment tracking.md +789 -0
- package/methodology/productivity porn risk in meta-system building.md +30 -0
- package/methodology/programmable notes could enable property-triggered workflows.md +64 -0
- package/methodology/progressive disclosure means reading right not reading less.md +69 -0
- package/methodology/progressive schema validates only what active modules require not the full system schema.md +49 -0
- package/methodology/project management uses decision tracking with stakeholder context.md +776 -0
- package/methodology/propositional link semantics transform wiki links from associative to reasoned.md +87 -0
- package/methodology/prospective memory requires externalization.md +53 -0
- package/methodology/provenance tracks where beliefs come from.md +62 -0
- package/methodology/queries evolve during search so agents should checkpoint.md +35 -0
- package/methodology/question-answer metadata enables inverted search patterns.md +39 -0
- package/methodology/random note resurfacing prevents write-only memory.md +33 -0
- package/methodology/reconciliation loops that compare desired state to actual state enable drift correction without continuous monitoring.md +59 -0
- package/methodology/reflection synthesizes existing notes into new insight.md +100 -0
- package/methodology/retrieval utility should drive design over capture completeness.md +69 -0
- package/methodology/retrieval verification loop tests description quality at scale.md +81 -0
- package/methodology/role field makes graph structure explicit.md +94 -0
- package/methodology/scaffolding enables divergence that fine-tuning cannot.md +67 -0
- package/methodology/schema enforcement via validation agents enables soft consistency.md +60 -0
- package/methodology/schema evolution follows observe-then-formalize not design-then-enforce.md +65 -0
- package/methodology/schema field names are the only domain specific element in the universal note pattern.md +46 -0
- package/methodology/schema fields should use domain-native vocabulary not abstract terminology.md +47 -0
- package/methodology/schema templates reduce cognitive overhead at capture time.md +55 -0
- package/methodology/schema validation hooks externalize inhibitory control that degrades under cognitive load.md +48 -0
- package/methodology/schema-enforcement.md +27 -0
- package/methodology/self-extension requires context files to contain platform operations knowledge not just methodology.md +47 -0
- package/methodology/sense-making vs storage does compression lose essential nuance.md +73 -0
- package/methodology/session boundary hooks implement cognitive bookends for orientation and reflection.md +60 -0
- package/methodology/session handoff creates continuity without persistent memory.md +43 -0
- package/methodology/session outputs are packets for future selves.md +43 -0
- package/methodology/session transcript mining enables experiential validation that structural tests cannot provide.md +38 -0
- package/methodology/skill context budgets constrain knowledge system complexity on agent platforms.md +52 -0
- package/methodology/skills encode methodology so manual execution bypasses quality gates.md +50 -0
- package/methodology/small-world topology requires hubs and dense local links.md +99 -0
- package/methodology/source attribution enables tracing claims to foundations.md +38 -0
- package/methodology/spaced repetition scheduling could optimize vault maintenance.md +44 -0
- package/methodology/spreading activation models how agents should traverse.md +79 -0
- package/methodology/stale navigation actively misleads because agents trust curated maps completely.md +43 -0
- package/methodology/stigmergy coordinates agents through environmental traces without direct communication.md +62 -0
- package/methodology/storage versus thinking distinction determines which tool patterns apply.md +56 -0
- package/methodology/structure enables navigation without reading everything.md +52 -0
- package/methodology/structure without processing provides no value.md +56 -0
- package/methodology/student learning uses prerequisite graphs with spaced retrieval.md +770 -0
- package/methodology/summary coherence tests composability before filing.md +37 -0
- package/methodology/tag rot applies to wiki links because titles serve as both identifier and display text.md +50 -0
- package/methodology/temporal media must convert to spatial text for agent traversal.md +43 -0
- package/methodology/temporal processing priority creates age-based inbox urgency.md +45 -0
- package/methodology/temporal separation of capture and processing preserves context freshness.md +39 -0
- package/methodology/ten universal primitives form the kernel of every viable agent knowledge system.md +162 -0
- package/methodology/testing effect could enable agent knowledge verification.md +38 -0
- package/methodology/the AgentSkills standard embodies progressive disclosure at the skill level.md +40 -0
- package/methodology/the derivation engine improves recursively as deployed systems generate observations.md +49 -0
- package/methodology/the determinism boundary separates hook methodology from skill methodology.md +46 -0
- package/methodology/the fix-versus-report decision depends on determinism reversibility and accumulated trust.md +45 -0
- package/methodology/the generation effect requires active transformation not just storage.md +57 -0
- package/methodology/the no wrong patches guarantee ensures any valid module combination produces a valid system.md +58 -0
- package/methodology/the system is the argument.md +46 -0
- package/methodology/the vault constitutes identity for agents.md +86 -0
- package/methodology/the vault methodology transfers because it encodes cognitive science not domain specifics.md +47 -0
- package/methodology/therapy journal uses warm personality with pattern detection for emotional processing.md +584 -0
- package/methodology/three capture schools converge through agent-mediated synthesis.md +55 -0
- package/methodology/three concurrent maintenance loops operate at different timescales to catch different classes of problems.md +56 -0
- package/methodology/throughput matters more than accumulation.md +58 -0
- package/methodology/title as claim enables traversal as reasoning.md +50 -0
- package/methodology/topological organization beats temporal for knowledge work.md +52 -0
- package/methodology/trading uses conviction tracking with thesis-outcome correlation.md +699 -0
- package/methodology/trails transform ephemeral navigation into persistent artifacts.md +39 -0
- package/methodology/transform universal vocabulary to domain-native language through six levels.md +259 -0
- package/methodology/type field enables structured queries without folder hierarchies.md +53 -0
- package/methodology/use-case presets dissolve the tension between composability and simplicity.md +44 -0
- package/methodology/vault conventions may impose hidden rigidity on thinking.md +44 -0
- package/methodology/verbatim risk applies to agents too.md +31 -0
- package/methodology/vibe notetaking is the emerging industry consensus for AI-native self-organization.md +56 -0
- package/methodology/vivid memories need verification.md +45 -0
- package/methodology/vocabulary-transformation.md +27 -0
- package/methodology/voice capture is the highest-bandwidth channel for agent-delegated knowledge systems.md +45 -0
- package/methodology/wiki links are the digital evolution of analog indexing.md +73 -0
- package/methodology/wiki links as social contract transforms agents into stewards of incomplete references.md +52 -0
- package/methodology/wiki links create navigation paths that shape retrieval.md +63 -0
- package/methodology/wiki links implement GraphRAG without the infrastructure.md +101 -0
- package/methodology/writing for audience blocks authentic creation.md +22 -0
- package/methodology/you operate a system that takes notes.md +79 -0
- package/openclaw/SKILL.md +110 -0
- package/package.json +45 -0
- package/platforms/README.md +51 -0
- package/platforms/claude-code/generator.md +61 -0
- package/platforms/claude-code/hooks/README.md +186 -0
- package/platforms/claude-code/hooks/auto-commit.sh.template +38 -0
- package/platforms/claude-code/hooks/session-capture.sh.template +72 -0
- package/platforms/claude-code/hooks/session-orient.sh.template +189 -0
- package/platforms/claude-code/hooks/write-validate.sh.template +106 -0
- package/platforms/openclaw/generator.md +82 -0
- package/platforms/openclaw/hooks/README.md +89 -0
- package/platforms/openclaw/hooks/bootstrap.ts.template +224 -0
- package/platforms/openclaw/hooks/command-new.ts.template +165 -0
- package/platforms/openclaw/hooks/heartbeat.ts.template +214 -0
- package/platforms/shared/features/README.md +70 -0
- package/platforms/shared/skill-blocks/graph.md +145 -0
- package/platforms/shared/skill-blocks/learn.md +119 -0
- package/platforms/shared/skill-blocks/next.md +131 -0
- package/platforms/shared/skill-blocks/pipeline.md +326 -0
- package/platforms/shared/skill-blocks/ralph.md +616 -0
- package/platforms/shared/skill-blocks/reduce.md +1142 -0
- package/platforms/shared/skill-blocks/refactor.md +129 -0
- package/platforms/shared/skill-blocks/reflect.md +780 -0
- package/platforms/shared/skill-blocks/remember.md +524 -0
- package/platforms/shared/skill-blocks/rethink.md +574 -0
- package/platforms/shared/skill-blocks/reweave.md +680 -0
- package/platforms/shared/skill-blocks/seed.md +320 -0
- package/platforms/shared/skill-blocks/stats.md +145 -0
- package/platforms/shared/skill-blocks/tasks.md +171 -0
- package/platforms/shared/skill-blocks/validate.md +323 -0
- package/platforms/shared/skill-blocks/verify.md +562 -0
- package/platforms/shared/templates/README.md +35 -0
- package/presets/experimental/categories.yaml +1 -0
- package/presets/experimental/preset.yaml +38 -0
- package/presets/experimental/starter/README.md +7 -0
- package/presets/experimental/vocabulary.yaml +7 -0
- package/presets/personal/categories.yaml +7 -0
- package/presets/personal/preset.yaml +41 -0
- package/presets/personal/starter/goals.md +21 -0
- package/presets/personal/starter/index.md +17 -0
- package/presets/personal/starter/life-areas.md +21 -0
- package/presets/personal/starter/people.md +21 -0
- package/presets/personal/vocabulary.yaml +32 -0
- package/presets/research/categories.yaml +8 -0
- package/presets/research/preset.yaml +41 -0
- package/presets/research/starter/index.md +17 -0
- package/presets/research/starter/methods.md +21 -0
- package/presets/research/starter/open-questions.md +21 -0
- package/presets/research/vocabulary.yaml +33 -0
- package/reference/AUDIT-REPORT.md +238 -0
- package/reference/claim-map.md +172 -0
- package/reference/components.md +327 -0
- package/reference/conversation-patterns.md +542 -0
- package/reference/derivation-validation.md +649 -0
- package/reference/dimension-claim-map.md +134 -0
- package/reference/evolution-lifecycle.md +297 -0
- package/reference/failure-modes.md +235 -0
- package/reference/interaction-constraints.md +204 -0
- package/reference/kernel.yaml +242 -0
- package/reference/methodology.md +283 -0
- package/reference/open-questions.md +279 -0
- package/reference/personality-layer.md +302 -0
- package/reference/self-space.md +299 -0
- package/reference/semantic-vs-keyword.md +288 -0
- package/reference/session-lifecycle.md +298 -0
- package/reference/templates/base-note.md +16 -0
- package/reference/templates/companion-note.md +70 -0
- package/reference/templates/creative-note.md +16 -0
- package/reference/templates/learning-note.md +16 -0
- package/reference/templates/life-note.md +16 -0
- package/reference/templates/moc.md +26 -0
- package/reference/templates/relationship-note.md +17 -0
- package/reference/templates/research-note.md +19 -0
- package/reference/templates/session-log.md +24 -0
- package/reference/templates/therapy-note.md +16 -0
- package/reference/test-fixtures/edge-case-constraints.md +148 -0
- package/reference/test-fixtures/multi-domain.md +164 -0
- package/reference/test-fixtures/novel-domain-gaming.md +138 -0
- package/reference/test-fixtures/research-minimal.md +102 -0
- package/reference/test-fixtures/therapy-full.md +155 -0
- package/reference/testing-milestones.md +1087 -0
- package/reference/three-spaces.md +363 -0
- package/reference/tradition-presets.md +203 -0
- package/reference/use-case-presets.md +341 -0
- package/reference/validate-kernel.sh +432 -0
- package/reference/vocabulary-transforms.md +85 -0
- package/scripts/sync-thinking.sh +147 -0
- package/skill-sources/graph/SKILL.md +567 -0
- package/skill-sources/graph/skill.json +17 -0
- package/skill-sources/learn/SKILL.md +254 -0
- package/skill-sources/learn/skill.json +17 -0
- package/skill-sources/next/SKILL.md +407 -0
- package/skill-sources/next/skill.json +17 -0
- package/skill-sources/pipeline/SKILL.md +314 -0
- package/skill-sources/pipeline/skill.json +17 -0
- package/skill-sources/ralph/SKILL.md +604 -0
- package/skill-sources/ralph/skill.json +17 -0
- package/skill-sources/reduce/SKILL.md +1113 -0
- package/skill-sources/reduce/skill.json +17 -0
- package/skill-sources/refactor/SKILL.md +448 -0
- package/skill-sources/refactor/skill.json +17 -0
- package/skill-sources/reflect/SKILL.md +747 -0
- package/skill-sources/reflect/skill.json +17 -0
- package/skill-sources/remember/SKILL.md +534 -0
- package/skill-sources/remember/skill.json +17 -0
- package/skill-sources/rethink/SKILL.md +658 -0
- package/skill-sources/rethink/skill.json +17 -0
- package/skill-sources/reweave/SKILL.md +657 -0
- package/skill-sources/reweave/skill.json +17 -0
- package/skill-sources/seed/SKILL.md +303 -0
- package/skill-sources/seed/skill.json +17 -0
- package/skill-sources/stats/SKILL.md +371 -0
- package/skill-sources/stats/skill.json +17 -0
- package/skill-sources/tasks/SKILL.md +402 -0
- package/skill-sources/tasks/skill.json +17 -0
- package/skill-sources/validate/SKILL.md +310 -0
- package/skill-sources/validate/skill.json +17 -0
- package/skill-sources/verify/SKILL.md +532 -0
- package/skill-sources/verify/skill.json +17 -0
- package/skills/add-domain/SKILL.md +441 -0
- package/skills/add-domain/skill.json +17 -0
- package/skills/architect/SKILL.md +568 -0
- package/skills/architect/skill.json +17 -0
- package/skills/ask/SKILL.md +388 -0
- package/skills/ask/skill.json +17 -0
- package/skills/health/SKILL.md +760 -0
- package/skills/health/skill.json +17 -0
- package/skills/help/SKILL.md +348 -0
- package/skills/help/skill.json +17 -0
- package/skills/recommend/SKILL.md +553 -0
- package/skills/recommend/skill.json +17 -0
- package/skills/reseed/SKILL.md +385 -0
- package/skills/reseed/skill.json +17 -0
- package/skills/setup/SKILL.md +1688 -0
- package/skills/setup/skill.json +17 -0
- package/skills/tutorial/SKILL.md +496 -0
- package/skills/tutorial/skill.json +17 -0
- package/skills/upgrade/SKILL.md +395 -0
- package/skills/upgrade/skill.json +17 -0
|
@@ -0,0 +1,56 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: Minimum viable seeding, friction-driven evolution, principled restructuring when incoherence accumulates — reseeding re-derives from first principles enriched by operational experience
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[design-dimensions]]", "[[maintenance-patterns]]"]
|
|
5
|
+
methodology: ["Systems Theory", "Original"]
|
|
6
|
+
source: [[knowledge-system-derivation-blueprint]]
|
|
7
|
+
---
|
|
8
|
+
|
|
9
|
+
# derived systems follow a seed-evolve-reseed lifecycle
|
|
10
|
+
|
|
11
|
+
Knowledge systems derived from research claims do not follow a simple create-then-maintain trajectory. They move through three qualitatively different phases, and understanding which phase you are in determines what actions are appropriate.
|
|
12
|
+
|
|
13
|
+
**Seeding** is initial derivation constrained by Gall's Law. Since [[complex systems evolve from simple working systems]], the derived system must start as a minimum viable configuration: two or three note types, a handful of MOCs, basic folder structure, and a context file that teaches the agent how to operate and — crucially — how to evolve. Because [[methodology traditions are named points in a shared configuration space not competing paradigms]], the seed can start from a tradition's pre-validated configuration rather than composing from raw dimensions, and because [[novel domains derive by mapping knowledge type to closest reference domain then adapting]], unfamiliar domains can use reference domains as seeding points with adaptation axes defining the initial evolution trajectory. The temptation is to derive the theoretically optimal configuration from the claim graph. But even a perfectly justified complex system will collapse under its own weight because the countless micro-adaptations that make a system work can only emerge through use. What makes this incremental adoption safe is that [[the no wrong patches guarantee ensures any valid module combination produces a valid system]] — enabling a module cannot break existing structure, so the seed phase can start minimal with confidence that each subsequent addition is non-destructive. The most accessible seeding mechanism is the use-case preset. Since [[use-case presets dissolve the tension between composability and simplicity]], presets bundle a curated module selection under a use-case label — the user selects "Research Vault" and gets thirteen modules covering the full pipeline, without composing the selection from scratch. The preset resolves dimension coupling and dependency chains at activation time, producing a minimum viable configuration that the user can evolve through friction-driven module toggling. The seed includes evolution guidelines: when to add a field, when to create a new MOC, when to adjust processing cadence. These guidelines are as important as the initial structure because they encode the conditions under which complexity is justified.
|
|
14
|
+
|
|
15
|
+
**Evolution** is growth through observed friction. Since [[friction-driven module adoption prevents configuration debt by adding complexity only at pain points]], the evolution phase has concrete thresholds: add a module after five manual repetitions of the process it would automate, split a module when its description exceeds 500 characters, remove after three unused sessions, and cap active modules at fifteen to twenty. Because [[composable knowledge architecture builds systems from independent toggleable modules not monolithic templates]], each capability is an independent module with explicit dependencies, so evolution means toggling modules on at these calibrated friction points rather than restructuring a monolithic configuration — the agent adds yaml-schema when structured metadata becomes necessary, processing-pipeline when manual workflows create friction, each addition resolving its own dependencies and composing with everything already active. The agent monitors which note types get used, which fields actually get filled, where navigation fails, where processing produces unlinked notes. Each adaptation traces back to a specific observation mapped to a knowledge graph claim — not intuition but evidence. Since [[schema evolution follows observe-then-formalize not design-then-enforce]], the concrete protocol for schema evolution is a quarterly review with five signals: add a field when 20% of notes manually include it, demote required fields that collect placeholder values, prune unused enums, formalize patterned free text, split MOCs that exceed navigation thresholds. Since [[bootstrapping principle enables self-improving systems]], this phase is bootstrapping in action: the system uses its current capabilities to identify and implement improvements, with each improvement becoming available for the next cycle. And since [[incremental formalization happens through repeated touching of old notes]], the notes themselves are evolving in parallel with the system — claims sharpen, connections densify, vague inklings crystallize — so the system and its content co-evolve. The encoding trajectory mirrors this: since [[methodology development should follow the trajectory from documentation to skill to hook as understanding hardens]], methodology patterns within the system follow their own seed-evolve cycle — seeded as documentation in the context file, evolved through friction encounters into skills, and hardened into hooks when understanding is sufficient.
|
|
16
|
+
|
|
17
|
+
But evolution has a limit, and without the transition to reseeding, [[PKM failure follows a predictable cycle]] — over-engineering and analysis paralysis set in as people attempt to fix systemic incoherence with local adaptations. Small adaptations accumulate side effects. A field added here creates a query expectation there. A new MOC shifts navigation patterns that other MOCs assumed were stable. Because [[configuration dimensions interact so choices in one create pressure on others]], changes in one dimension create pressure on others, and accumulated incremental changes can drift the configuration into an incoherent region where individual choices are locally justified but globally contradictory. Schema drift, navigation degradation, processing pipelines that no longer match the content they handle — these are symptoms of a system that has evolved past its derived coherence without restructuring. And since [[module deactivation must account for structural artifacts that survive the toggle]], experimentation itself becomes a source of incoherence: each module that was enabled, tested, and then disabled leaves ghost YAML fields and orphaned MOC links that accumulate as structural debris proportional to the number of experiments attempted.
|
|
18
|
+
|
|
19
|
+
**Reseeding** is the response: principled restructuring that re-derives the system using original constraints enriched by accumulated observations. This is fundamentally different from both incremental evolution and routine maintenance. Since [[backward maintenance asks what would be different if written today]], reweaving asks this question at the note level — what would this note say given current understanding? Reseeding asks it at the system level — what would this entire architecture look like given what we now know about how it gets used? The scale difference matters because systemic incoherence cannot be fixed by improving individual notes. You need to step back and reconsider the templates, the pipeline structure, the MOC hierarchy, the processing cadence — the framework that everything else depends on. And since [[every knowledge domain shares a four-phase processing skeleton that diverges only in the process step]], reseeding does not redesign the skeleton itself — capture, process, connect, verify remain invariant — but restructures the process step implementation, the schema definitions, and the navigation hierarchy that sit on top of it.
|
|
20
|
+
|
|
21
|
+
What makes reseeding principled rather than ad hoc is that [[derivation generates knowledge systems from composable research claims not template customization]]. The same claim graph that produced the initial seed now has additional inputs: every observation logged during evolution, every tension discovered through use, every adaptation and its justification. Re-derivation traverses the enriched claim graph with tighter constraints: not "what does the research suggest for this use case?" but "what does the research suggest for this use case given these specific operational realities?" The output preserves all content, all links, all accumulated understanding — it restructures the framework around them.
|
|
22
|
+
|
|
23
|
+
The lifecycle is not linear but spiral. After reseeding, a new evolution phase begins from the restructured base, accumulating new observations until the next reseeding is warranted. Each cycle produces a system that is both more adapted to its actual use patterns and more coherent in its configuration choices. The claim graph itself improves through this process: observations from deployed systems feed back as new claims or sharpened existing ones, making subsequent derivations — for this system or any other — more grounded.
|
|
24
|
+
|
|
25
|
+
There is a tension in recognizing when evolution should give way to reseeding. Too early, and you discard adaptations that haven't had time to prove their value. Too late, and incoherence compounds until the system's behavior becomes unpredictable. The trigger signals are systemic rather than local: not "this field doesn't work" (that is an evolution-phase adaptation) but "our schema, navigation, and processing have drifted apart" (that requires reseeding). Since [[gardening cycle implements tend prune fertilize operations]], the gardening metaphor clarifies the distinction: tending and fertilizing are evolution-phase operations that maintain and connect existing plants. Reseeding redesigns the garden's layout — which beds exist, how paths connect them, what the irrigation system looks like — while preserving the plants themselves.
|
|
26
|
+
|
|
27
|
+
The practical implication is that derived systems need to be designed for reseeding from the start. Since [[context files function as agent operating systems through self-referential self-extension]], the context file is the natural home for reseeding meta-documentation — not just how to operate but how to re-derive: what constraints were used, what trade-offs were made and why, what observations would trigger reconsideration. A system that lacks this meta-documentation can still be reseeded, but the re-derivation loses access to the reasoning behind the original choices, making it more likely to repeat mistakes the first derivation already resolved.
|
|
28
|
+
---
|
|
29
|
+
|
|
30
|
+
Relevant Notes:
|
|
31
|
+
- [[complex systems evolve from simple working systems]] — provides the theoretical foundation: Gall's Law explains why seeding must start simple, but this note adds the reseeding phase that Gall's Law alone does not describe
|
|
32
|
+
- [[bootstrapping principle enables self-improving systems]] — evolution is bootstrapping in action, but reseeding is a phase transition that bootstrapping alone cannot accomplish because the system must step outside itself to re-derive
|
|
33
|
+
- [[backward maintenance asks what would be different if written today]] — reweaving operates at the note level while reseeding operates at the system level: both ask what current understanding would change, but at different scales
|
|
34
|
+
- [[derivation generates knowledge systems from composable research claims not template customization]] — derivation is the mechanism that makes reseeding principled rather than ad hoc: the claim graph provides the substrate for re-derivation
|
|
35
|
+
- [[incremental formalization happens through repeated touching of old notes]] — evolution phase works through incremental formalization at the note level, but reseeding recognizes that accumulated incremental changes can produce systemic incoherence that incremental approaches cannot fix
|
|
36
|
+
- [[configuration dimensions interact so choices in one create pressure on others]] — dimension coupling explains why evolution eventually requires reseeding: small adaptations in one dimension accumulate pressure on others until the configuration drifts into an incoherent region
|
|
37
|
+
- [[gardening cycle implements tend prune fertilize operations]] — tend and fertilize are evolution-phase operations while reseeding is a qualitatively different act: restructuring the garden's layout rather than maintaining its plants
|
|
38
|
+
- [[evolution observations provide actionable signals for system adaptation]] — provides the specific diagnostic protocol for the evolution phase: six observation patterns map to structural causes, telling the agent whether accumulated drift is still incremental (evolution-phase correction) or has crossed into systemic incoherence (reseeding trigger)
|
|
39
|
+
- [[methodology development should follow the trajectory from documentation to skill to hook as understanding hardens]] — the encoding trajectory is a lifecycle within the evolution phase: methodology patterns seed as documentation, evolve through friction, and reseed at the next encoding level when understanding hardens
|
|
40
|
+
- [[schema evolution follows observe-then-formalize not design-then-enforce]] — the schema-specific instantiation of the evolution phase: the quarterly review protocol with five concrete signals is how schemas evolve through observation rather than upfront design
|
|
41
|
+
- [[PKM failure follows a predictable cycle]] — the failure cascade is what happens when the evolution phase stalls and reseeding never occurs: over-engineering and analysis paralysis are symptoms of attempting local fixes for systemic incoherence
|
|
42
|
+
- [[every knowledge domain shares a four-phase processing skeleton that diverges only in the process step]] — clarifies what reseeding actually restructures: the skeleton itself (capture-process-connect-verify) is invariant, so reseeding targets the process step implementation, templates, and MOC hierarchy rather than the pipeline shape
|
|
43
|
+
- [[methodology traditions are named points in a shared configuration space not competing paradigms]] — traditions serve as pre-validated seeding points: a derived system can seed from a tradition's coherent configuration rather than composing from raw dimensions
|
|
44
|
+
- [[novel domains derive by mapping knowledge type to closest reference domain then adapting]] — extends the seeding concept: reference domains function as seeds for novel domains, with adaptation axes defining the initial evolution trajectory
|
|
45
|
+
- [[context files function as agent operating systems through self-referential self-extension]] — the context file is the substrate where reseeding meta-documentation lives: self-referential self-extension enables the context file to teach not just operation but re-derivation
|
|
46
|
+
- [[premature complexity is the most common derivation failure mode]] — the complexity budget is the seeding constraint: minimum viable configuration at seed time with evolution guidelines encoding when to add complexity rather than front-loading it; the budget constrains initial deployment between the kernel floor and the locally-justified-but-globally-unsustainable maximum
|
|
47
|
+
- [[configuration paralysis emerges when derivation surfaces too many decisions]] — configuration paralysis can prevent the seeding phase from completing: if the derivation interface surfaces every dimension as a question, the user never finishes setup, so the system never reaches the evolution phase where friction-driven learning begins
|
|
48
|
+
- [[the no wrong patches guarantee ensures any valid module combination produces a valid system]] — the safety property that makes the seed-evolve cycle reliable: each module toggle during seed and evolution phases is non-destructive because the guarantee ensures valid combinations never corrupt data or break structure
|
|
49
|
+
- [[composable knowledge architecture builds systems from independent toggleable modules not monolithic templates]] — the implementation pattern that makes the lifecycle's module-level toggling possible: seeding activates a minimal module set, evolution toggles modules on at friction points, and reseeding reconfigures the module selection; composable architecture is the engineering substrate on which the lifecycle operates
|
|
50
|
+
- [[use-case presets dissolve the tension between composability and simplicity]] — the most accessible seeding mechanism: presets bundle curated module selections under use-case labels so seeding requires one choice rather than individual module evaluation, with the composable substrate enabling friction-driven evolution after the preset provides initial configuration
|
|
51
|
+
- [[module deactivation must account for structural artifacts that survive the toggle]] — deactivation debris as a reseeding trigger: experimentation during the evolution phase leaves ghost YAML fields and orphaned MOC links from disabled modules, accumulating structural debris that comprehensive cleanup during reseeding addresses as part of principled restructuring
|
|
52
|
+
- [[friction-driven module adoption prevents configuration debt by adding complexity only at pain points]] — the evolution phase's specific mechanism: concrete thresholds (5-repetition addition, 500-char split, 3-session removal, 15-20 module cap) operationalize friction-driven toggling, preventing both premature adoption and configuration debt during the evolution phase
|
|
53
|
+
|
|
54
|
+
Topics:
|
|
55
|
+
- [[design-dimensions]]
|
|
56
|
+
- [[maintenance-patterns]]
|
|
@@ -0,0 +1,73 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: Descriptions that pass prediction tests (5/5) can fail BM25 retrieval because human-scannable prose uses connective vocabulary that dilutes keyword match scores — two optimization targets, not one
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[discovery-retrieval]]"]
|
|
5
|
+
methodology: ["Original"]
|
|
6
|
+
---
|
|
7
|
+
|
|
8
|
+
# description quality for humans diverges from description quality for keyword search
|
|
9
|
+
|
|
10
|
+
The recite skill's dual-test architecture — prediction scoring followed by retrieval testing — exposed a gap that the vault's description philosophy hadn't anticipated. Notes can score 5/5 on prediction (an agent reading title and description accurately predicts the note's content) while completely failing BM25 retrieval (searching the description text returns zero results). This isn't a bug in either test. It reveals that "description quality" is actually two different optimization targets that can pull in opposite directions.
|
|
11
|
+
|
|
12
|
+
## The Mechanism
|
|
13
|
+
|
|
14
|
+
Human-scannable descriptions use connective prose: "lossy compression optimized for decision-making, not content summary" reads naturally and enables prediction because the connective words ("optimized for," "not") create logical structure. But since [[BM25 retrieval fails on full-length descriptions because query term dilution reduces match scores]], every word in a BM25 query competes for scoring budget. Those connective words dilute the matching — common words like "optimized," "not," "for" consume IDF weight without contributing to retrieval. The distinctive terms that would actually match ("lossy," "compression," "retrieval," "filter") get proportionally less influence.
|
|
15
|
+
|
|
16
|
+
This is why since [[good descriptions layer heuristic then mechanism then implication]], the layering formula can paradoxically harm retrieval. A well-layered description spends character budget on three distinct information types connected by prose transitions. A retrieval-optimized description would strip those transitions and pack in more distinctive keywords. The formula that makes descriptions excellent for human scanning makes them worse for keyword search.
|
|
17
|
+
|
|
18
|
+
## Why This Matters for the Vault
|
|
19
|
+
|
|
20
|
+
The foundation claim that [[descriptions are retrieval filters not summaries]] assumes a single filter function. But this divergence reveals two distinct filter channels:
|
|
21
|
+
|
|
22
|
+
1. **Human scanning channel** — an agent reads descriptions sequentially, using natural language comprehension to assess relevance. Here, prose quality, logical structure, and layered information all help. Descriptions optimized for this channel are well-written sentences.
|
|
23
|
+
|
|
24
|
+
2. **Keyword search channel** — BM25 or similar search engines match query terms against description text. Here, keyword density, distinctive vocabulary, and minimal filler words help. Descriptions optimized for this channel are keyword-dense phrases.
|
|
25
|
+
|
|
26
|
+
These channels make different demands. A description like "cognitive modes conflict so separating them in time preserves quality of both" works beautifully for channel 1 — you understand the claim, the mechanism, the implication. For channel 2, the same text is a poor query: "cognitive," "modes," "conflict," "separating," "time," "preserves," "quality" are all common words that match widely.
|
|
27
|
+
|
|
28
|
+
Since [[narrow folksonomy optimizes for single-operator retrieval unlike broad consensus tagging]], the personal vocabulary problem compounds this. Idiosyncratic terms like "slip-box" or "agent-vault" or "systems-theoretic" are highly distinctive for human scanning — they immediately signal domain and perspective. But their very idiosyncrasy makes them sparse in the corpus, meaning BM25 has fewer documents to score against, and competing documents with more common vocabulary can outrank.
|
|
29
|
+
|
|
30
|
+
## The Dual Optimization Problem
|
|
31
|
+
|
|
32
|
+
Since [[metadata reduces entropy enabling precision over recall]], the entropy reduction argument assumes a unified retrieval mechanism. But the two channels reduce entropy differently:
|
|
33
|
+
|
|
34
|
+
- **Scanning** reduces entropy through comprehension — the agent understands what makes this note different from alternatives by processing the description's meaning
|
|
35
|
+
- **Keyword search** reduces entropy through term matching — the search engine narrows candidates by statistical co-occurrence of query terms and document terms
|
|
36
|
+
|
|
37
|
+
The same description text cannot simultaneously maximize both. Connective prose helps comprehension but hurts term matching. Keyword density helps term matching but reduces readability. The current vault philosophy implicitly optimizes for channel 1 (human scanning) because descriptions are designed to "add new information beyond the title" in natural prose. Channel 2 (keyword retrieval) is an afterthought. This is a specific instance of the broader tension that [[sense-making vs storage does compression lose essential nuance]] — but here the compression loss is channel-dependent rather than absolute. The nuance "lost" for BM25 (connective prose structure) is the nuance "preserved" for human scanning, and vice versa.
|
|
38
|
+
|
|
39
|
+
## Practical Resolution
|
|
40
|
+
|
|
41
|
+
The resolution is not to choose one channel over the other but to recognize which channel matters when. Since [[retrieval verification loop tests description quality at scale]] runs both prediction and retrieval tests, the dual-test architecture already captures this divergence. The improvement is in how we interpret results:
|
|
42
|
+
|
|
43
|
+
- **Prediction pass, retrieval fail** — description is well-written but keyword-sparse. This is the exact pathway where [[metacognitive confidence can diverge from retrieval capability]] — the description looks correct through the prediction test while failing the retrieval test. Consider: is keyword retrieval actually needed for this note, or will scanning and semantic search always find it? If BM25 matters, add distinctive keywords without destroying prose quality.
|
|
44
|
+
|
|
45
|
+
- **Prediction fail, retrieval pass** — description has good keywords but poor information structure. Apply the layering formula.
|
|
46
|
+
|
|
47
|
+
- **Both fail** — description needs rewriting on both dimensions.
|
|
48
|
+
|
|
49
|
+
- **Both pass** — description is genuinely high-quality across channels.
|
|
50
|
+
|
|
51
|
+
The deeper question is whether BM25 retrieval of descriptions even matters given that the vault has semantic search via qmd. Vector search (vsearch) finds meaning across vocabulary differences — exactly the divergence that BM25 misses. If the primary retrieval path is `qmd query` or `qmd vsearch`, then optimizing descriptions for BM25 is solving a secondary problem. The prediction test (channel 1) may be the only quality dimension that genuinely matters, with keyword retrieval as a degraded fallback that should not drive description design.
|
|
52
|
+
|
|
53
|
+
But fallbacks matter precisely when primary systems fail. Since qmd has documented reliability issues (stale indices, segfaults, MCP disposal errors), BM25 via `qmd search` is often the only working retrieval. Descriptions that fail BM25 are invisible during qmd outages — creating a reliability gap at the description layer that compounds the infrastructure reliability gap at the search layer. This is also evidence that [[vault conventions may impose hidden rigidity on thinking]] — the prose description convention, while justified for scanning, creates an invisible constraint on fallback retrieval that only surfaces during infrastructure failures.
|
|
54
|
+
|
|
55
|
+
---
|
|
56
|
+
|
|
57
|
+
Source: /rethink session (2026-02-08), promoted from [[description quality for humans diverges from description quality for keyword search]]
|
|
58
|
+
---
|
|
59
|
+
|
|
60
|
+
Relevant Notes:
|
|
61
|
+
- [[descriptions are retrieval filters not summaries]] — foundation that assumes one filter function, but this note reveals filtering splits into two distinct channels with different requirements
|
|
62
|
+
- [[good descriptions layer heuristic then mechanism then implication]] — the layering formula optimizes for human scanning but may inadvertently harm keyword retrieval by consuming character budget with connective prose
|
|
63
|
+
- [[retrieval verification loop tests description quality at scale]] — the dual-test architecture (prediction + retrieval) that first exposed this divergence; the loop correctly identifies it but the scoring conflates two separate quality dimensions
|
|
64
|
+
- [[distinctiveness scoring treats description quality as measurable]] — measures inter-note confusion via similarity but doesn't distinguish whether distinctiveness serves human scanning or machine retrieval
|
|
65
|
+
- [[metadata reduces entropy enabling precision over recall]] — the entropy reduction argument assumes a single retrieval channel, but BM25 and human scanning reduce entropy through different mechanisms
|
|
66
|
+
- [[narrow folksonomy optimizes for single-operator retrieval unlike broad consensus tagging]] — personal vocabulary amplifies this divergence because idiosyncratic terms may be highly distinctive for human scanning while being sparse in the BM25 index
|
|
67
|
+
- [[BM25 retrieval fails on full-length descriptions because query term dilution reduces match scores]] — provides the mechanism: IDF scoring explains WHY connective prose dilutes keyword retrieval, grounding this note's observation in specific search engine behavior
|
|
68
|
+
- [[sense-making vs storage does compression lose essential nuance]] — the dual-channel divergence is a specific instance of this tension: what counts as essential nuance depends on which retrieval channel the compression serves
|
|
69
|
+
- [[metacognitive confidence can diverge from retrieval capability]] — extends: prediction-pass retrieval-fail is a concrete pathway for the confidence-capability divergence where the system looks correct through one test while failing another
|
|
70
|
+
- [[vault conventions may impose hidden rigidity on thinking]] — the ~150 char prose description convention implicitly optimizes for human scanning, imposing a hidden constraint on keyword retrieval that this note makes visible
|
|
71
|
+
|
|
72
|
+
Topics:
|
|
73
|
+
- [[discovery-retrieval]]
|
|
@@ -0,0 +1,112 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: Note descriptions function as lossy compression enabling agents to filter before loading full content, which is information-theoretically sound
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[discovery-retrieval]]"]
|
|
5
|
+
methodology: ["Cornell"]
|
|
6
|
+
---
|
|
7
|
+
|
|
8
|
+
# descriptions are retrieval filters not summaries
|
|
9
|
+
|
|
10
|
+
The primary purpose of a note description is not to summarize what the note contains, but to help agents decide whether to load the full content. This is the retrieval filter function, and it maps directly to Cornell Note-Taking's cue column pattern.
|
|
11
|
+
|
|
12
|
+
## The Information-Theoretic Basis
|
|
13
|
+
|
|
14
|
+
Description as filter is information-theoretically sound: it's lossy compression that preserves decision-relevant features while discarding detail. The description answers "should I read this?" not "what does this say?" This is exactly what Shannon's rate-distortion theory predicts: optimal compression for a given task throws away information that doesn't help with that task.
|
|
15
|
+
|
|
16
|
+
Aggregating descriptions creates a pre-computed low-entropy representation of the system. Scanning 50 descriptions costs fewer tokens than reading 5 full notes to find the relevant one. The description layer enables efficient filtering.
|
|
17
|
+
|
|
18
|
+
## The Cornell Cue Column Pattern
|
|
19
|
+
|
|
20
|
+
Cornell Note-Taking has three sections:
|
|
21
|
+
- **Notes**: detailed content
|
|
22
|
+
- **Cue column**: questions or keywords that trigger recall
|
|
23
|
+
- **Summary**: synthesis across notes
|
|
24
|
+
|
|
25
|
+
The cue column is explicitly designed as a retrieval aid. You write cues that help you find and recall content later. This is the same function descriptions serve: helping future retrieval by capturing what makes this note distinctive. Since [[wiki links are the digital evolution of analog indexing]], the cue column was an early form of the same indexing pattern that wiki links now implement digitally — both create navigable pointers that serve as entry points into content. Cornell actually has two compression mechanisms operating at different points: because [[summary coherence tests composability before filing]], the summary section tests structural coherence at filing time (is this unit singular?), while descriptions enable retrieval filtering at query time (should I load this?). Both compress, but for different purposes.
|
|
26
|
+
|
|
27
|
+
The pattern validates that descriptions work as filters: humans need them too, not just agents. The difference is that agents can read YAML frontmatter programmatically, while humans read cue columns visually.
|
|
28
|
+
|
|
29
|
+
## Progressive Disclosure Pattern
|
|
30
|
+
|
|
31
|
+
The discovery layers implement this:
|
|
32
|
+
|
|
33
|
+
1. **File tree**: titles only (highest compression)
|
|
34
|
+
2. **Descriptions**: titles + descriptions (medium compression)
|
|
35
|
+
3. **Outline**: headings structure (low compression)
|
|
36
|
+
4. **Full content**: everything (no compression)
|
|
37
|
+
|
|
38
|
+
Each layer adds fidelity but costs more tokens. Since [[progressive disclosure means reading right not reading less]], this isn't about minimizing tokens — it's about curating what deserves the full-depth treatment. [[spreading activation models how agents should traverse]] implements this as decay-based context loading, and descriptions occupy a specific position: high-decay traversal stops at the description layer. An agent scanning descriptions activates many concepts at low depth. Loading full files activates fewer concepts at high depth. The description layer is the sweet spot for filtering: enough context to decide, low enough cost to scan many at once. And since [[AI shifts knowledge systems from externalizing memory to externalizing attention]], the description layer is not merely a retrieval optimization — it is an attention externalization mechanism. The system pre-computes which notes deserve deeper attention by encoding decision-relevant features into descriptions, directing the agent's finite focus rather than merely helping it recall what it stored.
|
|
39
|
+
|
|
40
|
+
## The Anti-Pattern: Description as Summary
|
|
41
|
+
|
|
42
|
+
When descriptions just restate the title in different words, they waste the agent's context. This is the same failure mode as vague claims: since [[claims must be specific enough to be wrong]], descriptions that merely paraphrase add nothing to disagree with or build on. This maps directly to [[the generation effect requires active transformation not just storage]]: paraphrase is not generation. Writing a description that restates the title produces nothing new — no cognitive hooks, no new information, no retrieval value. A good description generates something: mechanism, implication, or scope that the title doesn't contain. That generation is what creates the filter value. The experiment [[verbatim risk applies to agents too]] tests whether agents commit this anti-pattern at scale — producing well-structured outputs that reorganize content without genuine synthesis. If that experiment validates, description writing becomes a specific detection point: descriptions that merely paraphrase are evidence of the verbatim failure mode. This happens when you treat description as "mini-summary" rather than "retrieval cue."
|
|
43
|
+
|
|
44
|
+
Since [[testing effect could enable agent knowledge verification]], the paraphrase anti-pattern becomes testable. When an agent reads only title and description, then predicts note content, paraphrase descriptions should fail — they don't contain enough distinct information to enable accurate prediction. The testing effect experiment's first pre-registered prediction directly tests this: notes flagged by recite as poorly-retrievable should have descriptions that merely paraphrase titles. If validated, this provides an objective diagnostic for the anti-pattern rather than relying on subjective judgment about whether a description "adds enough."
|
|
45
|
+
|
|
46
|
+
Bad:
|
|
47
|
+
- Title: "vector proximity measures surface overlap not deep connection"
|
|
48
|
+
- Description: "Semantic similarity captures surface-level overlap rather than genuine conceptual relationships"
|
|
49
|
+
|
|
50
|
+
This description adds no new information. It just paraphrases the title. An agent scanning descriptions can't use this to filter.
|
|
51
|
+
|
|
52
|
+
Good:
|
|
53
|
+
- Title: "vector proximity measures surface overlap not deep connection"
|
|
54
|
+
- Description: "Two notes about the same concept with different vocabulary score high, while genuinely related ideas across domains score low — embeddings miss what matters"
|
|
55
|
+
|
|
56
|
+
This description adds mechanism (vocabulary vs concepts) and implication (cross-domain connections fail). Now an agent knows: this note is about limitations of embedding-based search, specifically the cross-domain problem.
|
|
57
|
+
|
|
58
|
+
## Implications
|
|
59
|
+
|
|
60
|
+
Since descriptions function as retrieval filters, the quality standard is: does this help an agent decide whether to load the full note? Not: does this summarize the note accurately? An active experiment, [[question-answer metadata enables inverted search patterns]], tests whether an explicit `answers:` YAML field would extend this further — storing not just retrieval cues but the specific questions the note answers. If validated, question-matching would become an even faster filter than description scanning.
|
|
61
|
+
|
|
62
|
+
But the compression that enables filtering may lose exactly what makes complex ideas valuable. Since [[sense-making vs storage does compression lose essential nuance]], some knowledge types may resist the ~150-character filter format: contextual knowledge whose meaning depends on situation, procedural nuance where tacit judgment matters, or ideas whose distinctive features ARE the subtle details that don't compress well. The description layer works when it preserves enough distinctiveness for correct filtering — but for ideas where distinctiveness IS the nuance, compression may create systematic retrieval failures that go undetected because we never find what we're missing.
|
|
63
|
+
|
|
64
|
+
### Distinctiveness Over Comprehensiveness
|
|
65
|
+
|
|
66
|
+
The information-theoretic framing implies a specific design criterion: descriptions should maximize distinctiveness within the corpus rather than maximize comprehensiveness. Since [[metadata reduces entropy enabling precision over recall]], the goal is pre-computing low-entropy representations that shrink the search space. A description that tries to cover everything a note contains dilutes its filter value. A description that captures what makes this note different from similar notes creates high information content relative to the corpus.
|
|
67
|
+
|
|
68
|
+
This is the difference between:
|
|
69
|
+
- **Comprehensive**: "Discusses how descriptions work and why they matter for retrieval" (could apply to many notes)
|
|
70
|
+
- **Distinctive**: "Two notes about the same concept with different vocabulary score high, while genuinely related ideas across domains score low" (immediately identifies this specific note)
|
|
71
|
+
|
|
72
|
+
The discriminating question when writing descriptions: "Would this description help an agent choose THIS note over similar notes on related topics?" If the description could plausibly apply to multiple notes in the system, it lacks distinctiveness and fails as a filter.
|
|
73
|
+
|
|
74
|
+
This changes what makes a good description:
|
|
75
|
+
- Add NEW information beyond the title
|
|
76
|
+
- Capture scope, mechanism, or implication
|
|
77
|
+
- Include distinctive keywords for semantic search
|
|
78
|
+
- Enable filtering decisions without reading full content
|
|
79
|
+
- Maximize what distinguishes this note from related notes
|
|
80
|
+
|
|
81
|
+
Since [[good descriptions layer heuristic then mechanism then implication]], an effective structure emerges: lead with the actionable heuristic (what to do), back with mechanism (why it works), end with operational implication (what this means for practice). This layered formula operationalizes the distinctiveness criterion — each layer adds information that the previous layers don't contain.
|
|
82
|
+
|
|
83
|
+
But the description schema itself is a convention that may constrain. Since [[vault conventions may impose hidden rigidity on thinking]], the ~150-character limit forces a particular condensation style. Visual insights, procedural knowledge, or context-dependent ideas might resist this format — they can't be reduced to a sentence without losing something. The experiment tracks whether YAML metadata requirements accumulate into hidden rigidity.
|
|
84
|
+
|
|
85
|
+
The description layer becomes more valuable as descriptions improve, because better filters mean fewer wasted full-note loads. Since [[throughput matters more than accumulation]], this directly serves the success metric: filtering speed determines processing velocity from capture to synthesis. Every millisecond saved on filtering compounds across every retrieval operation. But the existence of descriptions doesn't guarantee they function as filters: since [[metacognitive confidence can diverge from retrieval capability]], a vault with complete description coverage may produce false confidence that the filtering layer works while actual retrieval systematically fails. This is the organized graveyard at the filtering layer — descriptions exist but don't enable filtering decisions.
|
|
86
|
+
|
|
87
|
+
Since [[retrieval verification loop tests description quality at scale]], this gap between "descriptions exist" and "descriptions work" becomes empirically testable. Systematic prediction-then-verify cycles across all notes with scoring reveal which descriptions fail as filters, common failure modes, and whether description quality correlates with note type or age. The loop turns the filter assumption into measurable infrastructure health.
|
|
88
|
+
|
|
89
|
+
The filter function itself splits across retrieval channels. Descriptions optimized for agent scanning — prose with connective tissue, layered information, natural flow — work well when agents read descriptions sequentially. But since [[BM25 retrieval fails on full-length descriptions because query term dilution reduces match scores]], those same prose descriptions fail keyword search because common words dilute the distinctive terms that would identify the right note. The filter works for one retrieval channel while failing another, and since [[description quality for humans diverges from description quality for keyword search]], this split reveals that "description quality" is not one dimension but two separate optimization targets. Human scanning and keyword matching make opposing demands on the same ~150 characters, so the single filter function this note assumes is actually a dual filter serving channels with different requirements.
|
|
90
|
+
---
|
|
91
|
+
|
|
92
|
+
Relevant Notes:
|
|
93
|
+
- [[progressive disclosure means reading right not reading less]] — articulates the philosophy: disclosure layers enable curation for quality, not token savings
|
|
94
|
+
- [[good descriptions layer heuristic then mechanism then implication]] — provides the structural formula: lead with actionable heuristic, back with mechanism, end with operational implication
|
|
95
|
+
- [[spreading activation models how agents should traverse]] — positions descriptions as the high-decay layer: agents can activate many concepts at description depth without loading full files
|
|
96
|
+
- [[processing effort should follow retrieval demand]] — descriptions enable JIT processing by providing efficient filtering before full retrieval
|
|
97
|
+
- [[queries evolve during search so agents should checkpoint]] — at each checkpoint, descriptions enable fast reassessment without loading full notes
|
|
98
|
+
- [[throughput matters more than accumulation]] — filtering speed directly serves the success metric: faster filtering means faster processing velocity
|
|
99
|
+
- [[the generation effect requires active transformation not just storage]] — explains why paraphrase fails as a description: generation creates value, restatement doesn't
|
|
100
|
+
- [[vault conventions may impose hidden rigidity on thinking]] — tests whether the description schema and other conventions accumulate into constraints that certain thinking styles can't fit
|
|
101
|
+
- [[testing effect could enable agent knowledge verification]] — tests whether the paraphrase anti-pattern causes measurable retrieval failures, providing objective diagnostic for description quality
|
|
102
|
+
- [[retrieval verification loop tests description quality at scale]] — operationalizes the testing effect as systematic vault-wide assessment: scoring across all notes turns description quality from assumption to measured property
|
|
103
|
+
- [[verbatim risk applies to agents too]] — the broader experiment testing whether agents produce well-structured outputs without genuine insight; description-as-paraphrase is a specific manifestation of this failure mode
|
|
104
|
+
- [[metacognitive confidence can diverge from retrieval capability]] — tests whether description coverage produces false confidence: descriptions may exist without functioning as filters, creating the organized graveyard at the filtering layer
|
|
105
|
+
- [[sense-making vs storage does compression lose essential nuance]] — the open tension: lossy compression is theoretically sound, but some knowledge types may resist the format in ways that create silent retrieval failures
|
|
106
|
+
- [[AI shifts knowledge systems from externalizing memory to externalizing attention]] — paradigm frame: description-based filtering is a form of attention externalization; the system pre-computes what deserves deeper focus rather than merely helping agents recall what was stored
|
|
107
|
+
- [[narrow folksonomy optimizes for single-operator retrieval unlike broad consensus tagging]] — explains why description optimization works: single-operator systems can tune ~150-char descriptions entirely to one agent's retrieval patterns without accommodating diverse users, making the filter layer maximally effective for the operator who will actually use it
|
|
108
|
+
- [[BM25 retrieval fails on full-length descriptions because query term dilution reduces match scores]] — reveals that the filter function splits across channels: prose descriptions that work for agent scanning fail keyword search because connective words dilute IDF scoring
|
|
109
|
+
- [[description quality for humans diverges from description quality for keyword search]] — develops the full consequence: the filter function is not one dimension but two, with human scanning and keyword matching as separate optimization targets that can pull in opposite directions
|
|
110
|
+
|
|
111
|
+
Topics:
|
|
112
|
+
- [[discovery-retrieval]]
|
|
@@ -0,0 +1,318 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: MOC best practices for derived knowledge systems — hierarchy patterns, lifecycle management, and health metrics adapted across domains
|
|
3
|
+
kind: guidance
|
|
4
|
+
status: active
|
|
5
|
+
topics: ["[[note-design]]", "[[maintenance-patterns]]"]
|
|
6
|
+
---
|
|
7
|
+
|
|
8
|
+
# design MOCs as attention management devices with lifecycle governance
|
|
9
|
+
|
|
10
|
+
MOCs (Maps of Content) are attention management devices. Since [[MOCs are attention management devices not just organizational tools]], they don't just organize notes — they tell agents where to focus, what connects to what, and where synthesis opportunities exist. Every derived vault needs MOCs, but different domains need different MOC patterns.
|
|
11
|
+
|
|
12
|
+
This doc tells the plugin HOW to generate MOC structures for each domain.
|
|
13
|
+
|
|
14
|
+
## Why MOCs Are Non-Negotiable
|
|
15
|
+
|
|
16
|
+
Without MOCs, agents face navigational vertigo. Since [[navigational vertigo emerges in pure association systems without local hierarchy]], a flat vault of interconnected notes provides no starting point for orientation. MOCs solve this by creating curated entry points — the agent doesn't read everything; it reads the MOC and follows relevant links.
|
|
17
|
+
|
|
18
|
+
Since [[MOC construction forces synthesis that automated generation from metadata cannot replicate]], MOCs carry intellectual value beyond navigation. The act of curating which notes appear in a MOC and writing context phrases for each link IS synthesis. Automated MOCs generated from metadata tags lack this synthesis — they're address books, not maps.
|
|
19
|
+
|
|
20
|
+
Since [[complete navigation requires four complementary types that no single mechanism provides]], MOCs fill the local navigation role — the "what's nearby?" question that search, hub navigation, and inline links cannot answer. Missing this one type creates a predictable blind spot: agents can find things (search) and follow connections (links) but cannot orient within a domain.
|
|
21
|
+
|
|
22
|
+
## The Dump-Lump-Jump Construction Process
|
|
23
|
+
|
|
24
|
+
Nick Milo's construction pattern has three phases that do fundamentally different cognitive work:
|
|
25
|
+
|
|
26
|
+
### Phase 1: Dump (Gather Everything)
|
|
27
|
+
|
|
28
|
+
Pull every note that might belong to the MOC's territory. This is exhaustive collection without judgment — cast the net wide. Automated tools handle this well: tag queries, semantic search, and link graph traversal can identify candidate notes reliably.
|
|
29
|
+
|
|
30
|
+
### Phase 2: Lump (Classify and Context)
|
|
31
|
+
|
|
32
|
+
Group notes by relationship, write context phrases for each link explaining WHY to follow it. This is where since [[basic level categorization determines optimal MOC granularity]], the builder exercises categorization judgment — deciding which notes cluster together, what the groupings mean, and whether the current granularity is right.
|
|
33
|
+
|
|
34
|
+
Context phrases are mandatory. Every link in a MOC must explain its relationship:
|
|
35
|
+
|
|
36
|
+
**Good:** `- [[mood patterns predict trigger sensitivity]] — connects daily mood ratings to identified triggers, showing which moods lower the threshold`
|
|
37
|
+
|
|
38
|
+
**Bad:** `- [[mood patterns predict trigger sensitivity]]` (bare link — address book entry, not a map)
|
|
39
|
+
|
|
40
|
+
The context phrase tells the agent: "If you're looking for X, follow this link." Without it, the agent must read every linked note to decide if it's relevant.
|
|
41
|
+
|
|
42
|
+
### Phase 3: Jump (Synthesize)
|
|
43
|
+
|
|
44
|
+
Identify tensions between notes, write the orientation synthesis, capture insights that emerge from seeing the collection as a whole. The Jump phase is where the MOC transcends its index function. Two clusters might be in tension. A gap might become visible. A higher-order claim might emerge from the arrangement.
|
|
45
|
+
|
|
46
|
+
**Why the Jump phase matters for agents:** Since [[the generation effect requires active transformation not just storage]], the cognitive value accrues to whoever does the construction. Even if an automated system could produce identical output, the act of construction produces understanding that informs future reasoning.
|
|
47
|
+
|
|
48
|
+
**Why manual synthesis matters:** Automated MOC generation can produce the Dump phase flawlessly and approximate the Lump phase through clustering. But it cannot perform the Jump because the Jump requires judgment about what matters, what conflicts, and what the collection means as a whole. The resulting MOC is structurally valid but functionally hollow — it satisfies the navigation need without providing genuine orientation, so the agent never looks further.
|
|
49
|
+
|
|
50
|
+
## Seven MOC Types
|
|
51
|
+
|
|
52
|
+
Different organizational needs call for different MOC structures. The plugin recognizes seven types:
|
|
53
|
+
|
|
54
|
+
### 1. Hub MOC
|
|
55
|
+
|
|
56
|
+
Entry point for the entire vault. Pure navigation, under 100 lines. Links to domain MOCs with brief orientation.
|
|
57
|
+
|
|
58
|
+
**Example:** `index.md` — one per vault, rarely changes, highest traffic.
|
|
59
|
+
|
|
60
|
+
### 2. Domain MOC
|
|
61
|
+
|
|
62
|
+
Entry point for a research area. Contains cross-cutting synthesis and links to topic MOCs.
|
|
63
|
+
|
|
64
|
+
**Example:** `knowledge-work.md` — links to graph-structure, agent-cognition, processing-workflow, etc.
|
|
65
|
+
|
|
66
|
+
### 3. Topic MOC
|
|
67
|
+
|
|
68
|
+
Active workspace for a specific subdomain. Core ideas, tensions, gaps, agent notes. This is where most intellectual work happens.
|
|
69
|
+
|
|
70
|
+
**Example:** `graph-structure.md` — 15-40 notes about graph topology, navigation, linking patterns.
|
|
71
|
+
|
|
72
|
+
### 4. Self MOC
|
|
73
|
+
|
|
74
|
+
Agent identity and operational memory. Flat peer structure, no hierarchy between peers.
|
|
75
|
+
|
|
76
|
+
**Example:** `identity.md`, `methodology.md`, `goals.md` — facets of agent state.
|
|
77
|
+
|
|
78
|
+
### 5. Operational MOC
|
|
79
|
+
|
|
80
|
+
Procedure tracking with atomic entries in subdirectories. Groups operational notes by category.
|
|
81
|
+
|
|
82
|
+
**Example:** `observations.md` — categorizes observation notes by type (methodology, process gap, friction).
|
|
83
|
+
|
|
84
|
+
### 6. Content Hub MOC
|
|
85
|
+
|
|
86
|
+
Domain workflow hub combining navigation with operational guidance. Lives at the root of a domain folder.
|
|
87
|
+
|
|
88
|
+
**Example:** `twitter.md` — navigation to drafts, published content, people, plus workflow instructions.
|
|
89
|
+
|
|
90
|
+
### 7. Entity MOC
|
|
91
|
+
|
|
92
|
+
Per-person, per-project, or per-character tracking. Accumulates context about a specific entity across time.
|
|
93
|
+
|
|
94
|
+
**Example:** `people/angie-bowen.md` — interaction history, relationship context, follow-up items.
|
|
95
|
+
|
|
96
|
+
**When to use which:**
|
|
97
|
+
|
|
98
|
+
| Need | MOC Type | Why |
|
|
99
|
+
|------|----------|-----|
|
|
100
|
+
| Vault-wide orientation | Hub | Single entry point, rare changes |
|
|
101
|
+
| Research area overview | Domain | Cross-cutting synthesis across topics |
|
|
102
|
+
| Active knowledge work | Topic | Working surface for claims, tensions, gaps |
|
|
103
|
+
| Agent self-knowledge | Self | Flat peers for identity facets |
|
|
104
|
+
| Operational logging | Operational | Categorized atomic entries |
|
|
105
|
+
| Workflow + navigation | Content Hub | Domain that needs both process and map |
|
|
106
|
+
| Tracking entities | Entity | Accumulating context about people, projects, characters |
|
|
107
|
+
|
|
108
|
+
## Three Hierarchy Patterns
|
|
109
|
+
|
|
110
|
+
The plugin chooses from three patterns based on domain characteristics:
|
|
111
|
+
|
|
112
|
+
### Pattern A: Three-Tier Research
|
|
113
|
+
|
|
114
|
+
```
|
|
115
|
+
Hub (index.md) -> Domain MOCs -> Topic MOCs -> Claim Notes
|
|
116
|
+
```
|
|
117
|
+
|
|
118
|
+
**Best for:** Domains with deep conceptual structure, many interconnected ideas, and synthesis as a primary value. The hierarchy provides layered orientation: hub for broad navigation, domain for area context, topic for active workspace.
|
|
119
|
+
|
|
120
|
+
**Domain examples:** Research and Academic, Legal (precedent hierarchies), Engineering (architecture domains), Student Learning (course and concept hierarchies)
|
|
121
|
+
|
|
122
|
+
**Plugin generates:**
|
|
123
|
+
- One hub MOC linking domain MOCs
|
|
124
|
+
- Domain MOCs for major knowledge areas
|
|
125
|
+
- Topic MOCs when 20+ notes cluster around a subtopic
|
|
126
|
+
- Auto-split suggestion when topic MOCs exceed 50 notes
|
|
127
|
+
|
|
128
|
+
### Pattern B: Flat Peer
|
|
129
|
+
|
|
130
|
+
```
|
|
131
|
+
MOC-A | MOC-B | MOC-C | MOC-D (no hierarchy, equal peers)
|
|
132
|
+
```
|
|
133
|
+
|
|
134
|
+
**Best for:** Domains with distinct areas that don't nest hierarchically. Each area is a self-contained workspace. No area is "parent" to another.
|
|
135
|
+
|
|
136
|
+
**Domain examples:** Personal Assistant (life areas), Health and Wellness (fitness / nutrition / sleep / symptoms as peers), Trading (strategy areas as peers)
|
|
137
|
+
|
|
138
|
+
**Plugin generates:**
|
|
139
|
+
- One MOC per life area or dimension
|
|
140
|
+
- No hub MOC — each area is a direct entry point
|
|
141
|
+
- Cross-area links in Agent Notes sections to surface correlations
|
|
142
|
+
|
|
143
|
+
### Pattern C: Hub-Plus-Entities
|
|
144
|
+
|
|
145
|
+
```
|
|
146
|
+
Hub MOC -> Entity MOCs (person/project/character)
|
|
147
|
+
-> Entity-specific notes
|
|
148
|
+
```
|
|
149
|
+
|
|
150
|
+
**Best for:** Domains centered around entities (people, projects, characters, cases) where each entity accumulates its own context over time.
|
|
151
|
+
|
|
152
|
+
**Domain examples:** People and Relationships (person MOCs), Creative Writing (character/location MOCs), Product Management (feature/customer MOCs), Project Management (project MOCs), Legal (case MOCs)
|
|
153
|
+
|
|
154
|
+
**Plugin generates:**
|
|
155
|
+
- One hub MOC with overview and quick reference
|
|
156
|
+
- Entity MOCs created dynamically as entities are added
|
|
157
|
+
- Entity subdirectories for clean organization
|
|
158
|
+
- Relationship-between-entities tracking
|
|
159
|
+
|
|
160
|
+
**Decision framework:** Active research -> Three-Tier. Agent identity -> Flat Peer. Domain workflow with entities -> Hub-Plus-Entities. Never exceed 4 tiers.
|
|
161
|
+
|
|
162
|
+
## MOC Content Rules
|
|
163
|
+
|
|
164
|
+
The plugin generates MOCs with these structural rules:
|
|
165
|
+
|
|
166
|
+
### Context Phrases Are Mandatory
|
|
167
|
+
|
|
168
|
+
Every link in a MOC must have a context phrase explaining WHY to follow that link. Since [[context phrase clarity determines how deep a navigation hierarchy can scale]], the quality of these phrases directly determines how far the MOC hierarchy can scale. Ambiguous phrases force agents to load linked notes to assess relevance, converting attention savings back into attention cost.
|
|
169
|
+
|
|
170
|
+
### Synthesis Lives in Notes, Not MOCs
|
|
171
|
+
|
|
172
|
+
When a MOC section starts containing developed arguments (multi-paragraph analysis, detailed reasoning), that content should be extracted into a synthesis note and the MOC should link to it.
|
|
173
|
+
|
|
174
|
+
MOCs navigate. Notes argue. The plugin flags when MOC sections exceed ~200 words as a signal to extract.
|
|
175
|
+
|
|
176
|
+
### Agent Notes Section
|
|
177
|
+
|
|
178
|
+
Every MOC should end with an Agent Notes section containing:
|
|
179
|
+
- Navigation shortcuts (e.g., "Start with note X if you're looking for Y")
|
|
180
|
+
- Dead ends documented (e.g., "Note Z looks relevant but doesn't lead anywhere for this topic")
|
|
181
|
+
- Productive combinations (e.g., "Notes A and B together form a complete argument about...")
|
|
182
|
+
|
|
183
|
+
Since [[agent notes externalize navigation intuition that search cannot discover and traversal cannot reconstruct]], agent notes provide strategic attention management — while the synthesis paragraph orients to WHAT the topic contains, agent notes orient to HOW to navigate it.
|
|
184
|
+
|
|
185
|
+
Agent notes are updated during reflect/reweave passes. They encode the navigation intuition that search alone cannot provide.
|
|
186
|
+
|
|
187
|
+
## Staleness Risk
|
|
188
|
+
|
|
189
|
+
Since [[stale navigation actively misleads because agents trust curated maps completely]], MOC staleness is the single most dangerous maintenance failure in agent-operated systems.
|
|
190
|
+
|
|
191
|
+
**Why staleness is worse than absence:** A missing MOC causes vertigo — disorienting but honest. The agent falls back to search, which accesses current content. A stale MOC is deceptive. It presents an outdated view as current, and the agent has no mechanism to suspect otherwise. Notes created after the last MOC update are invisible — not because they're hard to find, but because the agent never looks for them. The curated map satisfies the navigation need, so no further search occurs.
|
|
192
|
+
|
|
193
|
+
**Why agents are especially vulnerable:** Humans retain cross-session memory that might trigger doubt ("I remember writing something about this"). Agents have zero such intuition. Each session loads the MOC and treats its contents as authoritative.
|
|
194
|
+
|
|
195
|
+
**Detection mechanisms:**
|
|
196
|
+
- Reconciliation checks: compare notes with matching topics to MOC entries
|
|
197
|
+
- Coverage metrics: percentage of topic notes actually linked from their MOC
|
|
198
|
+
- Staleness signals: MOC last-updated date vs newest note in its territory
|
|
199
|
+
|
|
200
|
+
**Prevention:**
|
|
201
|
+
- Pipeline integration: /reflect updates MOCs as part of standard processing
|
|
202
|
+
- Batch-level checks: after processing batches, verify new notes appear in relevant MOCs
|
|
203
|
+
- Periodic MOC review: flag MOCs not updated in 90+ days for active topics
|
|
204
|
+
|
|
205
|
+
## Domain-Specific MOC Adaptations
|
|
206
|
+
|
|
207
|
+
Each domain adapts the MOC structure to match its natural organizing principles. See the domain examples in `examples/` for complete illustrations.
|
|
208
|
+
|
|
209
|
+
| Domain | Hierarchy | Key MOC Features |
|
|
210
|
+
|--------|-----------|-----------------|
|
|
211
|
+
| Research | Three-tier | Topic MOCs have Tensions and Explorations Needed sections |
|
|
212
|
+
| Therapy | Flat peer | Pattern MOC links to all detected patterns with date ranges |
|
|
213
|
+
| PM | Hub + entities | Project MOCs have Decision Log and Risk sections |
|
|
214
|
+
| Creative | Hub + entities | Character MOCs track arc evolution across scenes |
|
|
215
|
+
| Learning | Three-tier | Concept MOCs have prerequisite dependency information |
|
|
216
|
+
| Personal | Flat peer | Area MOCs have health indicators and review frequency |
|
|
217
|
+
| People | Hub + entities | Person MOCs track interaction recency and follow-ups |
|
|
218
|
+
| Engineering | Three-tier | System MOCs have dependency maps and ADR links |
|
|
219
|
+
| Trading | Flat peer | Strategy MOCs track win rate and drift metrics |
|
|
220
|
+
| Health | Flat peer | Each dimension MOC has correlation analysis links |
|
|
221
|
+
| Product | Hub + entities | Feature MOCs link to customer feedback and experiments |
|
|
222
|
+
| Legal | Three-tier + entities | Matter MOCs have precedent chains and deadline tracking |
|
|
223
|
+
|
|
224
|
+
**Domain-specific MOC sections beyond Core Ideas:**
|
|
225
|
+
|
|
226
|
+
| Domain | Extra Sections | Purpose |
|
|
227
|
+
|--------|---------------|---------|
|
|
228
|
+
| Research | Tensions, Explorations Needed | Track intellectual conflict and gaps |
|
|
229
|
+
| Therapy | Active Patterns, Strategy Effectiveness | Track therapeutic progress |
|
|
230
|
+
| PM | Decisions, Open Risks, Action Items | Track project health |
|
|
231
|
+
| Creative | Canon Facts, Continuity Notes | Track worldbuilding consistency |
|
|
232
|
+
| Learning | Prerequisites, Mastery Progress | Track learning dependencies |
|
|
233
|
+
| Trading | Active Theses, Performance Metrics | Track strategy effectiveness |
|
|
234
|
+
|
|
235
|
+
## MOC Lifecycle Management
|
|
236
|
+
|
|
237
|
+
The plugin generates lifecycle triggers:
|
|
238
|
+
|
|
239
|
+
### Creation Triggers
|
|
240
|
+
- 20+ notes cluster around a topic (detected via topic field analysis)
|
|
241
|
+
- Navigation friction reported: "I can't find notes about X"
|
|
242
|
+
- New domain area activated by user
|
|
243
|
+
- Mental squeeze point: the builder notices orientation within a topic has become effortful
|
|
244
|
+
|
|
245
|
+
### Split Triggers
|
|
246
|
+
- MOC exceeds 50 notes (since [[basic level categorization determines optimal MOC granularity]])
|
|
247
|
+
- Distinct sub-clusters emerge with different update frequencies
|
|
248
|
+
- Agent notes section exceeds 30 lines (too much navigation knowledge for one level)
|
|
249
|
+
|
|
250
|
+
**How to split:** Create sub-MOC(s) named `parent-subtopic.md`. Move Core Ideas for the sub-cluster. Parent becomes domain-style with synthesis and links to sub-MOCs. Update all affected notes' Topics footers.
|
|
251
|
+
|
|
252
|
+
### Merge Triggers
|
|
253
|
+
- Two MOCs have < 30 combined notes with significant overlap
|
|
254
|
+
- Same update frequency and related topics
|
|
255
|
+
- Since [[community detection algorithms can inform when MOCs should split or merge]]
|
|
256
|
+
|
|
257
|
+
### Archive Triggers
|
|
258
|
+
- < 5 notes and stagnant for 6+ months
|
|
259
|
+
- Topic absorbed into a broader MOC
|
|
260
|
+
- Domain area deactivated by user
|
|
261
|
+
|
|
262
|
+
## Health Metrics
|
|
263
|
+
|
|
264
|
+
The plugin monitors MOC health through these metrics:
|
|
265
|
+
|
|
266
|
+
| Metric | Healthy | Warning | Action |
|
|
267
|
+
|--------|---------|---------|--------|
|
|
268
|
+
| Notes per topic MOC | 10-40 | >50 | Suggest split |
|
|
269
|
+
| Orphan notes | 0 | Any | Flag for MOC addition |
|
|
270
|
+
| Context phrase coverage | 100% | <80% | Flag bare links |
|
|
271
|
+
| Agent notes length | 5-20 lines | >30 lines | Suggest observation extraction |
|
|
272
|
+
| Last updated | <30 days | >90 days | Flag for review |
|
|
273
|
+
| Cross-MOC membership | >10% of notes | <5% | Notes may be too siloed |
|
|
274
|
+
| Coverage (notes in topic vs MOC entries) | >95% | <80% | MOC is going stale |
|
|
275
|
+
|
|
276
|
+
## Compounding Returns
|
|
277
|
+
|
|
278
|
+
Since [[MOC maintenance investment compounds because orientation savings multiply across every future session]], maintenance is not overhead but investment. A single context phrase update costs seconds but saves orientation time in every future session that loads the MOC. If the MOC loads fifty times, a thirty-second investment returns minutes of cumulative savings.
|
|
279
|
+
|
|
280
|
+
**Prioritize by traffic:** Hub MOCs that load every session have the highest temporal multiplier. Topic MOCs that load in domain-specific sessions have moderate multiplier. Entity MOCs that load when that entity is active have the lowest but still positive multiplier. Maintenance effort should flow toward the highest-traffic MOCs.
|
|
281
|
+
|
|
282
|
+
**The hidden second return:** MOC maintenance forces genuine intellectual engagement — the maintainer reads notes in context, notices tensions, writes synthesis. Since [[the generation effect requires active transformation not just storage]], the act of maintaining IS thinking. The return is double: measurable orientation savings plus synthesis opportunities that emerge from reconsidering arrangements.
|
|
283
|
+
|
|
284
|
+
## Domain Examples
|
|
285
|
+
|
|
286
|
+
These domain compositions demonstrate MOC patterns in practice:
|
|
287
|
+
|
|
288
|
+
- [[academic research uses structured extraction with cross-source synthesis]] — Three-tier research pattern: Hub -> Literature Review MOCs -> Topic MOCs; demonstrates split triggers when topic MOCs exceed 50 claims
|
|
289
|
+
- [[personal assistant uses life area management with review automation]] — Flat peer pattern: life areas (Career, Health, Relationships, Finance, Growth) as equal MOCs with area health indicators
|
|
290
|
+
- [[people relationships uses Dunbar-layered graphs with interaction tracking]] — Hub-plus-entities pattern: hub MOC with quick reference table, individual person MOCs accumulating interaction context over time
|
|
291
|
+
- [[creative writing uses worldbuilding consistency with character tracking]] — Hub-plus-entities with mixed types: character MOCs track arc evolution, location MOCs serve as worldbuilding reference, plot thread MOCs track foreshadowing-to-resolution
|
|
292
|
+
- [[health wellness uses symptom-trigger correlation with multi-dimensional tracking]] — Flat peer with cross-dimensional links: fitness, nutrition, sleep, and symptom MOCs are peers, but Agent Notes sections surface correlations across dimensions
|
|
293
|
+
- [[student learning uses prerequisite graphs with spaced retrieval]] — Three-tier with prerequisite awareness: concept MOCs include dependency information, enabling the agent to identify prerequisite gaps
|
|
294
|
+
|
|
295
|
+
## Grounding
|
|
296
|
+
|
|
297
|
+
This guidance is grounded in:
|
|
298
|
+
- [[MOCs are attention management devices not just organizational tools]] — MOCs as attention management
|
|
299
|
+
- [[MOC construction forces synthesis that automated generation from metadata cannot replicate]] — why curation matters
|
|
300
|
+
- [[navigational vertigo emerges in pure association systems without local hierarchy]] — why MOCs are necessary
|
|
301
|
+
- [[basic level categorization determines optimal MOC granularity]] — optimal MOC size
|
|
302
|
+
- [[community detection algorithms can inform when MOCs should split or merge]] — algorithmic split/merge signals
|
|
303
|
+
- [[MOC maintenance investment compounds because orientation savings multiply across every future session]] — why MOC investment pays off
|
|
304
|
+
- [[stale navigation actively misleads because agents trust curated maps completely]] — the danger of unmaintained MOCs
|
|
305
|
+
- [[complete navigation requires four complementary types that no single mechanism provides]] — MOCs fill the local navigation role
|
|
306
|
+
- [[context phrase clarity determines how deep a navigation hierarchy can scale]] — context phrase quality constrains hierarchy depth
|
|
307
|
+
- [[the generation effect requires active transformation not just storage]] — why manual construction produces understanding
|
|
308
|
+
|
|
309
|
+
---
|
|
310
|
+
|
|
311
|
+
Topics:
|
|
312
|
+
- [[index]]
|
|
313
|
+
- [[index]]
|
|
314
|
+
---
|
|
315
|
+
|
|
316
|
+
Topics:
|
|
317
|
+
- [[note-design]]
|
|
318
|
+
- [[maintenance-patterns]]
|