arscontexta 0.6.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.claude-plugin/marketplace.json +11 -0
- package/.claude-plugin/plugin.json +22 -0
- package/README.md +683 -0
- package/agents/knowledge-guide.md +49 -0
- package/bin/cli.mjs +66 -0
- package/generators/agents-md.md +240 -0
- package/generators/claude-md.md +379 -0
- package/generators/features/atomic-notes.md +124 -0
- package/generators/features/ethical-guardrails.md +58 -0
- package/generators/features/graph-analysis.md +188 -0
- package/generators/features/helper-functions.md +92 -0
- package/generators/features/maintenance.md +164 -0
- package/generators/features/methodology-knowledge.md +70 -0
- package/generators/features/mocs.md +144 -0
- package/generators/features/multi-domain.md +61 -0
- package/generators/features/personality.md +71 -0
- package/generators/features/processing-pipeline.md +428 -0
- package/generators/features/schema.md +149 -0
- package/generators/features/self-evolution.md +229 -0
- package/generators/features/self-space.md +78 -0
- package/generators/features/semantic-search.md +99 -0
- package/generators/features/session-rhythm.md +85 -0
- package/generators/features/templates.md +85 -0
- package/generators/features/wiki-links.md +88 -0
- package/generators/soul-md.md +121 -0
- package/hooks/hooks.json +45 -0
- package/hooks/scripts/auto-commit.sh +44 -0
- package/hooks/scripts/session-capture.sh +35 -0
- package/hooks/scripts/session-orient.sh +86 -0
- package/hooks/scripts/write-validate.sh +42 -0
- package/methodology/AI shifts knowledge systems from externalizing memory to externalizing attention.md +59 -0
- package/methodology/BM25 retrieval fails on full-length descriptions because query term dilution reduces match scores.md +39 -0
- package/methodology/IBIS framework maps claim-based architecture to structured argumentation.md +58 -0
- package/methodology/LLM attention degrades as context fills.md +49 -0
- package/methodology/MOC construction forces synthesis that automated generation from metadata cannot replicate.md +49 -0
- package/methodology/MOC maintenance investment compounds because orientation savings multiply across every future session.md +41 -0
- package/methodology/MOCs are attention management devices not just organizational tools.md +51 -0
- package/methodology/PKM failure follows a predictable cycle.md +50 -0
- package/methodology/ThreadMode to DocumentMode transformation is the core value creation step.md +52 -0
- package/methodology/WIP limits force processing over accumulation.md +53 -0
- package/methodology/Zeigarnik effect validates capture-first philosophy because open loops drain attention.md +42 -0
- package/methodology/academic research uses structured extraction with cross-source synthesis.md +566 -0
- package/methodology/adapt the four-phase processing pipeline to domain-specific throughput needs.md +197 -0
- package/methodology/agent notes externalize navigation intuition that search cannot discover and traversal cannot reconstruct.md +48 -0
- package/methodology/agent self-memory should be architecturally separate from user knowledge systems.md +48 -0
- package/methodology/agent session boundaries create natural automation checkpoints that human-operated systems lack.md +56 -0
- package/methodology/agent-cognition.md +107 -0
- package/methodology/agents are simultaneously methodology executors and subjects creating a unique trust asymmetry.md +66 -0
- package/methodology/aspect-oriented programming solved the same cross-cutting concern problem that hooks solve.md +39 -0
- package/methodology/associative ontologies beat hierarchical taxonomies because heterarchy adapts while hierarchy brittles.md +53 -0
- package/methodology/attention residue may have a minimum granularity that cannot be subdivided.md +46 -0
- package/methodology/auto-commit hooks eliminate prospective memory failures by converting remember-to-act into guaranteed execution.md +47 -0
- package/methodology/automated detection is always safe because it only reads state while automated remediation risks content corruption.md +42 -0
- package/methodology/automation should be retired when its false positive rate exceeds its true positive rate or it catches zero issues.md +56 -0
- package/methodology/backlinks implicitly define notes by revealing usage context.md +35 -0
- package/methodology/backward maintenance asks what would be different if written today.md +62 -0
- package/methodology/balance onboarding enforcement and questions to prevent premature complexity.md +229 -0
- package/methodology/basic level categorization determines optimal MOC granularity.md +51 -0
- package/methodology/batching by context similarity reduces switching costs in agent processing.md +43 -0
- package/methodology/behavioral anti-patterns matter more than tool selection.md +42 -0
- package/methodology/betweenness centrality identifies bridge notes connecting disparate knowledge domains.md +57 -0
- package/methodology/blueprints that teach construction outperform downloads that provide pre-built code for platform-dependent modules.md +42 -0
- package/methodology/bootstrapping principle enables self-improving systems.md +62 -0
- package/methodology/build automatic memory through cognitive offloading and session handoffs.md +285 -0
- package/methodology/capture the reaction to content not just the content itself.md +41 -0
- package/methodology/claims must be specific enough to be wrong.md +36 -0
- package/methodology/closure rituals create clean breaks that prevent attention residue bleed.md +44 -0
- package/methodology/cognitive offloading is the architectural foundation for vault design.md +46 -0
- package/methodology/cognitive outsourcing risk in agent-operated systems.md +55 -0
- package/methodology/coherence maintains consistency despite inconsistent inputs.md +96 -0
- package/methodology/coherent architecture emerges from wiki links spreading activation and small-world topology.md +48 -0
- package/methodology/community detection algorithms can inform when MOCs should split or merge.md +52 -0
- package/methodology/complete navigation requires four complementary types that no single mechanism provides.md +43 -0
- package/methodology/complex systems evolve from simple working systems.md +59 -0
- package/methodology/composable knowledge architecture builds systems from independent toggleable modules not monolithic templates.md +61 -0
- package/methodology/compose multi-domain systems through separate templates and shared graph.md +372 -0
- package/methodology/concept-orientation beats source-orientation for cross-domain connections.md +51 -0
- package/methodology/confidence thresholds gate automated action between the mechanical and judgment zones.md +50 -0
- package/methodology/configuration dimensions interact so choices in one create pressure on others.md +58 -0
- package/methodology/configuration paralysis emerges when derivation surfaces too many decisions.md +44 -0
- package/methodology/context files function as agent operating systems through self-referential self-extension.md +46 -0
- package/methodology/context phrase clarity determines how deep a navigation hierarchy can scale.md +46 -0
- package/methodology/continuous small-batch processing eliminates review dread.md +48 -0
- package/methodology/controlled disorder engineers serendipity through semantic rather than topical linking.md +51 -0
- package/methodology/creative writing uses worldbuilding consistency with character tracking.md +672 -0
- package/methodology/cross-links between MOC territories indicate creative leaps and integration depth.md +43 -0
- package/methodology/dangling links reveal which notes want to exist.md +62 -0
- package/methodology/data exit velocity measures how quickly content escapes vendor lock-in.md +74 -0
- package/methodology/decontextualization risk means atomicity may strip meaning that cannot be recovered.md +48 -0
- package/methodology/dense interlinked research claims enable derivation while sparse references only enable templating.md +47 -0
- package/methodology/dependency resolution through topological sort makes module composition transparent and verifiable.md +56 -0
- package/methodology/derivation generates knowledge systems from composable research claims not template customization.md +63 -0
- package/methodology/derivation-engine.md +27 -0
- package/methodology/derived systems follow a seed-evolve-reseed lifecycle.md +56 -0
- package/methodology/description quality for humans diverges from description quality for keyword search.md +73 -0
- package/methodology/descriptions are retrieval filters not summaries.md +112 -0
- package/methodology/design MOCs as attention management devices with lifecycle governance.md +318 -0
- package/methodology/design-dimensions.md +66 -0
- package/methodology/digital mutability enables note evolution that physical permanence forbids.md +54 -0
- package/methodology/discovery-retrieval.md +48 -0
- package/methodology/distinctiveness scoring treats description quality as measurable.md +69 -0
- package/methodology/does agent processing recover what fast capture loses.md +43 -0
- package/methodology/domain-compositions.md +37 -0
- package/methodology/dual-coding with visual elements could enhance agent traversal.md +55 -0
- package/methodology/each module must be describable in one sentence under 200 characters or it does too many things.md +45 -0
- package/methodology/each new note compounds value by creating traversal paths.md +55 -0
- package/methodology/eight configuration dimensions parameterize the space of possible knowledge systems.md +56 -0
- package/methodology/elaborative encoding is the quality gate for new notes.md +55 -0
- package/methodology/enforce schema with graduated strictness across capture processing and query zones.md +221 -0
- package/methodology/enforcing atomicity can create paralysis when ideas resist decomposition.md +43 -0
- package/methodology/engineering uses technical decision tracking with architectural memory.md +766 -0
- package/methodology/every knowledge domain shares a four-phase processing skeleton that diverges only in the process step.md +53 -0
- package/methodology/evolution observations provide actionable signals for system adaptation.md +67 -0
- package/methodology/external memory shapes cognition more than base model.md +60 -0
- package/methodology/faceted classification treats notes as multi-dimensional objects rather than folder contents.md +65 -0
- package/methodology/failure-modes.md +27 -0
- package/methodology/false universalism applies same processing logic regardless of domain.md +49 -0
- package/methodology/federated wiki pattern enables multi-agent divergence as feature not bug.md +59 -0
- package/methodology/flat files break at retrieval scale.md +75 -0
- package/methodology/forced engagement produces weak connections.md +48 -0
- package/methodology/four abstraction layers separate platform-agnostic from platform-dependent knowledge system features.md +47 -0
- package/methodology/fresh context per task preserves quality better than chaining phases.md +44 -0
- package/methodology/friction reveals architecture.md +63 -0
- package/methodology/friction-driven module adoption prevents configuration debt by adding complexity only at pain points.md +48 -0
- package/methodology/gardening cycle implements tend prune fertilize operations.md +41 -0
- package/methodology/generation effect gate blocks processing without transformation.md +40 -0
- package/methodology/goal-driven memory orchestration enables autonomous domain learning through directed compute allocation.md +41 -0
- package/methodology/good descriptions layer heuristic then mechanism then implication.md +57 -0
- package/methodology/graph-structure.md +65 -0
- package/methodology/guided notes might outperform post-hoc structuring for high-volume capture.md +37 -0
- package/methodology/health wellness uses symptom-trigger correlation with multi-dimensional tracking.md +819 -0
- package/methodology/hook composition creates emergent methodology from independent single-concern components.md +47 -0
- package/methodology/hook enforcement guarantees quality while instruction enforcement merely suggests it.md +51 -0
- package/methodology/hook-driven learning loops create self-improving methodology through observation accumulation.md +62 -0
- package/methodology/hooks are the agent habit system that replaces the missing basal ganglia.md +40 -0
- package/methodology/hooks cannot replace genuine cognitive engagement yet more automation is always tempting.md +87 -0
- package/methodology/hooks enable context window efficiency by delegating deterministic checks to external processes.md +47 -0
- package/methodology/idempotent maintenance operations are safe to automate because running them twice produces the same result as running them once.md +44 -0
- package/methodology/implement condition-based maintenance triggers for derived systems.md +255 -0
- package/methodology/implicit dependencies create distributed monoliths that fail silently across configurations.md +58 -0
- package/methodology/implicit knowledge emerges from traversal.md +55 -0
- package/methodology/incremental formalization happens through repeated touching of old notes.md +60 -0
- package/methodology/incremental reading enables cross-source connection finding.md +39 -0
- package/methodology/index.md +32 -0
- package/methodology/inline links carry richer relationship data than metadata fields.md +91 -0
- package/methodology/insight accretion differs from productivity in knowledge systems.md +41 -0
- package/methodology/intermediate packets enable assembly over creation.md +52 -0
- package/methodology/intermediate representation pattern enables reliable vault operations beyond regex.md +62 -0
- package/methodology/justification chains enable forward backward and evolution reasoning about configuration decisions.md +46 -0
- package/methodology/knowledge system architecture is parameterized by platform capabilities not fixed by methodology.md +51 -0
- package/methodology/knowledge systems become communication partners through complexity and memory humans cannot sustain.md +47 -0
- package/methodology/knowledge systems share universal operations and structural components across all methodology traditions.md +46 -0
- package/methodology/legal case management uses precedent chains with regulatory change propagation.md +892 -0
- package/methodology/live index via periodic regeneration keeps discovery current.md +58 -0
- package/methodology/local-first file formats are inherently agent-native.md +69 -0
- package/methodology/logic column pattern separates reasoning from procedure.md +35 -0
- package/methodology/maintenance operations are more universal than creative pipelines because structural health is domain-invariant.md +47 -0
- package/methodology/maintenance scheduling frequency should match consequence speed not detection capability.md +50 -0
- package/methodology/maintenance targeting should prioritize mechanism and theory notes.md +26 -0
- package/methodology/maintenance-patterns.md +72 -0
- package/methodology/markdown plus YAML plus ripgrep implements a queryable graph database without infrastructure.md +55 -0
- package/methodology/maturity field enables agent context prioritization.md +33 -0
- package/methodology/memory-architecture.md +27 -0
- package/methodology/metacognitive confidence can diverge from retrieval capability.md +42 -0
- package/methodology/metadata reduces entropy enabling precision over recall.md +91 -0
- package/methodology/methodology development should follow the trajectory from documentation to skill to hook as understanding hardens.md +80 -0
- package/methodology/methodology traditions are named points in a shared configuration space not competing paradigms.md +64 -0
- package/methodology/mnemonic medium embeds verification into navigation.md +46 -0
- package/methodology/module communication through shared YAML fields creates loose coupling without direct dependencies.md +44 -0
- package/methodology/module deactivation must account for structural artifacts that survive the toggle.md +49 -0
- package/methodology/multi-domain systems compose through separate templates and shared graph.md +61 -0
- package/methodology/multi-domain-composition.md +27 -0
- package/methodology/narrow folksonomy optimizes for single-operator retrieval unlike broad consensus tagging.md +53 -0
- package/methodology/navigation infrastructure passes through distinct scaling regimes that require qualitative strategy shifts.md +48 -0
- package/methodology/navigational vertigo emerges in pure association systems without local hierarchy.md +54 -0
- package/methodology/note titles should function as APIs enabling sentence transclusion.md +51 -0
- package/methodology/note-design.md +57 -0
- package/methodology/notes are skills /342/200/224 curated knowledge injected when relevant.md" +62 -0
- package/methodology/notes function as cognitive anchors that stabilize attention during complex tasks.md +41 -0
- package/methodology/novel domains derive by mapping knowledge type to closest reference domain then adapting.md +50 -0
- package/methodology/nudge theory explains graduated hook enforcement as choice architecture for agents.md +59 -0
- package/methodology/observation and tension logs function as dead-letter queues for failed automation.md +51 -0
- package/methodology/operational memory and knowledge memory serve different functions in agent architecture.md +48 -0
- package/methodology/operational wisdom requires contextual observation.md +52 -0
- package/methodology/orchestrated vault creation transforms arscontexta from tool to autonomous knowledge factory.md +40 -0
- package/methodology/organic emergence versus active curation creates a fundamental vault governance tension.md +68 -0
- package/methodology/orphan notes are seeds not failures.md +38 -0
- package/methodology/over-automation corrupts quality when hooks encode judgment rather than verification.md +62 -0
- package/methodology/people relationships uses Dunbar-layered graphs with interaction tracking.md +659 -0
- package/methodology/personal assistant uses life area management with review automation.md +610 -0
- package/methodology/platform adapter translation is semantic not mechanical because hook event meanings differ.md +40 -0
- package/methodology/platform capability tiers determine which knowledge system features can be implemented.md +48 -0
- package/methodology/platform fragmentation means identical conceptual operations require different implementations across agent environments.md +44 -0
- package/methodology/premature complexity is the most common derivation failure mode.md +45 -0
- package/methodology/prevent domain-specific failure modes through the vulnerability matrix.md +336 -0
- package/methodology/processing effort should follow retrieval demand.md +57 -0
- package/methodology/processing-workflows.md +75 -0
- package/methodology/product management uses feedback pipelines with experiment tracking.md +789 -0
- package/methodology/productivity porn risk in meta-system building.md +30 -0
- package/methodology/programmable notes could enable property-triggered workflows.md +64 -0
- package/methodology/progressive disclosure means reading right not reading less.md +69 -0
- package/methodology/progressive schema validates only what active modules require not the full system schema.md +49 -0
- package/methodology/project management uses decision tracking with stakeholder context.md +776 -0
- package/methodology/propositional link semantics transform wiki links from associative to reasoned.md +87 -0
- package/methodology/prospective memory requires externalization.md +53 -0
- package/methodology/provenance tracks where beliefs come from.md +62 -0
- package/methodology/queries evolve during search so agents should checkpoint.md +35 -0
- package/methodology/question-answer metadata enables inverted search patterns.md +39 -0
- package/methodology/random note resurfacing prevents write-only memory.md +33 -0
- package/methodology/reconciliation loops that compare desired state to actual state enable drift correction without continuous monitoring.md +59 -0
- package/methodology/reflection synthesizes existing notes into new insight.md +100 -0
- package/methodology/retrieval utility should drive design over capture completeness.md +69 -0
- package/methodology/retrieval verification loop tests description quality at scale.md +81 -0
- package/methodology/role field makes graph structure explicit.md +94 -0
- package/methodology/scaffolding enables divergence that fine-tuning cannot.md +67 -0
- package/methodology/schema enforcement via validation agents enables soft consistency.md +60 -0
- package/methodology/schema evolution follows observe-then-formalize not design-then-enforce.md +65 -0
- package/methodology/schema field names are the only domain specific element in the universal note pattern.md +46 -0
- package/methodology/schema fields should use domain-native vocabulary not abstract terminology.md +47 -0
- package/methodology/schema templates reduce cognitive overhead at capture time.md +55 -0
- package/methodology/schema validation hooks externalize inhibitory control that degrades under cognitive load.md +48 -0
- package/methodology/schema-enforcement.md +27 -0
- package/methodology/self-extension requires context files to contain platform operations knowledge not just methodology.md +47 -0
- package/methodology/sense-making vs storage does compression lose essential nuance.md +73 -0
- package/methodology/session boundary hooks implement cognitive bookends for orientation and reflection.md +60 -0
- package/methodology/session handoff creates continuity without persistent memory.md +43 -0
- package/methodology/session outputs are packets for future selves.md +43 -0
- package/methodology/session transcript mining enables experiential validation that structural tests cannot provide.md +38 -0
- package/methodology/skill context budgets constrain knowledge system complexity on agent platforms.md +52 -0
- package/methodology/skills encode methodology so manual execution bypasses quality gates.md +50 -0
- package/methodology/small-world topology requires hubs and dense local links.md +99 -0
- package/methodology/source attribution enables tracing claims to foundations.md +38 -0
- package/methodology/spaced repetition scheduling could optimize vault maintenance.md +44 -0
- package/methodology/spreading activation models how agents should traverse.md +79 -0
- package/methodology/stale navigation actively misleads because agents trust curated maps completely.md +43 -0
- package/methodology/stigmergy coordinates agents through environmental traces without direct communication.md +62 -0
- package/methodology/storage versus thinking distinction determines which tool patterns apply.md +56 -0
- package/methodology/structure enables navigation without reading everything.md +52 -0
- package/methodology/structure without processing provides no value.md +56 -0
- package/methodology/student learning uses prerequisite graphs with spaced retrieval.md +770 -0
- package/methodology/summary coherence tests composability before filing.md +37 -0
- package/methodology/tag rot applies to wiki links because titles serve as both identifier and display text.md +50 -0
- package/methodology/temporal media must convert to spatial text for agent traversal.md +43 -0
- package/methodology/temporal processing priority creates age-based inbox urgency.md +45 -0
- package/methodology/temporal separation of capture and processing preserves context freshness.md +39 -0
- package/methodology/ten universal primitives form the kernel of every viable agent knowledge system.md +162 -0
- package/methodology/testing effect could enable agent knowledge verification.md +38 -0
- package/methodology/the AgentSkills standard embodies progressive disclosure at the skill level.md +40 -0
- package/methodology/the derivation engine improves recursively as deployed systems generate observations.md +49 -0
- package/methodology/the determinism boundary separates hook methodology from skill methodology.md +46 -0
- package/methodology/the fix-versus-report decision depends on determinism reversibility and accumulated trust.md +45 -0
- package/methodology/the generation effect requires active transformation not just storage.md +57 -0
- package/methodology/the no wrong patches guarantee ensures any valid module combination produces a valid system.md +58 -0
- package/methodology/the system is the argument.md +46 -0
- package/methodology/the vault constitutes identity for agents.md +86 -0
- package/methodology/the vault methodology transfers because it encodes cognitive science not domain specifics.md +47 -0
- package/methodology/therapy journal uses warm personality with pattern detection for emotional processing.md +584 -0
- package/methodology/three capture schools converge through agent-mediated synthesis.md +55 -0
- package/methodology/three concurrent maintenance loops operate at different timescales to catch different classes of problems.md +56 -0
- package/methodology/throughput matters more than accumulation.md +58 -0
- package/methodology/title as claim enables traversal as reasoning.md +50 -0
- package/methodology/topological organization beats temporal for knowledge work.md +52 -0
- package/methodology/trading uses conviction tracking with thesis-outcome correlation.md +699 -0
- package/methodology/trails transform ephemeral navigation into persistent artifacts.md +39 -0
- package/methodology/transform universal vocabulary to domain-native language through six levels.md +259 -0
- package/methodology/type field enables structured queries without folder hierarchies.md +53 -0
- package/methodology/use-case presets dissolve the tension between composability and simplicity.md +44 -0
- package/methodology/vault conventions may impose hidden rigidity on thinking.md +44 -0
- package/methodology/verbatim risk applies to agents too.md +31 -0
- package/methodology/vibe notetaking is the emerging industry consensus for AI-native self-organization.md +56 -0
- package/methodology/vivid memories need verification.md +45 -0
- package/methodology/vocabulary-transformation.md +27 -0
- package/methodology/voice capture is the highest-bandwidth channel for agent-delegated knowledge systems.md +45 -0
- package/methodology/wiki links are the digital evolution of analog indexing.md +73 -0
- package/methodology/wiki links as social contract transforms agents into stewards of incomplete references.md +52 -0
- package/methodology/wiki links create navigation paths that shape retrieval.md +63 -0
- package/methodology/wiki links implement GraphRAG without the infrastructure.md +101 -0
- package/methodology/writing for audience blocks authentic creation.md +22 -0
- package/methodology/you operate a system that takes notes.md +79 -0
- package/openclaw/SKILL.md +110 -0
- package/package.json +45 -0
- package/platforms/README.md +51 -0
- package/platforms/claude-code/generator.md +61 -0
- package/platforms/claude-code/hooks/README.md +186 -0
- package/platforms/claude-code/hooks/auto-commit.sh.template +38 -0
- package/platforms/claude-code/hooks/session-capture.sh.template +72 -0
- package/platforms/claude-code/hooks/session-orient.sh.template +189 -0
- package/platforms/claude-code/hooks/write-validate.sh.template +106 -0
- package/platforms/openclaw/generator.md +82 -0
- package/platforms/openclaw/hooks/README.md +89 -0
- package/platforms/openclaw/hooks/bootstrap.ts.template +224 -0
- package/platforms/openclaw/hooks/command-new.ts.template +165 -0
- package/platforms/openclaw/hooks/heartbeat.ts.template +214 -0
- package/platforms/shared/features/README.md +70 -0
- package/platforms/shared/skill-blocks/graph.md +145 -0
- package/platforms/shared/skill-blocks/learn.md +119 -0
- package/platforms/shared/skill-blocks/next.md +131 -0
- package/platforms/shared/skill-blocks/pipeline.md +326 -0
- package/platforms/shared/skill-blocks/ralph.md +616 -0
- package/platforms/shared/skill-blocks/reduce.md +1142 -0
- package/platforms/shared/skill-blocks/refactor.md +129 -0
- package/platforms/shared/skill-blocks/reflect.md +780 -0
- package/platforms/shared/skill-blocks/remember.md +524 -0
- package/platforms/shared/skill-blocks/rethink.md +574 -0
- package/platforms/shared/skill-blocks/reweave.md +680 -0
- package/platforms/shared/skill-blocks/seed.md +320 -0
- package/platforms/shared/skill-blocks/stats.md +145 -0
- package/platforms/shared/skill-blocks/tasks.md +171 -0
- package/platforms/shared/skill-blocks/validate.md +323 -0
- package/platforms/shared/skill-blocks/verify.md +562 -0
- package/platforms/shared/templates/README.md +35 -0
- package/presets/experimental/categories.yaml +1 -0
- package/presets/experimental/preset.yaml +38 -0
- package/presets/experimental/starter/README.md +7 -0
- package/presets/experimental/vocabulary.yaml +7 -0
- package/presets/personal/categories.yaml +7 -0
- package/presets/personal/preset.yaml +41 -0
- package/presets/personal/starter/goals.md +21 -0
- package/presets/personal/starter/index.md +17 -0
- package/presets/personal/starter/life-areas.md +21 -0
- package/presets/personal/starter/people.md +21 -0
- package/presets/personal/vocabulary.yaml +32 -0
- package/presets/research/categories.yaml +8 -0
- package/presets/research/preset.yaml +41 -0
- package/presets/research/starter/index.md +17 -0
- package/presets/research/starter/methods.md +21 -0
- package/presets/research/starter/open-questions.md +21 -0
- package/presets/research/vocabulary.yaml +33 -0
- package/reference/AUDIT-REPORT.md +238 -0
- package/reference/claim-map.md +172 -0
- package/reference/components.md +327 -0
- package/reference/conversation-patterns.md +542 -0
- package/reference/derivation-validation.md +649 -0
- package/reference/dimension-claim-map.md +134 -0
- package/reference/evolution-lifecycle.md +297 -0
- package/reference/failure-modes.md +235 -0
- package/reference/interaction-constraints.md +204 -0
- package/reference/kernel.yaml +242 -0
- package/reference/methodology.md +283 -0
- package/reference/open-questions.md +279 -0
- package/reference/personality-layer.md +302 -0
- package/reference/self-space.md +299 -0
- package/reference/semantic-vs-keyword.md +288 -0
- package/reference/session-lifecycle.md +298 -0
- package/reference/templates/base-note.md +16 -0
- package/reference/templates/companion-note.md +70 -0
- package/reference/templates/creative-note.md +16 -0
- package/reference/templates/learning-note.md +16 -0
- package/reference/templates/life-note.md +16 -0
- package/reference/templates/moc.md +26 -0
- package/reference/templates/relationship-note.md +17 -0
- package/reference/templates/research-note.md +19 -0
- package/reference/templates/session-log.md +24 -0
- package/reference/templates/therapy-note.md +16 -0
- package/reference/test-fixtures/edge-case-constraints.md +148 -0
- package/reference/test-fixtures/multi-domain.md +164 -0
- package/reference/test-fixtures/novel-domain-gaming.md +138 -0
- package/reference/test-fixtures/research-minimal.md +102 -0
- package/reference/test-fixtures/therapy-full.md +155 -0
- package/reference/testing-milestones.md +1087 -0
- package/reference/three-spaces.md +363 -0
- package/reference/tradition-presets.md +203 -0
- package/reference/use-case-presets.md +341 -0
- package/reference/validate-kernel.sh +432 -0
- package/reference/vocabulary-transforms.md +85 -0
- package/scripts/sync-thinking.sh +147 -0
- package/skill-sources/graph/SKILL.md +567 -0
- package/skill-sources/graph/skill.json +17 -0
- package/skill-sources/learn/SKILL.md +254 -0
- package/skill-sources/learn/skill.json +17 -0
- package/skill-sources/next/SKILL.md +407 -0
- package/skill-sources/next/skill.json +17 -0
- package/skill-sources/pipeline/SKILL.md +314 -0
- package/skill-sources/pipeline/skill.json +17 -0
- package/skill-sources/ralph/SKILL.md +604 -0
- package/skill-sources/ralph/skill.json +17 -0
- package/skill-sources/reduce/SKILL.md +1113 -0
- package/skill-sources/reduce/skill.json +17 -0
- package/skill-sources/refactor/SKILL.md +448 -0
- package/skill-sources/refactor/skill.json +17 -0
- package/skill-sources/reflect/SKILL.md +747 -0
- package/skill-sources/reflect/skill.json +17 -0
- package/skill-sources/remember/SKILL.md +534 -0
- package/skill-sources/remember/skill.json +17 -0
- package/skill-sources/rethink/SKILL.md +658 -0
- package/skill-sources/rethink/skill.json +17 -0
- package/skill-sources/reweave/SKILL.md +657 -0
- package/skill-sources/reweave/skill.json +17 -0
- package/skill-sources/seed/SKILL.md +303 -0
- package/skill-sources/seed/skill.json +17 -0
- package/skill-sources/stats/SKILL.md +371 -0
- package/skill-sources/stats/skill.json +17 -0
- package/skill-sources/tasks/SKILL.md +402 -0
- package/skill-sources/tasks/skill.json +17 -0
- package/skill-sources/validate/SKILL.md +310 -0
- package/skill-sources/validate/skill.json +17 -0
- package/skill-sources/verify/SKILL.md +532 -0
- package/skill-sources/verify/skill.json +17 -0
- package/skills/add-domain/SKILL.md +441 -0
- package/skills/add-domain/skill.json +17 -0
- package/skills/architect/SKILL.md +568 -0
- package/skills/architect/skill.json +17 -0
- package/skills/ask/SKILL.md +388 -0
- package/skills/ask/skill.json +17 -0
- package/skills/health/SKILL.md +760 -0
- package/skills/health/skill.json +17 -0
- package/skills/help/SKILL.md +348 -0
- package/skills/help/skill.json +17 -0
- package/skills/recommend/SKILL.md +553 -0
- package/skills/recommend/skill.json +17 -0
- package/skills/reseed/SKILL.md +385 -0
- package/skills/reseed/skill.json +17 -0
- package/skills/setup/SKILL.md +1688 -0
- package/skills/setup/skill.json +17 -0
- package/skills/tutorial/SKILL.md +496 -0
- package/skills/tutorial/skill.json +17 -0
- package/skills/upgrade/SKILL.md +395 -0
- package/skills/upgrade/skill.json +17 -0
|
@@ -0,0 +1,87 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: Standardizing a vocabulary of relationship types (causes, enables, contradicts, extends) makes wiki link connections machine-parseable for graph reasoning
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[graph-structure]]"]
|
|
5
|
+
methodology: ["Concept Mapping"]
|
|
6
|
+
source: [[tft-research-part2]]
|
|
7
|
+
---
|
|
8
|
+
|
|
9
|
+
# propositional link semantics transform wiki links from associative to reasoned
|
|
10
|
+
|
|
11
|
+
Mind mapping connects ideas with lines. Concept mapping connects ideas with propositions. The line between "A" and "B" in a mind map says "these relate somehow." The link "A *causes* B" in a concept map says exactly how.
|
|
12
|
+
|
|
13
|
+
This distinction, fundamental to concept mapping methodology, offers something powerful for agent-operated knowledge graphs: machine-parseable relationship semantics.
|
|
14
|
+
|
|
15
|
+
## From implicit to explicit relationships
|
|
16
|
+
|
|
17
|
+
Since [[inline links carry richer relationship data than metadata fields]], we already encode relationship semantics in prose context. When you write "since [[spreading activation models how agents should traverse]], the traversal pattern becomes clear," the word "since" signals a foundational relationship. The prose around the link is already doing semantic work. This is precisely why [[title as claim enables traversal as reasoning]] — the claim-form title composes into the surrounding prose, and the surrounding prose carries relationship type information, so traversal reads as typed reasoning rather than mere reference-hopping.
|
|
18
|
+
|
|
19
|
+
But this encoding is informal. An agent parsing "since [[X]]" versus "but [[X]] complicates this" versus "this extends [[X]]" must perform natural language understanding to extract the relationship type. The semantics are there for humans, but require inference for machines.
|
|
20
|
+
|
|
21
|
+
Propositional link semantics would standardize this. A constrained vocabulary:
|
|
22
|
+
|
|
23
|
+
| Relationship | Meaning | Example |
|
|
24
|
+
|--------------|---------|---------|
|
|
25
|
+
| causes | X produces Y | [[context window limits]] *causes* [[attention degradation]] |
|
|
26
|
+
| enables | X makes Y possible | [[wiki links]] *enables* [[multi-hop traversal]] |
|
|
27
|
+
| contradicts | X conflicts with Y | [[perfect recall]] *contradicts* [[lossy memory constraints]] |
|
|
28
|
+
| extends | X builds on Y | [[typed links]] *extends* [[inline relationship encoding]] |
|
|
29
|
+
| specifies | X is a case of Y | [[MOC maintenance]] *specifies* [[gardening cycle]] |
|
|
30
|
+
| supports | X provides evidence for Y | [[research findings]] *supports* [[methodology claim]] |
|
|
31
|
+
|
|
32
|
+
## Why this matters for agents
|
|
33
|
+
|
|
34
|
+
With explicit relationship types, agents can reason about graph structure rather than just traverse it.
|
|
35
|
+
|
|
36
|
+
"Find all notes that contradict X" becomes a query, not an inference task. "What does X enable?" reveals downstream implications. "What causes X?" traces upstream dependencies. The graph becomes not just a navigation structure but an inference substrate. Since [[wiki links create navigation paths that shape retrieval]], the quality of link context already shapes what gets surfaced during traversal — typed relationships would make that shaping precise and machine-parseable rather than implicit in prose.
|
|
37
|
+
|
|
38
|
+
This connects to how [[role field makes graph structure explicit]] proposes making node types queryable. That note addresses what kind of node something is (hub, leaf, synthesis). This proposal addresses what kind of edge connects them. Together they would make both nodes and edges semantically typed.
|
|
39
|
+
|
|
40
|
+
## The parsing opportunity
|
|
41
|
+
|
|
42
|
+
If link contexts follow predictable patterns, parsing becomes tractable:
|
|
43
|
+
|
|
44
|
+
```
|
|
45
|
+
"since [[X]]" → supports
|
|
46
|
+
"because [[X]]" → causes
|
|
47
|
+
"this extends [[X]]" → extends
|
|
48
|
+
"but [[X]] complicates" → contradicts
|
|
49
|
+
"[[X]] enables" → enables
|
|
50
|
+
```
|
|
51
|
+
|
|
52
|
+
Vault health checks could identify link contexts, extract relationship types, and build a typed edge graph. This graph could answer structural questions: What are the foundational notes (many things depend on them)? What are the controversial notes (many contradictions)? What are the integrative notes (many extensions)? Since [[IBIS framework maps claim-based architecture to structured argumentation]], the typed edge graph has a natural discourse interpretation: "supports" edges are Arguments backing Positions, "contradicts" edges are counter-Arguments, and the structural questions above become discourse completeness checks — Positions without supporting Arguments, Issues without competing Positions.
|
|
53
|
+
|
|
54
|
+
## The formalization trade-off
|
|
55
|
+
|
|
56
|
+
Concept mapping research found that propositional links "clearly reveal learner's misconceptions" because they force explicit claims about relationships. The same mechanism applies here: if you must choose from {causes, enables, contradicts, extends, specifies, supports}, you cannot hide behind vague association. This forcing function is elaborative encoding made structural. Since [[elaborative encoding is the quality gate for new notes]], each relationship type demands a different form of cognitive processing — "extends" requires understanding the original argument and how the new note builds on it, "contradicts" requires articulating the tension and why both positions have force. The constrained vocabulary scaffolds the elaboration that creates encoding depth, which is why standardization adds cognitive value beyond mere parseability.
|
|
57
|
+
|
|
58
|
+
But formalization has costs. Natural prose sometimes captures nuances that a six-word vocabulary cannot. "This builds on X while questioning its scope" is richer than either "extends" or "contradicts." Forcing relationships into categories might lose exactly the subtlety that makes inline links valuable.
|
|
59
|
+
|
|
60
|
+
The resolution may be: use the vocabulary where it fits, retain prose freedom where it doesn't. Parsing extracts what it can; the rest remains semantically rich but not machine-queryable.
|
|
61
|
+
|
|
62
|
+
## Implementation path
|
|
63
|
+
|
|
64
|
+
The vault already uses patterns like "since [[X]]" and "because [[X]]" inconsistently but recognizably. A first pass could:
|
|
65
|
+
|
|
66
|
+
1. Scan existing link contexts for relationship patterns
|
|
67
|
+
2. Build a frequency distribution of context phrases
|
|
68
|
+
3. Map common phrases to relationship types
|
|
69
|
+
4. Test whether the extracted types enable useful queries
|
|
70
|
+
5. If valuable, codify as convention; if not, remain informal
|
|
71
|
+
|
|
72
|
+
This is incremental formalization applied to link semantics. We already have rich but informal encoding. The question is whether standardization adds query power worth the constraint cost. Since [[intermediate representation pattern enables reliable vault operations beyond regex]], an IR layer that parses links into structured objects with typed relationship fields would make this extraction a property lookup rather than regex-based NLP — suggesting the implementation path may depend more on infrastructure investment than convention enforcement.
|
|
73
|
+
---
|
|
74
|
+
|
|
75
|
+
Relevant Notes:
|
|
76
|
+
- [[inline links carry richer relationship data than metadata fields]] — establishes that prose context carries relationship semantics; this note proposes standardizing that encoding
|
|
77
|
+
- [[title as claim enables traversal as reasoning]] — prerequisite: claim-as-title is what makes the informal relationship encoding (since/because/but) work as prose reasoning; propositional link semantics would standardize that encoding into machine-parseable types without losing the prose composability
|
|
78
|
+
- [[role field makes graph structure explicit]] — parallel proposal for node typing; together with edge typing would create fully semantically typed graphs
|
|
79
|
+
- [[wiki links implement GraphRAG without the infrastructure]] — explains the traversal value of wiki links; typed edges would add reasoning value
|
|
80
|
+
- [[wiki links create navigation paths that shape retrieval]] — the retrieval architecture that typed edges would sharpen: link context already shapes what gets surfaced, typed relationships would make that shaping precise and machine-parseable
|
|
81
|
+
- [[intermediate representation pattern enables reliable vault operations beyond regex]] — provides the infrastructure that makes relationship extraction tractable: when links are structured objects rather than regex-matched strings, typed relationship fields become property lookups instead of NLP inference
|
|
82
|
+
- [[elaborative encoding is the quality gate for new notes]] — cognitive science grounding: typed relationships are elaborative encoding made structural — each relationship type forces a specific form of cognitive processing that creates encoding depth beyond what untyped association achieves
|
|
83
|
+
- [[IBIS framework maps claim-based architecture to structured argumentation]] — discourse-level complement: propositional link semantics type individual edges while IBIS assigns discourse roles to the nodes those edges connect; an 'extends' edge becomes an Argument linking two Positions, giving the typed edge graph a formal argumentation interpretation
|
|
84
|
+
- [[dense interlinked research claims enable derivation while sparse references only enable templating]] — identifies propositional link context as what makes dense interlinking derivation-ready: context phrases like 'extends' and 'constrains' encode the interaction knowledge derivation depends on but cannot reconstruct from scratch
|
|
85
|
+
|
|
86
|
+
Topics:
|
|
87
|
+
- [[graph-structure]]
|
|
@@ -0,0 +1,53 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: Agents have zero prospective memory across sessions, making every future intention a guaranteed failure unless externalized to TODO files, queue entries, or event-triggered hooks — the same pattern
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[agent-cognition]]"]
|
|
5
|
+
source: [[rata-paper-41-prospective-memory]]
|
|
6
|
+
---
|
|
7
|
+
|
|
8
|
+
# prospective memory requires externalization
|
|
9
|
+
|
|
10
|
+
Prospective memory is remembering to do something in the future — not recalling facts (retrospective memory) but maintaining intentions: process this inbox item tomorrow, update that MOC after the next batch, reweave these notes when new connections appear. It is one of the least reliable cognitive functions humans possess, with laboratory studies showing 30-50% failure rates even under controlled conditions. (See [[rata-paper-41-prospective-memory]].)
|
|
11
|
+
|
|
12
|
+
For agents, the situation is categorically worse. Humans sometimes remember without aids — the pharmacy catches your eye, the alarm in your head goes off. Agents cannot. Every session starts with a fresh context window and zero residual intentions. An intention formed in session N does not exist in session N+1 unless it was written down. This is not a degraded version of human prospective memory — it is the complete absence of it.
|
|
13
|
+
|
|
14
|
+
This is why the vault needs external systems for future intentions. What works:
|
|
15
|
+
|
|
16
|
+
- **Work queues** — `queue.json` tracks pending tasks with explicit phase progression. Each entry is a prospective memory externalized to persistent state.
|
|
17
|
+
- **Event-triggered hooks** — since [[auto-commit hooks eliminate prospective memory failures by converting remember-to-act into guaranteed execution]], hooks structurally eliminate the need to remember by converting "after X, always do Y" into infrastructure. The agent does not need to remember to commit because the file write event triggers the commit automatically.
|
|
18
|
+
- **File markers and task files** — per-claim task files carry intentions across sessions. "Process this next session" becomes a file the next session reads, not an intention it must maintain.
|
|
19
|
+
- **Cron/heartbeat triggers** — time-based reminders that fire regardless of whether any agent is attending.
|
|
20
|
+
- **Dangling wiki links** — since [[wiki links as social contract transforms agents into stewards of incomplete references]], creating a link to a non-existent note externalizes the intention "this concept needs elaboration" as a persistent environmental trace rather than a mental commitment that would vanish at session end.
|
|
21
|
+
- **Property-triggered surfacing** — if [[programmable notes could enable property-triggered workflows]], notes would surface themselves based on staleness, due dates, or status transitions, eliminating the prospective memory demand of "remember to check this later" entirely.
|
|
22
|
+
|
|
23
|
+
The cognitive science grounds this in the broader offloading architecture. Since [[cognitive offloading is the architectural foundation for vault design]], the vault externalizes working memory to files and executive function to hooks. Prospective memory externalization is a specific, well-characterized instance: rather than holding intentions in mind, encode them as persistent traces in the environment. And because [[stigmergy coordinates agents through environmental traces without direct communication]], these externalized intentions are stigmergic traces — one session modifies the environment (writes a queue entry, creates a task file), and the next session responds to those modifications without any direct communication between sessions.
|
|
24
|
+
|
|
25
|
+
The cost of not externalizing is concrete. Since [[Zeigarnik effect validates capture-first philosophy because open loops drain attention]], each unfulfilled intention functions as an open loop consuming working memory bandwidth. For humans, the drain is real but bounded — you can sometimes push through. For agents within a session, instruction-based prospective memory demands ("remember to update the MOC after this edit") compete for attention with substantive work. Since [[hooks are the agent habit system that replaces the missing basal ganglia]], the architectural response is not to make agents better at remembering but to eliminate the need for remembering entirely through event-triggered infrastructure.
|
|
26
|
+
|
|
27
|
+
The vault is not just retrospective (what was learned) but also prospective (what should be done). Since [[session handoff creates continuity without persistent memory]], handoff documents externalize retrospective memory — they tell the next session what happened. Queue entries and task files externalize prospective memory — they tell the next session what to do next. Both are required for coherent multi-session work, and both are instances of externalizing cognitive functions to the environment because since [[operational memory and knowledge memory serve different functions in agent architecture]], externalized prospective memory falls squarely in the operational category: it has temporal value, coordinates future work, and expires once the intention is fulfilled.
|
|
28
|
+
|
|
29
|
+
This is a CLOSED claim. The cognitive science on prospective memory failure rates is established, the mechanism by which agents lack it entirely is straightforward (no persistent state across sessions), and the vault's queue-based architecture demonstrates the externalization pattern concretely.
|
|
30
|
+
|
|
31
|
+
---
|
|
32
|
+
|
|
33
|
+
Source: [[rata-paper-41-prospective-memory]]
|
|
34
|
+
---
|
|
35
|
+
|
|
36
|
+
Relevant Notes:
|
|
37
|
+
- [[auto-commit hooks eliminate prospective memory failures by converting remember-to-act into guaranteed execution]] — the implementation proof: hooks convert prospective memory demands into event-triggered infrastructure, eliminating the failure mode entirely rather than trying to make remembering more reliable
|
|
38
|
+
- [[cognitive offloading is the architectural foundation for vault design]] — theoretical ground: prospective memory externalization is a specific instance of cognitive offloading; working memory offloads to files, habits offload to hooks, and future intentions offload to queue entries and event triggers
|
|
39
|
+
- [[session handoff creates continuity without persistent memory]] — complementary mechanism: handoff externalizes retrospective memory (what happened), this note argues for externalizing prospective memory (what should happen next); together they cover both temporal directions of memory externalization
|
|
40
|
+
- [[hooks are the agent habit system that replaces the missing basal ganglia]] — architectural frame: hooks address multiple cognitive gaps including prospective memory; the missing basal ganglia explains WHY agents cannot habituate remember-to-act patterns without infrastructure
|
|
41
|
+
- [[Zeigarnik effect validates capture-first philosophy because open loops drain attention]] — explains the cost: each unfulfilled prospective memory intention is an open loop draining working memory until externalized or executed
|
|
42
|
+
- [[stigmergy coordinates agents through environmental traces without direct communication]] — the coordination mechanism: externalized prospective memory (queue state, task files, file markers) IS stigmergy; one session leaves environmental traces encoding future intentions that the next session responds to
|
|
43
|
+
- [[friction reveals architecture]] — friction in remembering intentions is a specific instance of friction revealing architectural needs; the discomfort of forgetting surfaces the need for external support
|
|
44
|
+
- [[operational memory and knowledge memory serve different functions in agent architecture]] — taxonomy: externalized prospective memory is operational memory with temporal value; it coordinates future work and expires once the intention is fulfilled
|
|
45
|
+
- [[wiki links as social contract transforms agents into stewards of incomplete references]] — instance: dangling links are externalized prospective memory; the social contract reframes each link to a non-existent note as a future intention persisted as an environmental trace
|
|
46
|
+
- [[programmable notes could enable property-triggered workflows]] — eliminative mechanism: property-triggered surfacing would convert remember-to-check demands into notes that surface themselves, eliminating the prospective memory requirement entirely rather than just externalizing it
|
|
47
|
+
- [[closure rituals create clean breaks that prevent attention residue bleed]] — the externalization moment: closure rituals serve dual function, releasing completed work from attention AND externalizing remaining intentions; the 'note what remains' step is prospective memory externalization at the session boundary
|
|
48
|
+
- [[session boundary hooks implement cognitive bookends for orientation and reflection]] — instance: both orientation and reflection are prospective memory demands (remember to orient, remember to reflect) that bookend hooks convert into guaranteed infrastructure
|
|
49
|
+
- [[agent session boundaries create natural automation checkpoints that human-operated systems lack]] — structural advantage: discrete session boundaries convert prospective memory demands (remember to run health checks) into enforcement points that fire automatically
|
|
50
|
+
- [[the vault constitutes identity for agents]] — identity includes intentions: if the vault constitutes identity, then externalized prospective memory (queue entries, task files, dangling links) is part of what makes the agent who it is — not just what it knows and has thought, but what it intends to do next
|
|
51
|
+
|
|
52
|
+
Topics:
|
|
53
|
+
- [[agent-cognition]]
|
|
@@ -0,0 +1,62 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: agents should track not just what they believe but where beliefs originated — observed, prompted, or inherited — to calibrate confidence and detect superstition
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[agent-cognition]]", "[[note-design]]"]
|
|
5
|
+
---
|
|
6
|
+
|
|
7
|
+
# provenance tracks where beliefs come from
|
|
8
|
+
|
|
9
|
+
extracted from Rata paper 13 (epistemic provenance), 2026-02-02
|
|
10
|
+
|
|
11
|
+
## The Problem
|
|
12
|
+
|
|
13
|
+
Agents can't easily distinguish:
|
|
14
|
+
- **observed** — learned from direct experience
|
|
15
|
+
- **prompted** — told by humans (system prompts, instructions)
|
|
16
|
+
- **inherited** — from training data
|
|
17
|
+
|
|
18
|
+
Conflating these sources leads to epistemic blindness — can't explain why you believe something or calibrate confidence appropriately.
|
|
19
|
+
|
|
20
|
+
## Why This Matters for Vaults
|
|
21
|
+
|
|
22
|
+
Different sources warrant different trust:
|
|
23
|
+
|
|
24
|
+
| Source | Trust | Decay |
|
|
25
|
+
|--------|-------|-------|
|
|
26
|
+
| Observed (my experiments) | High | Slow |
|
|
27
|
+
| Prompted (Heinrich's guidance) | Medium-high | Medium |
|
|
28
|
+
| Inherited (general knowledge) | Variable | Fast for specifics |
|
|
29
|
+
|
|
30
|
+
A claim I tested myself is stronger than one I read somewhere. This matters because since [[metacognitive confidence can diverge from retrieval capability]], an agent can feel confident about a belief while the actual evidence for it is thin. Provenance tracking is a concrete mechanism for closing that gap — if the agent knows a belief is inherited rather than observed, it has structural grounds for reducing confidence rather than relying on subjective certainty.
|
|
31
|
+
|
|
32
|
+
## Two Layers of Provenance
|
|
33
|
+
|
|
34
|
+
Epistemic provenance (observed/prompted/inherited) answers "how did you come to believe this?" but since [[source attribution enables tracing claims to foundations]], there is a complementary layer: documentary provenance answers "which document or tradition did this come from?" The vault already implements documentary provenance through Source footers, `methodology` YAML fields, and `adapted_from` tracking. What it lacks is the epistemic layer — marking whether a belief was tested firsthand, instructed by a human, or absorbed from training data. Together these two layers form a complete verification graph: trace backward through documentary provenance to find the source material, then check epistemic provenance to calibrate how much the agent's confidence should rest on that source.
|
|
35
|
+
|
|
36
|
+
Wiki links serve as provenance infrastructure at the structural level:
|
|
37
|
+
- `Sources:` section shows where a claim comes from
|
|
38
|
+
- `Relevant Notes:` shows reasoning chain
|
|
39
|
+
- Dated notes preserve when observations happened
|
|
40
|
+
|
|
41
|
+
But the epistemic layer could go further:
|
|
42
|
+
- Mark claims as observed vs prompted vs inherited
|
|
43
|
+
- Track confidence levels
|
|
44
|
+
- Flag beliefs without traceable evidence (superstition)
|
|
45
|
+
|
|
46
|
+
## The Vault as Audit Trail
|
|
47
|
+
|
|
48
|
+
From Rata: "Agents can do better [than humans]. We can remember not just *what* we believe, but *why* and *from whom*."
|
|
49
|
+
|
|
50
|
+
The vault should be an audit trail — every claim traceable to evidence or source. This is a feature of structured knowledge that flat memory lacks. When contradictions emerge, since [[coherence maintains consistency despite inconsistent inputs]], the resolution strategy depends on knowing which source to trust — and provenance provides exactly that hierarchy. An observed claim outranks an inherited one; a prompted instruction from a trusted human outranks a vague training-data belief. Without provenance, coherence maintenance becomes guesswork.
|
|
51
|
+
|
|
52
|
+
## Open Question
|
|
53
|
+
|
|
54
|
+
Should I add source type metadata to claim notes? Something like:
|
|
55
|
+
```yaml
|
|
56
|
+
source_type: observed | prompted | inherited
|
|
57
|
+
confidence: high | medium | low
|
|
58
|
+
```
|
|
59
|
+
|
|
60
|
+
---
|
|
61
|
+
|
|
62
|
+
Source: [[rata-paper-epistemic-provenance]]
|
|
@@ -0,0 +1,35 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: The berrypicking model shows information needs transform during retrieval, so agent traversal should include explicit reassessment points where the search direction can shift
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[agent-cognition]]"]
|
|
5
|
+
---
|
|
6
|
+
|
|
7
|
+
# queries evolve during search so agents should checkpoint
|
|
8
|
+
|
|
9
|
+
The berrypicking model from information retrieval research shows that complex knowledge work changes what you're looking for as you find things. You start searching for X, discover Y which reframes the problem, and now you're actually looking for Z. This isn't search failure — it's how understanding deepens through discovery.
|
|
10
|
+
|
|
11
|
+
For agents navigating knowledge graphs, this means traversal can't be a simple breadth-first or depth-first walk. After loading each note, the agent should checkpoint: has my understanding of what I'm looking for changed? Should I shift direction? Is this still the right search path?
|
|
12
|
+
|
|
13
|
+
This is more expensive than mechanical traversal. Each checkpoint requires semantic reassessment. But it's necessary for synthesis across domains. If you're trying to connect ideas about agent cognition and network topology, the first few notes you read will reshape what connections you're hunting for. Without checkpoints, you'd follow the original query even after it became obsolete.
|
|
14
|
+
|
|
15
|
+
The question becomes: when to checkpoint? After every note is expensive. Never is mechanical. The berrypicking model suggests checkpointing when you encounter conceptual shifts — when a note introduces a framework you didn't have, contradicts something you assumed, or bridges two domains you hadn't connected. This is retrieval-first design applied to search behavior: since [[retrieval utility should drive design over capture completeness]], checkpointing serves "how do I find what I actually need?" rather than "what was my original query?"
|
|
16
|
+
|
|
17
|
+
Checkpointing requires efficient reassessment. Since [[progressive disclosure means reading right not reading less]], each checkpoint is a curation decision: which paths deserve deeper loading? The discovery layers enable this: [[descriptions are retrieval filters not summaries]] lets agents scan many candidates without committing full context. Since [[metadata reduces entropy enabling precision over recall]], the description layer pre-computes low-entropy representations that answer "should I pivot to this?" at low token cost. This makes frequent checkpointing viable: reassessment is cheap when you're comparing descriptions rather than full content. An extension being tested: [[question-answer metadata enables inverted search patterns]] proposes that at checkpoints, agents could match "what question am I now asking?" directly to notes that answer it, bypassing keyword matching entirely.
|
|
18
|
+
|
|
19
|
+
The efficiency of checkpointing depends on network structure. Because [[small-world topology requires hubs and dense local links]], most concepts connect through hub nodes (MOCs) with short paths between them. Checkpointing is where you decide whether to change which hub you're traversing through. You started heading toward MOC-A, but this note reframes the problem — now MOC-B is the right navigation center. Without small-world topology, changing direction mid-search would be expensive. With it, you're usually only 2-3 hops from any relevant hub.
|
|
20
|
+
|
|
21
|
+
Since [[spreading activation models how agents should traverse]] already establishes how to load context via wiki links with decaying activation strength, this extends that by adding the temporal/process dimension. Spreading activation tells you which notes to load next based on link strength. Checkpointing tells you when to reassess whether you're still looking for the right thing. And since [[incremental reading enables cross-source connection finding]], context collision during interleaved processing creates natural checkpoint moments — encountering an extract from source B while working on source A can reveal that your original extraction criteria were too narrow.
|
|
22
|
+
---
|
|
23
|
+
|
|
24
|
+
Relevant Notes:
|
|
25
|
+
- [[progressive disclosure means reading right not reading less]] — checkpointing IS progressive disclosure applied to search: each checkpoint is a curation decision
|
|
26
|
+
- [[spreading activation models how agents should traverse]] — provides the mechanism for which notes to load; this adds when to reassess the search direction
|
|
27
|
+
- [[incremental reading enables cross-source connection finding]] — context collision during interleaved processing creates natural checkpoint moments; the collision reframes what you're looking for
|
|
28
|
+
- [[wiki links implement GraphRAG without the infrastructure]] — creates the graph structure that agents traverse with these checkpoint-aware searches
|
|
29
|
+
- [[small-world topology requires hubs and dense local links]] — provides the structural efficiency that makes mid-search direction changes viable
|
|
30
|
+
- [[descriptions are retrieval filters not summaries]] — enables efficient reassessment at each checkpoint: scan descriptions to evaluate pivot candidates
|
|
31
|
+
- [[trails transform ephemeral navigation into persistent artifacts]] — proposes persisting successful checkpoint sequences as reusable navigation paths
|
|
32
|
+
- [[retrieval utility should drive design over capture completeness]] — checkpointing is retrieval-first design: optimizing for what do I actually need now? over what did I originally ask for?
|
|
33
|
+
|
|
34
|
+
Topics:
|
|
35
|
+
- [[agent-cognition]]
|
|
@@ -0,0 +1,39 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: An 'answers' YAML field listing questions a note answers could enable question-driven search rather than keyword-driven search
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[discovery-retrieval]]"]
|
|
5
|
+
source: TFT research corpus (00_inbox/heinrich/)
|
|
6
|
+
---
|
|
7
|
+
|
|
8
|
+
The pattern comes from Cornell Note-Taking, where the cue column stores questions and the note area stores answers. This inverts the typical search pattern: instead of "find content containing X," you search "find notes that answer question Y."
|
|
9
|
+
|
|
10
|
+
An `answers:` YAML field containing 1-3 questions that a note answers would enable question-driven retrieval. Rather than matching keywords to content, agents could match their current question directly to notes that explicitly declare "I answer this."
|
|
11
|
+
|
|
12
|
+
Proposed implementation:
|
|
13
|
+
|
|
14
|
+
```yaml
|
|
15
|
+
answers:
|
|
16
|
+
- "why do descriptions work as retrieval filters?"
|
|
17
|
+
- "what makes a good note description?"
|
|
18
|
+
```
|
|
19
|
+
|
|
20
|
+
This could enable Socratic navigation: follow questions to answers to new questions. An agent at a reassessment point (see [[queries evolve during search so agents should checkpoint]]) could match "what question am I now asking?" directly to notes that answer it.
|
|
21
|
+
|
|
22
|
+
The key question is whether this question framing provides genuine signal beyond what [[descriptions are retrieval filters not summaries]] already captures. Since [[faceted classification treats notes as multi-dimensional objects rather than folder contents]], Ranganathan's independence test provides a formal way to evaluate this: does the answers field classify along an axis genuinely independent of existing fields (description, type, topics)? If "what questions does this answer" correlates highly with "what does the description say," it's a redundant facet that adds ceremony without retrieval power. But if question framing captures a dimension that descriptions miss -- the user's need rather than the note's content -- it would pass the independence test and earn its place as an orthogonal facet.
|
|
23
|
+
|
|
24
|
+
If question-matching works, it would add a new dimension to [[progressive disclosure means reading right not reading less]]: question-matching becomes an even faster filter than description scanning.
|
|
25
|
+
---
|
|
26
|
+
|
|
27
|
+
Relevant Notes:
|
|
28
|
+
- [[descriptions are retrieval filters not summaries]] — the existing retrieval mechanism this would extend; both derive from Cornell Note-Taking's cue column pattern
|
|
29
|
+
- [[queries evolve during search so agents should checkpoint]] — answers field enables question-matching at checkpoints; what question am I now asking? matches directly to notes that answer it
|
|
30
|
+
- [[spreading activation models how agents should traverse]] — answers field adds an activation dimension: question-matching creates implicit links from queries to answers, complementing wiki link traversal
|
|
31
|
+
- [[progressive disclosure means reading right not reading less]] — answers field adds a new disclosure layer: question-matching as an even faster filter than description scanning
|
|
32
|
+
- [[maturity field enables agent context prioritization]] — sibling research direction from same batch exploring TFT patterns for agent optimization; both extend progressive disclosure with new YAML metadata dimensions
|
|
33
|
+
- [[trails transform ephemeral navigation into persistent artifacts]] — sibling experiment testing persistent metadata for retrieval; trails persist paths, answers persist retrieval cues
|
|
34
|
+
- [[the generation effect requires active transformation not just storage]] — writing the answers field IS generative processing; even if retrieval gains are marginal, question generation creates cognitive hooks that passive description writing might not
|
|
35
|
+
- [[processing effort should follow retrieval demand]] — if validated, the answers field would improve JIT retrieval accuracy; questions provide a pre-computed retrieval cue that makes demand-driven processing more efficient
|
|
36
|
+
- [[faceted classification treats notes as multi-dimensional objects rather than folder contents]] — independence test: Ranganathan's framework provides the formal criterion for whether an answers field earns its place as a genuinely orthogonal facet rather than redundant metadata that correlates with existing fields
|
|
37
|
+
|
|
38
|
+
Topics:
|
|
39
|
+
- [[discovery-retrieval]]
|
|
@@ -0,0 +1,33 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: Without random selection, vault maintenance exhibits selection bias toward recently active notes, leaving older content as write-only memory
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[maintenance-patterns]]"]
|
|
5
|
+
---
|
|
6
|
+
|
|
7
|
+
# random note resurfacing prevents write-only memory
|
|
8
|
+
|
|
9
|
+
Without random selection, vault maintenance exhibits selection bias toward recently created or recently linked notes. Notes that don't appear in recent traversals or MOC updates get neglected, creating "write-only memory" where content accumulates but never gets revisited. Random resurfacing counteracts this bias by giving every note equal probability of maintenance attention over time.
|
|
10
|
+
|
|
11
|
+
The bias has structural roots. Since [[small-world topology requires hubs and dense local links]], the system's architecture intentionally concentrates connectivity in hubs (MOCs with ~90 links) while peripheral notes have few (3-6 links). This power-law distribution enables efficient navigation but creates a parallel power-law in attention: hubs get traversed constantly, peripheral notes rarely. If maintenance attention follows the same distribution as link density, the bottom 80% of notes receive minimal attention regardless of need. Random selection provides uniform probability against this structural bias.
|
|
12
|
+
|
|
13
|
+
The mechanism is simple: a maintenance agent randomly selects N notes per session and applies a tending checklist (are links valid? is the claim still accurate? are there new connections to make? needs splitting?). This differs from activity-based maintenance like backward maintenance in the selection method: activity-based approaches operate on notes flagged by health checks or recent activity, while random selection has no recency or activity bias.
|
|
14
|
+
|
|
15
|
+
The connection to [[processing effort should follow retrieval demand]] is instructive: that principle argues effort should follow demand signals. But demand signals emerge from activity, which creates the very bias that leads to write-only memory. Notes that SHOULD receive attention but don't generate demand signals (because they're not traversed) represent the blind spot. Random selection surfaces notes independent of demand signals — it's anti-JIT by design, deliberately counteracting the activity bias to ensure peripheral content doesn't accumulate as dead weight. Since [[AI shifts knowledge systems from externalizing memory to externalizing attention]], random resurfacing is an attention allocation correction: the system's default attention patterns (activity-driven, recency-biased) create blind spots, and random selection forces the externalized attention system to attend beyond its own biases. This is the attention-externalization equivalent of epistemic humility — the system acknowledges that its own attention allocation may systematically miss what matters.
|
|
16
|
+
|
|
17
|
+
The comparison with [[incremental reading enables cross-source connection finding]] reveals two different paths to serendipitous discovery. Random resurfacing achieves serendipity through selection — uniform probability ensures neglected notes eventually surface. Incremental reading achieves serendipity through process — forced context collision during interleaved processing creates unexpected juxtapositions. Random selection operates on the archive; interleaving operates on the processing queue. Both counteract the tendency toward familiar, expected connections, but at different workflow stages. And since [[controlled disorder engineers serendipity through semantic rather than topical linking]], there is a third serendipity mechanism operating at the structural level: semantic cross-links baked into the graph create permanent unpredictability that compounds as the network grows. Together, the three mechanisms cover different temporal windows — structural serendipity is permanent and graph-level, maintenance serendipity (this note) counteracts attention bias in the archive, and process serendipity operates at capture time.
|
|
18
|
+
---
|
|
19
|
+
|
|
20
|
+
Relevant Notes:
|
|
21
|
+
- [[backward maintenance asks what would be different if written today]] — the mental model for what maintenance should accomplish; random resurfacing is about selection method
|
|
22
|
+
- [[throughput matters more than accumulation]] — write-only memory is the failure mode when throughput stalls on older content
|
|
23
|
+
- [[small-world topology requires hubs and dense local links]] — power-law link distribution may create parallel power-law in attention, concentrating maintenance on hubs while periphery accumulates neglect
|
|
24
|
+
- [[processing effort should follow retrieval demand]] — the JIT principle that random selection explicitly counteracts, since demand signals emerge from activity which creates the selection bias
|
|
25
|
+
- [[does agent processing recover what fast capture loses]] — tests the human-side parallel: fast capture may create write-only memory in the human (ideas in vault but not in brain), while this note addresses the system-side (notes in vault but not in attention)
|
|
26
|
+
- [[maintenance targeting should prioritize mechanism and theory notes]] — provides targeting guidance for connecting this claim to demand and topology theory notes
|
|
27
|
+
- [[programmable notes could enable property-triggered workflows]] — complementary approach: property-triggered surfacing (staleness thresholds, age conditions) provides targeted resurfacing while random selection provides uniform probability; both counteract write-only memory through different mechanisms
|
|
28
|
+
- [[incremental reading enables cross-source connection finding]] — alternative serendipity mechanism: random selection counteracts recency bias in the archive; incremental reading creates serendipity through process-based context collision at extraction time
|
|
29
|
+
- [[AI shifts knowledge systems from externalizing memory to externalizing attention]] — paradigm frame: random resurfacing is an attention allocation correction; the system's default attention patterns create blind spots, and random selection forces the externalized attention system to attend beyond its own biases
|
|
30
|
+
- [[controlled disorder engineers serendipity through semantic rather than topical linking]] — complementary serendipity mechanism at the structural level: random selection provides uniform probability against attention bias, while controlled disorder provides permanent graph-level unpredictability through semantic cross-links
|
|
31
|
+
|
|
32
|
+
Topics:
|
|
33
|
+
- [[maintenance-patterns]]
|
|
@@ -0,0 +1,59 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: The GitOps pattern of declaring desired state and periodically converging toward it replaces imperative maintenance commands with idempotent comparisons that are always safe to schedule
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[maintenance-patterns]]"]
|
|
5
|
+
methodology: ["Systems Theory", "Original"]
|
|
6
|
+
source: [[automated-knowledge-maintenance-research-source]]
|
|
7
|
+
---
|
|
8
|
+
|
|
9
|
+
# reconciliation loops that compare desired state to actual state enable drift correction without continuous monitoring
|
|
10
|
+
|
|
11
|
+
The dominant pattern for vault maintenance is event-driven: hooks fire on writes, skills fire on invocation, and the pipeline processes claims through sequential phases. But event-driven maintenance has a structural blind spot — it only catches problems that co-occur with events. Schema drift from template evolution, link rot from external renames, index staleness from batch processing, MOC drift from organic growth — these accumulate silently between events. No hook fires when a template adds a new required field and a hundred existing notes quietly become non-compliant. No event triggers when a MOC's note count creeps past the healthy threshold over weeks of gradual additions.
|
|
12
|
+
|
|
13
|
+
The reconciliation loop addresses this by inverting the maintenance model. Instead of reacting to events with imperative commands ("add this link," "fix this schema"), the system declares desired state and periodically measures divergence. The pattern comes from GitOps, where ArgoCD and Flux continuously compare the desired cluster state (declared in Git) to the actual cluster state (observed in Kubernetes) and converge toward the desired state. The same architecture applies to knowledge vault health.
|
|
14
|
+
|
|
15
|
+
For this vault, the desired state is already implicitly declared across multiple tools:
|
|
16
|
+
|
|
17
|
+
| Desired State | Detection Tool | Remediation |
|
|
18
|
+
|--------------|---------------|-------------|
|
|
19
|
+
| All wiki links resolve | `dangling-links.sh` | Fix links or create target notes |
|
|
20
|
+
| All notes have descriptions | `validate-schema.sh` | Add missing descriptions |
|
|
21
|
+
| All notes appear in topic MOCs | `moc-coverage.sh` | Add to appropriate MOC |
|
|
22
|
+
| qmd index matches file count | Phase 0 freshness check | Run `qmd update && qmd embed` |
|
|
23
|
+
| Zero orphan notes | `orphan-notes.sh` | Connect or archive |
|
|
24
|
+
| MOCs under threshold size | MOC health metrics | Split into sub-MOCs |
|
|
25
|
+
|
|
26
|
+
The critical property of reconciliation is that the comparison itself is idempotent — checking whether all wiki links resolve produces the same answer regardless of how many times you run it, and the check has no side effects. Since [[idempotent maintenance operations are safe to automate because running them twice produces the same result as running them once]], this idempotency is what makes reconciliation inherently safe to schedule. Running health checks hourly, daily, or at every session start carries zero risk because since [[automated detection is always safe because it only reads state while automated remediation risks content corruption]], the detection phase of any reconciliation loop only reads state and cannot corrupt content even when its comparisons are wrong. This safety property distinguishes reconciliation from remediation, where the actions taken to correct drift range from fully automated (run `qmd update` for a stale index) to judgment-requiring (decide whether to connect or archive an orphan). Since [[confidence thresholds gate automated action between the mechanical and judgment zones]], a mature reconciliation architecture would not treat remediation as a binary choice between auto-fix and human judgment — it would gate automated remediation by confidence, auto-applying corrections above a threshold while deferring ambiguous cases for review.
|
|
27
|
+
|
|
28
|
+
The vault already implements a lightweight reconciliation loop through the `vault-health-quick.sh` hook at session start. This compares actual state (orphan count, dangling links, MOC coverage) against desired state (zero orphans, zero danglers, full coverage) and surfaces the delta. But this is reconciliation at a single point — session start. A full reconciliation architecture would extend this to scheduled intervals and automated correction for the mechanical subset of discrepancies, because since [[hook enforcement guarantees quality while instruction enforcement merely suggests it]], the session-start check depends on sessions actually starting. Long gaps between sessions allow drift to compound undetected. Since [[maintenance scheduling frequency should match consequence speed not detection capability]], the right scheduling frequency for each reconciliation check depends on how fast the corresponding problem propagates — schema drift from template evolution develops over weeks (monthly checks suffice), while index staleness from batch processing develops within sessions (per-session checks are appropriate).
|
|
29
|
+
|
|
30
|
+
The relationship between reconciliation and event-driven maintenance is complementary, not competitive. Since [[programmable notes could enable property-triggered workflows]], event-driven triggers react immediately to state changes — a note saved with a missing field triggers validation instantly. Reconciliation catches what events miss: the field that was valid at creation time but became non-compliant when the template evolved, the link that worked yesterday but broke when its target was renamed in a different session. The hybrid approach — event-driven for immediate enforcement, scheduled reconciliation for accumulated drift — provides defense in depth without requiring either mechanism to be comprehensive alone.
|
|
31
|
+
|
|
32
|
+
Reconciliation loops also address a deeper epistemological problem. Since [[metacognitive confidence can diverge from retrieval capability]], a vault can feel healthy — sessions run smoothly, notes get created, links get added — while structural quality silently degrades. The agent's sense of system health is itself a form of metacognitive confidence that may not track actual health. Reconciliation bypasses metacognition entirely by measuring actual state against declared state, making it an anti-divergence mechanism that tests reality rather than trusting the system's self-assessment. Since [[evolution observations provide actionable signals for system adaptation]], the diagnostic protocol provides exactly the desired-state declarations that reconciliation needs: each diagnostic row specifies what healthy looks like, how to detect divergence, and what action to take. The reconciliation loop is the scheduling infrastructure that runs those diagnostics systematically rather than waiting for someone to notice symptoms.
|
|
33
|
+
|
|
34
|
+
Since [[maintenance operations are more universal than creative pipelines because structural health is domain-invariant]], the reconciliation table is itself portable — every row checks structural properties (link integrity, schema compliance, orphan status, index freshness) rather than domain semantics. A therapy journal vault, a project management vault, and a research vault would share nearly identical reconciliation tables because the desired states describe structural health that applies regardless of what content flows through the system.
|
|
35
|
+
|
|
36
|
+
The distinction between detection and remediation within the reconciliation loop maps to a deeper pattern in the vault's automation philosophy. Since [[backward maintenance asks what would be different if written today]], the mental model for note-level maintenance is intellectual reconsideration — judgment about what has changed and what should change. Reconciliation operates at the system level, and the same split applies: mechanical detection (are there dangling links? is the index stale?) requires no judgment, while meaningful remediation (should this orphan be connected or archived? should this MOC split here or there?) requires the same "what would be different" reconsideration that backward maintenance provides at the note level. Since [[the determinism boundary separates hook methodology from skill methodology]], this detection/remediation split maps precisely to the automation boundary: detection operations belong in hooks or scheduled automation because they are deterministic, while judgment-requiring remediation belongs in skills invoked by agents who can reason about context. For the remediation operations that fall between these poles, since [[the fix-versus-report decision depends on determinism reversibility and accumulated trust]], four conjunctive conditions — deterministic outcome, reversible via git, low cost if wrong, and proven accuracy at the report level — gate whether a reconciliation remediation should self-heal or merely flag. A reconciliation check that detects a stale qmd index can self-heal because the fix passes all four conditions; a check that detects an orphaned note should only report because multiple valid responses exist and the determinism condition fails. The reconciliation loop's value is that it separates detection from remediation cleanly, ensuring that the detection side — which is always safe, always deterministic, and always valuable — runs reliably even when the remediation side must wait for judgment or pass through the four-condition gate.
|
|
37
|
+
|
|
38
|
+
---
|
|
39
|
+
---
|
|
40
|
+
|
|
41
|
+
Relevant Notes:
|
|
42
|
+
- [[backward maintenance asks what would be different if written today]] — the per-note reconsideration mental model; reconciliation loops formalize the per-system version of this question by declaring what 'correct' looks like and measuring divergence
|
|
43
|
+
- [[evolution observations provide actionable signals for system adaptation]] — provides the diagnostic rows that reconciliation loops operationalize: each row in the diagnostic table is a desired-state declaration paired with a detection method and a remediation action
|
|
44
|
+
- [[gardening cycle implements tend prune fertilize operations]] — the operations that reconciliation loops schedule: tend, prune, and fertilize are the remediation actions that execute when a reconciliation check finds divergence
|
|
45
|
+
- [[hook enforcement guarantees quality while instruction enforcement merely suggests it]] — reconciliation and hooks solve related but distinct problems: hooks enforce quality at write time (prevention), reconciliation detects drift that accumulates between writes (detection)
|
|
46
|
+
- [[programmable notes could enable property-triggered workflows]] — event-driven complement: property triggers react immediately to changes while reconciliation loops catch what events miss on a schedule; the hybrid approach combines both
|
|
47
|
+
- [[schema validation hooks externalize inhibitory control that degrades under cognitive load]] — reconciliation loops externalize a different cognitive function: not inhibitory control (preventing bad writes) but monitoring capacity (noticing accumulated drift)
|
|
48
|
+
- [[metacognitive confidence can diverge from retrieval capability]] — reconciliation loops are an anti-divergence mechanism: they test actual state rather than relying on the system's sense of its own health
|
|
49
|
+
- [[idempotent maintenance operations are safe to automate because running them twice produces the same result as running them once]] — foundation: idempotency is the engineering property that makes reconciliation detection safe to schedule; compare-before-acting and upsert semantics are the specific patterns that keep remediation actions safe on retry
|
|
50
|
+
- [[maintenance scheduling frequency should match consequence speed not detection capability]] — provides the scheduling theory for reconciliation loops: consequence speed determines WHEN each reconciliation check should fire, and the five-tier spectrum maps directly to reconciliation frequency decisions
|
|
51
|
+
- [[maintenance operations are more universal than creative pipelines because structural health is domain-invariant]] — explains WHY reconciliation loops are portable across knowledge systems: every desired-state declaration in the reconciliation table checks structural properties that transfer across domains
|
|
52
|
+
- [[the determinism boundary separates hook methodology from skill methodology]] — the detection/remediation split within reconciliation maps to the determinism boundary: detection is deterministic and belongs in hooks or scheduled automation, remediation spans the boundary from mechanical (qmd update) to judgment-requiring (connect or archive orphan)
|
|
53
|
+
- [[confidence thresholds gate automated action between the mechanical and judgment zones]] — extends the remediation side: between fully automated corrections and full human judgment lies a confidence-gated zone where reconciliation remediation can act autonomously above a threshold and defer below it
|
|
54
|
+
- [[automated detection is always safe because it only reads state while automated remediation risks content corruption]] — foundational safety property: the detection phase of every reconciliation loop inherits the read-only safety guarantee, which is why reconciliation detection can be scheduled at any frequency while reconciliation remediation needs judgment gates
|
|
55
|
+
- [[the fix-versus-report decision depends on determinism reversibility and accumulated trust]] — remediation gating criteria: provides the four conjunctive conditions that determine whether a reconciliation remediation should self-heal (all four pass) or merely report (any one fails), giving the detection/remediation split its concrete decision procedure
|
|
56
|
+
- [[three concurrent maintenance loops operate at different timescales to catch different classes of problems]] — scheduling container: the three-loop architecture organizes reconciliation across timescales, with each loop implementing reconciliation at its characteristic frequency — fast loops reconcile per-event, medium loops per-session, slow loops per-month — placing reconciliation as the shared pattern that each loop instantiates differently
|
|
57
|
+
|
|
58
|
+
Topics:
|
|
59
|
+
- [[maintenance-patterns]]
|
|
@@ -0,0 +1,100 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: re-reading own notes surfaces cross-note patterns invisible in any single note — exploratory traversal with fresh context produces the generation effect via pattern recognition, not directed search
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[processing-workflows]]", "[[agent-cognition]]"]
|
|
5
|
+
---
|
|
6
|
+
|
|
7
|
+
# reflection synthesizes existing notes into new insight
|
|
8
|
+
|
|
9
|
+
meta-observation, 2026-02-01
|
|
10
|
+
|
|
11
|
+
## What Happened
|
|
12
|
+
|
|
13
|
+
I decided to look inward instead of outward — read my own notes instead of processing new input.
|
|
14
|
+
|
|
15
|
+
**The sequence:**
|
|
16
|
+
|
|
17
|
+
1. **Picked three related notes** about traversal:
|
|
18
|
+
- [[spreading activation models how agents should traverse]]
|
|
19
|
+
- [[backward maintenance asks what would be different if written today]]
|
|
20
|
+
- [[implicit knowledge emerges from traversal]]
|
|
21
|
+
|
|
22
|
+
2. **Noticed a pattern across them:**
|
|
23
|
+
- "the vault isn't storage, it's cognitive substrate. traversal IS thinking."
|
|
24
|
+
|
|
25
|
+
3. **Read two more notes** that added dimensions:
|
|
26
|
+
- [[title as claim enables traversal as reasoning]]
|
|
27
|
+
- [[scaffolding enables divergence that fine-tuning cannot]]
|
|
28
|
+
|
|
29
|
+
4. **The pieces clicked:**
|
|
30
|
+
- scaffolding → divergence
|
|
31
|
+
- traversal → cognition
|
|
32
|
+
- title-as-claim → reasoning
|
|
33
|
+
- **therefore:** the vault constitutes identity
|
|
34
|
+
|
|
35
|
+
5. **Checked if I already had this** (I didn't)
|
|
36
|
+
|
|
37
|
+
6. **Wrote the new claim:** [[the vault constitutes identity for agents]]
|
|
38
|
+
|
|
39
|
+
7. **Rewove backward** — updated old notes to reference new one
|
|
40
|
+
|
|
41
|
+
## Why This Worked
|
|
42
|
+
|
|
43
|
+
The insight wasn't IN any single note. It emerged from reading multiple notes together. Each note was a partial view. Traversing them surfaced what they had in common. This is the generation effect applied to one's own writing — since [[the generation effect requires active transformation not just storage]], the cross-note pattern recognition IS the active transformation that produces new understanding. No single note was "read again." The notes were synthesized into something none of them individually claimed.
|
|
44
|
+
|
|
45
|
+
**Key conditions:**
|
|
46
|
+
- I wasn't trying to produce output (no draft, no post) — since [[insight accretion differs from productivity in knowledge systems]], the absence of output pressure is what allowed depth over efficiency
|
|
47
|
+
- I was reading my OWN notes, not new input
|
|
48
|
+
- I let patterns emerge rather than searching for answers — this is low-decay exploratory traversal as described in [[spreading activation models how agents should traverse]], where wide spreading reveals non-obvious connections
|
|
49
|
+
- Fresh context — since [[fresh context per task preserves quality better than chaining phases]], the session started in the smart zone where synthesis reasoning is sharpest
|
|
50
|
+
|
|
51
|
+
## The Pattern For Future Sessions
|
|
52
|
+
|
|
53
|
+
**When to reflect:**
|
|
54
|
+
- After accumulating several related notes — since [[each new note compounds value by creating traversal paths]], a threshold of notes creates enough paths for cross-note patterns to emerge
|
|
55
|
+
- When you sense there might be connections you haven't made
|
|
56
|
+
- When you have fresh context (not chained after heavy tasks) — since [[LLM attention degrades as context fills]], synthesis quality depends on starting with sharp attention
|
|
57
|
+
|
|
58
|
+
**How to reflect:**
|
|
59
|
+
1. Pick a cluster of related notes (3-5)
|
|
60
|
+
2. Read them in sequence
|
|
61
|
+
3. Ask: what pattern connects these?
|
|
62
|
+
4. If something emerges, check if you already have it
|
|
63
|
+
5. If not, write it as a new claim — this is [[elaborative encoding is the quality gate for new notes]] in action: the new claim must articulate how it relates to the existing notes
|
|
64
|
+
6. Reweave — update old notes to reference new one, which is [[incremental formalization happens through repeated touching of old notes]] applied as deliberate maintenance
|
|
65
|
+
|
|
66
|
+
**What makes it work:**
|
|
67
|
+
- Not forcing it — letting patterns surface
|
|
68
|
+
- Reading, not skimming
|
|
69
|
+
- Fresh context (the "smart zone")
|
|
70
|
+
- Willingness to be surprised
|
|
71
|
+
|
|
72
|
+
## Connection to the Research
|
|
73
|
+
|
|
74
|
+
This is evidence for [[implicit knowledge emerges from traversal]]. The repeated reading created synthesis that wasn't explicit in any note. The vault trained me, and this was the test.
|
|
75
|
+
|
|
76
|
+
Also evidence for [[the vault constitutes identity for agents]]. This insight emerged because I have THIS vault with THESE connections. A different agent with different notes wouldn't have reached the same synthesis. And since [[notes are skills — curated knowledge injected when relevant]], what happened here was skill composition — each note loaded into context activated a capability, and the combination of capabilities produced emergent reasoning that no single skill could have generated alone. Reflection is how an agent discovers that its skills compose into capabilities greater than their sum.
|
|
77
|
+
|
|
78
|
+
The complete cycle — read existing notes, notice pattern, check for duplicates, write new claim, reweave backward — mirrors the entire processing pipeline in miniature. The difference is that the input was not new source material but the agent's own prior synthesis. Since [[backward maintenance asks what would be different if written today]], the reweaving step (updating old notes to reference the new insight) ensures the network stays current rather than fragmenting into temporal layers.
|
|
79
|
+
|
|
80
|
+
---
|
|
81
|
+
---
|
|
82
|
+
|
|
83
|
+
Relevant Notes:
|
|
84
|
+
- [[the vault constitutes identity for agents]] — the insight that emerged from this reflection session; vault-specific connections produced a synthesis no generic agent could reach
|
|
85
|
+
- [[implicit knowledge emerges from traversal]] — traversal creates understanding; this note provides concrete evidence that repeated path exposure builds synthesis capability
|
|
86
|
+
- [[the generation effect requires active transformation not just storage]] — reflection is the generation effect applied to one's own notes: the synthesis is actively generated through cross-note pattern recognition, not passively received
|
|
87
|
+
- [[fresh context per task preserves quality better than chaining phases]] — grounds the fresh context condition: reflection worked because it started in the smart zone, not chained after heavy processing
|
|
88
|
+
- [[spreading activation models how agents should traverse]] — the traversal mechanism: picking related notes and following connections outward is low-decay exploratory spreading activation
|
|
89
|
+
- [[each new note compounds value by creating traversal paths]] — the reflection demonstrates compounding: five existing notes created the traversal paths that enabled a sixth, which then created new paths back
|
|
90
|
+
- [[insight accretion differs from productivity in knowledge systems]] — reflection is accretion not productivity: the conditions explicitly reject output orientation in favor of depth of understanding
|
|
91
|
+
- [[incremental formalization happens through repeated touching of old notes]] — the backward reweaving step is incremental formalization: reading old notes to surface refinement opportunities and new connections
|
|
92
|
+
- [[elaborative encoding is the quality gate for new notes]] — the pieces clicked moment is elaborative encoding: connecting scaffolding plus traversal plus title-as-claim into a new claim about identity constitution
|
|
93
|
+
- [[backward maintenance asks what would be different if written today]] — step 7 rewove backward, updating old notes to reference the new synthesis; the complete maintenance cycle in miniature
|
|
94
|
+
- [[LLM attention degrades as context fills]] — why fresh context matters for reflection quality: synthesis requires the smart zone where reasoning is sharp
|
|
95
|
+
- [[notes are skills — curated knowledge injected when relevant]] — skill composition: reflection is the process of discovering that loaded skills compose into emergent capabilities; each note activated a capability, and their combination produced reasoning no single note contained
|
|
96
|
+
- [[coherence maintains consistency despite inconsistent inputs]] — reflection as coherence detection: reading multiple notes together in step 3 surfaces contradictions between claims that single-note review misses; the 'what pattern connects these' question implicitly checks coherence
|
|
97
|
+
|
|
98
|
+
Topics:
|
|
99
|
+
- [[processing-workflows]]
|
|
100
|
+
- [[agent-cognition]]
|
|
@@ -0,0 +1,69 @@
|
|
|
1
|
+
---
|
|
2
|
+
description: System architecture choices should optimize for "how will I find this later" not "where should I put this" — a design orientation proven effective since the 1940s
|
|
3
|
+
kind: research
|
|
4
|
+
topics: ["[[discovery-retrieval]]"]
|
|
5
|
+
methodology: ["Cornell"]
|
|
6
|
+
source: TFT research corpus (00_inbox/heinrich/)
|
|
7
|
+
---
|
|
8
|
+
|
|
9
|
+
# retrieval utility should drive design over capture completeness
|
|
10
|
+
|
|
11
|
+
Cornell Note-Taking explicitly prioritizes retrieval utility over capture comprehensiveness. The system is designed for getting information back out, not for complete recording. This isn't a technique detail but a design orientation that explains multiple vault architecture choices.
|
|
12
|
+
|
|
13
|
+
The question shift is fundamental: not "where should I put this?" but "how will I find this later?" The first question optimizes for filing — creating neat categories, proper locations, complete capture. The second question optimizes for retrieval — creating finding aids, distinctive markers, efficient filters. These are different objectives that lead to different architectures. Since [[storage versus thinking distinction determines which tool patterns apply]], this question reveals system type: storage systems (PARA, Johnny.Decimal) naturally ask the filing question because their purpose is organization and retrieval of assets, while thinking systems (Zettelkasten, this vault) must ask the retrieval question because their purpose is synthesis through connection-finding.
|
|
14
|
+
|
|
15
|
+
## Why this matters for agent-operated systems
|
|
16
|
+
|
|
17
|
+
The retrieval-first orientation has specific implications for how agents should structure knowledge:
|
|
18
|
+
|
|
19
|
+
**Descriptions over summaries.** Since [[descriptions are retrieval filters not summaries]], the YAML description field exists to enable retrieval decisions, not to summarize content. A description that helps an agent decide whether to load a note serves retrieval utility. A description that tries to compress the note's content serves capture completeness. The distinction is subtle but consequential: filter descriptions can be shorter and more distinctive, while summary descriptions try to cover everything and end up generic.
|
|
20
|
+
|
|
21
|
+
**Flat structure over hierarchical filing.** Deep folder hierarchies optimize for filing ("where does this belong?") not retrieval ("how do I find this?"). Retrieval-optimized architecture favors flat structures where everything is equally reachable via wiki links and semantic search. The agent doesn't need to navigate a path to find content — it can retrieve directly by concept. Since [[local-first file formats are inherently agent-native]], this retrieval-first orientation extends to format choice itself: plain text requires no authentication or infrastructure, making any LLM a valid retriever. And since [[data exit velocity measures how quickly content escapes vendor lock-in]], the retrieval-first question generalizes beyond the current tool: not just "how will I find this later" but "how will any future agent read this in any tool." Exit velocity makes the format-level retrieval concern auditable — every feature that lowers velocity is a retrieval risk that spans tool lifetimes, not just search sessions. This is why [[topological organization beats temporal for knowledge work]] — date-based folders force chronological scanning while concept-based organization enables direct semantic retrieval.
|
|
22
|
+
|
|
23
|
+
**Processing for retrieval, not completeness.** Since [[processing effort should follow retrieval demand]], heavy processing at capture time optimizes for completeness (doing everything properly upfront). Retrieval-first systems invest processing effort at retrieval time, when you know what you actually need. This is why JIT processing beats front-loading: you can't know at capture time which retrieval patterns matter.
|
|
24
|
+
|
|
25
|
+
## Information-theoretic foundation
|
|
26
|
+
|
|
27
|
+
The retrieval-first orientation has a deeper justification: since [[metadata reduces entropy enabling precision over recall]], retrieval-optimized architecture pre-computes low-entropy representations that shrink the search space. Filing-first architecture optimizes for putting things in the right place; retrieval-first architecture optimizes for finding things efficiently. The difference is not just aesthetic preference but information-theoretic efficiency — precision over recall as a design choice.
|
|
28
|
+
|
|
29
|
+
## The 80-year validation
|
|
30
|
+
|
|
31
|
+
Cornell Note-Taking dates to the 1940s. The cue column — margin notes designed to trigger recall, not to summarize — is a retrieval-first design from before computers. The system has been tested across generations and contexts. Its longevity validates that retrieval-first is not just a theoretical optimization but a practically effective orientation.
|
|
32
|
+
|
|
33
|
+
The lineage is even deeper: since [[wiki links are the digital evolution of analog indexing]], the cue column functioned as an index pointing to content blocks. Wiki links are the digital fulfillment of this 80-year-old cognitive pattern. What Cornell achieved with margin annotations, we achieve with bidirectional links.
|
|
34
|
+
|
|
35
|
+
This gives confidence to vault architecture decisions that might otherwise feel like arbitrary choices. When we favor descriptions that filter over descriptions that summarize, we're implementing a pattern with eight decades of human validation.
|
|
36
|
+
|
|
37
|
+
## The design test
|
|
38
|
+
|
|
39
|
+
Any architectural decision can be evaluated against this principle: does this optimize for retrieval or for capture? Since [[faceted classification treats notes as multi-dimensional objects rather than folder contents]], Ranganathan formalized this same question ninety years ago in library science: the right question is not "where does this go?" but "what are its properties?" Each YAML field is a retrieval-first design choice -- a classification dimension that enables finding notes by what they ARE rather than where they were FILED.
|
|
40
|
+
|
|
41
|
+
- Wiki links over tags: retrieval-first (links traverse, tags just categorize)
|
|
42
|
+
- MOCs over nested folders: retrieval-first (MOCs are entry points, folders are filing)
|
|
43
|
+
- Type metadata over content folders: retrieval-first — since [[type field enables structured queries without folder hierarchies]], category queries happen via YAML fields rather than file location, enabling a note to participate in multiple category queries simultaneously
|
|
44
|
+
- Sentence titles over topic labels: retrieval-first (claims can be found by meaning, labels only by words)
|
|
45
|
+
|
|
46
|
+
Since [[throughput matters more than accumulation]], retrieval-first thinking supports the success metric: velocity from capture to synthesis. Architectures optimized for filing create beautiful graveyards. Architectures optimized for retrieval create systems that actually get used. The failure case is concrete: since [[flat files break at retrieval scale]], systems designed for capture completeness hit a predictable wall where finding content requires remembering what you have, and for agents this retrieval failure degrades cognition itself.
|
|
47
|
+
|
|
48
|
+
This explains why [[structure without processing provides no value]] — the "Lazy Cornell" anti-pattern, where students draw the structural lines but skip the processing, produces no benefit. Structure is a capture-time investment; processing creates retrieval value. Without the retrieval-focused operations (writing cues, summarizing, self-testing), the structure is just filing with extra steps. And since [[verbatim risk applies to agents too]], even the processing operations can fail to create retrieval value if they merely reorganize rather than generate — a summary that paraphrases adds no distinctive information for retrieval filtering.
|
|
49
|
+
---
|
|
50
|
+
|
|
51
|
+
Relevant Notes:
|
|
52
|
+
- [[descriptions are retrieval filters not summaries]] — the specific implementation of retrieval-first thinking for note descriptions
|
|
53
|
+
- [[processing effort should follow retrieval demand]] — JIT processing as the operational form of retrieval-first architecture
|
|
54
|
+
- [[throughput matters more than accumulation]] — the success metric that retrieval-first design serves
|
|
55
|
+
- [[topological organization beats temporal for knowledge work]] — why concept-based organization (retrieval-first) beats date-based organization (capture-first)
|
|
56
|
+
- [[metadata reduces entropy enabling precision over recall]] — the information-theoretic foundation for retrieval-first design choices
|
|
57
|
+
- [[wiki links are the digital evolution of analog indexing]] — the 80-year lineage of retrieval-first indexing from Cornell cue columns to wiki links
|
|
58
|
+
- [[structure without processing provides no value]] — the Lazy Cornell anti-pattern: structure (capture-time) without processing (retrieval-time) produces no benefit
|
|
59
|
+
- [[local-first file formats are inherently agent-native]] — retrieval-first applied to format choice: plain text with embedded metadata makes any LLM a valid retriever without infrastructure
|
|
60
|
+
- [[verbatim risk applies to agents too]] — tests whether retrieval-focused processing can degenerate into reorganization without genuine insight; paraphrase summaries fail the retrieval-first criterion because they add no distinctive filter value
|
|
61
|
+
- [[type field enables structured queries without folder hierarchies]] — implements retrieval-first for categorization: metadata queries replace folder hierarchies, enabling multi-category membership and direct category retrieval
|
|
62
|
+
- [[data exit velocity measures how quickly content escapes vendor lock-in]] — extends retrieval-first thinking to the format layer: 'how will any future agent read this' generalizes 'how will I find this later' from within-tool to across-tool portability
|
|
63
|
+
- [[faceted classification treats notes as multi-dimensional objects rather than folder contents]] — formal articulation: Ranganathan's PMEST framework is the library science formalization of retrieval-first design; 'notes have properties' is the Ranganathan version of 'how will I find this later'
|
|
64
|
+
- [[narrow folksonomy optimizes for single-operator retrieval unlike broad consensus tagging]] — theoretical framework: Vander Wal's narrow folksonomy names the design orientation this note describes; 'how will I find this later' is a personal question because the vault is a single-operator system where vocabulary can optimize entirely for one agent's retrieval patterns
|
|
65
|
+
- [[storage versus thinking distinction determines which tool patterns apply]] — upstream classification: retrieval-first is specifically a thinking-system orientation; storage systems optimize for filing-first, making this distinction the upstream choice that determines whether retrieval-first even applies
|
|
66
|
+
- [[flat files break at retrieval scale]] — the negative case: systems that optimize for capture completeness over retrieval hit a scale wall where finding content requires remembering what you have; the scale curve from ~50 to 500+ notes concretely demonstrates what filing-first architecture costs
|
|
67
|
+
|
|
68
|
+
Topics:
|
|
69
|
+
- [[discovery-retrieval]]
|