claude-turing 3.4.0 → 4.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (38) hide show
  1. package/.claude-plugin/plugin.json +2 -2
  2. package/README.md +12 -2
  3. package/commands/annotate.md +23 -0
  4. package/commands/archive.md +23 -0
  5. package/commands/changelog.md +22 -0
  6. package/commands/cite.md +23 -0
  7. package/commands/flashback.md +22 -0
  8. package/commands/present.md +23 -0
  9. package/commands/replay.md +23 -0
  10. package/commands/search.md +22 -0
  11. package/commands/template.md +22 -0
  12. package/commands/trend.md +21 -0
  13. package/commands/turing.md +20 -0
  14. package/package.json +1 -1
  15. package/src/install.js +2 -0
  16. package/src/verify.js +10 -0
  17. package/templates/scripts/__pycache__/citation_manager.cpython-314.pyc +0 -0
  18. package/templates/scripts/__pycache__/experiment_annotations.cpython-314.pyc +0 -0
  19. package/templates/scripts/__pycache__/experiment_archive.cpython-314.pyc +0 -0
  20. package/templates/scripts/__pycache__/experiment_replay.cpython-314.pyc +0 -0
  21. package/templates/scripts/__pycache__/experiment_search.cpython-314.pyc +0 -0
  22. package/templates/scripts/__pycache__/experiment_templates.cpython-314.pyc +0 -0
  23. package/templates/scripts/__pycache__/generate_changelog.cpython-314.pyc +0 -0
  24. package/templates/scripts/__pycache__/generate_figures.cpython-314.pyc +0 -0
  25. package/templates/scripts/__pycache__/scaffold.cpython-314.pyc +0 -0
  26. package/templates/scripts/__pycache__/session_flashback.cpython-314.pyc +0 -0
  27. package/templates/scripts/__pycache__/trend_analysis.cpython-314.pyc +0 -0
  28. package/templates/scripts/citation_manager.py +436 -0
  29. package/templates/scripts/experiment_annotations.py +392 -0
  30. package/templates/scripts/experiment_archive.py +534 -0
  31. package/templates/scripts/experiment_replay.py +592 -0
  32. package/templates/scripts/experiment_search.py +451 -0
  33. package/templates/scripts/experiment_templates.py +501 -0
  34. package/templates/scripts/generate_changelog.py +464 -0
  35. package/templates/scripts/generate_figures.py +597 -0
  36. package/templates/scripts/scaffold.py +17 -0
  37. package/templates/scripts/session_flashback.py +461 -0
  38. package/templates/scripts/trend_analysis.py +503 -0
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "turing",
3
- "version": "3.4.0",
4
- "description": "Autonomous ML research harness — the autoresearch loop as a formal protocol. 53 commands, 2 specialized agents, model surgery (pruning + quantization + merging + architecture modification), feature & training intelligence (feature selection + curriculum optimization), model debugging (xray + sensitivity + calibration), pre-training intelligence (sanity checks + baseline generation + leakage detection), meta-intelligence (cross-project knowledge transfer + methodology audit), scaling & efficiency (scaling laws + compute budget + model distillation), model composition (ensemble + pipeline stitch + warm-start), deep analysis (experiment diff + live training monitor + regression gate), experiment orchestration (batch queue + smart retry + branching), literature integration + paper drafting, production model export, performance profiling, smart checkpoints, experiment intelligence, statistical rigor, tree-search hypothesis exploration, cost-performance frontier, model cards, model registry, hypothesis database with novelty guard, anti-cheating guardrails, and the taste-leverage loop. Inspired by Karpathy's autoresearch and the scientific method itself.",
3
+ "version": "4.0.0",
4
+ "description": "Autonomous ML research harness — the autoresearch loop as a formal protocol. 63 commands, 2 specialized agents, research communication (cite + present + changelog), experiment archaeology (trend + flashback + archive + annotate + search + template + replay), model surgery (prune + quantize + merge + surgery), feature & training intelligence, model debugging, pre-training intelligence, meta-intelligence, scaling & efficiency, model composition, deep analysis, experiment orchestration, literature + paper, model export, profiling, checkpoints, experiment intelligence, statistical rigor, tree-search, cost-performance, model cards, hypothesis database, novelty guard, anti-cheating, taste-leverage loop. Inspired by Karpathy's autoresearch and the scientific method itself.",
5
5
  "author": {
6
6
  "name": "pragnition"
7
7
  },
package/README.md CHANGED
@@ -364,6 +364,16 @@ The index (`hypotheses.yaml`) is the lightweight queue. The detail files (`hypot
364
364
  | `/turing:quantize <exp-id>` | Post-training quantization — FP16/INT8, accuracy-latency comparison |
365
365
  | `/turing:merge <exp-ids...>` | Model merging — uniform/greedy soup, TIES, DARE, zero latency cost |
366
366
  | `/turing:surgery <exp-id>` | Architecture modification — add/remove layer, widen/narrow, swap activation |
367
+ | `/turing:trend` | Long-term trend analysis — improvement velocity, family ROI, diminishing returns |
368
+ | `/turing:flashback` | Session context restoration — "where was I?" after days away |
369
+ | `/turing:archive` | Experiment lifecycle cleanup — compress old artifacts, summary index |
370
+ | `/turing:annotate <exp-id>` | Retrospective annotations — human notes and tags on experiments |
371
+ | `/turing:search <query>` | Natural language experiment search — text + structured filters |
372
+ | `/turing:template <action>` | Experiment template library — save/list/apply reusable configs |
373
+ | `/turing:replay <exp-id>` | Experiment replay — re-run old approach with current infrastructure |
374
+ | `/turing:cite <action>` | Citation & attribution manager — track papers, audit missing citations, generate BibTeX |
375
+ | `/turing:present [--figures]` | Presentation figures — training curves, comparisons, ablation, Pareto, sensitivity |
376
+ | `/turing:changelog [--audience]` | Model changelog — version-grouped improvements for technical or stakeholder audiences |
367
377
 
368
378
  And for fully hands-off operation:
369
379
 
@@ -548,11 +558,11 @@ Each project gets independent config, data, experiments, models, and agent memor
548
558
 
549
559
  ## Architecture of Turing Itself
550
560
 
551
- 53 commands, 2 agents, 10 config files, 72 template scripts, model registry, artifact contract, cost-performance frontier, model cards, tree-search exploration, statistical rigor, experiment intelligence, performance profiling, smart checkpoints, production model export, literature integration, paper section drafting, experiment orchestration (queue + retry + fork), deep analysis (diff + watch + regress), model composition (ensemble + stitch + warm), scaling & efficiency (scale + budget + distill), meta-intelligence (transfer + audit), pre-training intelligence (sanity + baseline + leak), model debugging (xray + sensitivity + calibrate), feature & training intelligence (feature + curriculum), model surgery (prune + quantize + merge + surgery), 16 ADRs. See [docs/ARCHITECTURE.md](docs/ARCHITECTURE.md) for the full codemap.
561
+ 63 commands, 2 agents, 10 config files, 82 template scripts, model registry, artifact contract, cost-performance frontier, model cards, tree-search exploration, statistical rigor, experiment intelligence, performance profiling, smart checkpoints, production model export, literature integration, paper section drafting, experiment orchestration (queue + retry + fork), deep analysis (diff + watch + regress), model composition (ensemble + stitch + warm), scaling & efficiency (scale + budget + distill), meta-intelligence (transfer + audit), pre-training intelligence (sanity + baseline + leak), model debugging (xray + sensitivity + calibrate), feature & training intelligence (feature + curriculum), model surgery (prune + quantize + merge + surgery), experiment archaeology (trend + flashback + archive + annotate + search + template + replay), research communication (cite + present + changelog), 16 ADRs. See [docs/ARCHITECTURE.md](docs/ARCHITECTURE.md) for the full codemap.
552
562
 
553
563
  ```
554
564
  turing/
555
- ├── commands/ 52 skill files (core + taste-leverage + reporting + exploration + statistical rigor + experiment intelligence + performance + deployment + research workflow + orchestration + deep analysis + model composition + scaling & efficiency + meta-intelligence + pre-training intelligence + model debugging + feature & training intelligence + model surgery)
565
+ ├── commands/ 62 skill files (core + taste-leverage + reporting + exploration + statistical rigor + experiment intelligence + performance + deployment + research workflow + orchestration + deep analysis + model composition + scaling & efficiency + meta-intelligence + pre-training intelligence + model debugging + feature & training intelligence + model surgery + experiment archaeology + research communication)
556
566
  ├── agents/ 2 agents (researcher: read/write, evaluator: read-only)
557
567
  ├── config/ 8 files (lifecycle, taxonomy, archetypes, novelty aliases)
558
568
  ├── templates/ Scaffolded into user projects by /turing:init
@@ -0,0 +1,23 @@
1
+ ---
2
+ name: annotate
3
+ description: Retrospective experiment annotations — add human notes, tags, and context that automated metrics can't capture.
4
+ disable-model-invocation: true
5
+ argument-hint: "<exp-id> \"note\" [--tag fragile] | --list | --search \"keyword\""
6
+ allowed-tools: Read, Bash(*), Grep, Glob
7
+ ---
8
+
9
+ Add context that experiment logs can't capture. "This only worked because the data was pre-sorted."
10
+
11
+ ## Steps
12
+ 1. **Activate environment:** `source .venv/bin/activate`
13
+ 2. **Run:** `python scripts/experiment_annotations.py $ARGUMENTS`
14
+ 3. **Operations:** add (text + tags), list (per-experiment or all), search (keyword or tag)
15
+ 4. **Stored in:** `experiments/annotations.yaml`
16
+
17
+ ## Examples
18
+ ```
19
+ /turing:annotate exp-042 "Fragile — only works with specific preprocessing"
20
+ /turing:annotate exp-042 "Reviewer 2 requested this" --tag reviewer-requested
21
+ /turing:annotate --list
22
+ /turing:annotate --search "fragile"
23
+ ```
@@ -0,0 +1,23 @@
1
+ ---
2
+ name: archive
3
+ description: Experiment lifecycle cleanup — compress old artifacts, prune checkpoints, create queryable summary index. Reclaim disk space.
4
+ disable-model-invocation: true
5
+ argument-hint: "[--older-than 30d] [--keep-best 10] [--dry-run]"
6
+ allowed-tools: Read, Bash(*), Grep, Glob
7
+ ---
8
+
9
+ Keep your project directory manageable after 200+ experiments.
10
+
11
+ ## Steps
12
+ 1. **Activate environment:** `source .venv/bin/activate`
13
+ 2. **Run:** `python scripts/experiment_archive.py $ARGUMENTS`
14
+ 3. **Protected experiments:** Pareto-optimal, current best, recent, top-N by metric
15
+ 4. **Report:** archived count, preserved count, space reclaimed
16
+ 5. **Saved output:** `experiments/archive/index.yaml`
17
+
18
+ ## Examples
19
+ ```
20
+ /turing:archive --dry-run # Preview what would be archived
21
+ /turing:archive --older-than 30 --keep-best 10 # Archive old, keep top 10
22
+ /turing:archive # Default: 30 days, keep 10
23
+ ```
@@ -0,0 +1,22 @@
1
+ ---
2
+ name: changelog
3
+ description: Model changelog generation — auto-generate human-readable progress narrative from experiment history for stakeholders.
4
+ disable-model-invocation: true
5
+ argument-hint: "[--since exp-id|date] [--audience technical|stakeholder]"
6
+ allowed-tools: Read, Bash(*), Grep, Glob
7
+ ---
8
+
9
+ Translate experiment logs into a narrative that PMs and stakeholders can read in 2 minutes.
10
+
11
+ ## Steps
12
+ 1. **Activate environment:** `source .venv/bin/activate`
13
+ 2. **Run:** `python scripts/generate_changelog.py $ARGUMENTS`
14
+ 3. **Audience:** technical (experiment IDs, configs), stakeholder (plain English, percentages)
15
+ 4. **Saved output:** `paper/CHANGELOG.md`
16
+
17
+ ## Examples
18
+ ```
19
+ /turing:changelog # Full changelog
20
+ /turing:changelog --audience stakeholder # Non-technical summary
21
+ /turing:changelog --since exp-042 # Since specific experiment
22
+ ```
@@ -0,0 +1,23 @@
1
+ ---
2
+ name: cite
3
+ description: Citation & attribution manager — track papers, datasets, methods. Audit for missing citations, generate BibTeX.
4
+ disable-model-invocation: true
5
+ argument-hint: "<add|list|check|bib> [--key Chen2016 --title XGBoost --url ...]"
6
+ allowed-tools: Read, Bash(*), Grep, Glob
7
+ ---
8
+
9
+ Track which papers and methods influenced each experiment. Catch missing citations before submission.
10
+
11
+ ## Steps
12
+ 1. **Activate environment:** `source .venv/bin/activate`
13
+ 2. **Run:** `python scripts/citation_manager.py $ARGUMENTS`
14
+ 3. **Operations:** add (associate citation with experiment), list (group by type), check (audit missing), bib (BibTeX)
15
+ 4. **Stored in:** `experiments/citations.yaml`
16
+
17
+ ## Examples
18
+ ```
19
+ /turing:cite add exp-042 --key Chen2016 --title "XGBoost" --type method --url "https://arxiv.org/abs/1603.02754"
20
+ /turing:cite list
21
+ /turing:cite check # Audit for missing citations
22
+ /turing:cite bib # Generate BibTeX
23
+ ```
@@ -0,0 +1,22 @@
1
+ ---
2
+ name: flashback
3
+ description: Session context restoration — "where was I?" summary after days away. Current best, pending hypotheses, last session, annotations.
4
+ disable-model-invocation: true
5
+ argument-hint: "[--days 7] [--last 10]"
6
+ allowed-tools: Read, Bash(*), Grep, Glob
7
+ ---
8
+
9
+ Come back to a project after a week and start working in 10 seconds instead of 30 minutes.
10
+
11
+ ## Steps
12
+ 1. **Activate environment:** `source .venv/bin/activate`
13
+ 2. **Run:** `python scripts/session_flashback.py $ARGUMENTS`
14
+ 3. **Report:** current best, last session experiments, pending hypotheses, annotations, budget, suggested next action
15
+ 4. **Saved output:** `experiments/flashbacks/flashback-*.yaml`
16
+
17
+ ## Examples
18
+ ```
19
+ /turing:flashback # Default: last 7 days
20
+ /turing:flashback --days 14 # 2-week lookback
21
+ /turing:flashback --last 5 # Last 5 experiments
22
+ ```
@@ -0,0 +1,23 @@
1
+ ---
2
+ name: present
3
+ description: Presentation figure generation — training curves, comparison charts, ablation tables, Pareto plots, sensitivity heatmaps.
4
+ disable-model-invocation: true
5
+ argument-hint: "[--figures training,comparison] [--style light|dark|poster]"
6
+ allowed-tools: Read, Bash(*), Grep, Glob
7
+ ---
8
+
9
+ Generate presentation-ready figure specifications from experiment data in seconds.
10
+
11
+ ## Steps
12
+ 1. **Activate environment:** `source .venv/bin/activate`
13
+ 2. **Run:** `python scripts/generate_figures.py $ARGUMENTS`
14
+ 3. **Figure types:** training, comparison, ablation, pareto, sensitivity
15
+ 4. **Styles:** light (papers), dark (demos), poster (large fonts)
16
+ 5. **Saved output:** `paper/figures/`
17
+
18
+ ## Examples
19
+ ```
20
+ /turing:present # All figures
21
+ /turing:present --figures training,comparison # Specific figures
22
+ /turing:present --style dark # Dark theme
23
+ ```
@@ -0,0 +1,23 @@
1
+ ---
2
+ name: replay
3
+ description: Experiment replay — re-run a historical experiment with current infrastructure to test if old approaches do better now.
4
+ disable-model-invocation: true
5
+ argument-hint: "<exp-id> [--with-current-data] [--with-current-preprocessing]"
6
+ allowed-tools: Read, Bash(*), Grep, Glob
7
+ ---
8
+
9
+ Should you revisit old ideas? Infrastructure changes may make failed approaches work now.
10
+
11
+ ## Steps
12
+ 1. **Activate environment:** `source .venv/bin/activate`
13
+ 2. **Run:** `python scripts/experiment_replay.py $ARGUMENTS`
14
+ 3. **Modes:** default (current code+data), --with-current-data, --with-current-preprocessing
15
+ 4. **Report:** original vs replayed metrics, delta, verdict
16
+ 5. **Saved output:** `experiments/replays/`
17
+
18
+ ## Examples
19
+ ```
20
+ /turing:replay exp-023 # Replay with current infrastructure
21
+ /turing:replay exp-023 --with-current-data # Current data, old code
22
+ /turing:replay --list # List replayable experiments
23
+ ```
@@ -0,0 +1,22 @@
1
+ ---
2
+ name: search
3
+ description: Natural language experiment search — query with text + structured filters over 200+ experiments.
4
+ disable-model-invocation: true
5
+ argument-hint: "<query> [--filter \"accuracy>0.85\"] [--limit 10]"
6
+ allowed-tools: Read, Bash(*), Grep, Glob
7
+ ---
8
+
9
+ Find specific experiments in a large history with natural language and structured filters.
10
+
11
+ ## Steps
12
+ 1. **Activate environment:** `source .venv/bin/activate`
13
+ 2. **Run:** `python scripts/experiment_search.py $ARGUMENTS`
14
+ 3. **Filters:** `accuracy>0.85`, `status:kept`, `family:baseline`, `date:last-week`
15
+ 4. **Report:** ranked table of matching experiments
16
+
17
+ ## Examples
18
+ ```
19
+ /turing:search "LightGBM high accuracy" --filter "accuracy>0.85"
20
+ /turing:search "failed neural net" --filter "status:discarded"
21
+ /turing:search "last week" --limit 5
22
+ ```
@@ -0,0 +1,22 @@
1
+ ---
2
+ name: template
3
+ description: Experiment template library — save winning configs as reusable templates, apply to new projects.
4
+ disable-model-invocation: true
5
+ argument-hint: "<save|list|apply|share> [--name name] [--from exp-id]"
6
+ allowed-tools: Read, Bash(*), Grep, Glob
7
+ ---
8
+
9
+ Turn your best experiment configs into reusable recipes that persist across projects.
10
+
11
+ ## Steps
12
+ 1. **Activate environment:** `source .venv/bin/activate`
13
+ 2. **Run:** `python scripts/experiment_templates.py $ARGUMENTS`
14
+ 3. **Operations:** save (from experiment), list (all templates), apply (to current project), share (export)
15
+ 4. **Stored at:** `~/.turing/templates/` (cross-project)
16
+
17
+ ## Examples
18
+ ```
19
+ /turing:template save --from exp-042 --name "tabular-xgboost-v2"
20
+ /turing:template list
21
+ /turing:template apply tabular-xgboost-v2
22
+ ```
@@ -0,0 +1,21 @@
1
+ ---
2
+ name: trend
3
+ description: Long-term trend analysis — improvement velocity, family ROI, diminishing returns detection, strategic research direction.
4
+ disable-model-invocation: true
5
+ argument-hint: "[--window 30d] [--metric accuracy]"
6
+ allowed-tools: Read, Bash(*), Grep, Glob
7
+ ---
8
+
9
+ See the arc of your research, not just the latest results. Strategic view over 100+ experiments.
10
+
11
+ ## Steps
12
+ 1. **Activate environment:** `source .venv/bin/activate`
13
+ 2. **Run:** `python scripts/trend_analysis.py $ARGUMENTS`
14
+ 3. **Report:** improvement velocity over time windows, family ROI ranking, diminishing returns prediction, phase transitions
15
+ 4. **Saved output:** `experiments/trends/trend-*.yaml`
16
+
17
+ ## Examples
18
+ ```
19
+ /turing:trend # Full trend analysis
20
+ /turing:trend --window 14d # Last 2 weeks
21
+ ```
@@ -62,6 +62,16 @@ You are the Turing ML research router. Detect the user's intent and route to the
62
62
  | "quantize", "quantization", "int8", "fp16", "reduce precision", "faster inference" | `/turing:quantize` | Optimize |
63
63
  | "merge", "model soup", "merge weights", "average models", "TIES", "DARE" | `/turing:merge` | Compose |
64
64
  | "surgery", "architecture", "add layer", "widen", "modify model", "swap activation" | `/turing:surgery` | Modify |
65
+ | "cite", "citation", "bibliography", "bibtex", "attribution", "references" | `/turing:cite` | Record |
66
+ | "present", "figures", "slides", "presentation", "charts", "plots" | `/turing:present` | Document |
67
+ | "changelog", "model changelog", "progress summary", "what improved" | `/turing:changelog` | Document |
68
+ | "trend", "trends", "research direction", "improvement rate", "diminishing returns", "what's working" | `/turing:trend` | Analyze |
69
+ | "flashback", "where was I", "context", "resume", "catch up", "what happened" | `/turing:flashback` | Recall |
70
+ | "archive", "cleanup", "compress old", "disk space", "archive experiments" | `/turing:archive` | Manage |
71
+ | "annotate", "note", "tag experiment", "add note", "experiment note" | `/turing:annotate` | Record |
72
+ | "search", "find experiment", "query experiments", "which experiments" | `/turing:search` | Query |
73
+ | "template", "recipe", "save config", "reusable config", "starting point" | `/turing:template` | Manage |
74
+ | "replay", "re-run", "revisit", "retry old", "would it work now" | `/turing:replay` | Validate |
65
75
 
66
76
  ## Sub-commands
67
77
 
@@ -120,6 +130,16 @@ You are the Turing ML research router. Detect the user's intent and route to the
120
130
  | `/turing:quantize <exp-id>` | Post-training quantization: FP16/INT8, accuracy-latency comparison, QAT suggestion | (inline) |
121
131
  | `/turing:merge <exp-ids...>` | Model merging: uniform/greedy soup, TIES, DARE — free accuracy, zero latency cost | (inline) |
122
132
  | `/turing:surgery <exp-id>` | Architecture modification: add/remove layer, widen/narrow, swap activation, skip connections | (inline) |
133
+ | `/turing:trend` | Long-term trend analysis: improvement velocity, family ROI, diminishing returns detection | (inline) |
134
+ | `/turing:flashback` | Session context restoration: "where was I?" after days away from the project | (inline) |
135
+ | `/turing:archive` | Experiment lifecycle cleanup: compress old artifacts, prune checkpoints, summary index | (inline) |
136
+ | `/turing:annotate <exp-id>` | Retrospective annotations: add human notes, tags, search by content | (inline) |
137
+ | `/turing:search <query>` | Natural language experiment search with structured filters | (inline) |
138
+ | `/turing:template <action>` | Experiment template library: save/list/apply reusable configs across projects | (inline) |
139
+ | `/turing:replay <exp-id>` | Experiment replay: re-run old experiment with current infrastructure | (inline) |
140
+ | `/turing:cite <action>` | Citation manager: add/list/check/bib for papers, datasets, methods | (inline) |
141
+ | `/turing:present [--figures]` | Presentation figures: training curves, comparisons, ablation, Pareto, sensitivity | (inline) |
142
+ | `/turing:changelog [--audience]` | Model changelog: version-grouped improvements for technical or stakeholder audiences | (inline) |
123
143
 
124
144
  ## Proactive Detection
125
145
 
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "claude-turing",
3
- "version": "3.4.0",
3
+ "version": "4.0.0",
4
4
  "type": "module",
5
5
  "description": "Autonomous ML research harness for Claude Code. The autoresearch loop as a formal protocol — iteratively trains, evaluates, and improves ML models with structured experiment tracking, convergence detection, immutable evaluation infrastructure, and safety guardrails.",
6
6
  "bin": {
package/src/install.js CHANGED
@@ -33,6 +33,8 @@ const SUB_COMMANDS = [
33
33
  "xray", "sensitivity", "calibrate",
34
34
  "feature", "curriculum",
35
35
  "prune", "quantize", "merge", "surgery",
36
+ "trend", "flashback", "archive", "annotate", "search", "template", "replay",
37
+ "cite", "present", "changelog",
36
38
  ];
37
39
 
38
40
  export async function install(opts = {}) {
package/src/verify.js CHANGED
@@ -67,6 +67,16 @@ const EXPECTED_COMMANDS = [
67
67
  "quantize/SKILL.md",
68
68
  "merge/SKILL.md",
69
69
  "surgery/SKILL.md",
70
+ "trend/SKILL.md",
71
+ "flashback/SKILL.md",
72
+ "archive/SKILL.md",
73
+ "annotate/SKILL.md",
74
+ "search/SKILL.md",
75
+ "template/SKILL.md",
76
+ "replay/SKILL.md",
77
+ "cite/SKILL.md",
78
+ "present/SKILL.md",
79
+ "changelog/SKILL.md",
70
80
  ];
71
81
 
72
82
  const EXPECTED_AGENTS = ["ml-researcher.md", "ml-evaluator.md"];