@claude-flow/cli 3.7.0-alpha.1 → 3.7.0-alpha.11

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (476) hide show
  1. package/.claude/agents/analysis/analyze-code-quality.md +178 -178
  2. package/.claude/agents/analysis/code-analyzer.md +209 -209
  3. package/.claude/agents/analysis/code-review/analyze-code-quality.md +178 -178
  4. package/.claude/agents/architecture/arch-system-design.md +156 -156
  5. package/.claude/agents/architecture/system-design/arch-system-design.md +154 -154
  6. package/.claude/agents/browser/browser-agent.yaml +182 -182
  7. package/.claude/agents/consensus/byzantine-coordinator.md +62 -62
  8. package/.claude/agents/consensus/crdt-synchronizer.md +996 -996
  9. package/.claude/agents/consensus/gossip-coordinator.md +62 -62
  10. package/.claude/agents/consensus/performance-benchmarker.md +850 -850
  11. package/.claude/agents/consensus/quorum-manager.md +822 -822
  12. package/.claude/agents/consensus/raft-manager.md +62 -62
  13. package/.claude/agents/consensus/security-manager.md +621 -621
  14. package/.claude/agents/core/coder.md +452 -452
  15. package/.claude/agents/core/planner.md +374 -374
  16. package/.claude/agents/core/researcher.md +368 -368
  17. package/.claude/agents/core/reviewer.md +519 -519
  18. package/.claude/agents/core/tester.md +511 -511
  19. package/.claude/agents/custom/test-long-runner.md +44 -44
  20. package/.claude/agents/data/data-ml-model.md +444 -444
  21. package/.claude/agents/data/ml/data-ml-model.md +192 -192
  22. package/.claude/agents/development/backend/dev-backend-api.md +141 -141
  23. package/.claude/agents/development/dev-backend-api.md +344 -344
  24. package/.claude/agents/devops/ci-cd/ops-cicd-github.md +163 -163
  25. package/.claude/agents/devops/ops-cicd-github.md +164 -164
  26. package/.claude/agents/documentation/api-docs/docs-api-openapi.md +173 -173
  27. package/.claude/agents/documentation/docs-api-openapi.md +354 -354
  28. package/.claude/agents/flow-nexus/app-store.md +87 -87
  29. package/.claude/agents/flow-nexus/authentication.md +68 -68
  30. package/.claude/agents/flow-nexus/challenges.md +80 -80
  31. package/.claude/agents/flow-nexus/neural-network.md +87 -87
  32. package/.claude/agents/flow-nexus/payments.md +82 -82
  33. package/.claude/agents/flow-nexus/sandbox.md +75 -75
  34. package/.claude/agents/flow-nexus/swarm.md +75 -75
  35. package/.claude/agents/flow-nexus/user-tools.md +95 -95
  36. package/.claude/agents/flow-nexus/workflow.md +83 -83
  37. package/.claude/agents/github/code-review-swarm.md +377 -377
  38. package/.claude/agents/github/github-modes.md +172 -172
  39. package/.claude/agents/github/issue-tracker.md +575 -575
  40. package/.claude/agents/github/multi-repo-swarm.md +552 -552
  41. package/.claude/agents/github/pr-manager.md +437 -437
  42. package/.claude/agents/github/project-board-sync.md +508 -508
  43. package/.claude/agents/github/release-manager.md +604 -604
  44. package/.claude/agents/github/release-swarm.md +582 -582
  45. package/.claude/agents/github/repo-architect.md +397 -397
  46. package/.claude/agents/github/swarm-issue.md +572 -572
  47. package/.claude/agents/github/swarm-pr.md +427 -427
  48. package/.claude/agents/github/sync-coordinator.md +451 -451
  49. package/.claude/agents/github/workflow-automation.md +902 -902
  50. package/.claude/agents/goal/agent.md +815 -815
  51. package/.claude/agents/goal/goal-planner.md +72 -72
  52. package/.claude/agents/optimization/benchmark-suite.md +664 -664
  53. package/.claude/agents/optimization/load-balancer.md +430 -430
  54. package/.claude/agents/optimization/performance-monitor.md +671 -671
  55. package/.claude/agents/optimization/resource-allocator.md +673 -673
  56. package/.claude/agents/optimization/topology-optimizer.md +807 -807
  57. package/.claude/agents/payments/agentic-payments.md +126 -126
  58. package/.claude/agents/sona/sona-learning-optimizer.md +74 -74
  59. package/.claude/agents/sparc/architecture.md +698 -698
  60. package/.claude/agents/sparc/pseudocode.md +519 -519
  61. package/.claude/agents/sparc/refinement.md +801 -801
  62. package/.claude/agents/sparc/specification.md +477 -477
  63. package/.claude/agents/specialized/mobile/spec-mobile-react-native.md +224 -224
  64. package/.claude/agents/specialized/spec-mobile-react-native.md +226 -226
  65. package/.claude/agents/sublinear/consensus-coordinator.md +337 -337
  66. package/.claude/agents/sublinear/matrix-optimizer.md +184 -184
  67. package/.claude/agents/sublinear/pagerank-analyzer.md +298 -298
  68. package/.claude/agents/sublinear/performance-optimizer.md +367 -367
  69. package/.claude/agents/sublinear/trading-predictor.md +245 -245
  70. package/.claude/agents/swarm/adaptive-coordinator.md +1126 -1126
  71. package/.claude/agents/swarm/hierarchical-coordinator.md +709 -709
  72. package/.claude/agents/swarm/mesh-coordinator.md +962 -962
  73. package/.claude/agents/templates/automation-smart-agent.md +204 -204
  74. package/.claude/agents/templates/base-template-generator.md +289 -289
  75. package/.claude/agents/templates/coordinator-swarm-init.md +89 -89
  76. package/.claude/agents/templates/github-pr-manager.md +176 -176
  77. package/.claude/agents/templates/implementer-sparc-coder.md +258 -258
  78. package/.claude/agents/templates/memory-coordinator.md +186 -186
  79. package/.claude/agents/templates/orchestrator-task.md +138 -138
  80. package/.claude/agents/templates/performance-analyzer.md +198 -198
  81. package/.claude/agents/templates/sparc-coordinator.md +513 -513
  82. package/.claude/agents/testing/production-validator.md +394 -394
  83. package/.claude/agents/testing/tdd-london-swarm.md +243 -243
  84. package/.claude/agents/v3/adr-architect.md +184 -184
  85. package/.claude/agents/v3/aidefence-guardian.md +282 -282
  86. package/.claude/agents/v3/claims-authorizer.md +208 -208
  87. package/.claude/agents/v3/collective-intelligence-coordinator.md +993 -993
  88. package/.claude/agents/v3/ddd-domain-expert.md +220 -220
  89. package/.claude/agents/v3/injection-analyst.md +236 -236
  90. package/.claude/agents/v3/memory-specialist.md +995 -995
  91. package/.claude/agents/v3/performance-engineer.md +1233 -1233
  92. package/.claude/agents/v3/pii-detector.md +151 -151
  93. package/.claude/agents/v3/reasoningbank-learner.md +213 -213
  94. package/.claude/agents/v3/security-architect-aidefence.md +410 -410
  95. package/.claude/agents/v3/security-architect.md +867 -867
  96. package/.claude/agents/v3/security-auditor.md +771 -771
  97. package/.claude/agents/v3/sparc-orchestrator.md +182 -182
  98. package/.claude/agents/v3/swarm-memory-manager.md +157 -157
  99. package/.claude/agents/v3/v3-integration-architect.md +205 -205
  100. package/.claude/commands/agents/README.md +50 -50
  101. package/.claude/commands/agents/agent-capabilities.md +140 -140
  102. package/.claude/commands/agents/agent-coordination.md +28 -28
  103. package/.claude/commands/agents/agent-spawning.md +28 -28
  104. package/.claude/commands/agents/agent-types.md +216 -216
  105. package/.claude/commands/agents/health.md +139 -139
  106. package/.claude/commands/agents/list.md +100 -100
  107. package/.claude/commands/agents/logs.md +130 -130
  108. package/.claude/commands/agents/metrics.md +122 -122
  109. package/.claude/commands/agents/pool.md +127 -127
  110. package/.claude/commands/agents/spawn.md +140 -140
  111. package/.claude/commands/agents/status.md +115 -115
  112. package/.claude/commands/agents/stop.md +102 -102
  113. package/.claude/commands/analysis/COMMAND_COMPLIANCE_REPORT.md +53 -53
  114. package/.claude/commands/analysis/README.md +9 -9
  115. package/.claude/commands/analysis/bottleneck-detect.md +162 -162
  116. package/.claude/commands/analysis/performance-bottlenecks.md +58 -58
  117. package/.claude/commands/analysis/performance-report.md +25 -25
  118. package/.claude/commands/analysis/token-efficiency.md +44 -44
  119. package/.claude/commands/analysis/token-usage.md +25 -25
  120. package/.claude/commands/automation/README.md +9 -9
  121. package/.claude/commands/automation/auto-agent.md +122 -122
  122. package/.claude/commands/automation/self-healing.md +105 -105
  123. package/.claude/commands/automation/session-memory.md +89 -89
  124. package/.claude/commands/automation/smart-agents.md +72 -72
  125. package/.claude/commands/automation/smart-spawn.md +25 -25
  126. package/.claude/commands/automation/workflow-select.md +25 -25
  127. package/.claude/commands/claude-flow-help.md +103 -103
  128. package/.claude/commands/claude-flow-memory.md +107 -107
  129. package/.claude/commands/claude-flow-swarm.md +205 -205
  130. package/.claude/commands/coordination/README.md +9 -9
  131. package/.claude/commands/coordination/agent-spawn.md +25 -25
  132. package/.claude/commands/coordination/init.md +44 -44
  133. package/.claude/commands/coordination/orchestrate.md +43 -43
  134. package/.claude/commands/coordination/spawn.md +45 -45
  135. package/.claude/commands/coordination/swarm-init.md +85 -85
  136. package/.claude/commands/coordination/task-orchestrate.md +25 -25
  137. package/.claude/commands/flow-nexus/app-store.md +123 -123
  138. package/.claude/commands/flow-nexus/challenges.md +119 -119
  139. package/.claude/commands/flow-nexus/login-registration.md +64 -64
  140. package/.claude/commands/flow-nexus/neural-network.md +133 -133
  141. package/.claude/commands/flow-nexus/payments.md +115 -115
  142. package/.claude/commands/flow-nexus/sandbox.md +82 -82
  143. package/.claude/commands/flow-nexus/swarm.md +86 -86
  144. package/.claude/commands/flow-nexus/user-tools.md +151 -151
  145. package/.claude/commands/flow-nexus/workflow.md +114 -114
  146. package/.claude/commands/github/README.md +11 -11
  147. package/.claude/commands/github/code-review-swarm.md +513 -513
  148. package/.claude/commands/github/code-review.md +25 -25
  149. package/.claude/commands/github/github-modes.md +146 -146
  150. package/.claude/commands/github/github-swarm.md +121 -121
  151. package/.claude/commands/github/issue-tracker.md +291 -291
  152. package/.claude/commands/github/issue-triage.md +25 -25
  153. package/.claude/commands/github/multi-repo-swarm.md +518 -518
  154. package/.claude/commands/github/pr-enhance.md +26 -26
  155. package/.claude/commands/github/pr-manager.md +169 -169
  156. package/.claude/commands/github/project-board-sync.md +470 -470
  157. package/.claude/commands/github/release-manager.md +337 -337
  158. package/.claude/commands/github/release-swarm.md +543 -543
  159. package/.claude/commands/github/repo-analyze.md +25 -25
  160. package/.claude/commands/github/repo-architect.md +366 -366
  161. package/.claude/commands/github/swarm-issue.md +481 -481
  162. package/.claude/commands/github/swarm-pr.md +284 -284
  163. package/.claude/commands/github/sync-coordinator.md +300 -300
  164. package/.claude/commands/github/workflow-automation.md +441 -441
  165. package/.claude/commands/hive-mind/README.md +17 -17
  166. package/.claude/commands/hive-mind/hive-mind-consensus.md +8 -8
  167. package/.claude/commands/hive-mind/hive-mind-init.md +18 -18
  168. package/.claude/commands/hive-mind/hive-mind-memory.md +8 -8
  169. package/.claude/commands/hive-mind/hive-mind-metrics.md +8 -8
  170. package/.claude/commands/hive-mind/hive-mind-resume.md +8 -8
  171. package/.claude/commands/hive-mind/hive-mind-sessions.md +8 -8
  172. package/.claude/commands/hive-mind/hive-mind-spawn.md +21 -21
  173. package/.claude/commands/hive-mind/hive-mind-status.md +8 -8
  174. package/.claude/commands/hive-mind/hive-mind-stop.md +8 -8
  175. package/.claude/commands/hive-mind/hive-mind-wizard.md +8 -8
  176. package/.claude/commands/hive-mind/hive-mind.md +27 -27
  177. package/.claude/commands/hooks/README.md +11 -11
  178. package/.claude/commands/hooks/overview.md +57 -57
  179. package/.claude/commands/hooks/post-edit.md +117 -117
  180. package/.claude/commands/hooks/post-task.md +112 -112
  181. package/.claude/commands/hooks/pre-edit.md +113 -113
  182. package/.claude/commands/hooks/pre-task.md +111 -111
  183. package/.claude/commands/hooks/session-end.md +118 -118
  184. package/.claude/commands/hooks/setup.md +102 -102
  185. package/.claude/commands/memory/README.md +9 -9
  186. package/.claude/commands/memory/memory-persist.md +25 -25
  187. package/.claude/commands/memory/memory-search.md +25 -25
  188. package/.claude/commands/memory/memory-usage.md +25 -25
  189. package/.claude/commands/memory/neural.md +47 -47
  190. package/.claude/commands/monitoring/README.md +9 -9
  191. package/.claude/commands/monitoring/agent-metrics.md +25 -25
  192. package/.claude/commands/monitoring/agents.md +44 -44
  193. package/.claude/commands/monitoring/real-time-view.md +25 -25
  194. package/.claude/commands/monitoring/status.md +46 -46
  195. package/.claude/commands/monitoring/swarm-monitor.md +25 -25
  196. package/.claude/commands/optimization/README.md +9 -9
  197. package/.claude/commands/optimization/auto-topology.md +61 -61
  198. package/.claude/commands/optimization/cache-manage.md +25 -25
  199. package/.claude/commands/optimization/parallel-execute.md +25 -25
  200. package/.claude/commands/optimization/parallel-execution.md +49 -49
  201. package/.claude/commands/optimization/topology-optimize.md +25 -25
  202. package/.claude/commands/pair/README.md +260 -260
  203. package/.claude/commands/pair/commands.md +545 -545
  204. package/.claude/commands/pair/config.md +509 -509
  205. package/.claude/commands/pair/examples.md +511 -511
  206. package/.claude/commands/pair/modes.md +347 -347
  207. package/.claude/commands/pair/session.md +406 -406
  208. package/.claude/commands/pair/start.md +208 -208
  209. package/.claude/commands/sparc/analyzer.md +51 -51
  210. package/.claude/commands/sparc/architect.md +53 -53
  211. package/.claude/commands/sparc/ask.md +97 -97
  212. package/.claude/commands/sparc/batch-executor.md +54 -54
  213. package/.claude/commands/sparc/code.md +89 -89
  214. package/.claude/commands/sparc/coder.md +54 -54
  215. package/.claude/commands/sparc/debug.md +83 -83
  216. package/.claude/commands/sparc/debugger.md +54 -54
  217. package/.claude/commands/sparc/designer.md +53 -53
  218. package/.claude/commands/sparc/devops.md +109 -109
  219. package/.claude/commands/sparc/docs-writer.md +80 -80
  220. package/.claude/commands/sparc/documenter.md +54 -54
  221. package/.claude/commands/sparc/innovator.md +54 -54
  222. package/.claude/commands/sparc/integration.md +83 -83
  223. package/.claude/commands/sparc/mcp.md +117 -117
  224. package/.claude/commands/sparc/memory-manager.md +54 -54
  225. package/.claude/commands/sparc/optimizer.md +54 -54
  226. package/.claude/commands/sparc/orchestrator.md +131 -131
  227. package/.claude/commands/sparc/post-deployment-monitoring-mode.md +83 -83
  228. package/.claude/commands/sparc/refinement-optimization-mode.md +83 -83
  229. package/.claude/commands/sparc/researcher.md +54 -54
  230. package/.claude/commands/sparc/reviewer.md +54 -54
  231. package/.claude/commands/sparc/security-review.md +80 -80
  232. package/.claude/commands/sparc/sparc-modes.md +174 -174
  233. package/.claude/commands/sparc/sparc.md +111 -111
  234. package/.claude/commands/sparc/spec-pseudocode.md +80 -80
  235. package/.claude/commands/sparc/supabase-admin.md +348 -348
  236. package/.claude/commands/sparc/swarm-coordinator.md +54 -54
  237. package/.claude/commands/sparc/tdd.md +54 -54
  238. package/.claude/commands/sparc/tester.md +54 -54
  239. package/.claude/commands/sparc/tutorial.md +79 -79
  240. package/.claude/commands/sparc/workflow-manager.md +54 -54
  241. package/.claude/commands/sparc.md +166 -166
  242. package/.claude/commands/stream-chain/pipeline.md +120 -120
  243. package/.claude/commands/stream-chain/run.md +69 -69
  244. package/.claude/commands/swarm/README.md +15 -15
  245. package/.claude/commands/swarm/analysis.md +95 -95
  246. package/.claude/commands/swarm/development.md +96 -96
  247. package/.claude/commands/swarm/examples.md +168 -168
  248. package/.claude/commands/swarm/maintenance.md +102 -102
  249. package/.claude/commands/swarm/optimization.md +117 -117
  250. package/.claude/commands/swarm/research.md +136 -136
  251. package/.claude/commands/swarm/swarm-analysis.md +8 -8
  252. package/.claude/commands/swarm/swarm-background.md +8 -8
  253. package/.claude/commands/swarm/swarm-init.md +19 -19
  254. package/.claude/commands/swarm/swarm-modes.md +8 -8
  255. package/.claude/commands/swarm/swarm-monitor.md +8 -8
  256. package/.claude/commands/swarm/swarm-spawn.md +19 -19
  257. package/.claude/commands/swarm/swarm-status.md +8 -8
  258. package/.claude/commands/swarm/swarm-strategies.md +8 -8
  259. package/.claude/commands/swarm/swarm.md +87 -87
  260. package/.claude/commands/swarm/testing.md +131 -131
  261. package/.claude/commands/training/README.md +9 -9
  262. package/.claude/commands/training/model-update.md +25 -25
  263. package/.claude/commands/training/neural-patterns.md +107 -107
  264. package/.claude/commands/training/neural-train.md +75 -75
  265. package/.claude/commands/training/pattern-learn.md +25 -25
  266. package/.claude/commands/training/specialization.md +62 -62
  267. package/.claude/commands/truth/start.md +142 -142
  268. package/.claude/commands/verify/check.md +49 -49
  269. package/.claude/commands/verify/start.md +127 -127
  270. package/.claude/commands/workflows/README.md +9 -9
  271. package/.claude/commands/workflows/development.md +77 -77
  272. package/.claude/commands/workflows/research.md +62 -62
  273. package/.claude/commands/workflows/workflow-create.md +25 -25
  274. package/.claude/commands/workflows/workflow-execute.md +25 -25
  275. package/.claude/commands/workflows/workflow-export.md +25 -25
  276. package/.claude/helpers/README.md +96 -96
  277. package/.claude/helpers/adr-compliance.sh +186 -186
  278. package/.claude/helpers/auto-commit.sh +178 -178
  279. package/.claude/helpers/auto-memory-hook.mjs +368 -368
  280. package/.claude/helpers/checkpoint-manager.sh +251 -251
  281. package/.claude/helpers/daemon-manager.sh +252 -252
  282. package/.claude/helpers/ddd-tracker.sh +144 -144
  283. package/.claude/helpers/github-safe.js +121 -121
  284. package/.claude/helpers/github-setup.sh +28 -28
  285. package/.claude/helpers/guidance-hook.sh +13 -13
  286. package/.claude/helpers/guidance-hooks.sh +102 -102
  287. package/.claude/helpers/health-monitor.sh +108 -108
  288. package/.claude/helpers/hook-handler.cjs +278 -278
  289. package/.claude/helpers/intelligence.cjs +1031 -1031
  290. package/.claude/helpers/learning-hooks.sh +329 -329
  291. package/.claude/helpers/learning-optimizer.sh +127 -127
  292. package/.claude/helpers/learning-service.mjs +1144 -1144
  293. package/.claude/helpers/memory.js +83 -83
  294. package/.claude/helpers/metrics-db.mjs +488 -488
  295. package/.claude/helpers/pattern-consolidator.sh +86 -86
  296. package/.claude/helpers/perf-worker.sh +160 -160
  297. package/.claude/helpers/post-commit +16 -16
  298. package/.claude/helpers/pre-commit +26 -26
  299. package/.claude/helpers/quick-start.sh +19 -19
  300. package/.claude/helpers/router.js +66 -66
  301. package/.claude/helpers/security-scanner.sh +127 -127
  302. package/.claude/helpers/session.js +135 -135
  303. package/.claude/helpers/setup-mcp.sh +18 -18
  304. package/.claude/helpers/standard-checkpoint-hooks.sh +189 -189
  305. package/.claude/helpers/statusline-hook.sh +21 -21
  306. package/.claude/helpers/statusline.cjs +575 -575
  307. package/.claude/helpers/statusline.js +321 -321
  308. package/.claude/helpers/swarm-comms.sh +353 -353
  309. package/.claude/helpers/swarm-hooks.sh +761 -761
  310. package/.claude/helpers/swarm-monitor.sh +210 -210
  311. package/.claude/helpers/sync-v3-metrics.sh +245 -245
  312. package/.claude/helpers/update-v3-progress.sh +165 -165
  313. package/.claude/helpers/v3-quick-status.sh +57 -57
  314. package/.claude/helpers/v3.sh +110 -110
  315. package/.claude/helpers/validate-v3-config.sh +215 -215
  316. package/.claude/helpers/worker-manager.sh +170 -170
  317. package/.claude/settings.json +182 -182
  318. package/.claude/skills/agentdb-advanced/SKILL.md +550 -550
  319. package/.claude/skills/agentdb-learning/SKILL.md +545 -545
  320. package/.claude/skills/agentdb-memory-patterns/SKILL.md +339 -339
  321. package/.claude/skills/agentdb-optimization/SKILL.md +509 -509
  322. package/.claude/skills/agentdb-vector-search/SKILL.md +339 -339
  323. package/.claude/skills/agentic-jujutsu/SKILL.md +645 -645
  324. package/.claude/skills/aidefence-scan.md +151 -151
  325. package/.claude/skills/aidefence.yaml +297 -297
  326. package/.claude/skills/browser/SKILL.md +204 -204
  327. package/.claude/skills/flow-nexus-neural/SKILL.md +738 -738
  328. package/.claude/skills/flow-nexus-platform/SKILL.md +1157 -1157
  329. package/.claude/skills/flow-nexus-swarm/SKILL.md +610 -610
  330. package/.claude/skills/github-code-review/SKILL.md +1140 -1140
  331. package/.claude/skills/github-multi-repo/SKILL.md +874 -874
  332. package/.claude/skills/github-project-management/SKILL.md +1290 -1277
  333. package/.claude/skills/github-release-management/SKILL.md +1081 -1081
  334. package/.claude/skills/github-workflow-automation/SKILL.md +1065 -1065
  335. package/.claude/skills/hive-mind-advanced/SKILL.md +712 -712
  336. package/.claude/skills/hooks-automation/SKILL.md +1201 -1201
  337. package/.claude/skills/pair-programming/SKILL.md +1202 -1202
  338. package/.claude/skills/performance-analysis/SKILL.md +563 -563
  339. package/.claude/skills/reasoningbank-agentdb/SKILL.md +446 -446
  340. package/.claude/skills/reasoningbank-intelligence/SKILL.md +201 -201
  341. package/.claude/skills/secure-review.md +181 -181
  342. package/.claude/skills/skill-builder/SKILL.md +910 -910
  343. package/.claude/skills/sparc-methodology/SKILL.md +1115 -1115
  344. package/.claude/skills/stream-chain/SKILL.md +563 -563
  345. package/.claude/skills/swarm-advanced/SKILL.md +973 -973
  346. package/.claude/skills/swarm-orchestration/SKILL.md +179 -179
  347. package/.claude/skills/v3-cli-modernization/SKILL.md +871 -871
  348. package/.claude/skills/v3-core-implementation/SKILL.md +796 -796
  349. package/.claude/skills/v3-ddd-architecture/SKILL.md +441 -441
  350. package/.claude/skills/v3-integration-deep/SKILL.md +240 -240
  351. package/.claude/skills/v3-mcp-optimization/SKILL.md +776 -776
  352. package/.claude/skills/v3-memory-unification/SKILL.md +173 -173
  353. package/.claude/skills/v3-performance-optimization/SKILL.md +389 -389
  354. package/.claude/skills/v3-security-overhaul/SKILL.md +81 -81
  355. package/.claude/skills/v3-swarm-coordination/SKILL.md +339 -339
  356. package/.claude/skills/verification-quality/SKILL.md +649 -649
  357. package/.claude/skills/worker-benchmarks/skill.md +135 -135
  358. package/.claude/skills/worker-integration/skill.md +154 -154
  359. package/README.md +393 -391
  360. package/bin/cli.js +220 -220
  361. package/bin/mcp-server.js +224 -224
  362. package/bin/preinstall.cjs +2 -2
  363. package/dist/src/commands/agent-wasm.js +2 -2
  364. package/dist/src/commands/agent-wasm.js.map +1 -1
  365. package/dist/src/commands/completions.js +409 -409
  366. package/dist/src/commands/daemon.d.ts.map +1 -1
  367. package/dist/src/commands/daemon.js +19 -3
  368. package/dist/src/commands/daemon.js.map +1 -1
  369. package/dist/src/commands/doctor.d.ts.map +1 -1
  370. package/dist/src/commands/doctor.js +105 -23
  371. package/dist/src/commands/doctor.js.map +1 -1
  372. package/dist/src/commands/embeddings.js +26 -26
  373. package/dist/src/commands/hive-mind.d.ts.map +1 -1
  374. package/dist/src/commands/hive-mind.js +122 -104
  375. package/dist/src/commands/hive-mind.js.map +1 -1
  376. package/dist/src/commands/hooks.d.ts.map +1 -1
  377. package/dist/src/commands/hooks.js +34 -21
  378. package/dist/src/commands/hooks.js.map +1 -1
  379. package/dist/src/commands/memory.d.ts.map +1 -1
  380. package/dist/src/commands/memory.js +68 -0
  381. package/dist/src/commands/memory.js.map +1 -1
  382. package/dist/src/commands/ruvector/backup.js +23 -23
  383. package/dist/src/commands/ruvector/benchmark.js +31 -31
  384. package/dist/src/commands/ruvector/import.js +14 -14
  385. package/dist/src/commands/ruvector/init.js +115 -115
  386. package/dist/src/commands/ruvector/migrate.js +99 -99
  387. package/dist/src/commands/ruvector/optimize.js +51 -51
  388. package/dist/src/commands/ruvector/setup.js +624 -624
  389. package/dist/src/commands/ruvector/status.js +38 -38
  390. package/dist/src/index.d.ts +5 -1
  391. package/dist/src/index.d.ts.map +1 -1
  392. package/dist/src/index.js +59 -18
  393. package/dist/src/index.js.map +1 -1
  394. package/dist/src/init/claudemd-generator.js +226 -226
  395. package/dist/src/init/executor.d.ts.map +1 -1
  396. package/dist/src/init/executor.js +511 -453
  397. package/dist/src/init/executor.js.map +1 -1
  398. package/dist/src/init/helpers-generator.js +645 -645
  399. package/dist/src/init/settings-generator.d.ts.map +1 -1
  400. package/dist/src/init/settings-generator.js +11 -5
  401. package/dist/src/init/settings-generator.js.map +1 -1
  402. package/dist/src/init/statusline-generator.js +858 -858
  403. package/dist/src/init/types.d.ts +7 -0
  404. package/dist/src/init/types.d.ts.map +1 -1
  405. package/dist/src/init/types.js.map +1 -1
  406. package/dist/src/mcp-tools/agentdb-tools.d.ts +3 -0
  407. package/dist/src/mcp-tools/agentdb-tools.d.ts.map +1 -1
  408. package/dist/src/mcp-tools/agentdb-tools.js +108 -0
  409. package/dist/src/mcp-tools/agentdb-tools.js.map +1 -1
  410. package/dist/src/mcp-tools/hooks-tools.d.ts.map +1 -1
  411. package/dist/src/mcp-tools/hooks-tools.js +4 -2
  412. package/dist/src/mcp-tools/hooks-tools.js.map +1 -1
  413. package/dist/src/mcp-tools/memory-tools.d.ts.map +1 -1
  414. package/dist/src/mcp-tools/memory-tools.js +19 -0
  415. package/dist/src/mcp-tools/memory-tools.js.map +1 -1
  416. package/dist/src/mcp-tools/neural-tools.d.ts.map +1 -1
  417. package/dist/src/mcp-tools/neural-tools.js +14 -1
  418. package/dist/src/mcp-tools/neural-tools.js.map +1 -1
  419. package/dist/src/mcp-tools/security-tools.d.ts.map +1 -1
  420. package/dist/src/mcp-tools/security-tools.js +28 -3
  421. package/dist/src/mcp-tools/security-tools.js.map +1 -1
  422. package/dist/src/mcp-tools/swarm-tools.d.ts.map +1 -1
  423. package/dist/src/mcp-tools/swarm-tools.js +72 -3
  424. package/dist/src/mcp-tools/swarm-tools.js.map +1 -1
  425. package/dist/src/mcp-tools/wasm-agent-tools.js +1 -1
  426. package/dist/src/mcp-tools/wasm-agent-tools.js.map +1 -1
  427. package/dist/src/memory/intelligence.d.ts.map +1 -1
  428. package/dist/src/memory/intelligence.js +28 -3
  429. package/dist/src/memory/intelligence.js.map +1 -1
  430. package/dist/src/memory/memory-bridge.d.ts +69 -0
  431. package/dist/src/memory/memory-bridge.d.ts.map +1 -1
  432. package/dist/src/memory/memory-bridge.js +319 -66
  433. package/dist/src/memory/memory-bridge.js.map +1 -1
  434. package/dist/src/memory/memory-initializer.d.ts +5 -0
  435. package/dist/src/memory/memory-initializer.d.ts.map +1 -1
  436. package/dist/src/memory/memory-initializer.js +369 -363
  437. package/dist/src/memory/memory-initializer.js.map +1 -1
  438. package/dist/src/memory/neural-package-bridge.d.ts +48 -0
  439. package/dist/src/memory/neural-package-bridge.d.ts.map +1 -0
  440. package/dist/src/memory/neural-package-bridge.js +87 -0
  441. package/dist/src/memory/neural-package-bridge.js.map +1 -0
  442. package/dist/src/memory/rabitq-index.js +5 -5
  443. package/dist/src/memory/sona-optimizer.d.ts.map +1 -1
  444. package/dist/src/memory/sona-optimizer.js +1 -0
  445. package/dist/src/memory/sona-optimizer.js.map +1 -1
  446. package/dist/src/parser.d.ts +9 -0
  447. package/dist/src/parser.d.ts.map +1 -1
  448. package/dist/src/parser.js +11 -0
  449. package/dist/src/parser.js.map +1 -1
  450. package/dist/src/runtime/headless.js +28 -28
  451. package/dist/src/ruvector/agent-wasm.d.ts.map +1 -1
  452. package/dist/src/ruvector/agent-wasm.js +4 -1
  453. package/dist/src/ruvector/agent-wasm.js.map +1 -1
  454. package/dist/src/ruvector/index.d.ts +0 -2
  455. package/dist/src/ruvector/index.d.ts.map +1 -1
  456. package/dist/src/ruvector/index.js +8 -2
  457. package/dist/src/ruvector/index.js.map +1 -1
  458. package/dist/src/ruvector/model-router.d.ts +22 -1
  459. package/dist/src/ruvector/model-router.d.ts.map +1 -1
  460. package/dist/src/ruvector/model-router.js +125 -5
  461. package/dist/src/ruvector/model-router.js.map +1 -1
  462. package/dist/src/services/headless-worker-executor.js +84 -84
  463. package/dist/src/transfer/deploy-seraphine.js +23 -23
  464. package/dist/tsconfig.tsbuildinfo +1 -1
  465. package/package.json +5 -4
  466. package/scripts/deploy-ipfs-node.sh +153 -153
  467. package/scripts/postinstall.cjs +153 -153
  468. package/scripts/publish-registry.ts +345 -345
  469. package/scripts/publish.sh +57 -57
  470. package/scripts/setup-ipfs-registry.md +366 -366
  471. package/dist/src/services/event-stream.d.ts.map +0 -1
  472. package/dist/src/services/event-stream.js.map +0 -1
  473. package/dist/src/services/loop-worker-runner.d.ts.map +0 -1
  474. package/dist/src/services/loop-worker-runner.js.map +0 -1
  475. package/dist/src/services/runtime-capabilities.d.ts.map +0 -1
  476. package/dist/src/services/runtime-capabilities.js.map +0 -1
@@ -1,545 +1,545 @@
1
- ---
2
- name: "AgentDB Learning Plugins"
3
- description: "Create and train AI learning plugins with AgentDB's 9 reinforcement learning algorithms. Includes Decision Transformer, Q-Learning, SARSA, Actor-Critic, and more. Use when building self-learning agents, implementing RL, or optimizing agent behavior through experience."
4
- ---
5
-
6
- # AgentDB Learning Plugins
7
-
8
- ## What This Skill Does
9
-
10
- Provides access to 9 reinforcement learning algorithms via AgentDB's plugin system. Create, train, and deploy learning plugins for autonomous agents that improve through experience. Includes offline RL (Decision Transformer), value-based learning (Q-Learning), policy gradients (Actor-Critic), and advanced techniques.
11
-
12
- **Performance**: Train models 10-100x faster with WASM-accelerated neural inference.
13
-
14
- ## Prerequisites
15
-
16
- - Node.js 18+
17
- - AgentDB v1.0.7+ (via agentic-flow)
18
- - Basic understanding of reinforcement learning (recommended)
19
-
20
- ---
21
-
22
- ## Quick Start with CLI
23
-
24
- ### Create Learning Plugin
25
-
26
- ```bash
27
- # Interactive wizard
28
- npx agentdb@latest create-plugin
29
-
30
- # Use specific template
31
- npx agentdb@latest create-plugin -t decision-transformer -n my-agent
32
-
33
- # Preview without creating
34
- npx agentdb@latest create-plugin -t q-learning --dry-run
35
-
36
- # Custom output directory
37
- npx agentdb@latest create-plugin -t actor-critic -o ./plugins
38
- ```
39
-
40
- ### List Available Templates
41
-
42
- ```bash
43
- # Show all plugin templates
44
- npx agentdb@latest list-templates
45
-
46
- # Available templates:
47
- # - decision-transformer (sequence modeling RL - recommended)
48
- # - q-learning (value-based learning)
49
- # - sarsa (on-policy TD learning)
50
- # - actor-critic (policy gradient with baseline)
51
- # - curiosity-driven (exploration-based)
52
- ```
53
-
54
- ### Manage Plugins
55
-
56
- ```bash
57
- # List installed plugins
58
- npx agentdb@latest list-plugins
59
-
60
- # Get plugin information
61
- npx agentdb@latest plugin-info my-agent
62
-
63
- # Shows: algorithm, configuration, training status
64
- ```
65
-
66
- ---
67
-
68
- ## Quick Start with API
69
-
70
- ```typescript
71
- import { createAgentDBAdapter } from 'agentic-flow/reasoningbank';
72
-
73
- // Initialize with learning enabled
74
- const adapter = await createAgentDBAdapter({
75
- dbPath: '.agentdb/learning.db',
76
- enableLearning: true, // Enable learning plugins
77
- enableReasoning: true,
78
- cacheSize: 1000,
79
- });
80
-
81
- // Store training experience
82
- await adapter.insertPattern({
83
- id: '',
84
- type: 'experience',
85
- domain: 'game-playing',
86
- pattern_data: JSON.stringify({
87
- embedding: await computeEmbedding('state-action-reward'),
88
- pattern: {
89
- state: [0.1, 0.2, 0.3],
90
- action: 2,
91
- reward: 1.0,
92
- next_state: [0.15, 0.25, 0.35],
93
- done: false
94
- }
95
- }),
96
- confidence: 0.9,
97
- usage_count: 1,
98
- success_count: 1,
99
- created_at: Date.now(),
100
- last_used: Date.now(),
101
- });
102
-
103
- // Train learning model
104
- const metrics = await adapter.train({
105
- epochs: 50,
106
- batchSize: 32,
107
- });
108
-
109
- console.log('Training Loss:', metrics.loss);
110
- console.log('Duration:', metrics.duration, 'ms');
111
- ```
112
-
113
- ---
114
-
115
- ## Available Learning Algorithms (9 Total)
116
-
117
- ### 1. Decision Transformer (Recommended)
118
-
119
- **Type**: Offline Reinforcement Learning
120
- **Best For**: Learning from logged experiences, imitation learning
121
- **Strengths**: No online interaction needed, stable training
122
-
123
- ```bash
124
- npx agentdb@latest create-plugin -t decision-transformer -n dt-agent
125
- ```
126
-
127
- **Use Cases**:
128
- - Learn from historical data
129
- - Imitation learning from expert demonstrations
130
- - Safe learning without environment interaction
131
- - Sequence modeling tasks
132
-
133
- **Configuration**:
134
- ```json
135
- {
136
- "algorithm": "decision-transformer",
137
- "model_size": "base",
138
- "context_length": 20,
139
- "embed_dim": 128,
140
- "n_heads": 8,
141
- "n_layers": 6
142
- }
143
- ```
144
-
145
- ### 2. Q-Learning
146
-
147
- **Type**: Value-Based RL (Off-Policy)
148
- **Best For**: Discrete action spaces, sample efficiency
149
- **Strengths**: Proven, simple, works well for small/medium problems
150
-
151
- ```bash
152
- npx agentdb@latest create-plugin -t q-learning -n q-agent
153
- ```
154
-
155
- **Use Cases**:
156
- - Grid worlds, board games
157
- - Navigation tasks
158
- - Resource allocation
159
- - Discrete decision-making
160
-
161
- **Configuration**:
162
- ```json
163
- {
164
- "algorithm": "q-learning",
165
- "learning_rate": 0.001,
166
- "gamma": 0.99,
167
- "epsilon": 0.1,
168
- "epsilon_decay": 0.995
169
- }
170
- ```
171
-
172
- ### 3. SARSA
173
-
174
- **Type**: Value-Based RL (On-Policy)
175
- **Best For**: Safe exploration, risk-sensitive tasks
176
- **Strengths**: More conservative than Q-Learning, better for safety
177
-
178
- ```bash
179
- npx agentdb@latest create-plugin -t sarsa -n sarsa-agent
180
- ```
181
-
182
- **Use Cases**:
183
- - Safety-critical applications
184
- - Risk-sensitive decision-making
185
- - Online learning with exploration
186
-
187
- **Configuration**:
188
- ```json
189
- {
190
- "algorithm": "sarsa",
191
- "learning_rate": 0.001,
192
- "gamma": 0.99,
193
- "epsilon": 0.1
194
- }
195
- ```
196
-
197
- ### 4. Actor-Critic
198
-
199
- **Type**: Policy Gradient with Value Baseline
200
- **Best For**: Continuous actions, variance reduction
201
- **Strengths**: Stable, works for continuous/discrete actions
202
-
203
- ```bash
204
- npx agentdb@latest create-plugin -t actor-critic -n ac-agent
205
- ```
206
-
207
- **Use Cases**:
208
- - Continuous control (robotics, simulations)
209
- - Complex action spaces
210
- - Multi-agent coordination
211
-
212
- **Configuration**:
213
- ```json
214
- {
215
- "algorithm": "actor-critic",
216
- "actor_lr": 0.001,
217
- "critic_lr": 0.002,
218
- "gamma": 0.99,
219
- "entropy_coef": 0.01
220
- }
221
- ```
222
-
223
- ### 5. Active Learning
224
-
225
- **Type**: Query-Based Learning
226
- **Best For**: Label-efficient learning, human-in-the-loop
227
- **Strengths**: Minimizes labeling cost, focuses on uncertain samples
228
-
229
- **Use Cases**:
230
- - Human feedback incorporation
231
- - Label-efficient training
232
- - Uncertainty sampling
233
- - Annotation cost reduction
234
-
235
- ### 6. Adversarial Training
236
-
237
- **Type**: Robustness Enhancement
238
- **Best For**: Safety, robustness to perturbations
239
- **Strengths**: Improves model robustness, adversarial defense
240
-
241
- **Use Cases**:
242
- - Security applications
243
- - Robust decision-making
244
- - Adversarial defense
245
- - Safety testing
246
-
247
- ### 7. Curriculum Learning
248
-
249
- **Type**: Progressive Difficulty Training
250
- **Best For**: Complex tasks, faster convergence
251
- **Strengths**: Stable learning, faster convergence on hard tasks
252
-
253
- **Use Cases**:
254
- - Complex multi-stage tasks
255
- - Hard exploration problems
256
- - Skill composition
257
- - Transfer learning
258
-
259
- ### 8. Federated Learning
260
-
261
- **Type**: Distributed Learning
262
- **Best For**: Privacy, distributed data
263
- **Strengths**: Privacy-preserving, scalable
264
-
265
- **Use Cases**:
266
- - Multi-agent systems
267
- - Privacy-sensitive data
268
- - Distributed training
269
- - Collaborative learning
270
-
271
- ### 9. Multi-Task Learning
272
-
273
- **Type**: Transfer Learning
274
- **Best For**: Related tasks, knowledge sharing
275
- **Strengths**: Faster learning on new tasks, better generalization
276
-
277
- **Use Cases**:
278
- - Task families
279
- - Transfer learning
280
- - Domain adaptation
281
- - Meta-learning
282
-
283
- ---
284
-
285
- ## Training Workflow
286
-
287
- ### 1. Collect Experiences
288
-
289
- ```typescript
290
- // Store experiences during agent execution
291
- for (let i = 0; i < numEpisodes; i++) {
292
- const episode = runEpisode();
293
-
294
- for (const step of episode.steps) {
295
- await adapter.insertPattern({
296
- id: '',
297
- type: 'experience',
298
- domain: 'task-domain',
299
- pattern_data: JSON.stringify({
300
- embedding: await computeEmbedding(JSON.stringify(step)),
301
- pattern: {
302
- state: step.state,
303
- action: step.action,
304
- reward: step.reward,
305
- next_state: step.next_state,
306
- done: step.done
307
- }
308
- }),
309
- confidence: step.reward > 0 ? 0.9 : 0.5,
310
- usage_count: 1,
311
- success_count: step.reward > 0 ? 1 : 0,
312
- created_at: Date.now(),
313
- last_used: Date.now(),
314
- });
315
- }
316
- }
317
- ```
318
-
319
- ### 2. Train Model
320
-
321
- ```typescript
322
- // Train on collected experiences
323
- const trainingMetrics = await adapter.train({
324
- epochs: 100,
325
- batchSize: 64,
326
- learningRate: 0.001,
327
- validationSplit: 0.2,
328
- });
329
-
330
- console.log('Training Metrics:', trainingMetrics);
331
- // {
332
- // loss: 0.023,
333
- // valLoss: 0.028,
334
- // duration: 1523,
335
- // epochs: 100
336
- // }
337
- ```
338
-
339
- ### 3. Evaluate Performance
340
-
341
- ```typescript
342
- // Retrieve similar successful experiences
343
- const testQuery = await computeEmbedding(JSON.stringify(testState));
344
- const result = await adapter.retrieveWithReasoning(testQuery, {
345
- domain: 'task-domain',
346
- k: 10,
347
- synthesizeContext: true,
348
- });
349
-
350
- // Evaluate action quality
351
- const suggestedAction = result.memories[0].pattern.action;
352
- const confidence = result.memories[0].similarity;
353
-
354
- console.log('Suggested Action:', suggestedAction);
355
- console.log('Confidence:', confidence);
356
- ```
357
-
358
- ---
359
-
360
- ## Advanced Training Techniques
361
-
362
- ### Experience Replay
363
-
364
- ```typescript
365
- // Store experiences in buffer
366
- const replayBuffer = [];
367
-
368
- // Sample random batch for training
369
- const batch = sampleRandomBatch(replayBuffer, batchSize: 32);
370
-
371
- // Train on batch
372
- await adapter.train({
373
- data: batch,
374
- epochs: 1,
375
- batchSize: 32,
376
- });
377
- ```
378
-
379
- ### Prioritized Experience Replay
380
-
381
- ```typescript
382
- // Store experiences with priority (TD error)
383
- await adapter.insertPattern({
384
- // ... standard fields
385
- confidence: tdError, // Use TD error as confidence/priority
386
- // ...
387
- });
388
-
389
- // Retrieve high-priority experiences
390
- const highPriority = await adapter.retrieveWithReasoning(queryEmbedding, {
391
- domain: 'task-domain',
392
- k: 32,
393
- minConfidence: 0.7, // Only high TD-error experiences
394
- });
395
- ```
396
-
397
- ### Multi-Agent Training
398
-
399
- ```typescript
400
- // Collect experiences from multiple agents
401
- for (const agent of agents) {
402
- const experience = await agent.step();
403
-
404
- await adapter.insertPattern({
405
- // ... store experience with agent ID
406
- domain: `multi-agent/${agent.id}`,
407
- });
408
- }
409
-
410
- // Train shared model
411
- await adapter.train({
412
- epochs: 50,
413
- batchSize: 64,
414
- });
415
- ```
416
-
417
- ---
418
-
419
- ## Performance Optimization
420
-
421
- ### Batch Training
422
-
423
- ```typescript
424
- // Collect batch of experiences
425
- const experiences = collectBatch(size: 1000);
426
-
427
- // Batch insert (500x faster)
428
- for (const exp of experiences) {
429
- await adapter.insertPattern({ /* ... */ });
430
- }
431
-
432
- // Train on batch
433
- await adapter.train({
434
- epochs: 10,
435
- batchSize: 128, // Larger batch for efficiency
436
- });
437
- ```
438
-
439
- ### Incremental Learning
440
-
441
- ```typescript
442
- // Train incrementally as new data arrives
443
- setInterval(async () => {
444
- const newExperiences = getNewExperiences();
445
-
446
- if (newExperiences.length > 100) {
447
- await adapter.train({
448
- epochs: 5,
449
- batchSize: 32,
450
- });
451
- }
452
- }, 60000); // Every minute
453
- ```
454
-
455
- ---
456
-
457
- ## Integration with Reasoning Agents
458
-
459
- Combine learning with reasoning for better performance:
460
-
461
- ```typescript
462
- // Train learning model
463
- await adapter.train({ epochs: 50, batchSize: 32 });
464
-
465
- // Use reasoning agents for inference
466
- const result = await adapter.retrieveWithReasoning(queryEmbedding, {
467
- domain: 'decision-making',
468
- k: 10,
469
- useMMR: true, // Diverse experiences
470
- synthesizeContext: true, // Rich context
471
- optimizeMemory: true, // Consolidate patterns
472
- });
473
-
474
- // Make decision based on learned experiences + reasoning
475
- const decision = result.context.suggestedAction;
476
- const confidence = result.memories[0].similarity;
477
- ```
478
-
479
- ---
480
-
481
- ## CLI Operations
482
-
483
- ```bash
484
- # Create plugin
485
- npx agentdb@latest create-plugin -t decision-transformer -n my-plugin
486
-
487
- # List plugins
488
- npx agentdb@latest list-plugins
489
-
490
- # Get plugin info
491
- npx agentdb@latest plugin-info my-plugin
492
-
493
- # List templates
494
- npx agentdb@latest list-templates
495
- ```
496
-
497
- ---
498
-
499
- ## Troubleshooting
500
-
501
- ### Issue: Training not converging
502
- ```typescript
503
- // Reduce learning rate
504
- await adapter.train({
505
- epochs: 100,
506
- batchSize: 32,
507
- learningRate: 0.0001, // Lower learning rate
508
- });
509
- ```
510
-
511
- ### Issue: Overfitting
512
- ```typescript
513
- // Use validation split
514
- await adapter.train({
515
- epochs: 50,
516
- batchSize: 64,
517
- validationSplit: 0.2, // 20% validation
518
- });
519
-
520
- // Enable memory optimization
521
- await adapter.retrieveWithReasoning(queryEmbedding, {
522
- optimizeMemory: true, // Consolidate, reduce overfitting
523
- });
524
- ```
525
-
526
- ### Issue: Slow training
527
- ```bash
528
- # Enable quantization for faster inference
529
- # Use binary quantization (32x faster)
530
- ```
531
-
532
- ---
533
-
534
- ## Learn More
535
-
536
- - **Algorithm Papers**: See docs/algorithms/ for detailed papers
537
- - **GitHub**: https://github.com/ruvnet/agentic-flow/tree/main/packages/agentdb
538
- - **MCP Integration**: `npx agentdb@latest mcp`
539
- - **Website**: https://agentdb.ruv.io
540
-
541
- ---
542
-
543
- **Category**: Machine Learning / Reinforcement Learning
544
- **Difficulty**: Intermediate to Advanced
545
- **Estimated Time**: 30-60 minutes
1
+ ---
2
+ name: "AgentDB Learning Plugins"
3
+ description: "Create and train AI learning plugins with AgentDB's 9 reinforcement learning algorithms. Includes Decision Transformer, Q-Learning, SARSA, Actor-Critic, and more. Use when building self-learning agents, implementing RL, or optimizing agent behavior through experience."
4
+ ---
5
+
6
+ # AgentDB Learning Plugins
7
+
8
+ ## What This Skill Does
9
+
10
+ Provides access to 9 reinforcement learning algorithms via AgentDB's plugin system. Create, train, and deploy learning plugins for autonomous agents that improve through experience. Includes offline RL (Decision Transformer), value-based learning (Q-Learning), policy gradients (Actor-Critic), and advanced techniques.
11
+
12
+ **Performance**: Train models 10-100x faster with WASM-accelerated neural inference.
13
+
14
+ ## Prerequisites
15
+
16
+ - Node.js 18+
17
+ - AgentDB v1.0.7+ (via agentic-flow)
18
+ - Basic understanding of reinforcement learning (recommended)
19
+
20
+ ---
21
+
22
+ ## Quick Start with CLI
23
+
24
+ ### Create Learning Plugin
25
+
26
+ ```bash
27
+ # Interactive wizard
28
+ npx agentdb@latest create-plugin
29
+
30
+ # Use specific template
31
+ npx agentdb@latest create-plugin -t decision-transformer -n my-agent
32
+
33
+ # Preview without creating
34
+ npx agentdb@latest create-plugin -t q-learning --dry-run
35
+
36
+ # Custom output directory
37
+ npx agentdb@latest create-plugin -t actor-critic -o ./plugins
38
+ ```
39
+
40
+ ### List Available Templates
41
+
42
+ ```bash
43
+ # Show all plugin templates
44
+ npx agentdb@latest list-templates
45
+
46
+ # Available templates:
47
+ # - decision-transformer (sequence modeling RL - recommended)
48
+ # - q-learning (value-based learning)
49
+ # - sarsa (on-policy TD learning)
50
+ # - actor-critic (policy gradient with baseline)
51
+ # - curiosity-driven (exploration-based)
52
+ ```
53
+
54
+ ### Manage Plugins
55
+
56
+ ```bash
57
+ # List installed plugins
58
+ npx agentdb@latest list-plugins
59
+
60
+ # Get plugin information
61
+ npx agentdb@latest plugin-info my-agent
62
+
63
+ # Shows: algorithm, configuration, training status
64
+ ```
65
+
66
+ ---
67
+
68
+ ## Quick Start with API
69
+
70
+ ```typescript
71
+ import { createAgentDBAdapter } from 'agentic-flow/reasoningbank';
72
+
73
+ // Initialize with learning enabled
74
+ const adapter = await createAgentDBAdapter({
75
+ dbPath: '.agentdb/learning.db',
76
+ enableLearning: true, // Enable learning plugins
77
+ enableReasoning: true,
78
+ cacheSize: 1000,
79
+ });
80
+
81
+ // Store training experience
82
+ await adapter.insertPattern({
83
+ id: '',
84
+ type: 'experience',
85
+ domain: 'game-playing',
86
+ pattern_data: JSON.stringify({
87
+ embedding: await computeEmbedding('state-action-reward'),
88
+ pattern: {
89
+ state: [0.1, 0.2, 0.3],
90
+ action: 2,
91
+ reward: 1.0,
92
+ next_state: [0.15, 0.25, 0.35],
93
+ done: false
94
+ }
95
+ }),
96
+ confidence: 0.9,
97
+ usage_count: 1,
98
+ success_count: 1,
99
+ created_at: Date.now(),
100
+ last_used: Date.now(),
101
+ });
102
+
103
+ // Train learning model
104
+ const metrics = await adapter.train({
105
+ epochs: 50,
106
+ batchSize: 32,
107
+ });
108
+
109
+ console.log('Training Loss:', metrics.loss);
110
+ console.log('Duration:', metrics.duration, 'ms');
111
+ ```
112
+
113
+ ---
114
+
115
+ ## Available Learning Algorithms (9 Total)
116
+
117
+ ### 1. Decision Transformer (Recommended)
118
+
119
+ **Type**: Offline Reinforcement Learning
120
+ **Best For**: Learning from logged experiences, imitation learning
121
+ **Strengths**: No online interaction needed, stable training
122
+
123
+ ```bash
124
+ npx agentdb@latest create-plugin -t decision-transformer -n dt-agent
125
+ ```
126
+
127
+ **Use Cases**:
128
+ - Learn from historical data
129
+ - Imitation learning from expert demonstrations
130
+ - Safe learning without environment interaction
131
+ - Sequence modeling tasks
132
+
133
+ **Configuration**:
134
+ ```json
135
+ {
136
+ "algorithm": "decision-transformer",
137
+ "model_size": "base",
138
+ "context_length": 20,
139
+ "embed_dim": 128,
140
+ "n_heads": 8,
141
+ "n_layers": 6
142
+ }
143
+ ```
144
+
145
+ ### 2. Q-Learning
146
+
147
+ **Type**: Value-Based RL (Off-Policy)
148
+ **Best For**: Discrete action spaces, sample efficiency
149
+ **Strengths**: Proven, simple, works well for small/medium problems
150
+
151
+ ```bash
152
+ npx agentdb@latest create-plugin -t q-learning -n q-agent
153
+ ```
154
+
155
+ **Use Cases**:
156
+ - Grid worlds, board games
157
+ - Navigation tasks
158
+ - Resource allocation
159
+ - Discrete decision-making
160
+
161
+ **Configuration**:
162
+ ```json
163
+ {
164
+ "algorithm": "q-learning",
165
+ "learning_rate": 0.001,
166
+ "gamma": 0.99,
167
+ "epsilon": 0.1,
168
+ "epsilon_decay": 0.995
169
+ }
170
+ ```
171
+
172
+ ### 3. SARSA
173
+
174
+ **Type**: Value-Based RL (On-Policy)
175
+ **Best For**: Safe exploration, risk-sensitive tasks
176
+ **Strengths**: More conservative than Q-Learning, better for safety
177
+
178
+ ```bash
179
+ npx agentdb@latest create-plugin -t sarsa -n sarsa-agent
180
+ ```
181
+
182
+ **Use Cases**:
183
+ - Safety-critical applications
184
+ - Risk-sensitive decision-making
185
+ - Online learning with exploration
186
+
187
+ **Configuration**:
188
+ ```json
189
+ {
190
+ "algorithm": "sarsa",
191
+ "learning_rate": 0.001,
192
+ "gamma": 0.99,
193
+ "epsilon": 0.1
194
+ }
195
+ ```
196
+
197
+ ### 4. Actor-Critic
198
+
199
+ **Type**: Policy Gradient with Value Baseline
200
+ **Best For**: Continuous actions, variance reduction
201
+ **Strengths**: Stable, works for continuous/discrete actions
202
+
203
+ ```bash
204
+ npx agentdb@latest create-plugin -t actor-critic -n ac-agent
205
+ ```
206
+
207
+ **Use Cases**:
208
+ - Continuous control (robotics, simulations)
209
+ - Complex action spaces
210
+ - Multi-agent coordination
211
+
212
+ **Configuration**:
213
+ ```json
214
+ {
215
+ "algorithm": "actor-critic",
216
+ "actor_lr": 0.001,
217
+ "critic_lr": 0.002,
218
+ "gamma": 0.99,
219
+ "entropy_coef": 0.01
220
+ }
221
+ ```
222
+
223
+ ### 5. Active Learning
224
+
225
+ **Type**: Query-Based Learning
226
+ **Best For**: Label-efficient learning, human-in-the-loop
227
+ **Strengths**: Minimizes labeling cost, focuses on uncertain samples
228
+
229
+ **Use Cases**:
230
+ - Human feedback incorporation
231
+ - Label-efficient training
232
+ - Uncertainty sampling
233
+ - Annotation cost reduction
234
+
235
+ ### 6. Adversarial Training
236
+
237
+ **Type**: Robustness Enhancement
238
+ **Best For**: Safety, robustness to perturbations
239
+ **Strengths**: Improves model robustness, adversarial defense
240
+
241
+ **Use Cases**:
242
+ - Security applications
243
+ - Robust decision-making
244
+ - Adversarial defense
245
+ - Safety testing
246
+
247
+ ### 7. Curriculum Learning
248
+
249
+ **Type**: Progressive Difficulty Training
250
+ **Best For**: Complex tasks, faster convergence
251
+ **Strengths**: Stable learning, faster convergence on hard tasks
252
+
253
+ **Use Cases**:
254
+ - Complex multi-stage tasks
255
+ - Hard exploration problems
256
+ - Skill composition
257
+ - Transfer learning
258
+
259
+ ### 8. Federated Learning
260
+
261
+ **Type**: Distributed Learning
262
+ **Best For**: Privacy, distributed data
263
+ **Strengths**: Privacy-preserving, scalable
264
+
265
+ **Use Cases**:
266
+ - Multi-agent systems
267
+ - Privacy-sensitive data
268
+ - Distributed training
269
+ - Collaborative learning
270
+
271
+ ### 9. Multi-Task Learning
272
+
273
+ **Type**: Transfer Learning
274
+ **Best For**: Related tasks, knowledge sharing
275
+ **Strengths**: Faster learning on new tasks, better generalization
276
+
277
+ **Use Cases**:
278
+ - Task families
279
+ - Transfer learning
280
+ - Domain adaptation
281
+ - Meta-learning
282
+
283
+ ---
284
+
285
+ ## Training Workflow
286
+
287
+ ### 1. Collect Experiences
288
+
289
+ ```typescript
290
+ // Store experiences during agent execution
291
+ for (let i = 0; i < numEpisodes; i++) {
292
+ const episode = runEpisode();
293
+
294
+ for (const step of episode.steps) {
295
+ await adapter.insertPattern({
296
+ id: '',
297
+ type: 'experience',
298
+ domain: 'task-domain',
299
+ pattern_data: JSON.stringify({
300
+ embedding: await computeEmbedding(JSON.stringify(step)),
301
+ pattern: {
302
+ state: step.state,
303
+ action: step.action,
304
+ reward: step.reward,
305
+ next_state: step.next_state,
306
+ done: step.done
307
+ }
308
+ }),
309
+ confidence: step.reward > 0 ? 0.9 : 0.5,
310
+ usage_count: 1,
311
+ success_count: step.reward > 0 ? 1 : 0,
312
+ created_at: Date.now(),
313
+ last_used: Date.now(),
314
+ });
315
+ }
316
+ }
317
+ ```
318
+
319
+ ### 2. Train Model
320
+
321
+ ```typescript
322
+ // Train on collected experiences
323
+ const trainingMetrics = await adapter.train({
324
+ epochs: 100,
325
+ batchSize: 64,
326
+ learningRate: 0.001,
327
+ validationSplit: 0.2,
328
+ });
329
+
330
+ console.log('Training Metrics:', trainingMetrics);
331
+ // {
332
+ // loss: 0.023,
333
+ // valLoss: 0.028,
334
+ // duration: 1523,
335
+ // epochs: 100
336
+ // }
337
+ ```
338
+
339
+ ### 3. Evaluate Performance
340
+
341
+ ```typescript
342
+ // Retrieve similar successful experiences
343
+ const testQuery = await computeEmbedding(JSON.stringify(testState));
344
+ const result = await adapter.retrieveWithReasoning(testQuery, {
345
+ domain: 'task-domain',
346
+ k: 10,
347
+ synthesizeContext: true,
348
+ });
349
+
350
+ // Evaluate action quality
351
+ const suggestedAction = result.memories[0].pattern.action;
352
+ const confidence = result.memories[0].similarity;
353
+
354
+ console.log('Suggested Action:', suggestedAction);
355
+ console.log('Confidence:', confidence);
356
+ ```
357
+
358
+ ---
359
+
360
+ ## Advanced Training Techniques
361
+
362
+ ### Experience Replay
363
+
364
+ ```typescript
365
+ // Store experiences in buffer
366
+ const replayBuffer = [];
367
+
368
+ // Sample random batch for training
369
+ const batch = sampleRandomBatch(replayBuffer, batchSize: 32);
370
+
371
+ // Train on batch
372
+ await adapter.train({
373
+ data: batch,
374
+ epochs: 1,
375
+ batchSize: 32,
376
+ });
377
+ ```
378
+
379
+ ### Prioritized Experience Replay
380
+
381
+ ```typescript
382
+ // Store experiences with priority (TD error)
383
+ await adapter.insertPattern({
384
+ // ... standard fields
385
+ confidence: tdError, // Use TD error as confidence/priority
386
+ // ...
387
+ });
388
+
389
+ // Retrieve high-priority experiences
390
+ const highPriority = await adapter.retrieveWithReasoning(queryEmbedding, {
391
+ domain: 'task-domain',
392
+ k: 32,
393
+ minConfidence: 0.7, // Only high TD-error experiences
394
+ });
395
+ ```
396
+
397
+ ### Multi-Agent Training
398
+
399
+ ```typescript
400
+ // Collect experiences from multiple agents
401
+ for (const agent of agents) {
402
+ const experience = await agent.step();
403
+
404
+ await adapter.insertPattern({
405
+ // ... store experience with agent ID
406
+ domain: `multi-agent/${agent.id}`,
407
+ });
408
+ }
409
+
410
+ // Train shared model
411
+ await adapter.train({
412
+ epochs: 50,
413
+ batchSize: 64,
414
+ });
415
+ ```
416
+
417
+ ---
418
+
419
+ ## Performance Optimization
420
+
421
+ ### Batch Training
422
+
423
+ ```typescript
424
+ // Collect batch of experiences
425
+ const experiences = collectBatch(size: 1000);
426
+
427
+ // Batch insert (500x faster)
428
+ for (const exp of experiences) {
429
+ await adapter.insertPattern({ /* ... */ });
430
+ }
431
+
432
+ // Train on batch
433
+ await adapter.train({
434
+ epochs: 10,
435
+ batchSize: 128, // Larger batch for efficiency
436
+ });
437
+ ```
438
+
439
+ ### Incremental Learning
440
+
441
+ ```typescript
442
+ // Train incrementally as new data arrives
443
+ setInterval(async () => {
444
+ const newExperiences = getNewExperiences();
445
+
446
+ if (newExperiences.length > 100) {
447
+ await adapter.train({
448
+ epochs: 5,
449
+ batchSize: 32,
450
+ });
451
+ }
452
+ }, 60000); // Every minute
453
+ ```
454
+
455
+ ---
456
+
457
+ ## Integration with Reasoning Agents
458
+
459
+ Combine learning with reasoning for better performance:
460
+
461
+ ```typescript
462
+ // Train learning model
463
+ await adapter.train({ epochs: 50, batchSize: 32 });
464
+
465
+ // Use reasoning agents for inference
466
+ const result = await adapter.retrieveWithReasoning(queryEmbedding, {
467
+ domain: 'decision-making',
468
+ k: 10,
469
+ useMMR: true, // Diverse experiences
470
+ synthesizeContext: true, // Rich context
471
+ optimizeMemory: true, // Consolidate patterns
472
+ });
473
+
474
+ // Make decision based on learned experiences + reasoning
475
+ const decision = result.context.suggestedAction;
476
+ const confidence = result.memories[0].similarity;
477
+ ```
478
+
479
+ ---
480
+
481
+ ## CLI Operations
482
+
483
+ ```bash
484
+ # Create plugin
485
+ npx agentdb@latest create-plugin -t decision-transformer -n my-plugin
486
+
487
+ # List plugins
488
+ npx agentdb@latest list-plugins
489
+
490
+ # Get plugin info
491
+ npx agentdb@latest plugin-info my-plugin
492
+
493
+ # List templates
494
+ npx agentdb@latest list-templates
495
+ ```
496
+
497
+ ---
498
+
499
+ ## Troubleshooting
500
+
501
+ ### Issue: Training not converging
502
+ ```typescript
503
+ // Reduce learning rate
504
+ await adapter.train({
505
+ epochs: 100,
506
+ batchSize: 32,
507
+ learningRate: 0.0001, // Lower learning rate
508
+ });
509
+ ```
510
+
511
+ ### Issue: Overfitting
512
+ ```typescript
513
+ // Use validation split
514
+ await adapter.train({
515
+ epochs: 50,
516
+ batchSize: 64,
517
+ validationSplit: 0.2, // 20% validation
518
+ });
519
+
520
+ // Enable memory optimization
521
+ await adapter.retrieveWithReasoning(queryEmbedding, {
522
+ optimizeMemory: true, // Consolidate, reduce overfitting
523
+ });
524
+ ```
525
+
526
+ ### Issue: Slow training
527
+ ```bash
528
+ # Enable quantization for faster inference
529
+ # Use binary quantization (32x faster)
530
+ ```
531
+
532
+ ---
533
+
534
+ ## Learn More
535
+
536
+ - **Algorithm Papers**: See docs/algorithms/ for detailed papers
537
+ - **GitHub**: https://github.com/ruvnet/agentic-flow/tree/main/packages/agentdb
538
+ - **MCP Integration**: `npx agentdb@latest mcp`
539
+ - **Website**: https://agentdb.ruv.io
540
+
541
+ ---
542
+
543
+ **Category**: Machine Learning / Reinforcement Learning
544
+ **Difficulty**: Intermediate to Advanced
545
+ **Estimated Time**: 30-60 minutes