agentic-qe 1.1.0 → 1.2.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (297) hide show
  1. package/.claude/agents/qe-coverage-analyzer.md +8 -0
  2. package/.claude/agents/qe-flaky-test-hunter.md +9 -1
  3. package/.claude/agents/qe-test-generator.md +7 -0
  4. package/.claude/agents/reasoning/agent.md +816 -0
  5. package/.claude/agents/reasoning/goal-planner.md +73 -0
  6. package/.claude/settings.json +21 -20
  7. package/.claude/skills/README.md +124 -0
  8. package/.claude/skills/agentdb-advanced/SKILL.md +550 -0
  9. package/.claude/skills/agentdb-learning/SKILL.md +545 -0
  10. package/.claude/skills/agentdb-memory-patterns/SKILL.md +339 -0
  11. package/.claude/skills/agentdb-optimization/SKILL.md +509 -0
  12. package/.claude/skills/agentdb-vector-search/SKILL.md +339 -0
  13. package/.claude/skills/agentic-quality-engineering/SKILL.md +604 -0
  14. package/.claude/skills/api-testing-patterns/SKILL.md +686 -0
  15. package/.claude/skills/bug-reporting-excellence/SKILL.md +632 -0
  16. package/.claude/skills/code-review-quality/SKILL.md +683 -0
  17. package/.claude/skills/consultancy-practices/SKILL.md +540 -0
  18. package/.claude/skills/context-driven-testing/SKILL.md +466 -0
  19. package/.claude/skills/exploratory-testing-advanced/SKILL.md +676 -0
  20. package/.claude/skills/flow-nexus-neural/SKILL.md +738 -0
  21. package/.claude/skills/flow-nexus-platform/SKILL.md +1157 -0
  22. package/.claude/skills/flow-nexus-swarm/SKILL.md +610 -0
  23. package/.claude/skills/github-code-review/SKILL.md +1140 -0
  24. package/.claude/skills/github-multi-repo/SKILL.md +874 -0
  25. package/.claude/skills/github-project-management/SKILL.md +1277 -0
  26. package/.claude/skills/github-release-management/SKILL.md +1081 -0
  27. package/.claude/skills/github-workflow-automation/SKILL.md +1065 -0
  28. package/.claude/skills/hive-mind-advanced/SKILL.md +712 -0
  29. package/.claude/skills/holistic-testing-pact/SKILL.md +225 -0
  30. package/.claude/skills/hooks-automation/SKILL.md +1201 -0
  31. package/.claude/skills/pair-programming/SKILL.md +1202 -0
  32. package/.claude/skills/performance-analysis/SKILL.md +563 -0
  33. package/.claude/skills/performance-testing/SKILL.md +662 -0
  34. package/.claude/skills/quality-metrics/SKILL.md +592 -0
  35. package/.claude/skills/reasoningbank-agentdb/SKILL.md +446 -0
  36. package/.claude/skills/reasoningbank-intelligence/SKILL.md +201 -0
  37. package/.claude/skills/refactoring-patterns/SKILL.md +778 -0
  38. package/.claude/skills/risk-based-testing/SKILL.md +721 -0
  39. package/.claude/skills/security-testing/SKILL.md +651 -0
  40. package/.claude/skills/skill-builder/SKILL.md +910 -0
  41. package/.claude/skills/sparc-methodology/SKILL.md +1115 -0
  42. package/.claude/skills/stream-chain/SKILL.md +563 -0
  43. package/.claude/skills/swarm-advanced/SKILL.md +973 -0
  44. package/.claude/skills/swarm-orchestration/SKILL.md +179 -0
  45. package/.claude/skills/tdd-london-chicago/SKILL.md +567 -0
  46. package/.claude/skills/technical-writing/SKILL.md +235 -0
  47. package/.claude/skills/test-automation-strategy/SKILL.md +842 -0
  48. package/.claude/skills/verification-quality/SKILL.md +649 -0
  49. package/.claude/skills/xp-practices/SKILL.md +671 -0
  50. package/.claude/statusline-command.sh +176 -0
  51. package/CHANGELOG.md +536 -1
  52. package/README.md +92 -3
  53. package/config/improvement-loop.config.ts +323 -0
  54. package/config/neural-agent.config.ts +197 -0
  55. package/dist/adapters/MemoryStoreAdapter.d.ts +16 -16
  56. package/dist/adapters/MemoryStoreAdapter.d.ts.map +1 -1
  57. package/dist/adapters/MemoryStoreAdapter.js +16 -16
  58. package/dist/adapters/MemoryStoreAdapter.js.map +1 -1
  59. package/dist/agents/BaseAgent.d.ts +69 -0
  60. package/dist/agents/BaseAgent.d.ts.map +1 -1
  61. package/dist/agents/BaseAgent.js +382 -1
  62. package/dist/agents/BaseAgent.js.map +1 -1
  63. package/dist/agents/CoverageAnalyzerAgent.d.ts +13 -2
  64. package/dist/agents/CoverageAnalyzerAgent.d.ts.map +1 -1
  65. package/dist/agents/CoverageAnalyzerAgent.js +105 -6
  66. package/dist/agents/CoverageAnalyzerAgent.js.map +1 -1
  67. package/dist/agents/DeploymentReadinessAgent.d.ts.map +1 -1
  68. package/dist/agents/DeploymentReadinessAgent.js +13 -13
  69. package/dist/agents/DeploymentReadinessAgent.js.map +1 -1
  70. package/dist/agents/FlakyTestHunterAgent.d.ts +28 -0
  71. package/dist/agents/FlakyTestHunterAgent.d.ts.map +1 -1
  72. package/dist/agents/FlakyTestHunterAgent.js +159 -4
  73. package/dist/agents/FlakyTestHunterAgent.js.map +1 -1
  74. package/dist/agents/FleetCommanderAgent.d.ts.map +1 -1
  75. package/dist/agents/FleetCommanderAgent.js +2 -2
  76. package/dist/agents/FleetCommanderAgent.js.map +1 -1
  77. package/dist/agents/LearningAgent.d.ts +14 -5
  78. package/dist/agents/LearningAgent.d.ts.map +1 -1
  79. package/dist/agents/LearningAgent.js +36 -16
  80. package/dist/agents/LearningAgent.js.map +1 -1
  81. package/dist/agents/NeuralAgentExtension.d.ts +117 -0
  82. package/dist/agents/NeuralAgentExtension.d.ts.map +1 -0
  83. package/dist/agents/NeuralAgentExtension.js +288 -0
  84. package/dist/agents/NeuralAgentExtension.js.map +1 -0
  85. package/dist/agents/PerformanceTesterAgent.d.ts.map +1 -1
  86. package/dist/agents/PerformanceTesterAgent.js +4 -3
  87. package/dist/agents/PerformanceTesterAgent.js.map +1 -1
  88. package/dist/agents/ProductionIntelligenceAgent.js +7 -7
  89. package/dist/agents/ProductionIntelligenceAgent.js.map +1 -1
  90. package/dist/agents/QualityAnalyzerAgent.d.ts.map +1 -1
  91. package/dist/agents/QualityAnalyzerAgent.js +2 -2
  92. package/dist/agents/QualityAnalyzerAgent.js.map +1 -1
  93. package/dist/agents/QualityGateAgent.js +5 -5
  94. package/dist/agents/QualityGateAgent.js.map +1 -1
  95. package/dist/agents/RegressionRiskAnalyzerAgent.js +7 -7
  96. package/dist/agents/RegressionRiskAnalyzerAgent.js.map +1 -1
  97. package/dist/agents/RequirementsValidatorAgent.d.ts.map +1 -1
  98. package/dist/agents/RequirementsValidatorAgent.js +1 -1
  99. package/dist/agents/RequirementsValidatorAgent.js.map +1 -1
  100. package/dist/agents/SecurityScannerAgent.js +6 -6
  101. package/dist/agents/SecurityScannerAgent.js.map +1 -1
  102. package/dist/agents/TestExecutorAgent.d.ts.map +1 -1
  103. package/dist/agents/TestExecutorAgent.js +1 -3
  104. package/dist/agents/TestExecutorAgent.js.map +1 -1
  105. package/dist/agents/TestGeneratorAgent.d.ts +15 -4
  106. package/dist/agents/TestGeneratorAgent.d.ts.map +1 -1
  107. package/dist/agents/TestGeneratorAgent.js +165 -43
  108. package/dist/agents/TestGeneratorAgent.js.map +1 -1
  109. package/dist/agents/mixins/NeuralCapableMixin.d.ts +130 -0
  110. package/dist/agents/mixins/NeuralCapableMixin.d.ts.map +1 -0
  111. package/dist/agents/mixins/NeuralCapableMixin.js +358 -0
  112. package/dist/agents/mixins/NeuralCapableMixin.js.map +1 -0
  113. package/dist/agents/mixins/QUICCapableMixin.d.ts +34 -0
  114. package/dist/agents/mixins/QUICCapableMixin.d.ts.map +1 -0
  115. package/dist/agents/mixins/QUICCapableMixin.js +346 -0
  116. package/dist/agents/mixins/QUICCapableMixin.js.map +1 -0
  117. package/dist/cli/commands/agent/index.d.ts +5 -0
  118. package/dist/cli/commands/agent/index.d.ts.map +1 -1
  119. package/dist/cli/commands/agent/index.js +11 -6
  120. package/dist/cli/commands/agent/index.js.map +1 -1
  121. package/dist/cli/commands/agent/kill.d.ts +13 -0
  122. package/dist/cli/commands/agent/kill.d.ts.map +1 -0
  123. package/dist/cli/commands/agent/kill.js +65 -0
  124. package/dist/cli/commands/agent/kill.js.map +1 -0
  125. package/dist/cli/commands/agent/list.d.ts +19 -0
  126. package/dist/cli/commands/agent/list.d.ts.map +1 -0
  127. package/dist/cli/commands/agent/list.js +92 -0
  128. package/dist/cli/commands/agent/list.js.map +1 -0
  129. package/dist/cli/commands/agent/logs.d.ts +14 -0
  130. package/dist/cli/commands/agent/logs.d.ts.map +1 -0
  131. package/dist/cli/commands/agent/logs.js +77 -0
  132. package/dist/cli/commands/agent/logs.js.map +1 -0
  133. package/dist/cli/commands/agent/metrics.d.ts +21 -0
  134. package/dist/cli/commands/agent/metrics.d.ts.map +1 -0
  135. package/dist/cli/commands/agent/metrics.js +87 -0
  136. package/dist/cli/commands/agent/metrics.js.map +1 -0
  137. package/dist/cli/commands/agent/spawn.d.ts +28 -0
  138. package/dist/cli/commands/agent/spawn.d.ts.map +1 -0
  139. package/dist/cli/commands/agent/spawn.js +83 -0
  140. package/dist/cli/commands/agent/spawn.js.map +1 -0
  141. package/dist/cli/commands/init.d.ts +24 -0
  142. package/dist/cli/commands/init.d.ts.map +1 -1
  143. package/dist/cli/commands/init.js +475 -7
  144. package/dist/cli/commands/init.js.map +1 -1
  145. package/dist/cli/commands/skills/index.d.ts +51 -0
  146. package/dist/cli/commands/skills/index.d.ts.map +1 -0
  147. package/dist/cli/commands/skills/index.js +364 -0
  148. package/dist/cli/commands/skills/index.js.map +1 -0
  149. package/dist/cli/index.js +92 -1
  150. package/dist/cli/index.js.map +1 -1
  151. package/dist/core/EventBus.d.ts +38 -0
  152. package/dist/core/EventBus.d.ts.map +1 -1
  153. package/dist/core/EventBus.js +176 -31
  154. package/dist/core/EventBus.js.map +1 -1
  155. package/dist/core/FleetManager.d.ts +35 -1
  156. package/dist/core/FleetManager.d.ts.map +1 -1
  157. package/dist/core/FleetManager.js +121 -45
  158. package/dist/core/FleetManager.js.map +1 -1
  159. package/dist/core/MemoryManager.d.ts +19 -1
  160. package/dist/core/MemoryManager.d.ts.map +1 -1
  161. package/dist/core/MemoryManager.js +25 -1
  162. package/dist/core/MemoryManager.js.map +1 -1
  163. package/dist/core/embeddings/EmbeddingCache.d.ts +134 -0
  164. package/dist/core/embeddings/EmbeddingCache.d.ts.map +1 -0
  165. package/dist/core/embeddings/EmbeddingCache.js +239 -0
  166. package/dist/core/embeddings/EmbeddingCache.js.map +1 -0
  167. package/dist/core/embeddings/EmbeddingGenerator.d.ts +224 -0
  168. package/dist/core/embeddings/EmbeddingGenerator.d.ts.map +1 -0
  169. package/dist/core/embeddings/EmbeddingGenerator.js +459 -0
  170. package/dist/core/embeddings/EmbeddingGenerator.js.map +1 -0
  171. package/dist/core/embeddings/index.d.ts +15 -0
  172. package/dist/core/embeddings/index.d.ts.map +1 -0
  173. package/dist/core/embeddings/index.js +22 -0
  174. package/dist/core/embeddings/index.js.map +1 -0
  175. package/dist/core/memory/AgentDBIntegration.d.ts +35 -0
  176. package/dist/core/memory/AgentDBIntegration.d.ts.map +1 -0
  177. package/dist/core/memory/AgentDBIntegration.js +75 -0
  178. package/dist/core/memory/AgentDBIntegration.js.map +1 -0
  179. package/dist/core/memory/AgentDBManager.d.ts +200 -0
  180. package/dist/core/memory/AgentDBManager.d.ts.map +1 -0
  181. package/dist/core/memory/AgentDBManager.js +263 -0
  182. package/dist/core/memory/AgentDBManager.js.map +1 -0
  183. package/dist/core/memory/AgentDBService.d.ts +160 -0
  184. package/dist/core/memory/AgentDBService.d.ts.map +1 -0
  185. package/dist/core/memory/AgentDBService.js +450 -0
  186. package/dist/core/memory/AgentDBService.js.map +1 -0
  187. package/dist/core/memory/RealAgentDBAdapter.d.ts +51 -0
  188. package/dist/core/memory/RealAgentDBAdapter.d.ts.map +1 -0
  189. package/dist/core/memory/RealAgentDBAdapter.js +230 -0
  190. package/dist/core/memory/RealAgentDBAdapter.js.map +1 -0
  191. package/dist/core/memory/ReasoningBankAdapter.d.ts +58 -0
  192. package/dist/core/memory/ReasoningBankAdapter.d.ts.map +1 -0
  193. package/dist/core/memory/ReasoningBankAdapter.js +80 -0
  194. package/dist/core/memory/ReasoningBankAdapter.js.map +1 -0
  195. package/dist/core/memory/SwarmMemoryManager.d.ts +75 -3
  196. package/dist/core/memory/SwarmMemoryManager.d.ts.map +1 -1
  197. package/dist/core/memory/SwarmMemoryManager.js +236 -52
  198. package/dist/core/memory/SwarmMemoryManager.js.map +1 -1
  199. package/dist/core/memory/index.d.ts +4 -0
  200. package/dist/core/memory/index.d.ts.map +1 -1
  201. package/dist/core/memory/index.js +9 -1
  202. package/dist/core/memory/index.js.map +1 -1
  203. package/dist/core/neural/NeuralTrainer.d.ts +137 -0
  204. package/dist/core/neural/NeuralTrainer.d.ts.map +1 -0
  205. package/dist/core/neural/NeuralTrainer.js +543 -0
  206. package/dist/core/neural/NeuralTrainer.js.map +1 -0
  207. package/dist/core/neural/index.d.ts +8 -0
  208. package/dist/core/neural/index.d.ts.map +1 -0
  209. package/dist/core/neural/index.js +24 -0
  210. package/dist/core/neural/index.js.map +1 -0
  211. package/dist/core/neural/types.d.ts +216 -0
  212. package/dist/core/neural/types.d.ts.map +1 -0
  213. package/dist/core/neural/types.js +8 -0
  214. package/dist/core/neural/types.js.map +1 -0
  215. package/dist/core/security/CertificateValidator.d.ts +130 -0
  216. package/dist/core/security/CertificateValidator.d.ts.map +1 -0
  217. package/dist/core/security/CertificateValidator.js +376 -0
  218. package/dist/core/security/CertificateValidator.js.map +1 -0
  219. package/dist/core/transport/QUICTransport.d.ts +62 -0
  220. package/dist/core/transport/QUICTransport.d.ts.map +1 -0
  221. package/dist/core/transport/QUICTransport.js +381 -0
  222. package/dist/core/transport/QUICTransport.js.map +1 -0
  223. package/dist/core/transport/SecureQUICTransport.d.ts +71 -0
  224. package/dist/core/transport/SecureQUICTransport.d.ts.map +1 -0
  225. package/dist/core/transport/SecureQUICTransport.js +253 -0
  226. package/dist/core/transport/SecureQUICTransport.js.map +1 -0
  227. package/dist/learning/AdvancedFeatureExtractor.d.ts +123 -0
  228. package/dist/learning/AdvancedFeatureExtractor.d.ts.map +1 -0
  229. package/dist/learning/AdvancedFeatureExtractor.js +423 -0
  230. package/dist/learning/AdvancedFeatureExtractor.js.map +1 -0
  231. package/dist/learning/FlakyPredictionModel.d.ts +11 -1
  232. package/dist/learning/FlakyPredictionModel.d.ts.map +1 -1
  233. package/dist/learning/FlakyPredictionModel.js +82 -35
  234. package/dist/learning/FlakyPredictionModel.js.map +1 -1
  235. package/dist/learning/FlakyTestDetector.d.ts +9 -0
  236. package/dist/learning/FlakyTestDetector.d.ts.map +1 -1
  237. package/dist/learning/FlakyTestDetector.js +28 -6
  238. package/dist/learning/FlakyTestDetector.js.map +1 -1
  239. package/dist/learning/ImprovementLoop.d.ts +16 -2
  240. package/dist/learning/ImprovementLoop.d.ts.map +1 -1
  241. package/dist/learning/ImprovementLoop.js +67 -8
  242. package/dist/learning/ImprovementLoop.js.map +1 -1
  243. package/dist/learning/ImprovementWorker.d.ts +83 -0
  244. package/dist/learning/ImprovementWorker.d.ts.map +1 -0
  245. package/dist/learning/ImprovementWorker.js +164 -0
  246. package/dist/learning/ImprovementWorker.js.map +1 -0
  247. package/dist/learning/NeuralPatternMatcher.d.ts +184 -0
  248. package/dist/learning/NeuralPatternMatcher.d.ts.map +1 -0
  249. package/dist/learning/NeuralPatternMatcher.js +702 -0
  250. package/dist/learning/NeuralPatternMatcher.js.map +1 -0
  251. package/dist/learning/NeuralTrainer.d.ts +209 -0
  252. package/dist/learning/NeuralTrainer.d.ts.map +1 -0
  253. package/dist/learning/NeuralTrainer.js +478 -0
  254. package/dist/learning/NeuralTrainer.js.map +1 -0
  255. package/dist/learning/index.d.ts +13 -7
  256. package/dist/learning/index.d.ts.map +1 -1
  257. package/dist/learning/index.js +27 -11
  258. package/dist/learning/index.js.map +1 -1
  259. package/dist/learning/types.d.ts +1 -0
  260. package/dist/learning/types.d.ts.map +1 -1
  261. package/dist/learning/types.js +16 -0
  262. package/dist/learning/types.js.map +1 -1
  263. package/dist/mcp/MCPToolRegistry.d.ts +34 -0
  264. package/dist/mcp/MCPToolRegistry.d.ts.map +1 -0
  265. package/dist/mcp/MCPToolRegistry.js +48 -0
  266. package/dist/mcp/MCPToolRegistry.js.map +1 -0
  267. package/dist/mcp/server.d.ts +4 -4
  268. package/dist/mcp/services/AgentRegistry.d.ts.map +1 -1
  269. package/dist/mcp/services/AgentRegistry.js +7 -1
  270. package/dist/mcp/services/AgentRegistry.js.map +1 -1
  271. package/dist/transport/QUICTransport.d.ts +340 -0
  272. package/dist/transport/QUICTransport.d.ts.map +1 -0
  273. package/dist/transport/QUICTransport.js +814 -0
  274. package/dist/transport/QUICTransport.js.map +1 -0
  275. package/dist/transport/UDPTransport.d.ts +348 -0
  276. package/dist/transport/UDPTransport.d.ts.map +1 -0
  277. package/dist/transport/UDPTransport.js +820 -0
  278. package/dist/transport/UDPTransport.js.map +1 -0
  279. package/dist/types/errors.d.ts +1 -1
  280. package/dist/types/index.d.ts +37 -0
  281. package/dist/types/index.d.ts.map +1 -1
  282. package/dist/types/index.js +2 -0
  283. package/dist/types/index.js.map +1 -1
  284. package/dist/types/quic.d.ts +339 -0
  285. package/dist/types/quic.d.ts.map +1 -0
  286. package/dist/types/quic.js +48 -0
  287. package/dist/types/quic.js.map +1 -0
  288. package/dist/utils/Config.js +1 -1
  289. package/dist/utils/Config.js.map +1 -1
  290. package/dist/utils/Database.d.ts +14 -0
  291. package/dist/utils/Database.d.ts.map +1 -1
  292. package/dist/utils/Database.js +51 -4
  293. package/dist/utils/Database.js.map +1 -1
  294. package/dist/utils/Logger.d.ts.map +1 -1
  295. package/dist/utils/Logger.js +111 -26
  296. package/dist/utils/Logger.js.map +1 -1
  297. package/package.json +30 -7
@@ -0,0 +1,545 @@
1
+ ---
2
+ name: "AgentDB Learning Plugins"
3
+ description: "Create and train AI learning plugins with AgentDB's 9 reinforcement learning algorithms. Includes Decision Transformer, Q-Learning, SARSA, Actor-Critic, and more. Use when building self-learning agents, implementing RL, or optimizing agent behavior through experience."
4
+ ---
5
+
6
+ # AgentDB Learning Plugins
7
+
8
+ ## What This Skill Does
9
+
10
+ Provides access to 9 reinforcement learning algorithms via AgentDB's plugin system. Create, train, and deploy learning plugins for autonomous agents that improve through experience. Includes offline RL (Decision Transformer), value-based learning (Q-Learning), policy gradients (Actor-Critic), and advanced techniques.
11
+
12
+ **Performance**: Train models 10-100x faster with WASM-accelerated neural inference.
13
+
14
+ ## Prerequisites
15
+
16
+ - Node.js 18+
17
+ - AgentDB v1.0.7+ (via agentic-flow)
18
+ - Basic understanding of reinforcement learning (recommended)
19
+
20
+ ---
21
+
22
+ ## Quick Start with CLI
23
+
24
+ ### Create Learning Plugin
25
+
26
+ ```bash
27
+ # Interactive wizard
28
+ npx agentdb@latest create-plugin
29
+
30
+ # Use specific template
31
+ npx agentdb@latest create-plugin -t decision-transformer -n my-agent
32
+
33
+ # Preview without creating
34
+ npx agentdb@latest create-plugin -t q-learning --dry-run
35
+
36
+ # Custom output directory
37
+ npx agentdb@latest create-plugin -t actor-critic -o ./plugins
38
+ ```
39
+
40
+ ### List Available Templates
41
+
42
+ ```bash
43
+ # Show all plugin templates
44
+ npx agentdb@latest list-templates
45
+
46
+ # Available templates:
47
+ # - decision-transformer (sequence modeling RL - recommended)
48
+ # - q-learning (value-based learning)
49
+ # - sarsa (on-policy TD learning)
50
+ # - actor-critic (policy gradient with baseline)
51
+ # - curiosity-driven (exploration-based)
52
+ ```
53
+
54
+ ### Manage Plugins
55
+
56
+ ```bash
57
+ # List installed plugins
58
+ npx agentdb@latest list-plugins
59
+
60
+ # Get plugin information
61
+ npx agentdb@latest plugin-info my-agent
62
+
63
+ # Shows: algorithm, configuration, training status
64
+ ```
65
+
66
+ ---
67
+
68
+ ## Quick Start with API
69
+
70
+ ```typescript
71
+ import { createAgentDBAdapter } from 'agentic-flow/reasoningbank';
72
+
73
+ // Initialize with learning enabled
74
+ const adapter = await createAgentDBAdapter({
75
+ dbPath: '.agentdb/learning.db',
76
+ enableLearning: true, // Enable learning plugins
77
+ enableReasoning: true,
78
+ cacheSize: 1000,
79
+ });
80
+
81
+ // Store training experience
82
+ await adapter.insertPattern({
83
+ id: '',
84
+ type: 'experience',
85
+ domain: 'game-playing',
86
+ pattern_data: JSON.stringify({
87
+ embedding: await computeEmbedding('state-action-reward'),
88
+ pattern: {
89
+ state: [0.1, 0.2, 0.3],
90
+ action: 2,
91
+ reward: 1.0,
92
+ next_state: [0.15, 0.25, 0.35],
93
+ done: false
94
+ }
95
+ }),
96
+ confidence: 0.9,
97
+ usage_count: 1,
98
+ success_count: 1,
99
+ created_at: Date.now(),
100
+ last_used: Date.now(),
101
+ });
102
+
103
+ // Train learning model
104
+ const metrics = await adapter.train({
105
+ epochs: 50,
106
+ batchSize: 32,
107
+ });
108
+
109
+ console.log('Training Loss:', metrics.loss);
110
+ console.log('Duration:', metrics.duration, 'ms');
111
+ ```
112
+
113
+ ---
114
+
115
+ ## Available Learning Algorithms (9 Total)
116
+
117
+ ### 1. Decision Transformer (Recommended)
118
+
119
+ **Type**: Offline Reinforcement Learning
120
+ **Best For**: Learning from logged experiences, imitation learning
121
+ **Strengths**: No online interaction needed, stable training
122
+
123
+ ```bash
124
+ npx agentdb@latest create-plugin -t decision-transformer -n dt-agent
125
+ ```
126
+
127
+ **Use Cases**:
128
+ - Learn from historical data
129
+ - Imitation learning from expert demonstrations
130
+ - Safe learning without environment interaction
131
+ - Sequence modeling tasks
132
+
133
+ **Configuration**:
134
+ ```json
135
+ {
136
+ "algorithm": "decision-transformer",
137
+ "model_size": "base",
138
+ "context_length": 20,
139
+ "embed_dim": 128,
140
+ "n_heads": 8,
141
+ "n_layers": 6
142
+ }
143
+ ```
144
+
145
+ ### 2. Q-Learning
146
+
147
+ **Type**: Value-Based RL (Off-Policy)
148
+ **Best For**: Discrete action spaces, sample efficiency
149
+ **Strengths**: Proven, simple, works well for small/medium problems
150
+
151
+ ```bash
152
+ npx agentdb@latest create-plugin -t q-learning -n q-agent
153
+ ```
154
+
155
+ **Use Cases**:
156
+ - Grid worlds, board games
157
+ - Navigation tasks
158
+ - Resource allocation
159
+ - Discrete decision-making
160
+
161
+ **Configuration**:
162
+ ```json
163
+ {
164
+ "algorithm": "q-learning",
165
+ "learning_rate": 0.001,
166
+ "gamma": 0.99,
167
+ "epsilon": 0.1,
168
+ "epsilon_decay": 0.995
169
+ }
170
+ ```
171
+
172
+ ### 3. SARSA
173
+
174
+ **Type**: Value-Based RL (On-Policy)
175
+ **Best For**: Safe exploration, risk-sensitive tasks
176
+ **Strengths**: More conservative than Q-Learning, better for safety
177
+
178
+ ```bash
179
+ npx agentdb@latest create-plugin -t sarsa -n sarsa-agent
180
+ ```
181
+
182
+ **Use Cases**:
183
+ - Safety-critical applications
184
+ - Risk-sensitive decision-making
185
+ - Online learning with exploration
186
+
187
+ **Configuration**:
188
+ ```json
189
+ {
190
+ "algorithm": "sarsa",
191
+ "learning_rate": 0.001,
192
+ "gamma": 0.99,
193
+ "epsilon": 0.1
194
+ }
195
+ ```
196
+
197
+ ### 4. Actor-Critic
198
+
199
+ **Type**: Policy Gradient with Value Baseline
200
+ **Best For**: Continuous actions, variance reduction
201
+ **Strengths**: Stable, works for continuous/discrete actions
202
+
203
+ ```bash
204
+ npx agentdb@latest create-plugin -t actor-critic -n ac-agent
205
+ ```
206
+
207
+ **Use Cases**:
208
+ - Continuous control (robotics, simulations)
209
+ - Complex action spaces
210
+ - Multi-agent coordination
211
+
212
+ **Configuration**:
213
+ ```json
214
+ {
215
+ "algorithm": "actor-critic",
216
+ "actor_lr": 0.001,
217
+ "critic_lr": 0.002,
218
+ "gamma": 0.99,
219
+ "entropy_coef": 0.01
220
+ }
221
+ ```
222
+
223
+ ### 5. Active Learning
224
+
225
+ **Type**: Query-Based Learning
226
+ **Best For**: Label-efficient learning, human-in-the-loop
227
+ **Strengths**: Minimizes labeling cost, focuses on uncertain samples
228
+
229
+ **Use Cases**:
230
+ - Human feedback incorporation
231
+ - Label-efficient training
232
+ - Uncertainty sampling
233
+ - Annotation cost reduction
234
+
235
+ ### 6. Adversarial Training
236
+
237
+ **Type**: Robustness Enhancement
238
+ **Best For**: Safety, robustness to perturbations
239
+ **Strengths**: Improves model robustness, adversarial defense
240
+
241
+ **Use Cases**:
242
+ - Security applications
243
+ - Robust decision-making
244
+ - Adversarial defense
245
+ - Safety testing
246
+
247
+ ### 7. Curriculum Learning
248
+
249
+ **Type**: Progressive Difficulty Training
250
+ **Best For**: Complex tasks, faster convergence
251
+ **Strengths**: Stable learning, faster convergence on hard tasks
252
+
253
+ **Use Cases**:
254
+ - Complex multi-stage tasks
255
+ - Hard exploration problems
256
+ - Skill composition
257
+ - Transfer learning
258
+
259
+ ### 8. Federated Learning
260
+
261
+ **Type**: Distributed Learning
262
+ **Best For**: Privacy, distributed data
263
+ **Strengths**: Privacy-preserving, scalable
264
+
265
+ **Use Cases**:
266
+ - Multi-agent systems
267
+ - Privacy-sensitive data
268
+ - Distributed training
269
+ - Collaborative learning
270
+
271
+ ### 9. Multi-Task Learning
272
+
273
+ **Type**: Transfer Learning
274
+ **Best For**: Related tasks, knowledge sharing
275
+ **Strengths**: Faster learning on new tasks, better generalization
276
+
277
+ **Use Cases**:
278
+ - Task families
279
+ - Transfer learning
280
+ - Domain adaptation
281
+ - Meta-learning
282
+
283
+ ---
284
+
285
+ ## Training Workflow
286
+
287
+ ### 1. Collect Experiences
288
+
289
+ ```typescript
290
+ // Store experiences during agent execution
291
+ for (let i = 0; i < numEpisodes; i++) {
292
+ const episode = runEpisode();
293
+
294
+ for (const step of episode.steps) {
295
+ await adapter.insertPattern({
296
+ id: '',
297
+ type: 'experience',
298
+ domain: 'task-domain',
299
+ pattern_data: JSON.stringify({
300
+ embedding: await computeEmbedding(JSON.stringify(step)),
301
+ pattern: {
302
+ state: step.state,
303
+ action: step.action,
304
+ reward: step.reward,
305
+ next_state: step.next_state,
306
+ done: step.done
307
+ }
308
+ }),
309
+ confidence: step.reward > 0 ? 0.9 : 0.5,
310
+ usage_count: 1,
311
+ success_count: step.reward > 0 ? 1 : 0,
312
+ created_at: Date.now(),
313
+ last_used: Date.now(),
314
+ });
315
+ }
316
+ }
317
+ ```
318
+
319
+ ### 2. Train Model
320
+
321
+ ```typescript
322
+ // Train on collected experiences
323
+ const trainingMetrics = await adapter.train({
324
+ epochs: 100,
325
+ batchSize: 64,
326
+ learningRate: 0.001,
327
+ validationSplit: 0.2,
328
+ });
329
+
330
+ console.log('Training Metrics:', trainingMetrics);
331
+ // {
332
+ // loss: 0.023,
333
+ // valLoss: 0.028,
334
+ // duration: 1523,
335
+ // epochs: 100
336
+ // }
337
+ ```
338
+
339
+ ### 3. Evaluate Performance
340
+
341
+ ```typescript
342
+ // Retrieve similar successful experiences
343
+ const testQuery = await computeEmbedding(JSON.stringify(testState));
344
+ const result = await adapter.retrieveWithReasoning(testQuery, {
345
+ domain: 'task-domain',
346
+ k: 10,
347
+ synthesizeContext: true,
348
+ });
349
+
350
+ // Evaluate action quality
351
+ const suggestedAction = result.memories[0].pattern.action;
352
+ const confidence = result.memories[0].similarity;
353
+
354
+ console.log('Suggested Action:', suggestedAction);
355
+ console.log('Confidence:', confidence);
356
+ ```
357
+
358
+ ---
359
+
360
+ ## Advanced Training Techniques
361
+
362
+ ### Experience Replay
363
+
364
+ ```typescript
365
+ // Store experiences in buffer
366
+ const replayBuffer = [];
367
+
368
+ // Sample random batch for training
369
+ const batch = sampleRandomBatch(replayBuffer, batchSize: 32);
370
+
371
+ // Train on batch
372
+ await adapter.train({
373
+ data: batch,
374
+ epochs: 1,
375
+ batchSize: 32,
376
+ });
377
+ ```
378
+
379
+ ### Prioritized Experience Replay
380
+
381
+ ```typescript
382
+ // Store experiences with priority (TD error)
383
+ await adapter.insertPattern({
384
+ // ... standard fields
385
+ confidence: tdError, // Use TD error as confidence/priority
386
+ // ...
387
+ });
388
+
389
+ // Retrieve high-priority experiences
390
+ const highPriority = await adapter.retrieveWithReasoning(queryEmbedding, {
391
+ domain: 'task-domain',
392
+ k: 32,
393
+ minConfidence: 0.7, // Only high TD-error experiences
394
+ });
395
+ ```
396
+
397
+ ### Multi-Agent Training
398
+
399
+ ```typescript
400
+ // Collect experiences from multiple agents
401
+ for (const agent of agents) {
402
+ const experience = await agent.step();
403
+
404
+ await adapter.insertPattern({
405
+ // ... store experience with agent ID
406
+ domain: `multi-agent/${agent.id}`,
407
+ });
408
+ }
409
+
410
+ // Train shared model
411
+ await adapter.train({
412
+ epochs: 50,
413
+ batchSize: 64,
414
+ });
415
+ ```
416
+
417
+ ---
418
+
419
+ ## Performance Optimization
420
+
421
+ ### Batch Training
422
+
423
+ ```typescript
424
+ // Collect batch of experiences
425
+ const experiences = collectBatch(size: 1000);
426
+
427
+ // Batch insert (500x faster)
428
+ for (const exp of experiences) {
429
+ await adapter.insertPattern({ /* ... */ });
430
+ }
431
+
432
+ // Train on batch
433
+ await adapter.train({
434
+ epochs: 10,
435
+ batchSize: 128, // Larger batch for efficiency
436
+ });
437
+ ```
438
+
439
+ ### Incremental Learning
440
+
441
+ ```typescript
442
+ // Train incrementally as new data arrives
443
+ setInterval(async () => {
444
+ const newExperiences = getNewExperiences();
445
+
446
+ if (newExperiences.length > 100) {
447
+ await adapter.train({
448
+ epochs: 5,
449
+ batchSize: 32,
450
+ });
451
+ }
452
+ }, 60000); // Every minute
453
+ ```
454
+
455
+ ---
456
+
457
+ ## Integration with Reasoning Agents
458
+
459
+ Combine learning with reasoning for better performance:
460
+
461
+ ```typescript
462
+ // Train learning model
463
+ await adapter.train({ epochs: 50, batchSize: 32 });
464
+
465
+ // Use reasoning agents for inference
466
+ const result = await adapter.retrieveWithReasoning(queryEmbedding, {
467
+ domain: 'decision-making',
468
+ k: 10,
469
+ useMMR: true, // Diverse experiences
470
+ synthesizeContext: true, // Rich context
471
+ optimizeMemory: true, // Consolidate patterns
472
+ });
473
+
474
+ // Make decision based on learned experiences + reasoning
475
+ const decision = result.context.suggestedAction;
476
+ const confidence = result.memories[0].similarity;
477
+ ```
478
+
479
+ ---
480
+
481
+ ## CLI Operations
482
+
483
+ ```bash
484
+ # Create plugin
485
+ npx agentdb@latest create-plugin -t decision-transformer -n my-plugin
486
+
487
+ # List plugins
488
+ npx agentdb@latest list-plugins
489
+
490
+ # Get plugin info
491
+ npx agentdb@latest plugin-info my-plugin
492
+
493
+ # List templates
494
+ npx agentdb@latest list-templates
495
+ ```
496
+
497
+ ---
498
+
499
+ ## Troubleshooting
500
+
501
+ ### Issue: Training not converging
502
+ ```typescript
503
+ // Reduce learning rate
504
+ await adapter.train({
505
+ epochs: 100,
506
+ batchSize: 32,
507
+ learningRate: 0.0001, // Lower learning rate
508
+ });
509
+ ```
510
+
511
+ ### Issue: Overfitting
512
+ ```typescript
513
+ // Use validation split
514
+ await adapter.train({
515
+ epochs: 50,
516
+ batchSize: 64,
517
+ validationSplit: 0.2, // 20% validation
518
+ });
519
+
520
+ // Enable memory optimization
521
+ await adapter.retrieveWithReasoning(queryEmbedding, {
522
+ optimizeMemory: true, // Consolidate, reduce overfitting
523
+ });
524
+ ```
525
+
526
+ ### Issue: Slow training
527
+ ```bash
528
+ # Enable quantization for faster inference
529
+ # Use binary quantization (32x faster)
530
+ ```
531
+
532
+ ---
533
+
534
+ ## Learn More
535
+
536
+ - **Algorithm Papers**: See docs/algorithms/ for detailed papers
537
+ - **GitHub**: https://github.com/ruvnet/agentic-flow/tree/main/packages/agentdb
538
+ - **MCP Integration**: `npx agentdb@latest mcp`
539
+ - **Website**: https://agentdb.ruv.io
540
+
541
+ ---
542
+
543
+ **Category**: Machine Learning / Reinforcement Learning
544
+ **Difficulty**: Intermediate to Advanced
545
+ **Estimated Time**: 30-60 minutes