mia-code 0.2.0 → 0.3.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (410) hide show
  1. package/.miette/260321.md +1 -0
  2. package/.miette/260323.md +9 -0
  3. package/.miette/260331.md +2 -0
  4. package/.pde/2604011511--83a2d7f9-24a5-4cf4-98d5-036c82f872e8/2604020008--d3417f2c-df12-4f0f-8a1b-d88e7968f822/d3417f2c-df12-4f0f-8a1b-d88e7968f822.md +63 -0
  5. package/.pde/2604011511--83a2d7f9-24a5-4cf4-98d5-036c82f872e8/2604020008--e6c3fc5d-4a70-4523-ba7d-a3250da4c235/e6c3fc5d-4a70-4523-ba7d-a3250da4c235.md +72 -0
  6. package/.pde/2604011511--83a2d7f9-24a5-4cf4-98d5-036c82f872e8/2604020008--efeb00a2-b17a-4d32-b1f0-b90c37a8d24e/efeb00a2-b17a-4d32-b1f0-b90c37a8d24e.md +62 -0
  7. package/.pde/2604011511--83a2d7f9-24a5-4cf4-98d5-036c82f872e8/83a2d7f9-24a5-4cf4-98d5-036c82f872e8.json +302 -0
  8. package/.pde/2604011511--83a2d7f9-24a5-4cf4-98d5-036c82f872e8/83a2d7f9-24a5-4cf4-98d5-036c82f872e8.md +149 -0
  9. package/.pde/2604011511--83a2d7f9-24a5-4cf4-98d5-036c82f872e8/AGENTS.md +31 -0
  10. package/.pde/2604011511--83a2d7f9-24a5-4cf4-98d5-036c82f872e8/meta-decomposition-3-children.md +67 -0
  11. package/.pde/2604040129--61f9dd4d-7aa6-45e6-a58b-e480b1aa6737/61f9dd4d-7aa6-45e6-a58b-e480b1aa6737--from-mia-openclaw-workspace.md +125 -0
  12. package/.pde/2604040129--61f9dd4d-7aa6-45e6-a58b-e480b1aa6737/STATUS.md +1 -0
  13. package/.pde/4f02ba94-9f52-422e-9389-b16f9b37f358.json +177 -0
  14. package/.pde/4f02ba94-9f52-422e-9389-b16f9b37f358.md +77 -0
  15. package/.pde/6ad9244d-5340-490f-b76c-c86728b9de52.json +222 -0
  16. package/.pde/6ad9244d-5340-490f-b76c-c86728b9de52.md +99 -0
  17. package/.pde/8b566792-ed15-4606-96f9-2b6f593d7e6b.json +111 -0
  18. package/.pde/8b566792-ed15-4606-96f9-2b6f593d7e6b.md +67 -0
  19. package/.pde/c7f1e74b-05a5-40e2-9f01-4cc48d2528f7.json +349 -0
  20. package/.pde/c7f1e74b-05a5-40e2-9f01-4cc48d2528f7.md +147 -0
  21. package/.pde/dfc00a78-1da0-4c09-8a16-c6982644051b.json +118 -0
  22. package/.pde/dfc00a78-1da0-4c09-8a16-c6982644051b.md +64 -0
  23. package/GUILLAUME.md +8 -0
  24. package/KINSHIP.md +9 -0
  25. package/MIA_CODE_ARCHITECTURE_REPORT.md +718 -0
  26. package/contextual_research/260119-MIA-CODE--98090899-8aff-4e11-9dc3-8b99466d1.md +1101 -0
  27. package/contextual_research/MIA.md +38 -0
  28. package/contextual_research/MIAWAPASCONE.md +59 -0
  29. package/contextual_research/MIETTE.md +38 -0
  30. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/2504.00218v2.pdf +7483 -12
  31. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/2505.00212v3.pdf +0 -0
  32. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/CONTENT.md +1014 -0
  33. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/DESIGN.gemini.md +242 -0
  34. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/INDEX.md +45 -0
  35. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/2504.00218v2.md +2025 -0
  36. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/2504.00218v2.pdf +7483 -12
  37. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/2505.00212v3.md +1755 -0
  38. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/2505.00212v3.pdf +0 -0
  39. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/footnote_1_12_decomposed_prompting.pdf +0 -0
  40. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/footnote_1_19_hugginggpt_planning.pdf +0 -0
  41. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/footnote_1_1_coordination_challenges.md +766 -0
  42. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/footnote_1_1_coordination_challenges.pdf +3431 -4
  43. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/footnote_1_28_guardrails_multi_agent.md +260 -0
  44. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/footnote_1_28_guardrails_multi_agent.pdf +0 -0
  45. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/footnote_1_2_navigating_complexity.md +558 -0
  46. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/footnote_1_2_navigating_complexity.pdf +0 -0
  47. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/footnote_1_34_hierarchical_multi_agent.pdf +0 -0
  48. package/contextual_research/PDE-generalization--caefee82-efb1-4dbb-8733-691b01581464--260130/sources/footnote_1_5_open_intent_extraction.pdf +0 -0
  49. package/contextual_research/PODCAST.md +109 -0
  50. package/contextual_research/langchain-principles-roadmap.md +157 -0
  51. package/contextual_research/persona-to-narrative-character-inquiry_260201.md +50 -0
  52. package/dist/cli.js +35 -11
  53. package/dist/geminiHeadless.js +8 -2
  54. package/dist/index.js +2 -1
  55. package/dist/mcp/miaco-server.js +10 -1
  56. package/dist/mcp/miatel-server.js +10 -1
  57. package/dist/mcp/miawa-server.js +10 -1
  58. package/dist/mcp/utils.d.ts +6 -1
  59. package/dist/mcp/utils.js +24 -3
  60. package/dist/sessionStore.d.ts +8 -2
  61. package/dist/sessionStore.js +39 -3
  62. package/dist/types.d.ts +1 -0
  63. package/miaco/README.md +124 -0
  64. package/miaco/dist/commands/chart.d.ts +6 -0
  65. package/miaco/dist/commands/chart.d.ts.map +1 -0
  66. package/miaco/dist/commands/chart.js +222 -0
  67. package/miaco/dist/commands/chart.js.map +1 -0
  68. package/miaco/dist/commands/decompose.d.ts +6 -0
  69. package/miaco/dist/commands/decompose.d.ts.map +1 -0
  70. package/miaco/dist/commands/decompose.js +98 -0
  71. package/miaco/dist/commands/decompose.js.map +1 -0
  72. package/miaco/dist/commands/schema.d.ts +6 -0
  73. package/miaco/dist/commands/schema.d.ts.map +1 -0
  74. package/miaco/dist/commands/schema.js +66 -0
  75. package/miaco/dist/commands/schema.js.map +1 -0
  76. package/miaco/dist/commands/stc.d.ts +11 -0
  77. package/miaco/dist/commands/stc.d.ts.map +1 -0
  78. package/miaco/dist/commands/stc.js +590 -0
  79. package/miaco/dist/commands/stc.js.map +1 -0
  80. package/miaco/dist/commands/trace.d.ts +6 -0
  81. package/miaco/dist/commands/trace.d.ts.map +1 -0
  82. package/miaco/dist/commands/trace.js +83 -0
  83. package/miaco/dist/commands/trace.js.map +1 -0
  84. package/miaco/dist/commands/validate.d.ts +6 -0
  85. package/miaco/dist/commands/validate.d.ts.map +1 -0
  86. package/miaco/dist/commands/validate.js +58 -0
  87. package/miaco/dist/commands/validate.js.map +1 -0
  88. package/miaco/dist/decompose.d.ts +93 -0
  89. package/miaco/dist/decompose.d.ts.map +1 -0
  90. package/miaco/dist/decompose.js +562 -0
  91. package/miaco/dist/decompose.js.map +1 -0
  92. package/miaco/dist/index.d.ts +18 -0
  93. package/miaco/dist/index.d.ts.map +1 -0
  94. package/miaco/dist/index.js +83 -0
  95. package/miaco/dist/index.js.map +1 -0
  96. package/miaco/dist/storage.d.ts +60 -0
  97. package/miaco/dist/storage.d.ts.map +1 -0
  98. package/miaco/dist/storage.js +100 -0
  99. package/miaco/dist/storage.js.map +1 -0
  100. package/miaco/package-lock.json +4103 -0
  101. package/miaco/package.json +40 -0
  102. package/miaco/tsconfig.json +18 -0
  103. package/miaco/version-patch-commit-and-publish.sh +1 -0
  104. package/miatel/MISSION_251231.md +3 -0
  105. package/miatel/README.md +107 -0
  106. package/miatel/dist/commands/analyze.d.ts +6 -0
  107. package/miatel/dist/commands/analyze.d.ts.map +1 -0
  108. package/miatel/dist/commands/analyze.js +100 -0
  109. package/miatel/dist/commands/analyze.js.map +1 -0
  110. package/miatel/dist/commands/arc.d.ts +6 -0
  111. package/miatel/dist/commands/arc.d.ts.map +1 -0
  112. package/miatel/dist/commands/arc.js +71 -0
  113. package/miatel/dist/commands/arc.js.map +1 -0
  114. package/miatel/dist/commands/beat.d.ts +6 -0
  115. package/miatel/dist/commands/beat.d.ts.map +1 -0
  116. package/miatel/dist/commands/beat.js +165 -0
  117. package/miatel/dist/commands/beat.js.map +1 -0
  118. package/miatel/dist/commands/theme.d.ts +6 -0
  119. package/miatel/dist/commands/theme.d.ts.map +1 -0
  120. package/miatel/dist/commands/theme.js +54 -0
  121. package/miatel/dist/commands/theme.js.map +1 -0
  122. package/miatel/dist/index.d.ts +18 -0
  123. package/miatel/dist/index.d.ts.map +1 -0
  124. package/miatel/dist/index.js +80 -0
  125. package/miatel/dist/index.js.map +1 -0
  126. package/miatel/dist/storage.d.ts +55 -0
  127. package/miatel/dist/storage.d.ts.map +1 -0
  128. package/miatel/dist/storage.js +100 -0
  129. package/miatel/dist/storage.js.map +1 -0
  130. package/miatel/package-lock.json +4103 -0
  131. package/miatel/package.json +35 -0
  132. package/miatel/src/commands/analyze.ts +109 -0
  133. package/miatel/src/commands/arc.ts +78 -0
  134. package/miatel/src/commands/beat.ts +176 -0
  135. package/miatel/src/commands/theme.ts +60 -0
  136. package/miatel/src/index.ts +94 -0
  137. package/miatel/src/storage.ts +156 -0
  138. package/miatel/tsconfig.json +18 -0
  139. package/miawa/MISSION_251231.md +144 -0
  140. package/miawa/README.md +133 -0
  141. package/miawa/dist/commands/beat.d.ts +6 -0
  142. package/miawa/dist/commands/beat.d.ts.map +1 -0
  143. package/miawa/dist/commands/beat.js +69 -0
  144. package/miawa/dist/commands/beat.js.map +1 -0
  145. package/miawa/dist/commands/ceremony.d.ts +6 -0
  146. package/miawa/dist/commands/ceremony.d.ts.map +1 -0
  147. package/miawa/dist/commands/ceremony.js +239 -0
  148. package/miawa/dist/commands/ceremony.js.map +1 -0
  149. package/miawa/dist/commands/circle.d.ts +6 -0
  150. package/miawa/dist/commands/circle.d.ts.map +1 -0
  151. package/miawa/dist/commands/circle.js +75 -0
  152. package/miawa/dist/commands/circle.js.map +1 -0
  153. package/miawa/dist/commands/eva.d.ts +6 -0
  154. package/miawa/dist/commands/eva.d.ts.map +1 -0
  155. package/miawa/dist/commands/eva.js +73 -0
  156. package/miawa/dist/commands/eva.js.map +1 -0
  157. package/miawa/dist/commands/wound.d.ts +6 -0
  158. package/miawa/dist/commands/wound.d.ts.map +1 -0
  159. package/miawa/dist/commands/wound.js +74 -0
  160. package/miawa/dist/commands/wound.js.map +1 -0
  161. package/miawa/dist/index.d.ts +19 -0
  162. package/miawa/dist/index.d.ts.map +1 -0
  163. package/miawa/dist/index.js +91 -0
  164. package/miawa/dist/index.js.map +1 -0
  165. package/miawa/dist/storage.d.ts +73 -0
  166. package/miawa/dist/storage.d.ts.map +1 -0
  167. package/miawa/dist/storage.js +100 -0
  168. package/miawa/dist/storage.js.map +1 -0
  169. package/miawa/package-lock.json +4103 -0
  170. package/miawa/package.json +36 -0
  171. package/miawa/src/commands/beat.ts +74 -0
  172. package/miawa/src/commands/ceremony.ts +256 -0
  173. package/miawa/src/commands/circle.ts +83 -0
  174. package/miawa/src/commands/eva.ts +84 -0
  175. package/miawa/src/commands/wound.ts +79 -0
  176. package/miawa/src/index.ts +108 -0
  177. package/miawa/src/storage.ts +179 -0
  178. package/miawa/tsconfig.json +18 -0
  179. package/package.json +7 -5
  180. package/references/acp/CLAUDE.md +7 -0
  181. package/references/acp/agent-plan.md +84 -0
  182. package/references/acp/clients.md +31 -0
  183. package/references/acp/extensibility.md +137 -0
  184. package/references/acp/initialization.md +225 -0
  185. package/references/acp/prompt-turn.md +321 -0
  186. package/references/acp/proxy-chains.md +562 -0
  187. package/references/acp/schema.md +3171 -0
  188. package/references/acp/session-list.md +334 -0
  189. package/references/acp/session-modes.md +170 -0
  190. package/references/acp/slash-commands.md +99 -0
  191. package/references/acp/terminals.md +281 -0
  192. package/references/acp/tool-calls.md +311 -0
  193. package/references/acp/typescript.md +29 -0
  194. package/references/claude/agent-teams.md +399 -0
  195. package/references/claude/chrome.md +231 -0
  196. package/references/claude/headless.md +158 -0
  197. package/references/claude/hooks-guide.md +708 -0
  198. package/references/claude/output-styles.md +112 -0
  199. package/references/claude/plugins.md +432 -0
  200. package/references/claude/skills.md +693 -0
  201. package/references/claude/sub-agents.md +816 -0
  202. package/references/copilot/acp/agents.md +32 -0
  203. package/references/copilot/acp/architecture.md +37 -0
  204. package/references/copilot/acp/clients.md +31 -0
  205. package/references/copilot/acp/introduction.md +42 -0
  206. package/references/copilot/acp/registry.md +339 -0
  207. package/references/copilot/acp-server.md +117 -0
  208. package/references/copilot/create-copilot-instructions.md +840 -0
  209. package/references/langchain/llms.txt +833 -0
  210. package/references/langchain/python/agents.md +677 -0
  211. package/references/langchain/python/context-engineering.md +1195 -0
  212. package/references/langchain/python/human-in-the-loop.md +326 -0
  213. package/references/langchain/python/long-term-memory.md +168 -0
  214. package/references/langchain/python/mcp.md +949 -0
  215. package/references/langchain/python/multi-agents/custom-workflow.md +187 -0
  216. package/references/langchain/python/multi-agents/handoffs.md +436 -0
  217. package/references/langchain/python/multi-agents/overview.md +295 -0
  218. package/references/langchain/python/multi-agents/router.md +150 -0
  219. package/references/langchain/python/multi-agents/skills.md +92 -0
  220. package/references/langchain/python/multi-agents/subagents.md +486 -0
  221. package/references/langchain/python/retrieval.md +320 -0
  222. package/references/langchain/python/runtime.md +141 -0
  223. package/references/langchain/python/short-term-memory.md +658 -0
  224. package/references/langchain/python/structured-output.md +712 -0
  225. package/references/langfuse/llms.txt +148 -0
  226. package/references/langgraph/javascript/llms.txt +275 -0
  227. package/references/skills/home.md +259 -0
  228. package/references/skills/integrate-skills.md +103 -0
  229. package/references/skills/specification.md +254 -0
  230. package/references/skills/what-are-skills.md +74 -0
  231. package/rispecs/README.md +164 -0
  232. package/rispecs/_sync_/miadi-code/SPEC.md +313 -0
  233. package/rispecs/_sync_/miadi-code/STATUS.md +177 -0
  234. package/rispecs/_sync_/miadi-code/dashboard/SPEC.md +465 -0
  235. package/rispecs/_sync_/miadi-code/dashboard/STATUS.md +212 -0
  236. package/rispecs/_sync_/miadi-code/multiline-input/SPEC.md +232 -0
  237. package/rispecs/_sync_/miadi-code/multiline-input/STATUS.md +108 -0
  238. package/rispecs/_sync_/miadi-code/pde/SPEC.md +253 -0
  239. package/rispecs/_sync_/miadi-code/pde/STATUS.md +56 -0
  240. package/rispecs/_sync_/miadi-code/stc/SPEC.md +397 -0
  241. package/rispecs/_sync_/miadi-code/stc/STATUS.md +70 -0
  242. package/rispecs/ava-langstack/inquiry-routing-upgrade.spec.md +119 -0
  243. package/rispecs/borrowed_from_opencode/001-client-server-architecture.rispec.md +98 -0
  244. package/rispecs/borrowed_from_opencode/002-event-bus-system.rispec.md +125 -0
  245. package/rispecs/borrowed_from_opencode/003-instance-state-pattern.rispec.md +136 -0
  246. package/rispecs/borrowed_from_opencode/004-namespace-module-pattern.rispec.md +151 -0
  247. package/rispecs/borrowed_from_opencode/005-zod-schema-validation.rispec.md +139 -0
  248. package/rispecs/borrowed_from_opencode/006-named-error-system.rispec.md +155 -0
  249. package/rispecs/borrowed_from_opencode/007-structured-logging.rispec.md +138 -0
  250. package/rispecs/borrowed_from_opencode/008-lazy-initialization.rispec.md +127 -0
  251. package/rispecs/borrowed_from_opencode/009-multi-agent-system.rispec.md +97 -0
  252. package/rispecs/borrowed_from_opencode/010-agent-definition-config.rispec.md +135 -0
  253. package/rispecs/borrowed_from_opencode/011-agent-permission-rulesets.rispec.md +151 -0
  254. package/rispecs/borrowed_from_opencode/012-agent-prompt-templates.rispec.md +141 -0
  255. package/rispecs/borrowed_from_opencode/013-agent-generation.rispec.md +142 -0
  256. package/rispecs/borrowed_from_opencode/014-plan-build-mode-toggle.rispec.md +155 -0
  257. package/rispecs/borrowed_from_opencode/015-subagent-task-delegation.rispec.md +146 -0
  258. package/rispecs/borrowed_from_opencode/016-agent-model-selection.rispec.md +151 -0
  259. package/rispecs/borrowed_from_opencode/017-compaction-agent.rispec.md +150 -0
  260. package/rispecs/borrowed_from_opencode/018-session-persistence.rispec.md +125 -0
  261. package/rispecs/borrowed_from_opencode/019-session-compaction.rispec.md +132 -0
  262. package/rispecs/borrowed_from_opencode/020-session-forking.rispec.md +134 -0
  263. package/rispecs/borrowed_from_opencode/021-session-revert-snapshot.rispec.md +135 -0
  264. package/rispecs/borrowed_from_opencode/022-session-sharing.rispec.md +165 -0
  265. package/rispecs/borrowed_from_opencode/023-session-summary-diffs.rispec.md +165 -0
  266. package/rispecs/borrowed_from_opencode/024-child-sessions.rispec.md +164 -0
  267. package/rispecs/borrowed_from_opencode/025-session-title-generation.rispec.md +162 -0
  268. package/rispecs/borrowed_from_opencode/026-message-parts-model.rispec.md +201 -0
  269. package/rispecs/borrowed_from_opencode/027-streaming-message-deltas.rispec.md +212 -0
  270. package/rispecs/borrowed_from_opencode/028-multi-provider-architecture.rispec.md +184 -0
  271. package/rispecs/borrowed_from_opencode/029-provider-authentication.rispec.md +225 -0
  272. package/rispecs/borrowed_from_opencode/030-model-registry.rispec.md +222 -0
  273. package/rispecs/borrowed_from_opencode/031-cost-tracking.rispec.md +243 -0
  274. package/rispecs/borrowed_from_opencode/032-provider-transform-pipeline.rispec.md +282 -0
  275. package/rispecs/borrowed_from_opencode/033-provider-sdk-abstraction.rispec.md +338 -0
  276. package/rispecs/borrowed_from_opencode/034-tool-registry.rispec.md +110 -0
  277. package/rispecs/borrowed_from_opencode/035-tool-context-injection.rispec.md +155 -0
  278. package/rispecs/borrowed_from_opencode/036-tool-output-truncation.rispec.md +138 -0
  279. package/rispecs/borrowed_from_opencode/037-batch-tool.rispec.md +129 -0
  280. package/rispecs/borrowed_from_opencode/038-multi-edit-tool.rispec.md +167 -0
  281. package/rispecs/borrowed_from_opencode/039-apply-patch-tool.rispec.md +161 -0
  282. package/rispecs/borrowed_from_opencode/040-code-search-tool.rispec.md +143 -0
  283. package/rispecs/borrowed_from_opencode/041-web-fetch-tool.rispec.md +131 -0
  284. package/rispecs/borrowed_from_opencode/042-web-search-tool.rispec.md +159 -0
  285. package/rispecs/borrowed_from_opencode/043-todo-tool.rispec.md +156 -0
  286. package/rispecs/borrowed_from_opencode/044-plan-mode-tool.rispec.md +139 -0
  287. package/rispecs/borrowed_from_opencode/045-task-tool.rispec.md +146 -0
  288. package/rispecs/borrowed_from_opencode/046-question-tool.rispec.md +170 -0
  289. package/rispecs/borrowed_from_opencode/047-external-directory-tool.rispec.md +166 -0
  290. package/rispecs/borrowed_from_opencode/048-file-read-write-tools.rispec.md +205 -0
  291. package/rispecs/borrowed_from_opencode/049-lsp-server-management.rispec.md +104 -0
  292. package/rispecs/borrowed_from_opencode/050-lsp-hover-completion.rispec.md +102 -0
  293. package/rispecs/borrowed_from_opencode/051-lsp-diagnostics.rispec.md +86 -0
  294. package/rispecs/borrowed_from_opencode/052-lsp-root-detection.rispec.md +109 -0
  295. package/rispecs/borrowed_from_opencode/053-remote-mcp-servers.rispec.md +119 -0
  296. package/rispecs/borrowed_from_opencode/054-mcp-oauth-flow.rispec.md +107 -0
  297. package/rispecs/borrowed_from_opencode/055-mcp-tool-conversion.rispec.md +118 -0
  298. package/rispecs/borrowed_from_opencode/056-mcp-connection-monitoring.rispec.md +106 -0
  299. package/rispecs/borrowed_from_opencode/057-local-mcp-servers.rispec.md +116 -0
  300. package/rispecs/borrowed_from_opencode/058-rich-tui.rispec.md +108 -0
  301. package/rispecs/borrowed_from_opencode/059-streaming-display.rispec.md +116 -0
  302. package/rispecs/borrowed_from_opencode/060-permission-prompts.rispec.md +130 -0
  303. package/rispecs/borrowed_from_opencode/061-session-navigation.rispec.md +155 -0
  304. package/rispecs/borrowed_from_opencode/062-syntax-highlighting.rispec.md +151 -0
  305. package/rispecs/borrowed_from_opencode/063-keybinding-system.rispec.md +181 -0
  306. package/rispecs/borrowed_from_opencode/064-multi-level-config.rispec.md +155 -0
  307. package/rispecs/borrowed_from_opencode/065-jsonc-config.rispec.md +190 -0
  308. package/rispecs/borrowed_from_opencode/066-config-env-variables.rispec.md +153 -0
  309. package/rispecs/borrowed_from_opencode/067-config-deep-merging.rispec.md +178 -0
  310. package/rispecs/borrowed_from_opencode/068-remote-org-config.rispec.md +183 -0
  311. package/rispecs/borrowed_from_opencode/069-config-markdown-frontmatter.rispec.md +206 -0
  312. package/rispecs/borrowed_from_opencode/070-managed-config-directory.rispec.md +232 -0
  313. package/rispecs/borrowed_from_opencode/071-plugin-architecture.rispec.md +104 -0
  314. package/rispecs/borrowed_from_opencode/072-plugin-hooks.rispec.md +123 -0
  315. package/rispecs/borrowed_from_opencode/073-plugin-auto-install.rispec.md +115 -0
  316. package/rispecs/borrowed_from_opencode/074-permission-system.rispec.md +133 -0
  317. package/rispecs/borrowed_from_opencode/075-git-worktree-management.rispec.md +126 -0
  318. package/rispecs/borrowed_from_opencode/076-snapshot-system.rispec.md +124 -0
  319. package/rispecs/borrowed_from_opencode/077-snapshot-diff.rispec.md +117 -0
  320. package/rispecs/borrowed_from_opencode/078-snapshot-restore.rispec.md +128 -0
  321. package/rispecs/borrowed_from_opencode/079-worktree-branch-naming.rispec.md +122 -0
  322. package/rispecs/borrowed_from_opencode/080-sqlite-storage.rispec.md +134 -0
  323. package/rispecs/borrowed_from_opencode/081-database-migrations.rispec.md +148 -0
  324. package/rispecs/borrowed_from_opencode/082-database-transactions.rispec.md +138 -0
  325. package/rispecs/borrowed_from_opencode/083-deferred-effects.rispec.md +148 -0
  326. package/rispecs/borrowed_from_opencode/084-permission-rules.rispec.md +123 -0
  327. package/rispecs/borrowed_from_opencode/085-permission-glob-patterns.rispec.md +113 -0
  328. package/rispecs/borrowed_from_opencode/086-permission-merging.rispec.md +134 -0
  329. package/rispecs/borrowed_from_opencode/087-permission-modes.rispec.md +145 -0
  330. package/rispecs/borrowed_from_opencode/088-http-api-server.rispec.md +165 -0
  331. package/rispecs/borrowed_from_opencode/089-openapi-spec-generation.rispec.md +164 -0
  332. package/rispecs/borrowed_from_opencode/090-websocket-support.rispec.md +136 -0
  333. package/rispecs/borrowed_from_opencode/091-sse-streaming.rispec.md +168 -0
  334. package/rispecs/borrowed_from_opencode/092-mdns-discovery.rispec.md +145 -0
  335. package/rispecs/borrowed_from_opencode/093-javascript-sdk.rispec.md +200 -0
  336. package/rispecs/borrowed_from_opencode/094-skill-system.rispec.md +187 -0
  337. package/rispecs/borrowed_from_opencode/095-skill-discovery.rispec.md +182 -0
  338. package/rispecs/borrowed_from_opencode/096-desktop-remote-driving.rispec.md +175 -0
  339. package/rispecs/borrowed_from_opencode/INDEX.md +255 -0
  340. package/rispecs/core.rispecs.md +261 -0
  341. package/rispecs/engines.rispecs.md +241 -0
  342. package/rispecs/formatting.rispecs.md +252 -0
  343. package/rispecs/living-specifications.rispecs.md +361 -0
  344. package/rispecs/mcp.rispecs.md +197 -0
  345. package/rispecs/pde.rispecs.md +399 -0
  346. package/rispecs/pi-mono-envisionning/ENVISIONING.md +366 -0
  347. package/rispecs/pi-mono-envisionning/storytelling-horizon.rispecs.md +76 -0
  348. package/rispecs/pi-mono-envisionning/widget.rispecs.md +2 -0
  349. package/rispecs/relation-to-mcp-structural-thinking.kin.md +72 -0
  350. package/rispecs/research-for-better-framework/CLAUDE.md +7 -0
  351. package/rispecs/research-for-better-framework/survey-pi-openclaw-opencode-openhands.md +210 -0
  352. package/rispecs/session.rispecs.md +277 -0
  353. package/rispecs/stc.rispecs.md +138 -0
  354. package/rispecs/unifier.rispecs.md +317 -0
  355. package/scripts/LAUNCH--mcp-mia-code--testing--2603141315--ac705a66-2c15-4a1c-a26d-9491018c5ba8.sh +2 -0
  356. package/scripts/RESUME--mia-code--mcps--260313--ac705a66-2c15-4a1c-a26d-9491018c5ba8.sh +1 -0
  357. package/scripts/install-widget-in-home-pi-agent-extensions.sh +4 -0
  358. package/scripts/sample-decompose--2604011535-prompt.sh +1 -0
  359. package/skills/deep-search/AGENTS.md +17 -0
  360. package/skills/deep-search/SKILL.md +281 -0
  361. package/skills/deep-search/agent-templates.md +224 -0
  362. package/skills/deep-search/orchestration-patterns.md +95 -0
  363. package/skills/miaco-pde-inquiry-routing-deep-search/AGENTS.md +13 -0
  364. package/skills/miaco-pde-inquiry-routing-deep-search/SKILL.md +136 -0
  365. package/skills/miaco-pde-inquiry-routing-internal-external-relationship/AGENTS.md +4 -0
  366. package/skills/miaco-pde-inquiry-routing-internal-external-relationship/SKILL.md +157 -0
  367. package/skills/miaco-pde-inquiry-routing-local-qmd/AGENTS.md +42 -0
  368. package/skills/miaco-pde-inquiry-routing-local-qmd/SKILL.md +135 -0
  369. package/skills/qmd/AGENTS.md +3 -0
  370. package/skills/qmd/SKILL.md +144 -0
  371. package/skills/qmd/references/mcp-setup.md +102 -0
  372. package/skills/rise-pde-inquiry-session-multi-agents-v3/SKILL.md +234 -0
  373. package/skills/rise-pde-inquiry-session-multi-agents-v3/agent-templates.md +436 -0
  374. package/skills/rise-pde-inquiry-session-multi-agents-v3/orchestration-patterns.md +197 -0
  375. package/skills/rise-pde-inquiry-session-multi-agents-v3/references/ceremonial-technology.md +102 -0
  376. package/skills/rise-pde-inquiry-session-multi-agents-v3/references/creative-orientation.md +99 -0
  377. package/skills/rise-pde-inquiry-session-multi-agents-v3/references/prompt-decomposition.md +73 -0
  378. package/skills/rise-pde-inquiry-session-multi-agents-v3/references/rise-framework.md +74 -0
  379. package/skills/rise-pde-inquiry-session-multi-agents-v3/references/structural-tension.md +82 -0
  380. package/src/cli.ts +35 -11
  381. package/src/geminiHeadless.ts +7 -2
  382. package/src/index.ts +2 -1
  383. package/src/mcp/miaco-server.ts +13 -1
  384. package/src/mcp/miatel-server.ts +13 -1
  385. package/src/mcp/miawa-server.ts +13 -1
  386. package/src/mcp/utils.ts +41 -8
  387. package/src/sessionStore.ts +44 -4
  388. package/src/types.ts +2 -1
  389. package/widget/mia-ceremony/README.md +36 -0
  390. package/widget/mia-ceremony/index.ts +143 -0
  391. package/widget/mia-interceptor/README.md +39 -0
  392. package/widget/mia-interceptor/index.ts +221 -0
  393. package/widget/mia-tools/README.md +37 -0
  394. package/widget/mia-tools/index.ts +569 -0
  395. package/widget/miette-echo/README.md +44 -0
  396. package/widget/miette-echo/index.ts +164 -0
  397. package/.claude/settings.local.json +0 -9
  398. package/.hch/issue_.env +0 -4
  399. package/.hch/issue_add__2601211715.json +0 -77
  400. package/.hch/issue_add__2601211715.md +0 -4
  401. package/.hch/issue_add__2602242020.json +0 -78
  402. package/.hch/issue_add__2602242020.md +0 -7
  403. package/.hch/issues.json +0 -2312
  404. package/.hch/issues.md +0 -30
  405. package/WS__mia-code__260214__IAIP_PDE.code-workspace +0 -29
  406. package/WS__mia-code__src332__260122.code-workspace +0 -23
  407. package/samples/copilot/session-state/be76abaa-a27f-4725-b2a9-22fb45f7e0f7/checkpoints/index.md +0 -6
  408. package/samples/copilot/session-state/be76abaa-a27f-4725-b2a9-22fb45f7e0f7/events.jsonl +0 -213
  409. package/samples/copilot/session-state/be76abaa-a27f-4725-b2a9-22fb45f7e0f7/plan.md +0 -243
  410. package/samples/copilot/session-state/be76abaa-a27f-4725-b2a9-22fb45f7e0f7/workspace.yaml +0 -5
@@ -0,0 +1,558 @@
1
+ 4
2
+ 2
3
+ 0
4
+ 2
5
+
6
+ l
7
+ u
8
+ J
9
+
10
+ 0
11
+ 1
12
+
13
+ ]
14
+
15
+ A
16
+ M
17
+
18
+ .
19
+ s
20
+ c
21
+ [
22
+
23
+ 2
24
+ v
25
+ 3
26
+ 1
27
+ 7
28
+ 6
29
+ 1
30
+ .
31
+ 2
32
+ 0
33
+ 4
34
+ 2
35
+ :
36
+ v
37
+ i
38
+ X
39
+ r
40
+ a
41
+
42
+ NAVIGATING COMPLEXITY: ORCHESTRATED PROBLEM
43
+ SOLVING WITH MULTI-AGENT LLMS
44
+
45
+ Sumedh Rasal
46
+ Georgia Institute of Technology
47
+ Chicago, IL
48
+ srasal3@gatech.edu
49
+
50
+ E. J. Hauer
51
+ EJ@ThoughtCrafters.com
52
+
53
+ ABSTRACT
54
+
55
+ Large Language Models (LLMs) have demonstrated remarkable capabilities in solving various tasks,
56
+ yet they often struggle with comprehensively addressing complex and vague problems. Existing ap-
57
+ proaches, including multi-agent LLM systems, offer solutions to certain challenges but still require
58
+ manual setup and lack scalability. To address this gap, we propose a novel approach leveraging
59
+ decomposition to enable LLMs to tackle vague problems effectively.
60
+ Our approach involves an orchestrating LLM that interacts with users to understand the problem and
61
+ then decomposes it into tangible sub-problems. Instead of expecting the LLM to solve the entire
62
+ problem in one go, we train it to ask follow-up questions to gain a deeper understanding of the
63
+ user’s requirements. Once the problem is adequately understood, the orchestrating LLM divides
64
+ it into smaller, manageable sub-problems. Each sub-problem is then assigned to specialized LLM
65
+ agents or non-LLM functions for resolution. These agents work in parallel to solve their respective
66
+ sub-problems, with the orchestrating LLM overseeing the process and compiling the solutions into
67
+ a comprehensive answer for the user. By adopting this decomposition approach, we alleviate the
68
+ constraints imposed by token limitations on LLM outputs and empower them to provide nuanced
69
+ solutions to complex and ambiguous problems.
70
+ Through our approach, we aim to enable LLMs to think and operate more like humans, breaking
71
+ down complex problems into manageable parts and collaboratively solving them. This not only
72
+ enhances the problem-solving capabilities of LLMs but also offers a scalable and efficient method
73
+ for addressing a wide range of real-world challenges.
74
+
75
+ Keywords Decomposition · Multi-Agents
76
+
77
+ 1 Introduction
78
+
79
+ In recent years, the proliferation of Large Language Models (LLMs) has revolutionized problem-solving across various
80
+ domains [Devlin et al., 2018] [Patil et al., 2023] [Touvron et al., 2023]. These models possess the ability to compre-
81
+ hend user queries and generate responses, often providing novel insights and solutions [Wei et al., 2022a]. Notably,
82
+ LLMs exhibit exceptional performance in tasks that can be tackled through zero-shot or few-shot training techniques
83
+ [Brown et al., 2020], such as transforming questions into familiar formats or solving coding problems based on anal-
84
+ ogous examples [Chen et al., 2021] [Sanh et al., 2021]. As such, they hold immense potential for automating numer-
85
+ ous tasks and streamlining workflows across diverse professional domains [Liu et al., 2023b] [Chiang and Lee, 2023]
86
+ [Gao et al., 2023] [Shen et al., 2023] [Zheng et al., 2023].
87
+
88
+ However, despite their remarkable capabilities, LLMs are not immune to limitations. Instances of hallucination, con-
89
+ tradictory responses, or misinterpretation of user queries highlight the challenges inherent in relying solely on these
90
+ models for problem-solving [Azamfirei et al., 2023]. Existing frameworks designed to aid users in systematically ad-
91
+ dressing novel problems are constrained by token limits imposed by LLM models [Gur et al., 2023]. These limits
92
+ restrict the number of words or tokens that an LLM can process in a single message or query, often hindering the
93
+ comprehensiveness of responses and impeding the continuity of conversations. While approaches such as chaining
94
+
95
+ Navigating Complexity: Orchestrated Problem Solving with Multi-Agent LLMs
96
+
97
+ previous conversations or employing semantic search techniques mitigate some of these challenges, they are suscepti-
98
+ ble to errors, including redundancy or irrelevance in retrieved information [Ouyang et al., 2022] [Saunders et al., 2022]
99
+ [Liu et al., 2023a].
100
+
101
+ To address these limitations, a promising avenue lies in leveraging multi-agent frameworks to tackle lengthy or
102
+ complex problems [Du et al., 2023] [Cohen et al., 2023] [Dafoe et al., 2021] [Dafoe et al., 2020] [Kramár et al., 2022]
103
+ [Yao et al., 2022].
104
+ Such frameworks distribute problem-solving tasks among multiple LLM agents, offering
105
+ a more robust approach compared to heuristic methods [Li et al., 2023] [Liang et al., 2023] [Qian et al., 2023]
106
+ [Lazaridou et al., 2020] [Graesser et al., 2020] [Lee et al., 2018] [Wu et al., 2023]. However, the manual assignment
107
+ of personas or roles to each LLM agent presents a bottleneck, particularly when dealing with novel problems lacking
108
+ predefined roles [Rasal, 2024].
109
+
110
+ In this paper, we propose a novel solution that integrates the strengths of existing approaches while addressing their
111
+ limitations. Our approach centers on an orchestration strategy that employs problem decomposition to facilitate the
112
+ seamless resolution of lengthy or complex problems. Inspired by human problem-solving approaches, our framework
113
+ involves an orchestration LLM that engages in iterative conversations with users to ascertain problem requirements.
114
+ Once the requirements are elucidated, the orchestration LLM decomposes the problem into manageable subproblems
115
+ and assigns them to specialized agents, each tailored to address specific aspects of the problem. Notably, our approach
116
+ incorporates non-LLM functionalities where necessary, enhancing robustness and adaptability to real-world tasks
117
+ [Qin et al., 2023]. By leveraging problem decomposition and distributed task assignment, our framework offers a
118
+ comprehensive and scalable solution for problem-solving with LLMs.
119
+
120
+ This is what we propose:
121
+
122
+ • Our proposed solution integrates problem decomposition and multi-agent orchestration, inspired by human
123
+
124
+ problem-solving approaches, to address the limitations of existing frameworks.
125
+
126
+ • By leveraging problem decomposition and distributed task assignment, our framework offers a comprehensive
127
+
128
+ and scalable solution for problem-solving with LLMs, enhancing adaptability to real-world tasks.
129
+
130
+ 2 Methodology
131
+
132
+ In response to these challenges, our methodology adopts a novel approach that seeks to address the token limit issue
133
+ through algorithmic enhancements rather than relying solely on memory expansion. At the core of our methodol-
134
+ ogy lies the development of an Orchestration LLM, tasked with the responsibility of comprehending user problem
135
+ statements, decomposing them into manageable subproblems, and orchestrating interactions among specialized LLM
136
+ agents.
137
+
138
+ Inspired by human problem-solving strategies, our methodology mirrors the process of breaking down complex tasks
139
+ into smaller, more digestible components. We employ advanced natural language processing (NLP) techniques to
140
+ analyze and dissect complex queries, identifying distinct subproblems across various domains. These subproblems are
141
+ then mapped to specialized LLM agents, each pre-trained on specific tasks such as data analysis, creative writing, or
142
+ technical problem-solving, to ensure precision and relevance in their responses.
143
+
144
+ Analogous to seeking expertise in real-life scenarios, our methodology leverages collaboration with domain experts
145
+ to optimize problem-solving outcomes. For instance, just as individuals seek tax experts for efficient tax filing, our
146
+ approach integrates specialized agents to tackle domain-specific challenges effectively.
147
+
148
+ Furthermore, we develop an algorithmic framework that enables the Orchestration LLM to decompose complex prob-
149
+ lems based on predefined criteria such as problem domain, required expertise, and task complexity. Finally, we imple-
150
+ ment a task assignment system to allocate subproblems among selected specialized agents, considering factors such
151
+ as availability, expertise, and workload to streamline the problem-solving process and optimize resource utilization.
152
+
153
+ The high-level algorithm to our framework can be found 2
154
+
155
+ 2
156
+
157
+ Navigating Complexity: Orchestrated Problem Solving with Multi-Agent LLMs
158
+
159
+ Step Orchestration of Problem Solving with LLMs
160
+
161
+ 1
162
+ 2
163
+ 3
164
+ 4
165
+ 5
166
+ 6
167
+ 7
168
+ 8
169
+ 9
170
+ 10
171
+ 11
172
+ 12
173
+ 13
174
+
175
+ Require A complex problem statement
176
+ Ensure A comprehensive solution to the problem
177
+ Begin
178
+ Input the complex problem statement
179
+ Decompose the problem into subproblems
180
+ for each subproblem do
181
+ - Select a specialized LLM agent based on the subproblem’s domain and requirements
182
+ - Assign the subproblem to the selected agent
183
+ - Agent solves the subproblem and stores the solution
184
+ end for
185
+ - Aggregate the solutions of subproblems into a coherent overall solution
186
+ - Output the comprehensive solution
187
+ End
188
+
189
+ 2.1 Decomposition Approach
190
+
191
+ The implementation of our framework initiates with the user’s submission of a complex problem statement, facilitated
192
+ through a conversational interface within the LangChain framework [Chase, 2022]. Within the Orchestration LLM,
193
+ this input undergoes meticulous processing, wherein key information and requirements are extracted and analyzed with
194
+ precision. Subsequently, the Orchestration LLM seamlessly applies a sophisticated decomposition algorithm, system-
195
+ atically breaking down the problem into manageable subproblems, thereby ensuring clarity and precision throughout
196
+ the process.
197
+
198
+ The decomposition approach was crafted through a series of few-shot training iterations employing the GPT4 model
199
+ [OpenAI, 2023] [Bubeck et al., 2023]. Initially, the LLM model decomposed the main problem into vague subprob-
200
+ lems, which necessitated further refinement to ensure alignment with the expected decomposition format. Utilizing
201
+ few-shot training, we provided the model with chain-of-thought type reasoning to guide its decomposition process
202
+ effectively.
203
+
204
+ Upon successful decomposition, the Orchestration LLM assumes the pivotal role of assigning these subproblems to
205
+ the most suitable specialized LLM agents. This assignment process is meticulously orchestrated to ensure that each
206
+ subproblem is delegated to an agent possessing the requisite expertise and capabilities to effectively address it.
207
+
208
+ Central to the framework’s efficacy is the decomposition phase, which serves as its cornerstone. We deliberately crafted
209
+ the initial design to be adaptable and generalizable across diverse problem domains, prioritizing the demonstration of
210
+ decomposition capabilities over specialized use-case system development.
211
+
212
+ 3 Experiments
213
+
214
+ Our experiments center around the utilization of an orchestration LLM agent, employing OpenAI’s gpt-4 model
215
+ [OpenAI, 2023], which adeptly engages with users to elucidate requirements and subsequently decomposes problems
216
+ into manageable subproblems. This orchestration LLM also determines the most suitable specialized LLM agent to ad-
217
+ dress each subproblem. Each specialized LLM agent employed in our experiments is based on OpenAI’s gpt3.5-turbo
218
+ model, configured with a temperature parameter set to 0.0, reflecting the balance between coherence and creativity
219
+ in responses. Facilitating communication among LLM agents is the LangChain framework, enabling seamless multi-
220
+ agent interaction and facilitating focus on problem-solving.
221
+
222
+ 3.1 Experiment 1: Travel Planning with LLM Agents
223
+
224
+ In this experiment, users present their travel plans to the framework, initiating a dialogue with the orchestration LLM
225
+ to refine requirements and subsequently decompose the problem into subproblems. The orchestration LLM allocates
226
+ tasks to specialized agents, which upon completion, contribute their findings for collation into a final course of action.
227
+ Before execution, the orchestration LLM verifies details with the user to ensure accuracy.
228
+
229
+ Here’s the conversational output of the framework.
230
+
231
+ Initial Interaction: Gathering Requirements
232
+
233
+ 3
234
+
235
+ Navigating Complexity: Orchestrated Problem Solving with Multi-Agent LLMs
236
+
237
+ User: I need to book a return flight from Chicago Airport to Newark Airport from 3/22/24 to 3/26/24.
238
+ I want to fly in luxury and travel between the hours of 10 am and 4 pm.
239
+ Orchestration LLM Agent: Got it. To ensure I find the best options for you, could you specify
240
+ your preferences for seating class and any additional services you might require during your flight?
241
+ User: I prefer business class and would like access to in-flight Wi-Fi and a window seat.
242
+ Orchestration LLM Agent Decomposes the Problem
243
+ The Orchestration LLM Agent processes the user’s request and decomposes it into subproblems to
244
+ be handled by specialized agents:
245
+ Flight Search Agent: Find business class flights from Chicago to Newark that depart between 10
246
+ am and 4 pm on 3/22/24 and return flights on 3/26/24.
247
+ Amenity Preferences Agent: Identify airlines offering in-flight Wi-Fi and window seat availability
248
+ in business class.
249
+ Booking Agent: Handle the booking process once the flight options are selected.
250
+ Interaction with Specialized Agents
251
+ Orchestration LLM Agent to Flight Search Agent: Find business class flights from Chicago to
252
+ Newark departing between 10 am and 4 pm on 3/22/24 and returning on 3/26/24.
253
+ Flight Search Agent: Found several options.
254
+ Orchestration LLM Agent to Amenity Preferences Agent: Which of these options offers in-flight
255
+ Wi-Fi and window seats in business class?
256
+ Amenity Preferences Agent: [Lists airlines that meet the criteria] Finalizing the Option
257
+ Orchestration LLM Agent to User: Based on your preferences, here are the best flight options that
258
+ offer business class, in-flight Wi-Fi, and window seats. [Presents options]
259
+ User: I’ll take the option with Airline X on the 3/22 at 2 pm.
260
+ Orchestration LLM Agent to Booking Agent: Please book the flight with Airline X for 3/22,
261
+ departing at 2 pm in business class, with a window seat and in-flight Wi-Fi.
262
+ Booking Agent: Your flight is booked. Here are your confirmation details.
263
+
264
+ 3.2 Experiment 2: Exploratory Research and Summarization
265
+
266
+ Users pose exploratory research queries to the framework, prompting the orchestration LLM to segment the problem
267
+ into subproblems and assign tasks to specialized agents. Following the completion of tasks, the specialized agents
268
+ provide their outputs, which are synthesized by the orchestration LLM into a comprehensive summary. Feedback
269
+ from the user is sought to refine the final output as necessary.
270
+
271
+ Here’s the conversational output of the framework.
272
+
273
+ Initial Interaction: Gathering Requirements
274
+ User: I need to research topics about long-term memory management in LLMs, focusing on what’s
275
+ currently possible and what’s not. Then, I’d like a 5-page article summarizing these findings and
276
+ suggesting future research directions.
277
+ Orchestration LLM Agent Decomposes the Problem
278
+ The Orchestration LLM Agent analyzes the request and identifies several key tasks:
279
+ Literature Review Agent: Conduct a comprehensive review of existing research on long-term mem-
280
+ ory management in LLMs.
281
+ Analysis Agent: Analyze the collected research to distinguish between solved problems, ongoing
282
+ challenges, and unexplored areas.
283
+ Writing Agent: Synthesize the findings into a coherent 5-page article, highlighting current capabil-
284
+ ities, and limitations, and suggesting future research directions.
285
+ Interaction with Specialized Agents
286
+ Task 1: Conduct Literature Review
287
+ Orchestration LLM Agent to Literature Review Agent: Identify and summarize key research
288
+ articles on long-term memory management in LLMs, focusing on methodologies, findings, and
289
+ limitations.
290
+ Literature Review Agent: [Returns a summary of articles, including key methodologies, findings,
291
+ challenges, and limitations related to long-term memory management.]
292
+
293
+ 4
294
+
295
+ Navigating Complexity: Orchestrated Problem Solving with Multi-Agent LLMs
296
+
297
+ Table 1: GSM8K Data set Evaluation
298
+
299
+ Agent
300
+
301
+ Solve Rate
302
+
303
+ Single GPT3.5-turbo
304
+ Multi-Agent GPT3.5-turbo
305
+ Multi-Agent GPT3.5-turbo (LLM-Harmony)
306
+ Multi-Agent GPT4 (Our Approach)
307
+
308
+ 50%
309
+ 55%
310
+ 65%
311
+ 73%
312
+
313
+ Task 2: Analyze Research
314
+
315
+ Orchestration LLM Agent to Analysis Agent: Based on the literature review, analyze the cur-
316
+ rent state of long-term memory management in LLMs, identifying what is possible, what remains
317
+ challenging, and potential areas for future research.
318
+
319
+ Analysis Agent: [Provides an analysis distinguishing between the current capabilities and limita-
320
+ tions of long-term memory management in LLMs, along with suggestions for future research direc-
321
+ tions.]
322
+
323
+ Task 3: Write Article
324
+
325
+ Orchestration LLM Agent to Writing Agent: Using the information collected and analyzed, write
326
+ a 5-page article summarizing the findings on long-term memory management in LLMs. Highlight
327
+ what is currently possible, and the limitations, and suggest ideas for future research.
328
+
329
+ Writing Agent: [Drafts a 5-page article that summarizes the state of long-term memory manage-
330
+ ment in LLMs, including an overview of existing research, current limitations, and proposed future
331
+ directions.]
332
+
333
+ Finalizing the Article
334
+
335
+ Orchestration LLM Agent to User: Based on the comprehensive research and analysis conducted
336
+ by specialized agents, we have compiled a 5-page article summarizing the current state of long-term
337
+ memory management in LLMs. The article discusses what is currently possible, outlines the main
338
+ challenges and limitations, and suggests potential areas for future research.
339
+
340
+ User: Reviews the article and provides feedback or requests revisions if necessary.
341
+
342
+ 4 Results
343
+
344
+ Rather than relying solely on conventional datasets typically utilized in research, we opted to assess the efficacy of
345
+ our framework using the GSM8K dataset, chosen to specifically demonstrate the capabilities of our approach. We
346
+ employed four distinct types of agents in our experimentation: a single LLM agent powered by the GPT3.5-turbo
347
+ model, two LLM agents supported by GPT3.5-turbo models, two LLM agents utilizing the GPT3.5-turbo model
348
+ combined with the chain-of-thought and persona approach [Wei et al., 2022b] [Rasal, 2024], and finally, two LLM
349
+ agents operating with the GPT4 model [OpenAI, 2023] for orchestration and decomposition methodology and GPT3.5-
350
+ turbo model for the specialized LLM agents, which involves breaking down the primary problem into subproblems.
351
+ Our orchestration framework represents a notable advancement over existing methodologies, with detailed results
352
+ presented in Table 1.
353
+
354
+ "GSM8K consists of 8.5K high-quality grade school math problems created by human problem
355
+ writers. We segmented these into 7.5K training problems and 1K test problems. These problems take
356
+ between 2 and 8 steps to solve, and solutions primarily involve performing a sequence of elementary
357
+ calculations using basic arithmetic operations (+ - / *) to reach the final answer. A bright middle
358
+ school student should be able to solve every problem."[Cobbe et al., 2021]
359
+
360
+ Furthermore, our observations indicated a significant enhancement in accuracy when employing a calculator function
361
+ as opposed to a math-specialized LLM agent. This improvement stems from the inherent limitations of LLMs in
362
+ executing mathematical operations, despite their versatility across various domains. We contend that rather than train-
363
+ ing LLMs for math operations, our framework facilitates the orchestration LLM in discerning whether a subproblem
364
+ necessitates an LLM agent or an alternative non-LLM approach.
365
+
366
+ 5
367
+
368
+ Navigating Complexity: Orchestrated Problem Solving with Multi-Agent LLMs
369
+
370
+ 5 Limitations
371
+
372
+ While our decomposition framework represents a significant advancement, it possesses inherent limitations that war-
373
+ rant consideration. Foremost among these limitations is the framework’s lack of generalization to address all con-
374
+ ceivable problem scenarios, both novel and otherwise. We recognize the necessity of further training the orchestration
375
+ model, essentially serving as a foundational Large Language Model (LLM), to acquire a comprehensive understanding
376
+ of diverse problem-solving approaches. Presently, the model exhibits proficiency in decomposing various problems,
377
+ yet its scope may not extend to accommodate the complexities of future challenges.
378
+
379
+ For instance, consider a user’s request for transportation. While the current model adeptly decomposes the problem
380
+ into a specialized agent capable of booking a ride, scenarios may arise where multiple specialized agents possess this
381
+ capability. In such instances, the orchestration LLM faces the challenge of determining the most appropriate course
382
+ of action. Additionally, the framework lacks a mechanism to assimilate information about new specialized agents,
383
+ highlighting the need for a retraining mechanism to continually update the orchestration LLM’s knowledge base.
384
+
385
+ Furthermore, while the current model effectively decomposes problems, it lacks the capability to iteratively learn and
386
+ refine its problem-solving strategies based on the outcomes of decomposition steps. Addressing this limitation repre-
387
+ sents a key focus of our future research endeavors. Nonetheless, the primary objective of this paper is to disseminate
388
+ our decomposition approach to the readership, offering insights into its application in problem-solving contexts.
389
+
390
+ 6 Conclusion
391
+
392
+ In conclusion, our paper introduces a novel approach to problem-solving using a decomposition framework orches-
393
+ trated by Large Language Models (LLMs). Through meticulous decomposition of complex problem statements into
394
+ manageable subproblems, our framework demonstrates the potential to streamline problem-solving processes and
395
+ mitigate the limitations inherent in current LLM-based solutions. By leveraging the inherent capabilities of LLMs to
396
+ understand user queries and decompose problems, our approach offers a promising avenue for addressing a wide range
397
+ of challenges across various domains.
398
+
399
+ Our implementation of the framework highlights its efficacy in breaking down intricate problem statements and assign-
400
+ ing tasks to specialized LLM agents, resulting in coherent and comprehensive solutions. Moreover, our experiments
401
+ underscore the versatility of the framework in facilitating tasks such as travel planning and research summary genera-
402
+ tion, showcasing its applicability in real-world scenarios.
403
+
404
+ However, our work also reveals several limitations that necessitate further research and development. While the
405
+ decomposition framework represents a significant advancement, its generalization to novel problem scenarios remains
406
+ an ongoing challenge. Additionally, the framework lacks mechanisms for adaptive learning and integration of new
407
+ specialized agents, indicating avenues for future improvement.
408
+
409
+ Despite these limitations, our paper contributes to the growing body of literature on LLM-based problem-solving
410
+ approaches by introducing a systematic and scalable framework. By sharing our approach and insights, we aim to
411
+ stimulate further research and innovation in this domain, paving the way for enhanced problem-solving capabilities
412
+ and the realization of artificial general intelligence.
413
+
414
+ References
415
+
416
+ [Azamfirei et al., 2023] Azamfirei, R., Kudchadkar, S. R., and Fackler, J. (2023). Large language models and the
417
+
418
+ perils of their hallucinations. Critical Care, 27(1):1–2.
419
+
420
+ [Brown et al., 2020] Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A.,
421
+ Shyam, P., Sastry, G., Askell, A., et al. (2020). Language models are few-shot learners. Advances in neural
422
+ information processing systems, 33:1877–1901.
423
+
424
+ [Bubeck et al., 2023] Bubeck, S., Chandrasekaran, V., Eldan, R., Gehrke, J., Horvitz, E., Kamar, E., Lee, P., Lee, Y. T.,
425
+ Li, Y., Lundberg, S., et al. (2023). Sparks of artificial general intelligence: Early experiments with gpt-4. arXiv
426
+ preprint arXiv:2303.12712.
427
+
428
+ [Chase, 2022] Chase, H. (2022). Langchain, october 2022. URL https://github. com/langchain-ai/langchain.
429
+
430
+ [Chen et al., 2021] Chen, M., Tworek, J., Jun, H., Yuan, Q., Pinto, H. P. d. O., Kaplan, J., Edwards, H., Burda,
431
+ Y., Joseph, N., Brockman, G., et al. (2021). Evaluating large language models trained on code. arXiv preprint
432
+ arXiv:2107.03374.
433
+
434
+ 6
435
+
436
+ Navigating Complexity: Orchestrated Problem Solving with Multi-Agent LLMs
437
+
438
+ [Chiang and Lee, 2023] Chiang, C.-H. and Lee, H.-y. (2023). Can large language models be an alternative to human
439
+
440
+ evaluations? arXiv preprint arXiv:2305.01937.
441
+
442
+ [Cobbe et al., 2021] Cobbe, K., Kosaraju, V., Bavarian, M., Chen, M., Jun, H., Kaiser, L., Plappert, M., Tworek, J.,
443
+ Hilton, J., Nakano, R., Hesse, C., and Schulman, J. (2021). Training verifiers to solve math word problems. arXiv
444
+ preprint arXiv:2110.14168.
445
+
446
+ [Cohen et al., 2023] Cohen, R., Hamri, M., Geva, M., and Globerson, A. (2023). Lm vs lm: Detecting factual errors
447
+
448
+ via cross examination. arXiv preprint arXiv:2305.13281.
449
+
450
+ [Dafoe et al., 2021] Dafoe, A., Bachrach, Y., Hadfield, G., Horvitz, E., Larson, K., and Graepel, T. (2021). Coopera-
451
+
452
+ tive ai: machines must learn to find common ground. Nature, 593(7857):33–36.
453
+
454
+ [Dafoe et al., 2020] Dafoe, A., Hughes, E., Bachrach, Y., Collins, T., McKee, K. R., Leibo, J. Z., Larson, K., and
455
+
456
+ Graepel, T. (2020). Open problems in cooperative ai. arXiv preprint arXiv:2012.08630.
457
+
458
+ [Devlin et al., 2018] Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. (2018). Bert: Pre-training of deep bidirec-
459
+
460
+ tional transformers for language understanding. arXiv preprint arXiv:1810.04805.
461
+
462
+ [Du et al., 2023] Du, Y., Li, S., Torralba, A., Tenenbaum, J. B., and Mordatch, I. (2023). Improving factuality and
463
+
464
+ reasoning in language models through multiagent debate. arXiv preprint arXiv:2305.14325.
465
+
466
+ [Gao et al., 2023] Gao, M., Ruan, J., Sun, R., Yin, X., Yang, S., and Wan, X. (2023). Human-like summarization
467
+
468
+ evaluation with chatgpt. arXiv preprint arXiv:2304.02554.
469
+
470
+ [Graesser et al., 2020] Graesser, L., Cho, K., and Kiela, D. (2020). Emergent linguistic phenomena in multi-agent
471
+
472
+ communication games.
473
+
474
+ [Gur et al., 2023] Gur, I., Furuta, H., Huang, A., Safdari, M., Matsuo, Y., Eck, D., and Faust, A. (2023). A real-world
475
+ webagent with planning, long context understanding, and program synthesis. arXiv preprint arXiv:2307.12856.
476
+ [Kramár et al., 2022] Kramár, J., Eccles, T., Gemp, I., Tacchetti, A., McKee, K. R., Malinowski, M., Graepel, T., and
477
+ Bachrach, Y. (2022). Negotiation and honesty in artificial intelligence methods for the board game of diplomacy.
478
+ Nature Communications, 13(1):7214.
479
+
480
+ [Lazaridou et al., 2020] Lazaridou, A., Potapenko, A., and Tieleman, O. (2020). Multi-agent communication meets
481
+
482
+ natural language: Synergies between functional and structural language learning.
483
+
484
+ [Lee et al., 2018] Lee, J., Cho, K., Weston, J., and Kiela, D. (2018). Emergent translation in multi-agent communica-
485
+
486
+ tion.
487
+
488
+ [Li et al., 2023] Li, G., Hammoud, H. A. A. K., Itani, H., Khizbullin, D., and Ghanem, B. (2023). Camel: Commu-
489
+ nicative agents for" mind" exploration of large scale language model society. arXiv preprint arXiv:2303.17760.
490
+ [Liang et al., 2023] Liang, T., He, Z., Jiao, W., Wang, X., Wang, Y., Wang, R., Yang, Y., Tu, Z., and Shi, S.
491
+ (2023). Encouraging divergent thinking in large language models through multi-agent debate. arXiv preprint
492
+ arXiv:2305.19118.
493
+
494
+ [Liu et al., 2023a] Liu, J., Xia, C. S., Wang, Y., and Zhang, L. (2023a). Is your code generated by chatgpt really
495
+ correct? rigorous evaluation of large language models for code generation. arXiv preprint arXiv:2305.01210.
496
+ [Liu et al., 2023b] Liu, Y., Iter, D., Xu, Y., Wang, S., Xu, R., and Zhu, C. (2023b). Gpteval: Nlg evaluation using
497
+
498
+ gpt-4 with better human alignment. arXiv preprint arXiv:2303.16634.
499
+
500
+ [OpenAI, 2023] OpenAI (2023). Gpt-4 technical report. arxiv 2303.08774. View in Article, 2:13.
501
+ [Ouyang et al., 2022] Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C. L., Mishkin, P., Zhang, C., Agarwal,
502
+ S., Slama, K., Ray, A., et al. (2022). Training language models to follow instructions with human feedback, 2022.
503
+ URL https://arxiv. org/abs/2203.02155, 13.
504
+
505
+ [Patil et al., 2023] Patil, S. G., Zhang, T., Wang, X., and Gonzalez, J. E. (2023). Gorilla: Large language model
506
+
507
+ connected with massive apis. arXiv preprint arXiv:2305.15334.
508
+
509
+ [Qian et al., 2023] Qian, C., Cong, X., Yang, C., Chen, W., Su, Y., Xu, J., Liu, Z., and Sun, M. (2023). Communicative
510
+
511
+ agents for software development. arXiv preprint arXiv:2307.07924.
512
+
513
+ [Qin et al., 2023] Qin, Y., Liang, S., Ye, Y., Zhu, K., Yan, L., Lu, Y., Lin, Y., Cong, X., Tang, X., Qian, B., et al. (2023).
514
+ Toolllm: Facilitating large language models to master 16000+ real-world apis. arXiv preprint arXiv:2307.16789.
515
+ [Rasal, 2024] Rasal, S. (2024). Llm harmony: Multi-agent communication for problem solving. arXiv preprint
516
+
517
+ arXiv:2401.01312.
518
+
519
+ [Sanh et al., 2021] Sanh, V., Webson, A., Raffel, C., Bach, S. H., Sutawika, L., Alyafeai, Z., Chaffin, A., Stiegler,
520
+ A., Scao, T. L., Raja, A., et al. (2021). Multitask prompted training enables zero-shot task generalization. arXiv
521
+ preprint arXiv:2110.08207.
522
+
523
+ 7
524
+
525
+ Navigating Complexity: Orchestrated Problem Solving with Multi-Agent LLMs
526
+
527
+ [Saunders et al., 2022] Saunders, W., Yeh, C., Wu, J., Bills, S., Ouyang, L., Ward, J., and Leike, J. (2022). Self-
528
+
529
+ critiquing models for assisting human evaluators. arXiv preprint arXiv:2206.05802.
530
+
531
+ [Shen et al., 2023] Shen, C., Cheng, L., You, Y., and Bing, L. (2023). Are large language models good evaluators for
532
+
533
+ abstractive summarization? arXiv preprint arXiv:2305.13091.
534
+
535
+ [Touvron et al., 2023] Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M.-A., Lacroix, T., Rozière, B.,
536
+ Goyal, N., Hambro, E., Azhar, F., et al. (2023). Llama: Open and efficient foundation language models. arXiv
537
+ preprint arXiv:2302.13971.
538
+
539
+ [Wei et al., 2022a] Wei, J., Tay, Y., Bommasani, R., Raffel, C., Zoph, B., Borgeaud, S., Yogatama, D., Bosma, M.,
540
+ Zhou, D., Metzler, D., et al. (2022a). Emergent abilities of large language models. arXiv preprint arXiv:2206.07682.
541
+ [Wei et al., 2022b] Wei, J., Wang, X., Schuurmans, D., Bosma, M., Xia, F., Chi, E., Le, Q. V., Zhou, D., et al. (2022b).
542
+ Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing
543
+ Systems, 35:24824–24837.
544
+
545
+ [Wu et al., 2023] Wu, N., Gong, M., Shou, L., Liang, S., and Jiang, D. (2023). Large language models are diverse
546
+
547
+ role-players for summarization evaluation. arXiv preprint arXiv:2303.15078.
548
+
549
+ [Yao et al., 2022] Yao, S., Zhao, J., Yu, D., Du, N., Shafran, I., Narasimhan, K., and Cao, Y. (2022). React: Synergiz-
550
+
551
+ ing reasoning and acting in language models. arXiv preprint arXiv:2210.03629.
552
+
553
+ [Zheng et al., 2023] Zheng, L., Chiang, W.-L., Sheng, Y., Zhuang, S., Wu, Z., Zhuang, Y., Lin, Z., Li, Z., Li, D., Xing,
554
+
555
+ E., et al. (2023). Judging llm-as-a-judge with mt-bench and chatbot arena. arXiv preprint arXiv:2306.05685.
556
+
557
+ 8
558
+