dev-mcp-server 0.0.2 → 1.0.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.env.example +23 -55
- package/README.md +609 -219
- package/cli.js +486 -160
- package/package.json +2 -2
- package/src/agents/BaseAgent.js +113 -0
- package/src/agents/dreamer.js +165 -0
- package/src/agents/improver.js +175 -0
- package/src/agents/specialists.js +202 -0
- package/src/agents/taskDecomposer.js +176 -0
- package/src/agents/teamCoordinator.js +153 -0
- package/src/api/routes/agents.js +172 -0
- package/src/api/routes/extras.js +115 -0
- package/src/api/routes/git.js +72 -0
- package/src/api/routes/ingest.js +60 -40
- package/src/api/routes/knowledge.js +59 -41
- package/src/api/routes/memory.js +41 -0
- package/src/api/routes/newRoutes.js +168 -0
- package/src/api/routes/pipelines.js +41 -0
- package/src/api/routes/planner.js +54 -0
- package/src/api/routes/query.js +24 -0
- package/src/api/routes/sessions.js +54 -0
- package/src/api/routes/tasks.js +67 -0
- package/src/api/routes/tools.js +85 -0
- package/src/api/routes/v5routes.js +196 -0
- package/src/api/server.js +133 -5
- package/src/context/compactor.js +151 -0
- package/src/context/contextEngineer.js +181 -0
- package/src/context/contextVisualizer.js +140 -0
- package/src/core/conversationEngine.js +231 -0
- package/src/core/indexer.js +169 -143
- package/src/core/ingester.js +141 -126
- package/src/core/queryEngine.js +286 -236
- package/src/cron/cronScheduler.js +260 -0
- package/src/dashboard/index.html +1181 -0
- package/src/lsp/symbolNavigator.js +220 -0
- package/src/memory/memoryManager.js +186 -0
- package/src/memory/teamMemory.js +111 -0
- package/src/messaging/messageBus.js +177 -0
- package/src/monitor/proactiveMonitor.js +337 -0
- package/src/pipelines/pipelineEngine.js +230 -0
- package/src/planner/plannerEngine.js +202 -0
- package/src/plugins/builtin/stats-plugin.js +29 -0
- package/src/plugins/pluginManager.js +144 -0
- package/src/prompts/promptEngineer.js +289 -0
- package/src/sessions/sessionManager.js +166 -0
- package/src/skills/skillsManager.js +263 -0
- package/src/storage/store.js +127 -105
- package/src/tasks/taskManager.js +151 -0
- package/src/tools/BashTool.js +154 -0
- package/src/tools/FileEditTool.js +280 -0
- package/src/tools/GitTool.js +212 -0
- package/src/tools/GrepTool.js +199 -0
- package/src/tools/registry.js +1380 -0
- package/src/utils/costTracker.js +69 -0
- package/src/utils/fileParser.js +176 -153
- package/src/utils/llmClient.js +355 -206
- package/src/watcher/fileWatcher.js +137 -0
- package/src/worktrees/worktreeManager.js +176 -0
package/.env.example
CHANGED
|
@@ -1,68 +1,36 @@
|
|
|
1
|
-
#
|
|
2
|
-
#
|
|
3
|
-
# Copy this file to .env and fill in your values.
|
|
4
|
-
# ═══════════════════════════════════════════════════════════════
|
|
5
|
-
|
|
6
|
-
# ── LLM Provider ───────────────────────────────────────────────
|
|
7
|
-
# Which LLM backend to use.
|
|
8
|
-
# Options: anthropic | ollama | azure
|
|
9
|
-
# Default: anthropic
|
|
1
|
+
# ── LLM Provider ──────────────────────────────────────────────
|
|
2
|
+
# Options: anthropic | azure | ollama
|
|
10
3
|
LLM_PROVIDER=anthropic
|
|
11
4
|
|
|
12
|
-
# Model
|
|
13
|
-
|
|
14
|
-
# anthropic → claude-opus-4-5
|
|
15
|
-
# ollama → llama3
|
|
16
|
-
# azure → MUST be set (matches your Azure deployment name)
|
|
17
|
-
# LLM_MODEL=claude-opus-4-5
|
|
18
|
-
|
|
5
|
+
# Model override for ollama (e.g. llama3, mistral, codellama)
|
|
6
|
+
LLM_MODEL=
|
|
19
7
|
|
|
20
|
-
# ── Anthropic
|
|
21
|
-
# Required when LLM_PROVIDER=anthropic
|
|
8
|
+
# ── Anthropic ─────────────────────────────────────────────────
|
|
22
9
|
ANTHROPIC_API_KEY=sk-ant-your-key-here
|
|
23
10
|
|
|
11
|
+
# ── Azure OpenAI ──────────────────────────────────────────────
|
|
12
|
+
AZURE_OPENAI_ENDPOINT=
|
|
13
|
+
AZURE_OPENAI_API_KEY=
|
|
14
|
+
AZURE_OPENAI_DEPLOYMENT=clg-ms-gtp4turbo
|
|
15
|
+
AZURE_OPENAI_API_VERSION=2024-05-01-preview
|
|
24
16
|
|
|
25
|
-
# ── Ollama
|
|
26
|
-
|
|
27
|
-
#
|
|
28
|
-
|
|
29
|
-
#
|
|
30
|
-
# OLLAMA_BASE_URL=http://localhost:11434
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
# ── Azure OpenAI ───────────────────────────────────────────────
|
|
34
|
-
# Required when LLM_PROVIDER=azure
|
|
35
|
-
#
|
|
36
|
-
# Your Azure OpenAI resource endpoint:
|
|
37
|
-
# AZURE_OPENAI_ENDPOINT=https://<your-resource>.openai.azure.com
|
|
38
|
-
#
|
|
39
|
-
# API key from Azure portal → Keys and Endpoint:
|
|
40
|
-
# AZURE_OPENAI_API_KEY=your-azure-key-here
|
|
41
|
-
#
|
|
42
|
-
# The name of the deployment you created in Azure AI Studio
|
|
43
|
-
# (this is also used as the model parameter):
|
|
44
|
-
# AZURE_OPENAI_DEPLOYMENT=gpt-4o
|
|
45
|
-
#
|
|
46
|
-
# API version — default is fine unless you need a specific one:
|
|
47
|
-
# AZURE_OPENAI_API_VERSION=2024-05-01-preview
|
|
17
|
+
# ── Ollama (local) ────────────────────────────────────────────
|
|
18
|
+
OLLAMA_BASE_URL=http://localhost:11434/
|
|
19
|
+
# Set to true if your Ollama model supports tool/function calling
|
|
20
|
+
OLLAMA_TOOLS=false
|
|
48
21
|
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
# Whether to read the project's .gitignore and add its patterns
|
|
52
|
-
# to the ingest ignore list. Set to "false" to disable.
|
|
53
|
-
# Default: true
|
|
54
|
-
# INGEST_USE_GITIGNORE=true
|
|
55
|
-
|
|
56
|
-
# Extra glob patterns to ignore during ingest (comma-separated).
|
|
57
|
-
# Added on top of the built-in baseline and .gitignore patterns.
|
|
58
|
-
# Example: INGEST_EXTRA_IGNORE=**/fixtures/**,**/__snapshots__/**
|
|
22
|
+
# ── Ingest ────────────────────────────────────────────────────
|
|
23
|
+
INGEST_USE_GITIGNORE=true
|
|
59
24
|
# INGEST_EXTRA_IGNORE=
|
|
60
25
|
|
|
61
|
-
|
|
62
|
-
# ── Server ─────────────────────────────────────────────────────
|
|
26
|
+
# ── Server ────────────────────────────────────────────────────
|
|
63
27
|
PORT=3000
|
|
64
28
|
|
|
29
|
+
# ── Intelligence ──────────────────────────────────────────────
|
|
30
|
+
ENABLE_DREAMER=true
|
|
31
|
+
DREAM_INTERVAL_MINUTES=30
|
|
32
|
+
ENABLE_MONITOR=true
|
|
33
|
+
ENABLE_CRON=true
|
|
65
34
|
|
|
66
|
-
# ── Logging
|
|
67
|
-
# Options: error | warn | info | debug
|
|
35
|
+
# ── Logging ───────────────────────────────────────────────────
|
|
68
36
|
LOG_LEVEL=info
|