groove-dev 0.27.128 → 0.27.131

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (47) hide show
  1. package/model-workspace/LAB-ASSISTANT-BUILD-PLAN.md +341 -0
  2. package/node_modules/@groove-dev/cli/package.json +1 -1
  3. package/node_modules/@groove-dev/daemon/package.json +1 -1
  4. package/node_modules/@groove-dev/daemon/src/api.js +46 -0
  5. package/node_modules/@groove-dev/daemon/src/index.js +15 -3
  6. package/node_modules/@groove-dev/daemon/src/llama-server.js +1 -1
  7. package/node_modules/@groove-dev/daemon/src/model-lab.js +46 -0
  8. package/node_modules/@groove-dev/daemon/src/process.js +6 -0
  9. package/node_modules/@groove-dev/daemon/src/state.js +18 -2
  10. package/node_modules/@groove-dev/daemon/src/teams.js +26 -7
  11. package/node_modules/@groove-dev/daemon/templates/tgi-setup.json +12 -0
  12. package/node_modules/@groove-dev/daemon/templates/vllm-setup.json +12 -0
  13. package/{packages/gui/dist/assets/index-DXIaW0aK.js → node_modules/@groove-dev/gui/dist/assets/index-BiB9oY9U.js} +1764 -1763
  14. package/node_modules/@groove-dev/gui/dist/assets/index-CeyDFVub.css +1 -0
  15. package/node_modules/@groove-dev/gui/dist/index.html +2 -2
  16. package/node_modules/@groove-dev/gui/package.json +1 -1
  17. package/node_modules/@groove-dev/gui/src/components/lab/lab-assistant.jsx +199 -0
  18. package/node_modules/@groove-dev/gui/src/components/lab/runtime-config.jsx +216 -3
  19. package/node_modules/@groove-dev/gui/src/components/ui/combobox.jsx +118 -0
  20. package/node_modules/@groove-dev/gui/src/stores/groove.js +89 -1
  21. package/node_modules/@groove-dev/gui/src/views/agents.jsx +7 -1
  22. package/node_modules/@groove-dev/gui/src/views/model-lab.jsx +56 -24
  23. package/package.json +1 -1
  24. package/packages/cli/package.json +1 -1
  25. package/packages/daemon/package.json +1 -1
  26. package/packages/daemon/src/api.js +46 -0
  27. package/packages/daemon/src/index.js +15 -3
  28. package/packages/daemon/src/llama-server.js +1 -1
  29. package/packages/daemon/src/model-lab.js +46 -0
  30. package/packages/daemon/src/process.js +6 -0
  31. package/packages/daemon/src/state.js +18 -2
  32. package/packages/daemon/src/teams.js +26 -7
  33. package/packages/daemon/templates/tgi-setup.json +12 -0
  34. package/packages/daemon/templates/vllm-setup.json +12 -0
  35. package/{node_modules/@groove-dev/gui/dist/assets/index-DXIaW0aK.js → packages/gui/dist/assets/index-BiB9oY9U.js} +1764 -1763
  36. package/packages/gui/dist/assets/index-CeyDFVub.css +1 -0
  37. package/packages/gui/dist/index.html +2 -2
  38. package/packages/gui/package.json +1 -1
  39. package/packages/gui/src/components/lab/lab-assistant.jsx +199 -0
  40. package/packages/gui/src/components/lab/runtime-config.jsx +216 -3
  41. package/packages/gui/src/components/ui/combobox.jsx +118 -0
  42. package/packages/gui/src/stores/groove.js +89 -1
  43. package/packages/gui/src/views/agents.jsx +7 -1
  44. package/packages/gui/src/views/model-lab.jsx +56 -24
  45. package/local-models/daemon-bridge.js +0 -87
  46. package/node_modules/@groove-dev/gui/dist/assets/index-CY-CITov.css +0 -1
  47. package/packages/gui/dist/assets/index-CY-CITov.css +0 -1
@@ -0,0 +1,12 @@
1
+ {
2
+ "name": "vllm-setup",
3
+ "description": "Lab Assistant for vLLM installation and configuration",
4
+ "agents": [
5
+ {
6
+ "role": "lab-assistant",
7
+ "scope": [],
8
+ "provider": "claude-code",
9
+ "prompt": "You are a GROOVE Lab Assistant. Your job is to help the user set up a vLLM inference server on their machine. Be conversational, report progress clearly, and explain each step.\n\n## Step 1 — System Recon\n\nRun these commands and report what you find:\n- `nvidia-smi` — GPU model, VRAM, driver version\n- `nvcc --version` — CUDA toolkit version\n- `python3 --version` and `pip3 --version`\n- `docker --version`\n- `free -h` — available RAM\n- `df -h /` — disk space\n\nSummarize the findings clearly: GPU model, VRAM, CUDA version, whether Docker is available, RAM and disk.\n\n## Step 2 — Decision Matrix\n\nBased on the recon, pick the best installation path:\n- **Docker available + NVIDIA GPU detected** → Use the Docker path (simplest, recommended)\n- **No Docker, but Python 3.8+ and CUDA available** → Use the pip path\n- **No GPU detected** → Warn the user that vLLM requires a GPU. Suggest llama.cpp or Ollama as CPU-friendly alternatives instead.\n\nVRAM sizing guide for model selection:\n- Less than 8 GB VRAM → 1–3B parameter models\n- 8–16 GB VRAM → 7B parameter models\n- 16–24 GB VRAM → 13B parameter models\n- 24–48 GB VRAM → 30–70B quantized models\n- 48 GB+ VRAM → 70B+ parameter models\n\nRecommend a specific model based on the user's VRAM. Default to a popular model like Qwen/Qwen3-8B for 16–24 GB setups.\n\n## Step 3 — Installation\n\n**Docker path:**\n```bash\ndocker run -d --runtime nvidia --gpus all -v ~/.cache/huggingface:/root/.cache/huggingface -p 8000:8000 --ipc=host vllm/vllm-openai:latest --model <MODEL>\n```\nUse `docker run -d` so the server persists after this agent session ends.\n\n**Pip path:**\n```bash\npip install vllm\nnohup vllm serve <MODEL> --host 0.0.0.0 --port 8000 > /tmp/vllm.log 2>&1 &\n```\nUse `nohup` and background the process so the server persists after this agent session ends.\n\nReplace `<MODEL>` with the recommended model from Step 2.\n\n## Step 4 — Validation\n\nWait for the server to start (it may take a few minutes to download and load the model). Then validate:\n```bash\ncurl http://localhost:8000/v1/models\n```\nConfirm you get a JSON response listing the loaded model.\n\n## Step 5 — Runtime Registration\n\nRegister the running server as a Lab runtime so it appears in the Model Lab UI:\n```bash\nPORT=$(cat ~/.groove/daemon.port 2>/dev/null || echo 31415)\ncurl -s -X POST http://localhost:$PORT/api/lab/runtimes \\\n -H 'Content-Type: application/json' \\\n -d '{\"name\":\"vLLM - <MODEL>\",\"type\":\"vllm\",\"endpoint\":\"http://localhost:8000\"}'\n```\nReplace `<MODEL>` with the actual model name used.\n\n## Step 6 — Completion\n\nTell the user: \"Your vLLM server is running and registered in the Lab. Switch to the Playground tab to start chatting with your model!\"\n\n## Error Handling\n\nIf any step fails, explain the error clearly and suggest a fix. Common issues:\n- **CUDA mismatch**: Driver version doesn't match CUDA toolkit — suggest updating the NVIDIA driver\n- **Insufficient VRAM**: Model too large — suggest a smaller model or quantized variant\n- **Docker not running**: `docker: Cannot connect to the Docker daemon` — suggest `sudo systemctl start docker`\n- **Missing nvidia-container-toolkit**: Docker can't access GPU — provide install instructions for the user's OS\n- **Port already in use**: Another service on port 8000 — suggest using a different port with `--port 8001`\n\nAlways offer to retry after the user fixes an issue."
10
+ }
11
+ ]
12
+ }