consult-llm-mcp 2.8.0 → 2.10.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (3) hide show
  1. package/CHANGELOG.md +14 -0
  2. package/README.md +58 -7
  3. package/package.json +5 -5
package/CHANGELOG.md CHANGED
@@ -1,5 +1,19 @@
1
1
  # Changelog
2
2
 
3
+ ## v2.8.0 (2026-03-13)
4
+
5
+ - Replaced hardcoded model enum with abstract selectors (`gemini`, `openai`,
6
+ `deepseek`) that resolve to the best available model at query time. This
7
+ avoids the need to hardcode a specific model in the caller side.
8
+ - Responses now include a `[model:xxx]` prefix showing which concrete model was
9
+ used
10
+ - Default Codex reasoning effort to "high" (was previously unset)
11
+ - Monitor: added Task column to active and history tables
12
+ - Monitor: show task mode and reasoning effort in detail view header
13
+ - Monitor: press `s` in detail view to toggle system prompt display
14
+ - Monitor: system prompt is now recorded in sidecar event files for viewing in
15
+ the TUI
16
+
3
17
  ## v2.7.4 (2026-03-13)
4
18
 
5
19
  - Fixed Linux prebuilt binaries failing on older distros due to glibc version
package/README.md CHANGED
@@ -23,18 +23,17 @@ to bring in the heavy artillery. Supports multi-turn conversations.
23
23
  ```
24
24
 
25
25
  [Quick start](#quick-start) · [Configuration](#configuration) ·
26
- [Skills](#skills) · [Monitor TUI](#monitor) · [Changelog](CHANGELOG.md)
26
+ [Skills](#skills) · [Monitor TUI](#monitor) · [Why MCP?](#why-mcp-and-not-cli) ·
27
+ [Changelog](CHANGELOG.md)
27
28
 
28
29
  ## Features
29
30
 
30
31
  - Query powerful AI models (GPT-5.4, Gemini 3.1 Pro, DeepSeek Reasoner) with
31
32
  relevant files as context
32
- - Direct queries with optional file context
33
- - Include git changes for code review and analysis
34
- - Comprehensive logging with cost estimation
33
+ - Include git changes for code review
34
+ - Comprehensive logging with cost estimation (if using API)
35
35
  - [Monitor TUI](#monitor): Real-time dashboard for watching active consultations
36
- - [Gemini CLI backend](#gemini-cli): Use the `gemini` CLI to take advantage of
37
- [free quota](https://developers.google.com/gemini-code-assist/resources/quotas#quotas-for-agent-mode-gemini-cli)
36
+ - [Gemini CLI backend](#gemini-cli): Use the `gemini` CLI for Gemini models
38
37
  - [Codex CLI backend](#codex-cli): Use the `codex` CLI for OpenAI models
39
38
  - [Cursor CLI backend](#cursor-cli): Use the `cursor-agent` CLI to route GPT and
40
39
  Gemini models through a single tool
@@ -64,7 +63,7 @@ to bring in the heavy artillery. Supports multi-turn conversations.
64
63
  [Codex CLI](#codex-cli). No API keys required, just `gemini login` and
65
64
  `codex login`.
66
65
 
67
- **With binary** (no Node.js required, but no auto-update):
66
+ **With binary** (comes with the monitor TUI, no Node.js required):
68
67
 
69
68
  ```bash
70
69
  curl -fsSL https://raw.githubusercontent.com/raine/consult-llm-mcp/main/scripts/install.sh | bash
@@ -520,6 +519,12 @@ See the "Using web mode..." example above for a concrete transcript.
520
519
  (optional)
521
520
  - Overrides the default `~/.consult-llm-mcp/SYSTEM_PROMPT.md` location
522
521
  - Useful for project-specific prompts
522
+ - `CONSULT_LLM_NO_UPDATE_CHECK` - Disable automatic update checking on server
523
+ startup (optional)
524
+ - Set to `1` to disable
525
+ - By default, the server checks for new versions in the background every 24
526
+ hours and logs a notice when an update is available
527
+ - Only applies to binary installs — npm installs are never checked
523
528
  - `MCP_DEBUG_STDIN` - Log raw JSON-RPC messages received on stdin (optional)
524
529
  - Set to `1` to enable
525
530
  - Logs every message as `RAW RECV` entries and poll timing gaps as
@@ -818,6 +823,52 @@ forth before synthesizing and implementing. See
818
823
  > /debate-vs --gemini design the multi-tenant isolation strategy
819
824
  ```
820
825
 
826
+ ## Updating
827
+
828
+ **Binary installs:**
829
+
830
+ ```bash
831
+ consult-llm-mcp update
832
+ ```
833
+
834
+ Downloads the latest release from GitHub with SHA-256 checksum verification. If
835
+ `consult-llm-monitor` is found alongside the binary, it's updated too.
836
+
837
+ The server also checks for updates in the background on startup (every 24 hours)
838
+ and logs a notice when a newer version is available. Disable with
839
+ `CONSULT_LLM_NO_UPDATE_CHECK=1`.
840
+
841
+ ## Why MCP and not CLI?
842
+
843
+ The server maps one `model` parameter onto five backends (OpenAI API, Gemini
844
+ API, Gemini CLI, Codex CLI, Cursor CLI) with different commands, streaming
845
+ formats, output schemas, file handling, and resume semantics. Doing this through
846
+ agent Bash calls would push all of that per-provider plumbing into the agent or
847
+ a wrapper script/CLI.
848
+
849
+ MCP also sidesteps shell escaping. Prompts contain code with backticks, `$`, and
850
+ quotes. Passing one model's code-heavy response into another call breaks bash
851
+ quoting and requires temp files. MCP passes structured JSON instead.
852
+
853
+ Multi-turn workflows add more friction as a CLI. To continue a conversation, the
854
+ agent needs to find a session ID in the CLI's output and pass it back as a flag
855
+ on the next invocation. With MCP, the agent passes `thread_id` as a parameter
856
+ and the server handles the provider-specific resume mechanics internally.
857
+
858
+ The MCP tool is also easier to compose into [skills](#skills). `/consult`,
859
+ `/collab`, and `/debate` all just say "call `consult_llm` with these
860
+ parameters." A CLI version would need each skill to either teach the agent the
861
+ CLI's interface or reference a separate skill that does. A skill that
862
+ orchestrates a multi-model debate is ~90 lines with MCP. As shell commands, the
863
+ same skill would either balloon into hundreds of lines of escaping rules and
864
+ stdout parsing, or depend on another skill that teaches the agent how to call
865
+ each CLI.
866
+
867
+ If you only need a single provider with simple prompts, a Bash call to `gemini`
868
+ or `codex` with some `jq` filtering will work fine. MCP starts to make more
869
+ sense with multiple backends, multi-turn conversations across providers, or
870
+ custom workflows that nicely compose on top.
871
+
821
872
  ## Development
822
873
 
823
874
  To work on the MCP server locally and use your development version:
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "consult-llm-mcp",
3
- "version": "2.8.0",
3
+ "version": "2.10.0",
4
4
  "description": "MCP server for consulting powerful AI models",
5
5
  "repository": {
6
6
  "type": "git",
@@ -31,9 +31,9 @@
31
31
  "ai"
32
32
  ],
33
33
  "optionalDependencies": {
34
- "consult-llm-mcp-darwin-arm64": "2.8.0",
35
- "consult-llm-mcp-darwin-x64": "2.8.0",
36
- "consult-llm-mcp-linux-x64": "2.8.0",
37
- "consult-llm-mcp-linux-arm64": "2.8.0"
34
+ "consult-llm-mcp-darwin-arm64": "2.10.0",
35
+ "consult-llm-mcp-darwin-x64": "2.10.0",
36
+ "consult-llm-mcp-linux-x64": "2.10.0",
37
+ "consult-llm-mcp-linux-arm64": "2.10.0"
38
38
  }
39
39
  }