consult-llm-mcp 2.8.0 → 2.10.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +14 -0
- package/README.md +58 -7
- package/package.json +5 -5
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,19 @@
|
|
|
1
1
|
# Changelog
|
|
2
2
|
|
|
3
|
+
## v2.8.0 (2026-03-13)
|
|
4
|
+
|
|
5
|
+
- Replaced hardcoded model enum with abstract selectors (`gemini`, `openai`,
|
|
6
|
+
`deepseek`) that resolve to the best available model at query time. This
|
|
7
|
+
avoids the need to hardcode a specific model in the caller side.
|
|
8
|
+
- Responses now include a `[model:xxx]` prefix showing which concrete model was
|
|
9
|
+
used
|
|
10
|
+
- Default Codex reasoning effort to "high" (was previously unset)
|
|
11
|
+
- Monitor: added Task column to active and history tables
|
|
12
|
+
- Monitor: show task mode and reasoning effort in detail view header
|
|
13
|
+
- Monitor: press `s` in detail view to toggle system prompt display
|
|
14
|
+
- Monitor: system prompt is now recorded in sidecar event files for viewing in
|
|
15
|
+
the TUI
|
|
16
|
+
|
|
3
17
|
## v2.7.4 (2026-03-13)
|
|
4
18
|
|
|
5
19
|
- Fixed Linux prebuilt binaries failing on older distros due to glibc version
|
package/README.md
CHANGED
|
@@ -23,18 +23,17 @@ to bring in the heavy artillery. Supports multi-turn conversations.
|
|
|
23
23
|
```
|
|
24
24
|
|
|
25
25
|
[Quick start](#quick-start) · [Configuration](#configuration) ·
|
|
26
|
-
[Skills](#skills) · [Monitor TUI](#monitor) · [
|
|
26
|
+
[Skills](#skills) · [Monitor TUI](#monitor) · [Why MCP?](#why-mcp-and-not-cli) ·
|
|
27
|
+
[Changelog](CHANGELOG.md)
|
|
27
28
|
|
|
28
29
|
## Features
|
|
29
30
|
|
|
30
31
|
- Query powerful AI models (GPT-5.4, Gemini 3.1 Pro, DeepSeek Reasoner) with
|
|
31
32
|
relevant files as context
|
|
32
|
-
-
|
|
33
|
-
-
|
|
34
|
-
- Comprehensive logging with cost estimation
|
|
33
|
+
- Include git changes for code review
|
|
34
|
+
- Comprehensive logging with cost estimation (if using API)
|
|
35
35
|
- [Monitor TUI](#monitor): Real-time dashboard for watching active consultations
|
|
36
|
-
- [Gemini CLI backend](#gemini-cli): Use the `gemini` CLI
|
|
37
|
-
[free quota](https://developers.google.com/gemini-code-assist/resources/quotas#quotas-for-agent-mode-gemini-cli)
|
|
36
|
+
- [Gemini CLI backend](#gemini-cli): Use the `gemini` CLI for Gemini models
|
|
38
37
|
- [Codex CLI backend](#codex-cli): Use the `codex` CLI for OpenAI models
|
|
39
38
|
- [Cursor CLI backend](#cursor-cli): Use the `cursor-agent` CLI to route GPT and
|
|
40
39
|
Gemini models through a single tool
|
|
@@ -64,7 +63,7 @@ to bring in the heavy artillery. Supports multi-turn conversations.
|
|
|
64
63
|
[Codex CLI](#codex-cli). No API keys required, just `gemini login` and
|
|
65
64
|
`codex login`.
|
|
66
65
|
|
|
67
|
-
**With binary** (no Node.js required
|
|
66
|
+
**With binary** (comes with the monitor TUI, no Node.js required):
|
|
68
67
|
|
|
69
68
|
```bash
|
|
70
69
|
curl -fsSL https://raw.githubusercontent.com/raine/consult-llm-mcp/main/scripts/install.sh | bash
|
|
@@ -520,6 +519,12 @@ See the "Using web mode..." example above for a concrete transcript.
|
|
|
520
519
|
(optional)
|
|
521
520
|
- Overrides the default `~/.consult-llm-mcp/SYSTEM_PROMPT.md` location
|
|
522
521
|
- Useful for project-specific prompts
|
|
522
|
+
- `CONSULT_LLM_NO_UPDATE_CHECK` - Disable automatic update checking on server
|
|
523
|
+
startup (optional)
|
|
524
|
+
- Set to `1` to disable
|
|
525
|
+
- By default, the server checks for new versions in the background every 24
|
|
526
|
+
hours and logs a notice when an update is available
|
|
527
|
+
- Only applies to binary installs — npm installs are never checked
|
|
523
528
|
- `MCP_DEBUG_STDIN` - Log raw JSON-RPC messages received on stdin (optional)
|
|
524
529
|
- Set to `1` to enable
|
|
525
530
|
- Logs every message as `RAW RECV` entries and poll timing gaps as
|
|
@@ -818,6 +823,52 @@ forth before synthesizing and implementing. See
|
|
|
818
823
|
> /debate-vs --gemini design the multi-tenant isolation strategy
|
|
819
824
|
```
|
|
820
825
|
|
|
826
|
+
## Updating
|
|
827
|
+
|
|
828
|
+
**Binary installs:**
|
|
829
|
+
|
|
830
|
+
```bash
|
|
831
|
+
consult-llm-mcp update
|
|
832
|
+
```
|
|
833
|
+
|
|
834
|
+
Downloads the latest release from GitHub with SHA-256 checksum verification. If
|
|
835
|
+
`consult-llm-monitor` is found alongside the binary, it's updated too.
|
|
836
|
+
|
|
837
|
+
The server also checks for updates in the background on startup (every 24 hours)
|
|
838
|
+
and logs a notice when a newer version is available. Disable with
|
|
839
|
+
`CONSULT_LLM_NO_UPDATE_CHECK=1`.
|
|
840
|
+
|
|
841
|
+
## Why MCP and not CLI?
|
|
842
|
+
|
|
843
|
+
The server maps one `model` parameter onto five backends (OpenAI API, Gemini
|
|
844
|
+
API, Gemini CLI, Codex CLI, Cursor CLI) with different commands, streaming
|
|
845
|
+
formats, output schemas, file handling, and resume semantics. Doing this through
|
|
846
|
+
agent Bash calls would push all of that per-provider plumbing into the agent or
|
|
847
|
+
a wrapper script/CLI.
|
|
848
|
+
|
|
849
|
+
MCP also sidesteps shell escaping. Prompts contain code with backticks, `$`, and
|
|
850
|
+
quotes. Passing one model's code-heavy response into another call breaks bash
|
|
851
|
+
quoting and requires temp files. MCP passes structured JSON instead.
|
|
852
|
+
|
|
853
|
+
Multi-turn workflows add more friction as a CLI. To continue a conversation, the
|
|
854
|
+
agent needs to find a session ID in the CLI's output and pass it back as a flag
|
|
855
|
+
on the next invocation. With MCP, the agent passes `thread_id` as a parameter
|
|
856
|
+
and the server handles the provider-specific resume mechanics internally.
|
|
857
|
+
|
|
858
|
+
The MCP tool is also easier to compose into [skills](#skills). `/consult`,
|
|
859
|
+
`/collab`, and `/debate` all just say "call `consult_llm` with these
|
|
860
|
+
parameters." A CLI version would need each skill to either teach the agent the
|
|
861
|
+
CLI's interface or reference a separate skill that does. A skill that
|
|
862
|
+
orchestrates a multi-model debate is ~90 lines with MCP. As shell commands, the
|
|
863
|
+
same skill would either balloon into hundreds of lines of escaping rules and
|
|
864
|
+
stdout parsing, or depend on another skill that teaches the agent how to call
|
|
865
|
+
each CLI.
|
|
866
|
+
|
|
867
|
+
If you only need a single provider with simple prompts, a Bash call to `gemini`
|
|
868
|
+
or `codex` with some `jq` filtering will work fine. MCP starts to make more
|
|
869
|
+
sense with multiple backends, multi-turn conversations across providers, or
|
|
870
|
+
custom workflows that nicely compose on top.
|
|
871
|
+
|
|
821
872
|
## Development
|
|
822
873
|
|
|
823
874
|
To work on the MCP server locally and use your development version:
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "consult-llm-mcp",
|
|
3
|
-
"version": "2.
|
|
3
|
+
"version": "2.10.0",
|
|
4
4
|
"description": "MCP server for consulting powerful AI models",
|
|
5
5
|
"repository": {
|
|
6
6
|
"type": "git",
|
|
@@ -31,9 +31,9 @@
|
|
|
31
31
|
"ai"
|
|
32
32
|
],
|
|
33
33
|
"optionalDependencies": {
|
|
34
|
-
"consult-llm-mcp-darwin-arm64": "2.
|
|
35
|
-
"consult-llm-mcp-darwin-x64": "2.
|
|
36
|
-
"consult-llm-mcp-linux-x64": "2.
|
|
37
|
-
"consult-llm-mcp-linux-arm64": "2.
|
|
34
|
+
"consult-llm-mcp-darwin-arm64": "2.10.0",
|
|
35
|
+
"consult-llm-mcp-darwin-x64": "2.10.0",
|
|
36
|
+
"consult-llm-mcp-linux-x64": "2.10.0",
|
|
37
|
+
"consult-llm-mcp-linux-arm64": "2.10.0"
|
|
38
38
|
}
|
|
39
39
|
}
|