consult-llm-mcp 2.11.0 → 2.12.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +17 -0
- package/README.md +74 -6
- package/package.json +5 -5
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,22 @@
|
|
|
1
1
|
# Changelog
|
|
2
2
|
|
|
3
|
+
## v2.11.0 (2026-03-27)
|
|
4
|
+
|
|
5
|
+
- Added multi-turn thread support for API backends. Threads are stored as JSON
|
|
6
|
+
files under `$XDG_STATE_HOME/consult-llm-mcp/threads/` and replayed as the
|
|
7
|
+
messages array on each call. Expired threads (>7 days) are cleaned up
|
|
8
|
+
automatically. All backends now support `thread_id`.
|
|
9
|
+
- Fixed API cost tracking undercounting tokens for thinking models (e.g.
|
|
10
|
+
gemini-3.1-pro-preview). Thinking tokens excluded from `completion_tokens` are
|
|
11
|
+
now derived from `total_tokens`.
|
|
12
|
+
- Monitor: show cost information in history table, detail view header, usage
|
|
13
|
+
separator lines, and thread detail header. Cost is only shown for API backend
|
|
14
|
+
consultations.
|
|
15
|
+
- Monitor: show files as compact path list in detail view instead of inlined
|
|
16
|
+
file contents
|
|
17
|
+
- Fixed `reasoning_effort` incorrectly showing for non-codex models on
|
|
18
|
+
cursor_cli backend
|
|
19
|
+
|
|
3
20
|
## v2.10.0 (2026-03-15)
|
|
4
21
|
|
|
5
22
|
- Monitor: cycle between sibling consultations (started around the same time)
|
package/README.md
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
# consult-llm-mcp
|
|
2
2
|
|
|
3
3
|
An MCP server that lets Claude Code consult stronger AI models (GPT-5.4, Gemini
|
|
4
|
-
3.1 Pro, DeepSeek Reasoner) when Sonnet has you running in circles and you need
|
|
4
|
+
3.1 Pro, DeepSeek Reasoner, MiniMax M2.7) when Sonnet has you running in circles and you need
|
|
5
5
|
to bring in the heavy artillery. Supports multi-turn conversations.
|
|
6
6
|
|
|
7
7
|
```
|
|
@@ -28,7 +28,8 @@ to bring in the heavy artillery. Supports multi-turn conversations.
|
|
|
28
28
|
|
|
29
29
|
## Features
|
|
30
30
|
|
|
31
|
-
- Query powerful AI models (GPT-5.4, Gemini 3.1 Pro, DeepSeek Reasoner
|
|
31
|
+
- Query powerful AI models (GPT-5.4, Gemini 3.1 Pro, DeepSeek Reasoner, MiniMax
|
|
32
|
+
M2.7) with
|
|
32
33
|
relevant files as context
|
|
33
34
|
- Include git changes for code review
|
|
34
35
|
- Comprehensive logging with cost estimation (if using API)
|
|
@@ -37,6 +38,8 @@ to bring in the heavy artillery. Supports multi-turn conversations.
|
|
|
37
38
|
- [Codex CLI backend](#codex-cli): Use the `codex` CLI for OpenAI models
|
|
38
39
|
- [Cursor CLI backend](#cursor-cli): Use the `cursor-agent` CLI to route GPT and
|
|
39
40
|
Gemini models through a single tool
|
|
41
|
+
- [OpenCode CLI backend](#opencode-cli): Use `opencode` CLI with Copilot, OpenRouter,
|
|
42
|
+
or any of 75+ providers
|
|
40
43
|
- [Multi-turn conversations](#multi-turn-conversations): Resume CLI sessions
|
|
41
44
|
across requests with `thread_id`
|
|
42
45
|
- [Web mode](#web-mode): Copy formatted prompts to clipboard for browser-based
|
|
@@ -85,6 +88,7 @@ to bring in the heavy artillery. Supports multi-turn conversations.
|
|
|
85
88
|
-e OPENAI_API_KEY=your_openai_key \
|
|
86
89
|
-e GEMINI_API_KEY=your_gemini_key \
|
|
87
90
|
-e DEEPSEEK_API_KEY=your_deepseek_key \
|
|
91
|
+
-e MINIMAX_API_KEY=your_minimax_key \
|
|
88
92
|
-- npx -y consult-llm-mcp
|
|
89
93
|
```
|
|
90
94
|
|
|
@@ -335,6 +339,7 @@ Each model is routed to a **backend** — either an API endpoint or a CLI tool.
|
|
|
335
339
|
| **Gemini CLI** | Shells out to `gemini` CLI | Free quota (Gemini), existing subscriptions, or prefer CLI tools |
|
|
336
340
|
| **Codex CLI** | Shells out to `codex` CLI | OpenAI models via Codex subscription |
|
|
337
341
|
| **Cursor CLI** | Shells out to `cursor-agent` CLI | Route GPT and Gemini through one tool |
|
|
342
|
+
| **OpenCode CLI** | Shells out to `opencode` CLI | Use Copilot subscription, OpenCode's 75+ providers |
|
|
338
343
|
| **Web** | Copies prompt to clipboard | You prefer browser UIs or want to review prompts |
|
|
339
344
|
|
|
340
345
|
### API (default)
|
|
@@ -429,6 +434,58 @@ review), allow them in `~/.cursor/cli-config.json`:
|
|
|
429
434
|
|
|
430
435
|
Glob patterns are supported. The `deny` list takes precedence over `allow`.
|
|
431
436
|
|
|
437
|
+
#### OpenCode CLI
|
|
438
|
+
|
|
439
|
+
Use [OpenCode](https://opencode.ai) as a backend to route models through any of
|
|
440
|
+
its 75+ supported providers — including GitHub Copilot, OpenRouter, and local
|
|
441
|
+
models via Ollama.
|
|
442
|
+
|
|
443
|
+
**Requirements:**
|
|
444
|
+
|
|
445
|
+
1. Install [OpenCode](https://opencode.ai/docs/installation/)
|
|
446
|
+
2. Configure providers via `opencode providers`
|
|
447
|
+
|
|
448
|
+
**Setup:**
|
|
449
|
+
|
|
450
|
+
```bash
|
|
451
|
+
# Route MiniMax models through OpenCode
|
|
452
|
+
claude mcp add consult-llm \
|
|
453
|
+
-e CONSULT_LLM_MINIMAX_BACKEND=opencode \
|
|
454
|
+
-- npx -y consult-llm-mcp
|
|
455
|
+
|
|
456
|
+
# Route OpenAI models through Copilot subscription
|
|
457
|
+
claude mcp add consult-llm \
|
|
458
|
+
-e CONSULT_LLM_OPENAI_BACKEND=opencode \
|
|
459
|
+
-e CONSULT_LLM_OPENCODE_OPENAI_PROVIDER=copilot \
|
|
460
|
+
-- npx -y consult-llm-mcp
|
|
461
|
+
|
|
462
|
+
# Route everything through OpenCode
|
|
463
|
+
claude mcp add consult-llm \
|
|
464
|
+
-e CONSULT_LLM_OPENAI_BACKEND=opencode \
|
|
465
|
+
-e CONSULT_LLM_GEMINI_BACKEND=opencode \
|
|
466
|
+
-e CONSULT_LLM_DEEPSEEK_BACKEND=opencode \
|
|
467
|
+
-e CONSULT_LLM_MINIMAX_BACKEND=opencode \
|
|
468
|
+
-- npx -y consult-llm-mcp
|
|
469
|
+
```
|
|
470
|
+
|
|
471
|
+
The executor maps model IDs to OpenCode's `provider/model` format automatically.
|
|
472
|
+
For example, `MiniMax-M2.7` becomes `opencode run --model minimax/MiniMax-M2.7`.
|
|
473
|
+
|
|
474
|
+
**Provider prefix overrides:**
|
|
475
|
+
|
|
476
|
+
By default, each provider family maps to its natural OpenCode provider ID
|
|
477
|
+
(`openai`, `google`, `deepseek`, `minimax`). Override with per-family env vars
|
|
478
|
+
when you want to route through a different OpenCode provider:
|
|
479
|
+
|
|
480
|
+
- `CONSULT_LLM_OPENCODE_OPENAI_PROVIDER` — default: `openai`
|
|
481
|
+
- `CONSULT_LLM_OPENCODE_GEMINI_PROVIDER` — default: `google`
|
|
482
|
+
- `CONSULT_LLM_OPENCODE_DEEPSEEK_PROVIDER` — default: `deepseek`
|
|
483
|
+
- `CONSULT_LLM_OPENCODE_MINIMAX_PROVIDER` — default: `minimax`
|
|
484
|
+
- `CONSULT_LLM_OPENCODE_PROVIDER` — global fallback for all families
|
|
485
|
+
|
|
486
|
+
For example, `CONSULT_LLM_OPENCODE_OPENAI_PROVIDER=copilot` turns
|
|
487
|
+
`gpt-5.2` into `opencode run --model copilot/gpt-5.2`.
|
|
488
|
+
|
|
432
489
|
#### Multi-turn conversations
|
|
433
490
|
|
|
434
491
|
CLI backends support multi-turn conversations via the `thread_id` parameter. The
|
|
@@ -492,14 +549,20 @@ See the "Using web mode..." example above for a concrete transcript.
|
|
|
492
549
|
- `GEMINI_API_KEY` - Your Google AI API key (required for Gemini models in API
|
|
493
550
|
mode)
|
|
494
551
|
- `DEEPSEEK_API_KEY` - Your DeepSeek API key (required for DeepSeek models)
|
|
552
|
+
- `MINIMAX_API_KEY` - Your MiniMax API key (required for MiniMax models)
|
|
495
553
|
- `CONSULT_LLM_DEFAULT_MODEL` - Override the default model (optional)
|
|
496
|
-
- Accepts selectors (`gemini`, `openai`, `deepseek`) or exact model
|
|
554
|
+
- Accepts selectors (`gemini`, `openai`, `deepseek`, `minimax`) or exact model
|
|
555
|
+
IDs
|
|
497
556
|
(`gpt-5.4`, `gemini-3.1-pro-preview`, etc.)
|
|
498
557
|
- Selectors are resolved to the best available model at startup
|
|
499
558
|
- `CONSULT_LLM_GEMINI_BACKEND` - Backend for Gemini models (optional)
|
|
500
|
-
- Options: `api` (default), `gemini-cli`, `cursor-cli`
|
|
559
|
+
- Options: `api` (default), `gemini-cli`, `cursor-cli`, `opencode`
|
|
501
560
|
- `CONSULT_LLM_OPENAI_BACKEND` - Backend for OpenAI models (optional)
|
|
502
|
-
- Options: `api` (default), `codex-cli`, `cursor-cli`
|
|
561
|
+
- Options: `api` (default), `codex-cli`, `cursor-cli`, `opencode`
|
|
562
|
+
- `CONSULT_LLM_DEEPSEEK_BACKEND` - Backend for DeepSeek models (optional)
|
|
563
|
+
- Options: `api` (default), `opencode`
|
|
564
|
+
- `CONSULT_LLM_MINIMAX_BACKEND` - Backend for MiniMax models (optional)
|
|
565
|
+
- Options: `api` (default), `opencode`
|
|
503
566
|
- `CONSULT_LLM_ALLOWED_MODELS` - Restrict which concrete models can be used
|
|
504
567
|
(optional)
|
|
505
568
|
- Comma-separated list, e.g., `gpt-5.4,gemini-3.1-pro-preview`
|
|
@@ -511,10 +574,14 @@ See the "Using web mode..." example above for a concrete transcript.
|
|
|
511
574
|
- Comma-separated list, e.g., `grok-3,kimi-k2.5`
|
|
512
575
|
- Merged with built-in models and included in the tool schema
|
|
513
576
|
- Useful for newly released models with a known provider prefix (`gpt-`,
|
|
514
|
-
`gemini-`, `deepseek-`)
|
|
577
|
+
`gemini-`, `deepseek-`, `MiniMax-`)
|
|
515
578
|
- `CONSULT_LLM_CODEX_REASONING_EFFORT` - Configure reasoning effort for Codex
|
|
516
579
|
CLI (optional, default: `high`)
|
|
517
580
|
- See [Codex CLI](#codex-cli) for details and available options
|
|
581
|
+
- `CONSULT_LLM_OPENCODE_PROVIDER` - Global OpenCode provider prefix (optional)
|
|
582
|
+
- Overrides the default provider ID for all families when using the `opencode`
|
|
583
|
+
backend
|
|
584
|
+
- See [OpenCode CLI](#opencode-cli) for details and per-family overrides
|
|
518
585
|
- `CONSULT_LLM_SYSTEM_PROMPT_PATH` - Custom path to system prompt file
|
|
519
586
|
(optional)
|
|
520
587
|
- Overrides the default `~/.consult-llm-mcp/SYSTEM_PROMPT.md` location
|
|
@@ -645,6 +712,7 @@ models complex questions.
|
|
|
645
712
|
- **gemini-3-pro-preview**: Google's Gemini 3 Pro Preview
|
|
646
713
|
- **gemini-3.1-pro-preview**: Google's Gemini 3.1 Pro Preview
|
|
647
714
|
- **deepseek-reasoner**: DeepSeek's reasoning model
|
|
715
|
+
- **MiniMax-M2.7**: MiniMax's M2.7 reasoning model (204K context)
|
|
648
716
|
- **gpt-5.4**: OpenAI's GPT-5.4 model
|
|
649
717
|
- **gpt-5.2**: OpenAI's GPT-5.2 model
|
|
650
718
|
- **gpt-5.3-codex**: OpenAI's Codex model based on GPT-5.3
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "consult-llm-mcp",
|
|
3
|
-
"version": "2.
|
|
3
|
+
"version": "2.12.0",
|
|
4
4
|
"description": "MCP server for consulting powerful AI models",
|
|
5
5
|
"repository": {
|
|
6
6
|
"type": "git",
|
|
@@ -31,9 +31,9 @@
|
|
|
31
31
|
"ai"
|
|
32
32
|
],
|
|
33
33
|
"optionalDependencies": {
|
|
34
|
-
"consult-llm-mcp-darwin-arm64": "2.
|
|
35
|
-
"consult-llm-mcp-darwin-x64": "2.
|
|
36
|
-
"consult-llm-mcp-linux-x64": "2.
|
|
37
|
-
"consult-llm-mcp-linux-arm64": "2.
|
|
34
|
+
"consult-llm-mcp-darwin-arm64": "2.12.0",
|
|
35
|
+
"consult-llm-mcp-darwin-x64": "2.12.0",
|
|
36
|
+
"consult-llm-mcp-linux-x64": "2.12.0",
|
|
37
|
+
"consult-llm-mcp-linux-arm64": "2.12.0"
|
|
38
38
|
}
|
|
39
39
|
}
|