consult-llm-mcp 2.11.0 → 2.12.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (3) hide show
  1. package/CHANGELOG.md +17 -0
  2. package/README.md +74 -6
  3. package/package.json +5 -5
package/CHANGELOG.md CHANGED
@@ -1,5 +1,22 @@
1
1
  # Changelog
2
2
 
3
+ ## v2.11.0 (2026-03-27)
4
+
5
+ - Added multi-turn thread support for API backends. Threads are stored as JSON
6
+ files under `$XDG_STATE_HOME/consult-llm-mcp/threads/` and replayed as the
7
+ messages array on each call. Expired threads (>7 days) are cleaned up
8
+ automatically. All backends now support `thread_id`.
9
+ - Fixed API cost tracking undercounting tokens for thinking models (e.g.
10
+ gemini-3.1-pro-preview). Thinking tokens excluded from `completion_tokens` are
11
+ now derived from `total_tokens`.
12
+ - Monitor: show cost information in history table, detail view header, usage
13
+ separator lines, and thread detail header. Cost is only shown for API backend
14
+ consultations.
15
+ - Monitor: show files as compact path list in detail view instead of inlined
16
+ file contents
17
+ - Fixed `reasoning_effort` incorrectly showing for non-codex models on
18
+ cursor_cli backend
19
+
3
20
  ## v2.10.0 (2026-03-15)
4
21
 
5
22
  - Monitor: cycle between sibling consultations (started around the same time)
package/README.md CHANGED
@@ -1,7 +1,7 @@
1
1
  # consult-llm-mcp
2
2
 
3
3
  An MCP server that lets Claude Code consult stronger AI models (GPT-5.4, Gemini
4
- 3.1 Pro, DeepSeek Reasoner) when Sonnet has you running in circles and you need
4
+ 3.1 Pro, DeepSeek Reasoner, MiniMax M2.7) when Sonnet has you running in circles and you need
5
5
  to bring in the heavy artillery. Supports multi-turn conversations.
6
6
 
7
7
  ```
@@ -28,7 +28,8 @@ to bring in the heavy artillery. Supports multi-turn conversations.
28
28
 
29
29
  ## Features
30
30
 
31
- - Query powerful AI models (GPT-5.4, Gemini 3.1 Pro, DeepSeek Reasoner) with
31
+ - Query powerful AI models (GPT-5.4, Gemini 3.1 Pro, DeepSeek Reasoner, MiniMax
32
+ M2.7) with
32
33
  relevant files as context
33
34
  - Include git changes for code review
34
35
  - Comprehensive logging with cost estimation (if using API)
@@ -37,6 +38,8 @@ to bring in the heavy artillery. Supports multi-turn conversations.
37
38
  - [Codex CLI backend](#codex-cli): Use the `codex` CLI for OpenAI models
38
39
  - [Cursor CLI backend](#cursor-cli): Use the `cursor-agent` CLI to route GPT and
39
40
  Gemini models through a single tool
41
+ - [OpenCode CLI backend](#opencode-cli): Use `opencode` CLI with Copilot, OpenRouter,
42
+ or any of 75+ providers
40
43
  - [Multi-turn conversations](#multi-turn-conversations): Resume CLI sessions
41
44
  across requests with `thread_id`
42
45
  - [Web mode](#web-mode): Copy formatted prompts to clipboard for browser-based
@@ -85,6 +88,7 @@ to bring in the heavy artillery. Supports multi-turn conversations.
85
88
  -e OPENAI_API_KEY=your_openai_key \
86
89
  -e GEMINI_API_KEY=your_gemini_key \
87
90
  -e DEEPSEEK_API_KEY=your_deepseek_key \
91
+ -e MINIMAX_API_KEY=your_minimax_key \
88
92
  -- npx -y consult-llm-mcp
89
93
  ```
90
94
 
@@ -335,6 +339,7 @@ Each model is routed to a **backend** — either an API endpoint or a CLI tool.
335
339
  | **Gemini CLI** | Shells out to `gemini` CLI | Free quota (Gemini), existing subscriptions, or prefer CLI tools |
336
340
  | **Codex CLI** | Shells out to `codex` CLI | OpenAI models via Codex subscription |
337
341
  | **Cursor CLI** | Shells out to `cursor-agent` CLI | Route GPT and Gemini through one tool |
342
+ | **OpenCode CLI** | Shells out to `opencode` CLI | Use Copilot subscription, OpenCode's 75+ providers |
338
343
  | **Web** | Copies prompt to clipboard | You prefer browser UIs or want to review prompts |
339
344
 
340
345
  ### API (default)
@@ -429,6 +434,58 @@ review), allow them in `~/.cursor/cli-config.json`:
429
434
 
430
435
  Glob patterns are supported. The `deny` list takes precedence over `allow`.
431
436
 
437
+ #### OpenCode CLI
438
+
439
+ Use [OpenCode](https://opencode.ai) as a backend to route models through any of
440
+ its 75+ supported providers — including GitHub Copilot, OpenRouter, and local
441
+ models via Ollama.
442
+
443
+ **Requirements:**
444
+
445
+ 1. Install [OpenCode](https://opencode.ai/docs/installation/)
446
+ 2. Configure providers via `opencode providers`
447
+
448
+ **Setup:**
449
+
450
+ ```bash
451
+ # Route MiniMax models through OpenCode
452
+ claude mcp add consult-llm \
453
+ -e CONSULT_LLM_MINIMAX_BACKEND=opencode \
454
+ -- npx -y consult-llm-mcp
455
+
456
+ # Route OpenAI models through Copilot subscription
457
+ claude mcp add consult-llm \
458
+ -e CONSULT_LLM_OPENAI_BACKEND=opencode \
459
+ -e CONSULT_LLM_OPENCODE_OPENAI_PROVIDER=copilot \
460
+ -- npx -y consult-llm-mcp
461
+
462
+ # Route everything through OpenCode
463
+ claude mcp add consult-llm \
464
+ -e CONSULT_LLM_OPENAI_BACKEND=opencode \
465
+ -e CONSULT_LLM_GEMINI_BACKEND=opencode \
466
+ -e CONSULT_LLM_DEEPSEEK_BACKEND=opencode \
467
+ -e CONSULT_LLM_MINIMAX_BACKEND=opencode \
468
+ -- npx -y consult-llm-mcp
469
+ ```
470
+
471
+ The executor maps model IDs to OpenCode's `provider/model` format automatically.
472
+ For example, `MiniMax-M2.7` becomes `opencode run --model minimax/MiniMax-M2.7`.
473
+
474
+ **Provider prefix overrides:**
475
+
476
+ By default, each provider family maps to its natural OpenCode provider ID
477
+ (`openai`, `google`, `deepseek`, `minimax`). Override with per-family env vars
478
+ when you want to route through a different OpenCode provider:
479
+
480
+ - `CONSULT_LLM_OPENCODE_OPENAI_PROVIDER` — default: `openai`
481
+ - `CONSULT_LLM_OPENCODE_GEMINI_PROVIDER` — default: `google`
482
+ - `CONSULT_LLM_OPENCODE_DEEPSEEK_PROVIDER` — default: `deepseek`
483
+ - `CONSULT_LLM_OPENCODE_MINIMAX_PROVIDER` — default: `minimax`
484
+ - `CONSULT_LLM_OPENCODE_PROVIDER` — global fallback for all families
485
+
486
+ For example, `CONSULT_LLM_OPENCODE_OPENAI_PROVIDER=copilot` turns
487
+ `gpt-5.2` into `opencode run --model copilot/gpt-5.2`.
488
+
432
489
  #### Multi-turn conversations
433
490
 
434
491
  CLI backends support multi-turn conversations via the `thread_id` parameter. The
@@ -492,14 +549,20 @@ See the "Using web mode..." example above for a concrete transcript.
492
549
  - `GEMINI_API_KEY` - Your Google AI API key (required for Gemini models in API
493
550
  mode)
494
551
  - `DEEPSEEK_API_KEY` - Your DeepSeek API key (required for DeepSeek models)
552
+ - `MINIMAX_API_KEY` - Your MiniMax API key (required for MiniMax models)
495
553
  - `CONSULT_LLM_DEFAULT_MODEL` - Override the default model (optional)
496
- - Accepts selectors (`gemini`, `openai`, `deepseek`) or exact model IDs
554
+ - Accepts selectors (`gemini`, `openai`, `deepseek`, `minimax`) or exact model
555
+ IDs
497
556
  (`gpt-5.4`, `gemini-3.1-pro-preview`, etc.)
498
557
  - Selectors are resolved to the best available model at startup
499
558
  - `CONSULT_LLM_GEMINI_BACKEND` - Backend for Gemini models (optional)
500
- - Options: `api` (default), `gemini-cli`, `cursor-cli`
559
+ - Options: `api` (default), `gemini-cli`, `cursor-cli`, `opencode`
501
560
  - `CONSULT_LLM_OPENAI_BACKEND` - Backend for OpenAI models (optional)
502
- - Options: `api` (default), `codex-cli`, `cursor-cli`
561
+ - Options: `api` (default), `codex-cli`, `cursor-cli`, `opencode`
562
+ - `CONSULT_LLM_DEEPSEEK_BACKEND` - Backend for DeepSeek models (optional)
563
+ - Options: `api` (default), `opencode`
564
+ - `CONSULT_LLM_MINIMAX_BACKEND` - Backend for MiniMax models (optional)
565
+ - Options: `api` (default), `opencode`
503
566
  - `CONSULT_LLM_ALLOWED_MODELS` - Restrict which concrete models can be used
504
567
  (optional)
505
568
  - Comma-separated list, e.g., `gpt-5.4,gemini-3.1-pro-preview`
@@ -511,10 +574,14 @@ See the "Using web mode..." example above for a concrete transcript.
511
574
  - Comma-separated list, e.g., `grok-3,kimi-k2.5`
512
575
  - Merged with built-in models and included in the tool schema
513
576
  - Useful for newly released models with a known provider prefix (`gpt-`,
514
- `gemini-`, `deepseek-`)
577
+ `gemini-`, `deepseek-`, `MiniMax-`)
515
578
  - `CONSULT_LLM_CODEX_REASONING_EFFORT` - Configure reasoning effort for Codex
516
579
  CLI (optional, default: `high`)
517
580
  - See [Codex CLI](#codex-cli) for details and available options
581
+ - `CONSULT_LLM_OPENCODE_PROVIDER` - Global OpenCode provider prefix (optional)
582
+ - Overrides the default provider ID for all families when using the `opencode`
583
+ backend
584
+ - See [OpenCode CLI](#opencode-cli) for details and per-family overrides
518
585
  - `CONSULT_LLM_SYSTEM_PROMPT_PATH` - Custom path to system prompt file
519
586
  (optional)
520
587
  - Overrides the default `~/.consult-llm-mcp/SYSTEM_PROMPT.md` location
@@ -645,6 +712,7 @@ models complex questions.
645
712
  - **gemini-3-pro-preview**: Google's Gemini 3 Pro Preview
646
713
  - **gemini-3.1-pro-preview**: Google's Gemini 3.1 Pro Preview
647
714
  - **deepseek-reasoner**: DeepSeek's reasoning model
715
+ - **MiniMax-M2.7**: MiniMax's M2.7 reasoning model (204K context)
648
716
  - **gpt-5.4**: OpenAI's GPT-5.4 model
649
717
  - **gpt-5.2**: OpenAI's GPT-5.2 model
650
718
  - **gpt-5.3-codex**: OpenAI's Codex model based on GPT-5.3
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "consult-llm-mcp",
3
- "version": "2.11.0",
3
+ "version": "2.12.0",
4
4
  "description": "MCP server for consulting powerful AI models",
5
5
  "repository": {
6
6
  "type": "git",
@@ -31,9 +31,9 @@
31
31
  "ai"
32
32
  ],
33
33
  "optionalDependencies": {
34
- "consult-llm-mcp-darwin-arm64": "2.11.0",
35
- "consult-llm-mcp-darwin-x64": "2.11.0",
36
- "consult-llm-mcp-linux-x64": "2.11.0",
37
- "consult-llm-mcp-linux-arm64": "2.11.0"
34
+ "consult-llm-mcp-darwin-arm64": "2.12.0",
35
+ "consult-llm-mcp-darwin-x64": "2.12.0",
36
+ "consult-llm-mcp-linux-x64": "2.12.0",
37
+ "consult-llm-mcp-linux-arm64": "2.12.0"
38
38
  }
39
39
  }