consult-llm-mcp 2.7.3 → 2.8.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +5 -0
- package/README.md +65 -63
- package/package.json +5 -5
package/CHANGELOG.md
CHANGED
|
@@ -1,5 +1,10 @@
|
|
|
1
1
|
# Changelog
|
|
2
2
|
|
|
3
|
+
## v2.7.4 (2026-03-13)
|
|
4
|
+
|
|
5
|
+
- Fixed Linux prebuilt binaries failing on older distros due to glibc version
|
|
6
|
+
mismatch by switching to musl static linking
|
|
7
|
+
|
|
3
8
|
## v2.7.1 (2026-03-09)
|
|
4
9
|
|
|
5
10
|
- Monitor: show "Thinking..." spinner when thinking events are streaming
|
package/README.md
CHANGED
|
@@ -55,37 +55,40 @@ to bring in the heavy artillery. Supports multi-turn conversations.
|
|
|
55
55
|
|
|
56
56
|
```bash
|
|
57
57
|
claude mcp add consult-llm \
|
|
58
|
-
-e
|
|
59
|
-
-e
|
|
58
|
+
-e CONSULT_LLM_GEMINI_BACKEND=gemini-cli \
|
|
59
|
+
-e CONSULT_LLM_OPENAI_BACKEND=codex-cli \
|
|
60
60
|
-- npx -y consult-llm-mcp
|
|
61
61
|
```
|
|
62
62
|
|
|
63
|
-
|
|
63
|
+
This is the recommended setup. Uses [Gemini CLI](#gemini-cli) and
|
|
64
|
+
[Codex CLI](#codex-cli). No API keys required, just `gemini login` and
|
|
65
|
+
`codex login`.
|
|
66
|
+
|
|
67
|
+
**With binary** (no Node.js required, but no auto-update):
|
|
64
68
|
|
|
65
69
|
```bash
|
|
66
70
|
curl -fsSL https://raw.githubusercontent.com/raine/consult-llm-mcp/main/scripts/install.sh | bash
|
|
71
|
+
```
|
|
72
|
+
|
|
73
|
+
```bash
|
|
67
74
|
claude mcp add consult-llm \
|
|
68
|
-
-e
|
|
69
|
-
-e
|
|
75
|
+
-e CONSULT_LLM_GEMINI_BACKEND=gemini-cli \
|
|
76
|
+
-e CONSULT_LLM_OPENAI_BACKEND=codex-cli \
|
|
70
77
|
-- consult-llm-mcp
|
|
71
78
|
```
|
|
72
79
|
|
|
73
80
|
For global availability across projects, add `--scope user`.
|
|
74
81
|
|
|
75
|
-
|
|
76
|
-
<summary>Using multiple API keys or CLI backends</summary>
|
|
82
|
+
**Using API keys instead of CLI backends:**
|
|
77
83
|
|
|
78
84
|
```bash
|
|
79
85
|
claude mcp add consult-llm \
|
|
80
86
|
-e OPENAI_API_KEY=your_openai_key \
|
|
81
87
|
-e GEMINI_API_KEY=your_gemini_key \
|
|
82
88
|
-e DEEPSEEK_API_KEY=your_deepseek_key \
|
|
83
|
-
-e CONSULT_LLM_GEMINI_BACKEND=gemini-cli \
|
|
84
89
|
-- npx -y consult-llm-mcp
|
|
85
90
|
```
|
|
86
91
|
|
|
87
|
-
</details>
|
|
88
|
-
|
|
89
92
|
2. **Verify connection** with `/mcp`:
|
|
90
93
|
|
|
91
94
|
```
|
|
@@ -380,8 +383,9 @@ claude mcp add consult-llm -e CONSULT_LLM_OPENAI_BACKEND=codex-cli -- npx -y con
|
|
|
380
383
|
|
|
381
384
|
<!-- prettier-ignore -->
|
|
382
385
|
> [!TIP]
|
|
383
|
-
>
|
|
384
|
-
>
|
|
386
|
+
> Reasoning effort defaults to `high`. Override with
|
|
387
|
+
> `-e CONSULT_LLM_CODEX_REASONING_EFFORT=xhigh`. Options: `none`, `minimal`,
|
|
388
|
+
> `low`, `medium`, `high`, `xhigh`.
|
|
385
389
|
|
|
386
390
|
#### Cursor CLI
|
|
387
391
|
|
|
@@ -405,8 +409,7 @@ claude mcp add consult-llm -e CONSULT_LLM_GEMINI_BACKEND=cursor-cli -- npx -y co
|
|
|
405
409
|
claude mcp add consult-llm \
|
|
406
410
|
-e CONSULT_LLM_OPENAI_BACKEND=cursor-cli \
|
|
407
411
|
-e CONSULT_LLM_GEMINI_BACKEND=cursor-cli \
|
|
408
|
-
-e
|
|
409
|
-
-e CONSULT_LLM_ALLOWED_MODELS="gemini-3-pro-preview,gpt-5.3-codex" \
|
|
412
|
+
-e CONSULT_LLM_ALLOWED_MODELS="gemini-3.1-pro-preview,gpt-5.3-codex" \
|
|
410
413
|
-- npx -y consult-llm-mcp
|
|
411
414
|
```
|
|
412
415
|
|
|
@@ -419,11 +422,7 @@ review), allow them in `~/.cursor/cli-config.json`:
|
|
|
419
422
|
```json
|
|
420
423
|
{
|
|
421
424
|
"permissions": {
|
|
422
|
-
"allow": [
|
|
423
|
-
"Shell(git diff*)",
|
|
424
|
-
"Shell(git log*)",
|
|
425
|
-
"Shell(git show*)"
|
|
426
|
-
],
|
|
425
|
+
"allow": ["Shell(git diff*)", "Shell(git log*)", "Shell(git show*)"],
|
|
427
426
|
"deny": []
|
|
428
427
|
}
|
|
429
428
|
}
|
|
@@ -495,27 +494,28 @@ See the "Using web mode..." example above for a concrete transcript.
|
|
|
495
494
|
mode)
|
|
496
495
|
- `DEEPSEEK_API_KEY` - Your DeepSeek API key (required for DeepSeek models)
|
|
497
496
|
- `CONSULT_LLM_DEFAULT_MODEL` - Override the default model (optional)
|
|
498
|
-
-
|
|
499
|
-
`
|
|
500
|
-
|
|
497
|
+
- Accepts selectors (`gemini`, `openai`, `deepseek`) or exact model IDs
|
|
498
|
+
(`gpt-5.4`, `gemini-3.1-pro-preview`, etc.)
|
|
499
|
+
- Selectors are resolved to the best available model at startup
|
|
501
500
|
- `CONSULT_LLM_GEMINI_BACKEND` - Backend for Gemini models (optional)
|
|
502
501
|
- Options: `api` (default), `gemini-cli`, `cursor-cli`
|
|
503
502
|
- `CONSULT_LLM_OPENAI_BACKEND` - Backend for OpenAI models (optional)
|
|
504
503
|
- Options: `api` (default), `codex-cli`, `cursor-cli`
|
|
505
|
-
- `
|
|
506
|
-
|
|
507
|
-
-
|
|
504
|
+
- `CONSULT_LLM_ALLOWED_MODELS` - Restrict which concrete models can be used
|
|
505
|
+
(optional)
|
|
506
|
+
- Comma-separated list, e.g., `gpt-5.4,gemini-3.1-pro-preview`
|
|
507
|
+
- Selectors resolve against this list — e.g., if only `gemini-2.5-pro` is
|
|
508
|
+
allowed, the `gemini` selector resolves to it
|
|
509
|
+
- Useful when a backend doesn't support all models (e.g., Cursor CLI)
|
|
510
|
+
- See [Tips](#controlling-which-models-are-used) for usage examples
|
|
508
511
|
- `CONSULT_LLM_EXTRA_MODELS` - Add models not in the built-in list (optional)
|
|
509
512
|
- Comma-separated list, e.g., `grok-3,kimi-k2.5`
|
|
510
513
|
- Merged with built-in models and included in the tool schema
|
|
511
514
|
- Useful for newly released models with a known provider prefix (`gpt-`,
|
|
512
515
|
`gemini-`, `deepseek-`)
|
|
513
|
-
- `
|
|
514
|
-
|
|
515
|
-
-
|
|
516
|
-
- Filters the combined catalog (built-in + extra models)
|
|
517
|
-
- If `CONSULT_LLM_DEFAULT_MODEL` is set, it must be in this list
|
|
518
|
-
- See [Tips](#controlling-which-models-claude-uses) for usage examples
|
|
516
|
+
- `CONSULT_LLM_CODEX_REASONING_EFFORT` - Configure reasoning effort for Codex
|
|
517
|
+
CLI (optional, default: `high`)
|
|
518
|
+
- See [Codex CLI](#codex-cli) for details and available options
|
|
519
519
|
- `CONSULT_LLM_SYSTEM_PROMPT_PATH` - Custom path to system prompt file
|
|
520
520
|
(optional)
|
|
521
521
|
- Overrides the default `~/.consult-llm-mcp/SYSTEM_PROMPT.md` location
|
|
@@ -558,30 +558,33 @@ claude mcp add consult-llm \
|
|
|
558
558
|
|
|
559
559
|
## Tips
|
|
560
560
|
|
|
561
|
-
### Controlling which models
|
|
561
|
+
### Controlling which models are used
|
|
562
562
|
|
|
563
|
-
|
|
564
|
-
|
|
565
|
-
|
|
563
|
+
The `model` parameter accepts **selectors** (`gemini`, `openai`, `deepseek`)
|
|
564
|
+
that the server resolves to the best available concrete model. When no model is
|
|
565
|
+
specified, the server uses `CONSULT_LLM_DEFAULT_MODEL` or its built-in fallback.
|
|
566
566
|
|
|
567
|
-
|
|
568
|
-
`CONSULT_LLM_ALLOWED_MODELS`:
|
|
567
|
+
**Selector resolution order** (first available wins):
|
|
569
568
|
|
|
570
|
-
|
|
571
|
-
|
|
572
|
-
|
|
573
|
-
|
|
574
|
-
|
|
575
|
-
|
|
569
|
+
| Selector | Priority |
|
|
570
|
+
| ---------- | -------------------------------------------------------------- |
|
|
571
|
+
| `gemini` | gemini-3.1-pro-preview → gemini-3-pro-preview → gemini-2.5-pro |
|
|
572
|
+
| `openai` | gpt-5.4 → gpt-5.3-codex → gpt-5.2 → gpt-5.2-codex |
|
|
573
|
+
| `deepseek` | deepseek-reasoner |
|
|
574
|
+
|
|
575
|
+
**Restricting models with `CONSULT_LLM_ALLOWED_MODELS`:**
|
|
576
576
|
|
|
577
|
-
|
|
578
|
-
|
|
577
|
+
If your backend doesn't support all models (e.g., Cursor CLI can't use
|
|
578
|
+
`gpt-5.4`), use `CONSULT_LLM_ALLOWED_MODELS` to filter. Selectors will
|
|
579
|
+
automatically resolve to the best model within the allowed list:
|
|
579
580
|
|
|
580
581
|
```bash
|
|
582
|
+
# Only allow codex models through Cursor CLI
|
|
581
583
|
claude mcp add consult-llm \
|
|
582
|
-
-e
|
|
583
|
-
-e CONSULT_LLM_ALLOWED_MODELS='gemini-3-pro-preview' \
|
|
584
|
+
-e CONSULT_LLM_OPENAI_BACKEND=cursor-cli \
|
|
585
|
+
-e CONSULT_LLM_ALLOWED_MODELS='gpt-5.3-codex,gemini-3.1-pro-preview' \
|
|
584
586
|
-- npx -y consult-llm-mcp
|
|
587
|
+
# "openai" selector → gpt-5.3-codex (gpt-5.4 filtered out)
|
|
585
588
|
```
|
|
586
589
|
|
|
587
590
|
## MCP tool: consult_llm
|
|
@@ -596,10 +599,12 @@ models complex questions.
|
|
|
596
599
|
- **files** (optional): Array of file paths to include as context
|
|
597
600
|
- All files are added as context with file paths and code blocks
|
|
598
601
|
|
|
599
|
-
- **model** (optional):
|
|
600
|
-
-
|
|
601
|
-
|
|
602
|
-
|
|
602
|
+
- **model** (optional): Model selector or exact model ID
|
|
603
|
+
- Selectors: `gemini`, `openai`, `deepseek` — the server resolves to the best
|
|
604
|
+
available model for each family
|
|
605
|
+
- Exact model IDs (`gpt-5.4`, `gemini-3.1-pro-preview`, etc.) are also
|
|
606
|
+
accepted as an advanced override
|
|
607
|
+
- When omitted, the server uses the configured default
|
|
603
608
|
|
|
604
609
|
- **task_mode** (optional): Controls the system prompt persona. The calling LLM
|
|
605
610
|
should choose based on the task:
|
|
@@ -631,15 +636,12 @@ models complex questions.
|
|
|
631
636
|
|
|
632
637
|
## Supported models
|
|
633
638
|
|
|
634
|
-
- **gemini-2.5-pro**: Google's Gemini 2.5 Pro
|
|
635
|
-
- **gemini-3-pro-preview**: Google's Gemini 3 Pro Preview
|
|
636
|
-
|
|
637
|
-
- **
|
|
638
|
-
|
|
639
|
-
- **
|
|
640
|
-
tokens)
|
|
641
|
-
- **gpt-5.4**: OpenAI's GPT-5.4 model ($2.50/$15 per million tokens)
|
|
642
|
-
- **gpt-5.2**: OpenAI's GPT-5.2 model ($1.75/$14 per million tokens)
|
|
639
|
+
- **gemini-2.5-pro**: Google's Gemini 2.5 Pro
|
|
640
|
+
- **gemini-3-pro-preview**: Google's Gemini 3 Pro Preview
|
|
641
|
+
- **gemini-3.1-pro-preview**: Google's Gemini 3.1 Pro Preview
|
|
642
|
+
- **deepseek-reasoner**: DeepSeek's reasoning model
|
|
643
|
+
- **gpt-5.4**: OpenAI's GPT-5.4 model
|
|
644
|
+
- **gpt-5.2**: OpenAI's GPT-5.2 model
|
|
643
645
|
- **gpt-5.3-codex**: OpenAI's Codex model based on GPT-5.3
|
|
644
646
|
- **gpt-5.2-codex**: OpenAI's Codex model based on GPT-5.2
|
|
645
647
|
|
|
@@ -739,7 +741,9 @@ always reliably triggered. See the [consult skill](#consult) below.
|
|
|
739
741
|
**Recommendation:** Start with no custom activation. Use skills if you need
|
|
740
742
|
custom instructions for how the MCP is invoked.
|
|
741
743
|
|
|
742
|
-
##
|
|
744
|
+
## Skills
|
|
745
|
+
|
|
746
|
+
### Installing skills
|
|
743
747
|
|
|
744
748
|
Install all skills globally with a single command:
|
|
745
749
|
|
|
@@ -759,8 +763,6 @@ To uninstall:
|
|
|
759
763
|
curl -fsSL https://raw.githubusercontent.com/raine/consult-llm-mcp/main/scripts/install-skills | bash -s uninstall
|
|
760
764
|
```
|
|
761
765
|
|
|
762
|
-
## Skills
|
|
763
|
-
|
|
764
766
|
### consult
|
|
765
767
|
|
|
766
768
|
An example [Claude Code skill](https://code.claude.com/docs/en/skills) that uses
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "consult-llm-mcp",
|
|
3
|
-
"version": "2.
|
|
3
|
+
"version": "2.8.0",
|
|
4
4
|
"description": "MCP server for consulting powerful AI models",
|
|
5
5
|
"repository": {
|
|
6
6
|
"type": "git",
|
|
@@ -31,9 +31,9 @@
|
|
|
31
31
|
"ai"
|
|
32
32
|
],
|
|
33
33
|
"optionalDependencies": {
|
|
34
|
-
"consult-llm-mcp-darwin-arm64": "2.
|
|
35
|
-
"consult-llm-mcp-darwin-x64": "2.
|
|
36
|
-
"consult-llm-mcp-linux-x64": "2.
|
|
37
|
-
"consult-llm-mcp-linux-arm64": "2.
|
|
34
|
+
"consult-llm-mcp-darwin-arm64": "2.8.0",
|
|
35
|
+
"consult-llm-mcp-darwin-x64": "2.8.0",
|
|
36
|
+
"consult-llm-mcp-linux-x64": "2.8.0",
|
|
37
|
+
"consult-llm-mcp-linux-arm64": "2.8.0"
|
|
38
38
|
}
|
|
39
39
|
}
|