consult-llm-mcp 2.7.3 → 2.8.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (3) hide show
  1. package/CHANGELOG.md +5 -0
  2. package/README.md +65 -63
  3. package/package.json +5 -5
package/CHANGELOG.md CHANGED
@@ -1,5 +1,10 @@
1
1
  # Changelog
2
2
 
3
+ ## v2.7.4 (2026-03-13)
4
+
5
+ - Fixed Linux prebuilt binaries failing on older distros due to glibc version
6
+ mismatch by switching to musl static linking
7
+
3
8
  ## v2.7.1 (2026-03-09)
4
9
 
5
10
  - Monitor: show "Thinking..." spinner when thinking events are streaming
package/README.md CHANGED
@@ -55,37 +55,40 @@ to bring in the heavy artillery. Supports multi-turn conversations.
55
55
 
56
56
  ```bash
57
57
  claude mcp add consult-llm \
58
- -e OPENAI_API_KEY=your_key \
59
- -e GEMINI_API_KEY=your_key \
58
+ -e CONSULT_LLM_GEMINI_BACKEND=gemini-cli \
59
+ -e CONSULT_LLM_OPENAI_BACKEND=codex-cli \
60
60
  -- npx -y consult-llm-mcp
61
61
  ```
62
62
 
63
- **With binary** (no Node.js required):
63
+ This is the recommended setup. Uses [Gemini CLI](#gemini-cli) and
64
+ [Codex CLI](#codex-cli). No API keys required, just `gemini login` and
65
+ `codex login`.
66
+
67
+ **With binary** (no Node.js required, but no auto-update):
64
68
 
65
69
  ```bash
66
70
  curl -fsSL https://raw.githubusercontent.com/raine/consult-llm-mcp/main/scripts/install.sh | bash
71
+ ```
72
+
73
+ ```bash
67
74
  claude mcp add consult-llm \
68
- -e OPENAI_API_KEY=your_key \
69
- -e GEMINI_API_KEY=your_key \
75
+ -e CONSULT_LLM_GEMINI_BACKEND=gemini-cli \
76
+ -e CONSULT_LLM_OPENAI_BACKEND=codex-cli \
70
77
  -- consult-llm-mcp
71
78
  ```
72
79
 
73
80
  For global availability across projects, add `--scope user`.
74
81
 
75
- <details>
76
- <summary>Using multiple API keys or CLI backends</summary>
82
+ **Using API keys instead of CLI backends:**
77
83
 
78
84
  ```bash
79
85
  claude mcp add consult-llm \
80
86
  -e OPENAI_API_KEY=your_openai_key \
81
87
  -e GEMINI_API_KEY=your_gemini_key \
82
88
  -e DEEPSEEK_API_KEY=your_deepseek_key \
83
- -e CONSULT_LLM_GEMINI_BACKEND=gemini-cli \
84
89
  -- npx -y consult-llm-mcp
85
90
  ```
86
91
 
87
- </details>
88
-
89
92
  2. **Verify connection** with `/mcp`:
90
93
 
91
94
  ```
@@ -380,8 +383,9 @@ claude mcp add consult-llm -e CONSULT_LLM_OPENAI_BACKEND=codex-cli -- npx -y con
380
383
 
381
384
  <!-- prettier-ignore -->
382
385
  > [!TIP]
383
- > Set reasoning effort with `-e CONSULT_LLM_CODEX_REASONING_EFFORT=high`. Options:
384
- > `none`, `minimal`, `low`, `medium`, `high`, `xhigh`.
386
+ > Reasoning effort defaults to `high`. Override with
387
+ > `-e CONSULT_LLM_CODEX_REASONING_EFFORT=xhigh`. Options: `none`, `minimal`,
388
+ > `low`, `medium`, `high`, `xhigh`.
385
389
 
386
390
  #### Cursor CLI
387
391
 
@@ -405,8 +409,7 @@ claude mcp add consult-llm -e CONSULT_LLM_GEMINI_BACKEND=cursor-cli -- npx -y co
405
409
  claude mcp add consult-llm \
406
410
  -e CONSULT_LLM_OPENAI_BACKEND=cursor-cli \
407
411
  -e CONSULT_LLM_GEMINI_BACKEND=cursor-cli \
408
- -e CONSULT_LLM_CODEX_REASONING_EFFORT=high \
409
- -e CONSULT_LLM_ALLOWED_MODELS="gemini-3-pro-preview,gpt-5.3-codex" \
412
+ -e CONSULT_LLM_ALLOWED_MODELS="gemini-3.1-pro-preview,gpt-5.3-codex" \
410
413
  -- npx -y consult-llm-mcp
411
414
  ```
412
415
 
@@ -419,11 +422,7 @@ review), allow them in `~/.cursor/cli-config.json`:
419
422
  ```json
420
423
  {
421
424
  "permissions": {
422
- "allow": [
423
- "Shell(git diff*)",
424
- "Shell(git log*)",
425
- "Shell(git show*)"
426
- ],
425
+ "allow": ["Shell(git diff*)", "Shell(git log*)", "Shell(git show*)"],
427
426
  "deny": []
428
427
  }
429
428
  }
@@ -495,27 +494,28 @@ See the "Using web mode..." example above for a concrete transcript.
495
494
  mode)
496
495
  - `DEEPSEEK_API_KEY` - Your DeepSeek API key (required for DeepSeek models)
497
496
  - `CONSULT_LLM_DEFAULT_MODEL` - Override the default model (optional)
498
- - Options: `gpt-5.2` (default), `gpt-5.4`, `gemini-2.5-pro`,
499
- `gemini-3-pro-preview`, `gemini-3.1-pro-preview`, `deepseek-reasoner`,
500
- `gpt-5.3-codex`, `gpt-5.2-codex`
497
+ - Accepts selectors (`gemini`, `openai`, `deepseek`) or exact model IDs
498
+ (`gpt-5.4`, `gemini-3.1-pro-preview`, etc.)
499
+ - Selectors are resolved to the best available model at startup
501
500
  - `CONSULT_LLM_GEMINI_BACKEND` - Backend for Gemini models (optional)
502
501
  - Options: `api` (default), `gemini-cli`, `cursor-cli`
503
502
  - `CONSULT_LLM_OPENAI_BACKEND` - Backend for OpenAI models (optional)
504
503
  - Options: `api` (default), `codex-cli`, `cursor-cli`
505
- - `CONSULT_LLM_CODEX_REASONING_EFFORT` - Configure reasoning effort for Codex
506
- CLI (optional)
507
- - See [Codex CLI](#codex-cli) for details and available options
504
+ - `CONSULT_LLM_ALLOWED_MODELS` - Restrict which concrete models can be used
505
+ (optional)
506
+ - Comma-separated list, e.g., `gpt-5.4,gemini-3.1-pro-preview`
507
+ - Selectors resolve against this list — e.g., if only `gemini-2.5-pro` is
508
+ allowed, the `gemini` selector resolves to it
509
+ - Useful when a backend doesn't support all models (e.g., Cursor CLI)
510
+ - See [Tips](#controlling-which-models-are-used) for usage examples
508
511
  - `CONSULT_LLM_EXTRA_MODELS` - Add models not in the built-in list (optional)
509
512
  - Comma-separated list, e.g., `grok-3,kimi-k2.5`
510
513
  - Merged with built-in models and included in the tool schema
511
514
  - Useful for newly released models with a known provider prefix (`gpt-`,
512
515
  `gemini-`, `deepseek-`)
513
- - `CONSULT_LLM_ALLOWED_MODELS` - List of models to advertise (optional)
514
- - Comma-separated list, e.g., `gpt-5.2,gemini-3-pro-preview`
515
- - When set, only these models appear in the tool schema
516
- - Filters the combined catalog (built-in + extra models)
517
- - If `CONSULT_LLM_DEFAULT_MODEL` is set, it must be in this list
518
- - See [Tips](#controlling-which-models-claude-uses) for usage examples
516
+ - `CONSULT_LLM_CODEX_REASONING_EFFORT` - Configure reasoning effort for Codex
517
+ CLI (optional, default: `high`)
518
+ - See [Codex CLI](#codex-cli) for details and available options
519
519
  - `CONSULT_LLM_SYSTEM_PROMPT_PATH` - Custom path to system prompt file
520
520
  (optional)
521
521
  - Overrides the default `~/.consult-llm-mcp/SYSTEM_PROMPT.md` location
@@ -558,30 +558,33 @@ claude mcp add consult-llm \
558
558
 
559
559
  ## Tips
560
560
 
561
- ### Controlling which models Claude uses
561
+ ### Controlling which models are used
562
562
 
563
- When you ask Claude to "consult an LLM" without specifying a model, it picks one
564
- from the available options in the tool schema. The `CONSULT_LLM_DEFAULT_MODEL`
565
- only affects the fallback when no model is specified in the tool call.
563
+ The `model` parameter accepts **selectors** (`gemini`, `openai`, `deepseek`)
564
+ that the server resolves to the best available concrete model. When no model is
565
+ specified, the server uses `CONSULT_LLM_DEFAULT_MODEL` or its built-in fallback.
566
566
 
567
- To control which models Claude can choose from, use
568
- `CONSULT_LLM_ALLOWED_MODELS`:
567
+ **Selector resolution order** (first available wins):
569
568
 
570
- ```bash
571
- claude mcp add consult-llm \
572
- -e GEMINI_API_KEY=your_key \
573
- -e CONSULT_LLM_ALLOWED_MODELS='gemini-3-pro-preview,gpt-5.2-codex' \
574
- -- npx -y consult-llm-mcp
575
- ```
569
+ | Selector | Priority |
570
+ | ---------- | -------------------------------------------------------------- |
571
+ | `gemini` | gemini-3.1-pro-preview gemini-3-pro-preview → gemini-2.5-pro |
572
+ | `openai` | gpt-5.4 → gpt-5.3-codex → gpt-5.2 → gpt-5.2-codex |
573
+ | `deepseek` | deepseek-reasoner |
574
+
575
+ **Restricting models with `CONSULT_LLM_ALLOWED_MODELS`:**
576
576
 
577
- This restricts the tool schema to only advertise these models. For example, to
578
- ensure Claude always uses Gemini 3 Pro:
577
+ If your backend doesn't support all models (e.g., Cursor CLI can't use
578
+ `gpt-5.4`), use `CONSULT_LLM_ALLOWED_MODELS` to filter. Selectors will
579
+ automatically resolve to the best model within the allowed list:
579
580
 
580
581
  ```bash
582
+ # Only allow codex models through Cursor CLI
581
583
  claude mcp add consult-llm \
582
- -e GEMINI_API_KEY=your_key \
583
- -e CONSULT_LLM_ALLOWED_MODELS='gemini-3-pro-preview' \
584
+ -e CONSULT_LLM_OPENAI_BACKEND=cursor-cli \
585
+ -e CONSULT_LLM_ALLOWED_MODELS='gpt-5.3-codex,gemini-3.1-pro-preview' \
584
586
  -- npx -y consult-llm-mcp
587
+ # "openai" selector → gpt-5.3-codex (gpt-5.4 filtered out)
585
588
  ```
586
589
 
587
590
  ## MCP tool: consult_llm
@@ -596,10 +599,12 @@ models complex questions.
596
599
  - **files** (optional): Array of file paths to include as context
597
600
  - All files are added as context with file paths and code blocks
598
601
 
599
- - **model** (optional): LLM model to use
600
- - Options: `gpt-5.2` (default), `gpt-5.4`, `gemini-2.5-pro`,
601
- `gemini-3-pro-preview`, `gemini-3.1-pro-preview`, `deepseek-reasoner`,
602
- `gpt-5.3-codex`, `gpt-5.2-codex`
602
+ - **model** (optional): Model selector or exact model ID
603
+ - Selectors: `gemini`, `openai`, `deepseek` — the server resolves to the best
604
+ available model for each family
605
+ - Exact model IDs (`gpt-5.4`, `gemini-3.1-pro-preview`, etc.) are also
606
+ accepted as an advanced override
607
+ - When omitted, the server uses the configured default
603
608
 
604
609
  - **task_mode** (optional): Controls the system prompt persona. The calling LLM
605
610
  should choose based on the task:
@@ -631,15 +636,12 @@ models complex questions.
631
636
 
632
637
  ## Supported models
633
638
 
634
- - **gemini-2.5-pro**: Google's Gemini 2.5 Pro ($1.25/$10 per million tokens)
635
- - **gemini-3-pro-preview**: Google's Gemini 3 Pro Preview ($2/$12 per million
636
- tokens for prompts ≤200k tokens, $4/$18 for prompts >200k tokens)
637
- - **gemini-3.1-pro-preview**: Google's Gemini 3.1 Pro Preview ($2/$12 per
638
- million tokens for prompts ≤200k tokens, $4/$18 for prompts >200k tokens)
639
- - **deepseek-reasoner**: DeepSeek's reasoning model ($0.55/$2.19 per million
640
- tokens)
641
- - **gpt-5.4**: OpenAI's GPT-5.4 model ($2.50/$15 per million tokens)
642
- - **gpt-5.2**: OpenAI's GPT-5.2 model ($1.75/$14 per million tokens)
639
+ - **gemini-2.5-pro**: Google's Gemini 2.5 Pro
640
+ - **gemini-3-pro-preview**: Google's Gemini 3 Pro Preview
641
+ - **gemini-3.1-pro-preview**: Google's Gemini 3.1 Pro Preview
642
+ - **deepseek-reasoner**: DeepSeek's reasoning model
643
+ - **gpt-5.4**: OpenAI's GPT-5.4 model
644
+ - **gpt-5.2**: OpenAI's GPT-5.2 model
643
645
  - **gpt-5.3-codex**: OpenAI's Codex model based on GPT-5.3
644
646
  - **gpt-5.2-codex**: OpenAI's Codex model based on GPT-5.2
645
647
 
@@ -739,7 +741,9 @@ always reliably triggered. See the [consult skill](#consult) below.
739
741
  **Recommendation:** Start with no custom activation. Use skills if you need
740
742
  custom instructions for how the MCP is invoked.
741
743
 
742
- ## Installing skills
744
+ ## Skills
745
+
746
+ ### Installing skills
743
747
 
744
748
  Install all skills globally with a single command:
745
749
 
@@ -759,8 +763,6 @@ To uninstall:
759
763
  curl -fsSL https://raw.githubusercontent.com/raine/consult-llm-mcp/main/scripts/install-skills | bash -s uninstall
760
764
  ```
761
765
 
762
- ## Skills
763
-
764
766
  ### consult
765
767
 
766
768
  An example [Claude Code skill](https://code.claude.com/docs/en/skills) that uses
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "consult-llm-mcp",
3
- "version": "2.7.3",
3
+ "version": "2.8.0",
4
4
  "description": "MCP server for consulting powerful AI models",
5
5
  "repository": {
6
6
  "type": "git",
@@ -31,9 +31,9 @@
31
31
  "ai"
32
32
  ],
33
33
  "optionalDependencies": {
34
- "consult-llm-mcp-darwin-arm64": "2.7.3",
35
- "consult-llm-mcp-darwin-x64": "2.7.3",
36
- "consult-llm-mcp-linux-x64": "2.7.3",
37
- "consult-llm-mcp-linux-arm64": "2.7.3"
34
+ "consult-llm-mcp-darwin-arm64": "2.8.0",
35
+ "consult-llm-mcp-darwin-x64": "2.8.0",
36
+ "consult-llm-mcp-linux-x64": "2.8.0",
37
+ "consult-llm-mcp-linux-arm64": "2.8.0"
38
38
  }
39
39
  }