claudish 3.4.0 → 3.5.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/AI_AGENT_GUIDE.md CHANGED
@@ -1,8 +1,8 @@
1
1
  # Claudish AI Agent Usage Guide
2
2
 
3
- **Version:** 2.0.0
3
+ **Version:** 2.2.0
4
4
  **Target Audience:** AI Agents running within Claude Code
5
- **Purpose:** Quick reference for using Claudish CLI in agentic workflows
5
+ **Purpose:** Quick reference for using Claudish CLI and MCP server in agentic workflows
6
6
 
7
7
  ---
8
8
 
@@ -32,6 +32,7 @@ Claudish = Claude Code + Any AI Model
32
32
  - ✅ Run Claude Code with **any AI model** via prefix-based routing
33
33
  - ✅ Supports OpenRouter (100+ models), direct Gemini API, direct OpenAI API
34
34
  - ✅ Supports local models (Ollama, LM Studio, vLLM, MLX)
35
+ - ✅ **MCP Server mode** - expose models as tools for Claude Code
35
36
  - ✅ 100% Claude Code feature compatibility
36
37
  - ✅ Local proxy server (no data sent to Claudish servers)
37
38
  - ✅ Cost tracking and model selection
@@ -42,11 +43,28 @@ Claudish = Claude Code + Any AI Model
42
43
  |--------|---------|---------|
43
44
  | _(none)_ | OpenRouter | `openai/gpt-5.2` |
44
45
  | `g/` `gemini/` | Google Gemini | `g/gemini-2.0-flash` |
46
+ | `v/` `vertex/` | Vertex AI | `v/gemini-2.5-flash` |
45
47
  | `oai/` `openai/` | OpenAI | `oai/gpt-4o` |
46
48
  | `ollama/` | Ollama | `ollama/llama3.2` |
47
49
  | `lmstudio/` | LM Studio | `lmstudio/model` |
48
50
  | `http://...` | Custom | `http://localhost:8000/model` |
49
51
 
52
+ ### Vertex AI Partner Models
53
+
54
+ Vertex AI supports Google + partner models (MaaS):
55
+
56
+ ```bash
57
+ # Google Gemini on Vertex
58
+ claudish --model v/gemini-2.5-flash "task"
59
+
60
+ # Partner models (MiniMax, Mistral, DeepSeek, Qwen, OpenAI OSS)
61
+ claudish --model vertex/minimax/minimax-m2-maas "task"
62
+ claudish --model vertex/mistralai/codestral-2 "write code"
63
+ claudish --model vertex/deepseek/deepseek-v3-2-maas "analyze"
64
+ claudish --model vertex/qwen/qwen3-coder-480b-a35b-instruct-maas "implement"
65
+ claudish --model vertex/openai/gpt-oss-120b-maas "reason"
66
+ ```
67
+
50
68
  ## Prerequisites
51
69
 
52
70
  1. **Install Claudish:**
@@ -54,9 +72,19 @@ Claudish = Claude Code + Any AI Model
54
72
  npm install -g claudish
55
73
  ```
56
74
 
57
- 2. **Set OpenRouter API Key:**
75
+ 2. **Set API Key (at least one):**
58
76
  ```bash
77
+ # OpenRouter (100+ models)
59
78
  export OPENROUTER_API_KEY='sk-or-v1-...'
79
+
80
+ # OR Gemini direct
81
+ export GEMINI_API_KEY='...'
82
+
83
+ # OR Vertex AI (Express mode)
84
+ export VERTEX_API_KEY='...'
85
+
86
+ # OR Vertex AI (OAuth mode - uses gcloud ADC)
87
+ export VERTEX_PROJECT='your-gcp-project-id'
60
88
  ```
61
89
 
62
90
  3. **Optional but recommended:**
@@ -81,9 +109,20 @@ Claudish = Claude Code + Any AI Model
81
109
  | Model ID | Backend | Best For |
82
110
  |----------|---------|----------|
83
111
  | `g/gemini-2.0-flash` | Gemini | Fast tasks, large context |
112
+ | `v/gemini-2.5-flash` | Vertex AI | Enterprise, GCP billing |
84
113
  | `oai/gpt-4o` | OpenAI | General purpose |
85
114
  | `ollama/llama3.2` | Local | Free, private |
86
115
 
116
+ **Vertex AI Partner Models (MaaS):**
117
+
118
+ | Model ID | Provider | Best For |
119
+ |----------|----------|----------|
120
+ | `vertex/minimax/minimax-m2-maas` | MiniMax | Fast, budget-friendly |
121
+ | `vertex/mistralai/codestral-2` | Mistral | Code specialist |
122
+ | `vertex/deepseek/deepseek-v3-2-maas` | DeepSeek | Deep reasoning |
123
+ | `vertex/qwen/qwen3-coder-480b-a35b-instruct-maas` | Qwen | Agentic coding |
124
+ | `vertex/openai/gpt-oss-120b-maas` | OpenAI | Open-weight reasoning |
125
+
87
126
  **Update models:**
88
127
  ```bash
89
128
  claudish --models --force-update
@@ -247,6 +286,120 @@ git diff | claudish --stdin --model google/gemini-2.5-flash "review for bugs"
247
286
  claudish --model qwen/qwen3-vl-235b-a22b-instruct "implement dashboard from design"
248
287
  ```
249
288
 
289
+ ## MCP Server Mode
290
+
291
+ Claudish can run as an MCP (Model Context Protocol) server, exposing OpenRouter models as tools that Claude Code can call mid-conversation. This is useful when you want to:
292
+
293
+ - Query external models without spawning a subprocess
294
+ - Compare responses from multiple models
295
+ - Use specific models for specific subtasks
296
+
297
+ ### Starting MCP Server
298
+
299
+ ```bash
300
+ # Start MCP server (stdio transport)
301
+ claudish --mcp
302
+ ```
303
+
304
+ ### Claude Code Configuration
305
+
306
+ Add to `~/.claude/settings.json`:
307
+
308
+ ```json
309
+ {
310
+ "mcpServers": {
311
+ "claudish": {
312
+ "command": "claudish",
313
+ "args": ["--mcp"],
314
+ "env": {
315
+ "OPENROUTER_API_KEY": "sk-or-v1-..."
316
+ }
317
+ }
318
+ }
319
+ }
320
+ ```
321
+
322
+ Or use npx (no installation needed):
323
+
324
+ ```json
325
+ {
326
+ "mcpServers": {
327
+ "claudish": {
328
+ "command": "npx",
329
+ "args": ["claudish@latest", "--mcp"]
330
+ }
331
+ }
332
+ }
333
+ ```
334
+
335
+ ### Available MCP Tools
336
+
337
+ | Tool | Description | Example Use |
338
+ |------|-------------|-------------|
339
+ | `run_prompt` | Execute prompt on any model | Get a second opinion from Grok |
340
+ | `list_models` | Show recommended models | Find models with tool support |
341
+ | `search_models` | Fuzzy search all models | Find vision-capable models |
342
+ | `compare_models` | Run same prompt on multiple models | Compare reasoning approaches |
343
+
344
+ ### Using MCP Tools from Claude Code
345
+
346
+ Once configured, Claude Code can use these tools directly:
347
+
348
+ ```
349
+ User: "Use Grok to review this code"
350
+ Claude: [calls run_prompt tool with model="x-ai/grok-code-fast-1"]
351
+
352
+ User: "What models support vision?"
353
+ Claude: [calls search_models tool with query="vision"]
354
+
355
+ User: "Compare how GPT-5 and Gemini explain this concept"
356
+ Claude: [calls compare_models tool with models=["openai/gpt-5.2", "google/gemini-3-pro-preview"]]
357
+ ```
358
+
359
+ ### MCP vs CLI Mode
360
+
361
+ | Feature | CLI Mode | MCP Mode |
362
+ |---------|----------|----------|
363
+ | Use case | Replace Claude Code model | Call models as tools |
364
+ | Context | Full Claude Code session | Single prompt/response |
365
+ | Streaming | Full streaming | Buffered response |
366
+ | Best for | Primary model replacement | Second opinions, comparisons |
367
+
368
+ ### MCP Tool Details
369
+
370
+ **run_prompt**
371
+ ```typescript
372
+ {
373
+ model: string, // e.g., "x-ai/grok-code-fast-1"
374
+ prompt: string, // The prompt to send
375
+ system_prompt?: string, // Optional system prompt
376
+ max_tokens?: number // Default: 4096
377
+ }
378
+ ```
379
+
380
+ **list_models**
381
+ ```typescript
382
+ // No parameters - returns curated list of recommended models
383
+ {}
384
+ ```
385
+
386
+ **search_models**
387
+ ```typescript
388
+ {
389
+ query: string, // e.g., "grok", "vision", "free"
390
+ limit?: number // Default: 10
391
+ }
392
+ ```
393
+
394
+ **compare_models**
395
+ ```typescript
396
+ {
397
+ models: string[], // e.g., ["openai/gpt-5.2", "x-ai/grok-code-fast-1"]
398
+ prompt: string, // Prompt to send to all models
399
+ system_prompt?: string // Optional system prompt
400
+ }
401
+ ```
402
+
250
403
  ## Getting Model List
251
404
 
252
405
  ### JSON Output (Recommended)
@@ -558,6 +711,6 @@ claudish --help-ai > claudish-agent-guide.md
558
711
 
559
712
  ---
560
713
 
561
- **Version:** 2.0.0
562
- **Last Updated:** January 5, 2026
714
+ **Version:** 2.2.0
715
+ **Last Updated:** January 22, 2026
563
716
  **Maintained by:** MadAppGang