claudish 3.4.0 → 3.5.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/AI_AGENT_GUIDE.md +158 -5
- package/dist/index.js +1165 -190
- package/package.json +1 -1
package/AI_AGENT_GUIDE.md
CHANGED
|
@@ -1,8 +1,8 @@
|
|
|
1
1
|
# Claudish AI Agent Usage Guide
|
|
2
2
|
|
|
3
|
-
**Version:** 2.
|
|
3
|
+
**Version:** 2.2.0
|
|
4
4
|
**Target Audience:** AI Agents running within Claude Code
|
|
5
|
-
**Purpose:** Quick reference for using Claudish CLI in agentic workflows
|
|
5
|
+
**Purpose:** Quick reference for using Claudish CLI and MCP server in agentic workflows
|
|
6
6
|
|
|
7
7
|
---
|
|
8
8
|
|
|
@@ -32,6 +32,7 @@ Claudish = Claude Code + Any AI Model
|
|
|
32
32
|
- ✅ Run Claude Code with **any AI model** via prefix-based routing
|
|
33
33
|
- ✅ Supports OpenRouter (100+ models), direct Gemini API, direct OpenAI API
|
|
34
34
|
- ✅ Supports local models (Ollama, LM Studio, vLLM, MLX)
|
|
35
|
+
- ✅ **MCP Server mode** - expose models as tools for Claude Code
|
|
35
36
|
- ✅ 100% Claude Code feature compatibility
|
|
36
37
|
- ✅ Local proxy server (no data sent to Claudish servers)
|
|
37
38
|
- ✅ Cost tracking and model selection
|
|
@@ -42,11 +43,28 @@ Claudish = Claude Code + Any AI Model
|
|
|
42
43
|
|--------|---------|---------|
|
|
43
44
|
| _(none)_ | OpenRouter | `openai/gpt-5.2` |
|
|
44
45
|
| `g/` `gemini/` | Google Gemini | `g/gemini-2.0-flash` |
|
|
46
|
+
| `v/` `vertex/` | Vertex AI | `v/gemini-2.5-flash` |
|
|
45
47
|
| `oai/` `openai/` | OpenAI | `oai/gpt-4o` |
|
|
46
48
|
| `ollama/` | Ollama | `ollama/llama3.2` |
|
|
47
49
|
| `lmstudio/` | LM Studio | `lmstudio/model` |
|
|
48
50
|
| `http://...` | Custom | `http://localhost:8000/model` |
|
|
49
51
|
|
|
52
|
+
### Vertex AI Partner Models
|
|
53
|
+
|
|
54
|
+
Vertex AI supports Google + partner models (MaaS):
|
|
55
|
+
|
|
56
|
+
```bash
|
|
57
|
+
# Google Gemini on Vertex
|
|
58
|
+
claudish --model v/gemini-2.5-flash "task"
|
|
59
|
+
|
|
60
|
+
# Partner models (MiniMax, Mistral, DeepSeek, Qwen, OpenAI OSS)
|
|
61
|
+
claudish --model vertex/minimax/minimax-m2-maas "task"
|
|
62
|
+
claudish --model vertex/mistralai/codestral-2 "write code"
|
|
63
|
+
claudish --model vertex/deepseek/deepseek-v3-2-maas "analyze"
|
|
64
|
+
claudish --model vertex/qwen/qwen3-coder-480b-a35b-instruct-maas "implement"
|
|
65
|
+
claudish --model vertex/openai/gpt-oss-120b-maas "reason"
|
|
66
|
+
```
|
|
67
|
+
|
|
50
68
|
## Prerequisites
|
|
51
69
|
|
|
52
70
|
1. **Install Claudish:**
|
|
@@ -54,9 +72,19 @@ Claudish = Claude Code + Any AI Model
|
|
|
54
72
|
npm install -g claudish
|
|
55
73
|
```
|
|
56
74
|
|
|
57
|
-
2. **Set
|
|
75
|
+
2. **Set API Key (at least one):**
|
|
58
76
|
```bash
|
|
77
|
+
# OpenRouter (100+ models)
|
|
59
78
|
export OPENROUTER_API_KEY='sk-or-v1-...'
|
|
79
|
+
|
|
80
|
+
# OR Gemini direct
|
|
81
|
+
export GEMINI_API_KEY='...'
|
|
82
|
+
|
|
83
|
+
# OR Vertex AI (Express mode)
|
|
84
|
+
export VERTEX_API_KEY='...'
|
|
85
|
+
|
|
86
|
+
# OR Vertex AI (OAuth mode - uses gcloud ADC)
|
|
87
|
+
export VERTEX_PROJECT='your-gcp-project-id'
|
|
60
88
|
```
|
|
61
89
|
|
|
62
90
|
3. **Optional but recommended:**
|
|
@@ -81,9 +109,20 @@ Claudish = Claude Code + Any AI Model
|
|
|
81
109
|
| Model ID | Backend | Best For |
|
|
82
110
|
|----------|---------|----------|
|
|
83
111
|
| `g/gemini-2.0-flash` | Gemini | Fast tasks, large context |
|
|
112
|
+
| `v/gemini-2.5-flash` | Vertex AI | Enterprise, GCP billing |
|
|
84
113
|
| `oai/gpt-4o` | OpenAI | General purpose |
|
|
85
114
|
| `ollama/llama3.2` | Local | Free, private |
|
|
86
115
|
|
|
116
|
+
**Vertex AI Partner Models (MaaS):**
|
|
117
|
+
|
|
118
|
+
| Model ID | Provider | Best For |
|
|
119
|
+
|----------|----------|----------|
|
|
120
|
+
| `vertex/minimax/minimax-m2-maas` | MiniMax | Fast, budget-friendly |
|
|
121
|
+
| `vertex/mistralai/codestral-2` | Mistral | Code specialist |
|
|
122
|
+
| `vertex/deepseek/deepseek-v3-2-maas` | DeepSeek | Deep reasoning |
|
|
123
|
+
| `vertex/qwen/qwen3-coder-480b-a35b-instruct-maas` | Qwen | Agentic coding |
|
|
124
|
+
| `vertex/openai/gpt-oss-120b-maas` | OpenAI | Open-weight reasoning |
|
|
125
|
+
|
|
87
126
|
**Update models:**
|
|
88
127
|
```bash
|
|
89
128
|
claudish --models --force-update
|
|
@@ -247,6 +286,120 @@ git diff | claudish --stdin --model google/gemini-2.5-flash "review for bugs"
|
|
|
247
286
|
claudish --model qwen/qwen3-vl-235b-a22b-instruct "implement dashboard from design"
|
|
248
287
|
```
|
|
249
288
|
|
|
289
|
+
## MCP Server Mode
|
|
290
|
+
|
|
291
|
+
Claudish can run as an MCP (Model Context Protocol) server, exposing OpenRouter models as tools that Claude Code can call mid-conversation. This is useful when you want to:
|
|
292
|
+
|
|
293
|
+
- Query external models without spawning a subprocess
|
|
294
|
+
- Compare responses from multiple models
|
|
295
|
+
- Use specific models for specific subtasks
|
|
296
|
+
|
|
297
|
+
### Starting MCP Server
|
|
298
|
+
|
|
299
|
+
```bash
|
|
300
|
+
# Start MCP server (stdio transport)
|
|
301
|
+
claudish --mcp
|
|
302
|
+
```
|
|
303
|
+
|
|
304
|
+
### Claude Code Configuration
|
|
305
|
+
|
|
306
|
+
Add to `~/.claude/settings.json`:
|
|
307
|
+
|
|
308
|
+
```json
|
|
309
|
+
{
|
|
310
|
+
"mcpServers": {
|
|
311
|
+
"claudish": {
|
|
312
|
+
"command": "claudish",
|
|
313
|
+
"args": ["--mcp"],
|
|
314
|
+
"env": {
|
|
315
|
+
"OPENROUTER_API_KEY": "sk-or-v1-..."
|
|
316
|
+
}
|
|
317
|
+
}
|
|
318
|
+
}
|
|
319
|
+
}
|
|
320
|
+
```
|
|
321
|
+
|
|
322
|
+
Or use npx (no installation needed):
|
|
323
|
+
|
|
324
|
+
```json
|
|
325
|
+
{
|
|
326
|
+
"mcpServers": {
|
|
327
|
+
"claudish": {
|
|
328
|
+
"command": "npx",
|
|
329
|
+
"args": ["claudish@latest", "--mcp"]
|
|
330
|
+
}
|
|
331
|
+
}
|
|
332
|
+
}
|
|
333
|
+
```
|
|
334
|
+
|
|
335
|
+
### Available MCP Tools
|
|
336
|
+
|
|
337
|
+
| Tool | Description | Example Use |
|
|
338
|
+
|------|-------------|-------------|
|
|
339
|
+
| `run_prompt` | Execute prompt on any model | Get a second opinion from Grok |
|
|
340
|
+
| `list_models` | Show recommended models | Find models with tool support |
|
|
341
|
+
| `search_models` | Fuzzy search all models | Find vision-capable models |
|
|
342
|
+
| `compare_models` | Run same prompt on multiple models | Compare reasoning approaches |
|
|
343
|
+
|
|
344
|
+
### Using MCP Tools from Claude Code
|
|
345
|
+
|
|
346
|
+
Once configured, Claude Code can use these tools directly:
|
|
347
|
+
|
|
348
|
+
```
|
|
349
|
+
User: "Use Grok to review this code"
|
|
350
|
+
Claude: [calls run_prompt tool with model="x-ai/grok-code-fast-1"]
|
|
351
|
+
|
|
352
|
+
User: "What models support vision?"
|
|
353
|
+
Claude: [calls search_models tool with query="vision"]
|
|
354
|
+
|
|
355
|
+
User: "Compare how GPT-5 and Gemini explain this concept"
|
|
356
|
+
Claude: [calls compare_models tool with models=["openai/gpt-5.2", "google/gemini-3-pro-preview"]]
|
|
357
|
+
```
|
|
358
|
+
|
|
359
|
+
### MCP vs CLI Mode
|
|
360
|
+
|
|
361
|
+
| Feature | CLI Mode | MCP Mode |
|
|
362
|
+
|---------|----------|----------|
|
|
363
|
+
| Use case | Replace Claude Code model | Call models as tools |
|
|
364
|
+
| Context | Full Claude Code session | Single prompt/response |
|
|
365
|
+
| Streaming | Full streaming | Buffered response |
|
|
366
|
+
| Best for | Primary model replacement | Second opinions, comparisons |
|
|
367
|
+
|
|
368
|
+
### MCP Tool Details
|
|
369
|
+
|
|
370
|
+
**run_prompt**
|
|
371
|
+
```typescript
|
|
372
|
+
{
|
|
373
|
+
model: string, // e.g., "x-ai/grok-code-fast-1"
|
|
374
|
+
prompt: string, // The prompt to send
|
|
375
|
+
system_prompt?: string, // Optional system prompt
|
|
376
|
+
max_tokens?: number // Default: 4096
|
|
377
|
+
}
|
|
378
|
+
```
|
|
379
|
+
|
|
380
|
+
**list_models**
|
|
381
|
+
```typescript
|
|
382
|
+
// No parameters - returns curated list of recommended models
|
|
383
|
+
{}
|
|
384
|
+
```
|
|
385
|
+
|
|
386
|
+
**search_models**
|
|
387
|
+
```typescript
|
|
388
|
+
{
|
|
389
|
+
query: string, // e.g., "grok", "vision", "free"
|
|
390
|
+
limit?: number // Default: 10
|
|
391
|
+
}
|
|
392
|
+
```
|
|
393
|
+
|
|
394
|
+
**compare_models**
|
|
395
|
+
```typescript
|
|
396
|
+
{
|
|
397
|
+
models: string[], // e.g., ["openai/gpt-5.2", "x-ai/grok-code-fast-1"]
|
|
398
|
+
prompt: string, // Prompt to send to all models
|
|
399
|
+
system_prompt?: string // Optional system prompt
|
|
400
|
+
}
|
|
401
|
+
```
|
|
402
|
+
|
|
250
403
|
## Getting Model List
|
|
251
404
|
|
|
252
405
|
### JSON Output (Recommended)
|
|
@@ -558,6 +711,6 @@ claudish --help-ai > claudish-agent-guide.md
|
|
|
558
711
|
|
|
559
712
|
---
|
|
560
713
|
|
|
561
|
-
**Version:** 2.
|
|
562
|
-
**Last Updated:** January
|
|
714
|
+
**Version:** 2.2.0
|
|
715
|
+
**Last Updated:** January 22, 2026
|
|
563
716
|
**Maintained by:** MadAppGang
|