claudish 3.1.1 → 3.1.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/AI_AGENT_GUIDE.md +45 -16
- package/README.md +95 -37
- package/dist/index.js +49 -31
- package/package.json +1 -1
- package/skills/claudish-usage/SKILL.md +58 -44
package/AI_AGENT_GUIDE.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
# Claudish AI Agent Usage Guide
|
|
2
2
|
|
|
3
|
-
**Version:**
|
|
3
|
+
**Version:** 2.0.0
|
|
4
4
|
**Target Audience:** AI Agents running within Claude Code
|
|
5
5
|
**Purpose:** Quick reference for using Claudish CLI in agentic workflows
|
|
6
6
|
|
|
@@ -10,24 +10,43 @@
|
|
|
10
10
|
|
|
11
11
|
```bash
|
|
12
12
|
# 1. Get available models
|
|
13
|
-
claudish --
|
|
13
|
+
claudish --models --json
|
|
14
|
+
|
|
15
|
+
# 2. Run task with specific model (OpenRouter)
|
|
16
|
+
claudish --model openai/gpt-5.2 "your task here"
|
|
17
|
+
|
|
18
|
+
# 3. Run with direct Gemini API
|
|
19
|
+
claudish --model g/gemini-2.0-flash "your task here"
|
|
14
20
|
|
|
15
|
-
#
|
|
16
|
-
claudish --model
|
|
21
|
+
# 4. Run with local model
|
|
22
|
+
claudish --model ollama/llama3.2 "your task here"
|
|
17
23
|
|
|
18
|
-
#
|
|
19
|
-
echo "your task" | claudish --stdin --model
|
|
24
|
+
# 5. For large prompts, use stdin
|
|
25
|
+
echo "your task" | claudish --stdin --model openai/gpt-5.2
|
|
20
26
|
```
|
|
21
27
|
|
|
22
28
|
## What is Claudish?
|
|
23
29
|
|
|
24
|
-
Claudish = Claude Code +
|
|
30
|
+
Claudish = Claude Code + Any AI Model
|
|
25
31
|
|
|
26
|
-
- ✅ Run Claude Code with **any
|
|
32
|
+
- ✅ Run Claude Code with **any AI model** via prefix-based routing
|
|
33
|
+
- ✅ Supports OpenRouter (100+ models), direct Gemini API, direct OpenAI API
|
|
34
|
+
- ✅ Supports local models (Ollama, LM Studio, vLLM, MLX)
|
|
27
35
|
- ✅ 100% Claude Code feature compatibility
|
|
28
36
|
- ✅ Local proxy server (no data sent to Claudish servers)
|
|
29
37
|
- ✅ Cost tracking and model selection
|
|
30
38
|
|
|
39
|
+
## Model Routing
|
|
40
|
+
|
|
41
|
+
| Prefix | Backend | Example |
|
|
42
|
+
|--------|---------|---------|
|
|
43
|
+
| _(none)_ | OpenRouter | `openai/gpt-5.2` |
|
|
44
|
+
| `g/` `gemini/` | Google Gemini | `g/gemini-2.0-flash` |
|
|
45
|
+
| `oai/` `openai/` | OpenAI | `oai/gpt-4o` |
|
|
46
|
+
| `ollama/` | Ollama | `ollama/llama3.2` |
|
|
47
|
+
| `lmstudio/` | LM Studio | `lmstudio/model` |
|
|
48
|
+
| `http://...` | Custom | `http://localhost:8000/model` |
|
|
49
|
+
|
|
31
50
|
## Prerequisites
|
|
32
51
|
|
|
33
52
|
1. **Install Claudish:**
|
|
@@ -49,15 +68,25 @@ Claudish = Claude Code + OpenRouter models
|
|
|
49
68
|
|
|
50
69
|
| Model ID | Provider | Category | Best For |
|
|
51
70
|
|----------|----------|----------|----------|
|
|
52
|
-
| `
|
|
53
|
-
| `
|
|
54
|
-
| `
|
|
55
|
-
| `
|
|
56
|
-
| `
|
|
71
|
+
| `openai/gpt-5.2` | OpenAI | Reasoning | **Default** - Most advanced reasoning |
|
|
72
|
+
| `minimax/minimax-m2.1` | MiniMax | Coding | Budget-friendly, fast |
|
|
73
|
+
| `z-ai/glm-4.7` | Z.AI | Coding | Balanced performance |
|
|
74
|
+
| `google/gemini-3-pro-preview` | Google | Reasoning | 1M context window |
|
|
75
|
+
| `moonshotai/kimi-k2-thinking` | MoonShot | Reasoning | Extended thinking |
|
|
76
|
+
| `deepseek/deepseek-v3.2` | DeepSeek | Coding | Code specialist |
|
|
77
|
+
| `qwen/qwen3-vl-235b-a22b-thinking` | Alibaba | Vision | Vision + reasoning |
|
|
78
|
+
|
|
79
|
+
**Direct API Options (lower latency):**
|
|
80
|
+
|
|
81
|
+
| Model ID | Backend | Best For |
|
|
82
|
+
|----------|---------|----------|
|
|
83
|
+
| `g/gemini-2.0-flash` | Gemini | Fast tasks, large context |
|
|
84
|
+
| `oai/gpt-4o` | OpenAI | General purpose |
|
|
85
|
+
| `ollama/llama3.2` | Local | Free, private |
|
|
57
86
|
|
|
58
87
|
**Update models:**
|
|
59
88
|
```bash
|
|
60
|
-
claudish --
|
|
89
|
+
claudish --models --force-update
|
|
61
90
|
```
|
|
62
91
|
|
|
63
92
|
## Critical: File-Based Pattern for Sub-Agents
|
|
@@ -529,6 +558,6 @@ claudish --help-ai > claudish-agent-guide.md
|
|
|
529
558
|
|
|
530
559
|
---
|
|
531
560
|
|
|
532
|
-
**Version:**
|
|
533
|
-
**Last Updated:**
|
|
561
|
+
**Version:** 2.0.0
|
|
562
|
+
**Last Updated:** January 5, 2026
|
|
534
563
|
**Maintained by:** MadAppGang
|
package/README.md
CHANGED
|
@@ -1,17 +1,19 @@
|
|
|
1
1
|
# Claudish
|
|
2
2
|
|
|
3
|
-
> Run Claude Code with OpenRouter
|
|
3
|
+
> Run Claude Code with any AI model - OpenRouter, Gemini, OpenAI, or local models
|
|
4
4
|
|
|
5
|
-
**Claudish** (Claude-ish) is a CLI tool that allows you to run Claude Code with any
|
|
5
|
+
**Claudish** (Claude-ish) is a CLI tool that allows you to run Claude Code with any AI model by proxying requests through a local Anthropic API-compatible server. Supports OpenRouter (100+ models), direct Google Gemini API, direct OpenAI API, and local models (Ollama, LM Studio, vLLM, MLX).
|
|
6
6
|
|
|
7
7
|
## Features
|
|
8
8
|
|
|
9
|
+
- ✅ **Multi-provider support** - OpenRouter, Gemini, OpenAI, and local models via prefix routing
|
|
10
|
+
- ✅ **Direct API access** - Use `g/gemini-2.0-flash` or `oai/gpt-4o` for direct API calls
|
|
11
|
+
- ✅ **Local model support** - Ollama, LM Studio, vLLM, MLX with `ollama/`, `lmstudio/` prefixes
|
|
9
12
|
- ✅ **Cross-platform** - Works with both Node.js and Bun (v1.3.0+)
|
|
10
13
|
- ✅ **Universal compatibility** - Use with `npx` or `bunx` - no installation required
|
|
11
14
|
- ✅ **Interactive setup** - Prompts for API key and model if not provided (zero config!)
|
|
12
15
|
- ✅ **Monitor mode** - Proxy to real Anthropic API and log all traffic (for debugging)
|
|
13
16
|
- ✅ **Protocol compliance** - 1:1 compatibility with Claude Code communication protocol
|
|
14
|
-
- ✅ **Snapshot testing** - Comprehensive test suite with 13/13 passing tests
|
|
15
17
|
- ✅ **Headless mode** - Automatic print mode for non-interactive execution
|
|
16
18
|
- ✅ **Quiet mode** - Clean output by default (no log pollution)
|
|
17
19
|
- ✅ **JSON output** - Structured data for tool integration
|
|
@@ -19,7 +21,6 @@
|
|
|
19
21
|
- ✅ **Parallel runs** - Each instance gets isolated proxy
|
|
20
22
|
- ✅ **Autonomous mode** - Bypass all prompts with flags
|
|
21
23
|
- ✅ **Context inheritance** - Runs in current directory with same `.claude` settings
|
|
22
|
-
- ✅ **Multiple models** - 10+ prioritized OpenRouter models
|
|
23
24
|
- ✅ **Agent support** - Use Claude Code agents in headless mode with `--agent`
|
|
24
25
|
|
|
25
26
|
## Installation
|
|
@@ -43,7 +44,11 @@ bun install -g claudish
|
|
|
43
44
|
### Prerequisites
|
|
44
45
|
|
|
45
46
|
- [Claude Code](https://claude.com/claude-code) - Claude CLI must be installed
|
|
46
|
-
-
|
|
47
|
+
- At least one API key:
|
|
48
|
+
- [OpenRouter API Key](https://openrouter.ai/keys) - Access 100+ models (free tier available)
|
|
49
|
+
- [Google Gemini API Key](https://aistudio.google.com/apikey) - For direct Gemini access
|
|
50
|
+
- [OpenAI API Key](https://platform.openai.com/api-keys) - For direct OpenAI access
|
|
51
|
+
- Or local models (Ollama, LM Studio) - No API key needed
|
|
47
52
|
|
|
48
53
|
### Other Install Options
|
|
49
54
|
|
|
@@ -228,46 +233,94 @@ claudish [OPTIONS] <claude-args...>
|
|
|
228
233
|
|
|
229
234
|
### Environment Variables
|
|
230
235
|
|
|
231
|
-
|
|
236
|
+
#### API Keys (at least one required)
|
|
237
|
+
|
|
238
|
+
| Variable | Description | Used For |
|
|
232
239
|
|----------|-------------|----------|
|
|
233
|
-
| `OPENROUTER_API_KEY` |
|
|
234
|
-
| `
|
|
235
|
-
| `
|
|
236
|
-
| `
|
|
237
|
-
|
|
240
|
+
| `OPENROUTER_API_KEY` | OpenRouter API key | Default backend (100+ models) |
|
|
241
|
+
| `GEMINI_API_KEY` | Google Gemini API key | Direct Gemini access (`g/` prefix) |
|
|
242
|
+
| `OPENAI_API_KEY` | OpenAI API key | Direct OpenAI access (`oai/` prefix) |
|
|
243
|
+
| `ANTHROPIC_API_KEY` | Placeholder (any value) | Prevents Claude Code dialog |
|
|
244
|
+
|
|
245
|
+
#### Custom Endpoints (optional)
|
|
246
|
+
|
|
247
|
+
| Variable | Description | Default |
|
|
248
|
+
|----------|-------------|---------|
|
|
249
|
+
| `GEMINI_BASE_URL` | Custom Gemini endpoint | `https://generativelanguage.googleapis.com` |
|
|
250
|
+
| `OPENAI_BASE_URL` | Custom OpenAI/Azure endpoint | `https://api.openai.com` |
|
|
251
|
+
| `OLLAMA_BASE_URL` | Ollama server URL | `http://localhost:11434` |
|
|
252
|
+
| `LMSTUDIO_BASE_URL` | LM Studio server URL | `http://localhost:1234` |
|
|
253
|
+
| `VLLM_BASE_URL` | vLLM server URL | `http://localhost:8000` |
|
|
254
|
+
| `MLX_BASE_URL` | MLX server URL | `http://127.0.0.1:8080` |
|
|
255
|
+
|
|
256
|
+
#### Other Settings
|
|
257
|
+
|
|
258
|
+
| Variable | Description | Default |
|
|
259
|
+
|----------|-------------|---------|
|
|
260
|
+
| `CLAUDISH_MODEL` | Default model to use | `openai/gpt-5.2` |
|
|
261
|
+
| `CLAUDISH_PORT` | Default proxy port | Random (3000-9000) |
|
|
262
|
+
| `CLAUDISH_CONTEXT_WINDOW` | Override context window size | Auto-detected |
|
|
238
263
|
|
|
239
264
|
**Important Notes:**
|
|
240
|
-
- **NEW in v1.3.0:** In interactive mode, if `OPENROUTER_API_KEY` is not set, you'll be prompted to enter it
|
|
241
265
|
- You MUST set `ANTHROPIC_API_KEY=sk-ant-api03-placeholder` (or any value). Without it, Claude Code will show a dialog
|
|
266
|
+
- In interactive mode, if no API key is set, you'll be prompted to enter one
|
|
267
|
+
|
|
268
|
+
## Model Routing (v3.1.0+)
|
|
269
|
+
|
|
270
|
+
Claudish uses **prefix-based routing** to determine which API backend to use:
|
|
271
|
+
|
|
272
|
+
| Prefix | Backend | API Key | Example |
|
|
273
|
+
|--------|---------|---------|---------|
|
|
274
|
+
| _(none)_ | OpenRouter | `OPENROUTER_API_KEY` | `openai/gpt-5.2` |
|
|
275
|
+
| `or/` | OpenRouter | `OPENROUTER_API_KEY` | `or/anthropic/claude-3.5-sonnet` |
|
|
276
|
+
| `g/` `gemini/` `google/` | Google Gemini | `GEMINI_API_KEY` | `g/gemini-2.0-flash` |
|
|
277
|
+
| `oai/` `openai/` | OpenAI | `OPENAI_API_KEY` | `oai/gpt-4o` |
|
|
278
|
+
| `ollama/` | Ollama | _(none)_ | `ollama/llama3.2` |
|
|
279
|
+
| `lmstudio/` | LM Studio | _(none)_ | `lmstudio/qwen2.5-coder` |
|
|
280
|
+
| `vllm/` | vLLM | _(none)_ | `vllm/mistral-7b` |
|
|
281
|
+
| `mlx/` | MLX | _(none)_ | `mlx/llama-3.2-3b` |
|
|
282
|
+
| `http://...` | Custom | _(none)_ | `http://localhost:8000/model` |
|
|
242
283
|
|
|
243
|
-
|
|
284
|
+
### Examples
|
|
244
285
|
|
|
245
|
-
|
|
286
|
+
```bash
|
|
287
|
+
# OpenRouter (default) - 100+ models via unified API
|
|
288
|
+
claudish --model openai/gpt-5.2 "implement feature"
|
|
289
|
+
claudish --model anthropic/claude-3.5-sonnet "review code"
|
|
290
|
+
|
|
291
|
+
# Direct Gemini API - lower latency, direct billing
|
|
292
|
+
claudish --model g/gemini-2.0-flash "quick task"
|
|
293
|
+
claudish --model gemini/gemini-2.5-pro "complex analysis"
|
|
246
294
|
|
|
247
|
-
|
|
248
|
-
|
|
249
|
-
|
|
295
|
+
# Direct OpenAI API - lower latency, direct billing
|
|
296
|
+
claudish --model oai/gpt-4o "implement feature"
|
|
297
|
+
claudish --model openai/o1 "complex reasoning"
|
|
250
298
|
|
|
251
|
-
|
|
252
|
-
|
|
253
|
-
|
|
299
|
+
# Local models - free, private, no API key needed
|
|
300
|
+
claudish --model ollama/llama3.2 "code review"
|
|
301
|
+
claudish --model lmstudio/qwen2.5-coder "refactor"
|
|
302
|
+
```
|
|
254
303
|
|
|
255
|
-
|
|
256
|
-
- High-performance model from MiniMax
|
|
257
|
-
- Good for general coding tasks
|
|
304
|
+
## Curated Models
|
|
258
305
|
|
|
259
|
-
|
|
260
|
-
- Advanced model from Zhipu AI
|
|
261
|
-
- Good for multilingual code
|
|
306
|
+
Top recommended models for development (v3.1.1):
|
|
262
307
|
|
|
263
|
-
|
|
264
|
-
|
|
265
|
-
|
|
308
|
+
| Model | Provider | Best For |
|
|
309
|
+
|-------|----------|----------|
|
|
310
|
+
| `openai/gpt-5.2` | OpenAI | **Default** - Most advanced reasoning |
|
|
311
|
+
| `minimax/minimax-m2.1` | MiniMax | Budget-friendly, fast |
|
|
312
|
+
| `z-ai/glm-4.7` | Z.AI | Balanced performance |
|
|
313
|
+
| `google/gemini-3-pro-preview` | Google | 1M context window |
|
|
314
|
+
| `moonshotai/kimi-k2-thinking` | MoonShot | Extended reasoning |
|
|
315
|
+
| `deepseek/deepseek-v3.2` | DeepSeek | Code specialist |
|
|
316
|
+
| `qwen/qwen3-vl-235b-a22b-thinking` | Alibaba | Vision + reasoning |
|
|
266
317
|
|
|
267
|
-
List models
|
|
318
|
+
List all models:
|
|
268
319
|
|
|
269
320
|
```bash
|
|
270
|
-
claudish --models
|
|
321
|
+
claudish --models # List all OpenRouter models
|
|
322
|
+
claudish --models gemini # Search for specific models
|
|
323
|
+
claudish --top-models # Show curated recommendations
|
|
271
324
|
```
|
|
272
325
|
|
|
273
326
|
## Agent Support (NEW in v2.1.0)
|
|
@@ -363,14 +416,19 @@ claudish "implement user authentication with JWT tokens"
|
|
|
363
416
|
### With Specific Model
|
|
364
417
|
|
|
365
418
|
```bash
|
|
366
|
-
# Use
|
|
367
|
-
claudish --model
|
|
419
|
+
# Use OpenRouter models (default)
|
|
420
|
+
claudish --model openai/gpt-5.2 "refactor entire API layer"
|
|
421
|
+
claudish --model deepseek/deepseek-v3.2 "add error handling"
|
|
422
|
+
|
|
423
|
+
# Use direct Gemini API (faster, direct billing)
|
|
424
|
+
claudish --model g/gemini-2.0-flash "quick fix"
|
|
368
425
|
|
|
369
|
-
# Use
|
|
370
|
-
claudish --model
|
|
426
|
+
# Use direct OpenAI API
|
|
427
|
+
claudish --model oai/gpt-4o "implement feature"
|
|
371
428
|
|
|
372
|
-
# Use
|
|
373
|
-
claudish --model
|
|
429
|
+
# Use local models (free, private)
|
|
430
|
+
claudish --model ollama/llama3.2 "code review"
|
|
431
|
+
claudish --model lmstudio/qwen2.5-coder "implement dashboard UI"
|
|
374
432
|
```
|
|
375
433
|
|
|
376
434
|
### Autonomous Mode
|
package/dist/index.js
CHANGED
|
@@ -35367,12 +35367,22 @@ function printVersion() {
|
|
|
35367
35367
|
}
|
|
35368
35368
|
function printHelp() {
|
|
35369
35369
|
console.log(`
|
|
35370
|
-
claudish - Run Claude Code with OpenRouter
|
|
35370
|
+
claudish - Run Claude Code with any AI model (OpenRouter, Gemini, OpenAI, Local)
|
|
35371
35371
|
|
|
35372
35372
|
USAGE:
|
|
35373
35373
|
claudish # Interactive mode (default, shows model selector)
|
|
35374
35374
|
claudish [OPTIONS] <claude-args...> # Single-shot mode (requires --model)
|
|
35375
35375
|
|
|
35376
|
+
MODEL ROUTING (prefix-based):
|
|
35377
|
+
(no prefix) OpenRouter (default) claudish --model openai/gpt-5.2 "task"
|
|
35378
|
+
g/, gemini/ Google Gemini API claudish --model g/gemini-2.0-flash "task"
|
|
35379
|
+
oai/, openai/ OpenAI API claudish --model oai/gpt-4o "task"
|
|
35380
|
+
ollama/ Ollama (local) claudish --model ollama/llama3.2 "task"
|
|
35381
|
+
lmstudio/ LM Studio (local) claudish --model lmstudio/qwen "task"
|
|
35382
|
+
vllm/ vLLM (local) claudish --model vllm/model "task"
|
|
35383
|
+
mlx/ MLX (local) claudish --model mlx/model "task"
|
|
35384
|
+
http://... Custom endpoint claudish --model http://localhost:8000/model "task"
|
|
35385
|
+
|
|
35376
35386
|
OPTIONS:
|
|
35377
35387
|
-i, --interactive Run in interactive mode (default when no prompt given)
|
|
35378
35388
|
-m, --model <model> OpenRouter model to use (required for single-shot mode)
|
|
@@ -35433,27 +35443,33 @@ NOTES:
|
|
|
35433
35443
|
ENVIRONMENT VARIABLES:
|
|
35434
35444
|
Claudish automatically loads .env file from current directory.
|
|
35435
35445
|
|
|
35436
|
-
|
|
35437
|
-
|
|
35438
|
-
|
|
35446
|
+
API Keys (at least one required for cloud models):
|
|
35447
|
+
OPENROUTER_API_KEY OpenRouter API key (default backend)
|
|
35448
|
+
GEMINI_API_KEY Google Gemini API key (for g/ prefix)
|
|
35449
|
+
OPENAI_API_KEY OpenAI API key (for oai/ prefix)
|
|
35450
|
+
ANTHROPIC_API_KEY Placeholder (prevents Claude Code dialog)
|
|
35451
|
+
|
|
35452
|
+
Custom endpoints:
|
|
35453
|
+
GEMINI_BASE_URL Custom Gemini endpoint
|
|
35454
|
+
OPENAI_BASE_URL Custom OpenAI/Azure endpoint
|
|
35455
|
+
|
|
35456
|
+
Local providers:
|
|
35457
|
+
OLLAMA_BASE_URL Ollama server (default: http://localhost:11434)
|
|
35458
|
+
OLLAMA_HOST Alias for OLLAMA_BASE_URL
|
|
35459
|
+
LMSTUDIO_BASE_URL LM Studio server (default: http://localhost:1234)
|
|
35460
|
+
VLLM_BASE_URL vLLM server (default: http://localhost:8000)
|
|
35461
|
+
MLX_BASE_URL MLX server (default: http://127.0.0.1:8080)
|
|
35462
|
+
|
|
35463
|
+
Model settings:
|
|
35464
|
+
CLAUDISH_MODEL Default model to use (default: openai/gpt-5.2)
|
|
35439
35465
|
CLAUDISH_PORT Default port for proxy
|
|
35440
|
-
|
|
35466
|
+
CLAUDISH_CONTEXT_WINDOW Override context window size
|
|
35441
35467
|
|
|
35442
|
-
Model mapping (
|
|
35468
|
+
Model mapping (per-role):
|
|
35443
35469
|
CLAUDISH_MODEL_OPUS Override model for Opus role
|
|
35444
35470
|
CLAUDISH_MODEL_SONNET Override model for Sonnet role
|
|
35445
35471
|
CLAUDISH_MODEL_HAIKU Override model for Haiku role
|
|
35446
35472
|
CLAUDISH_MODEL_SUBAGENT Override model for sub-agents
|
|
35447
|
-
ANTHROPIC_DEFAULT_OPUS_MODEL Claude Code standard: Opus model (fallback)
|
|
35448
|
-
ANTHROPIC_DEFAULT_SONNET_MODEL Claude Code standard: Sonnet model (fallback)
|
|
35449
|
-
ANTHROPIC_DEFAULT_HAIKU_MODEL Claude Code standard: Haiku model (fallback)
|
|
35450
|
-
CLAUDE_CODE_SUBAGENT_MODEL Claude Code standard: sub-agent model (fallback)
|
|
35451
|
-
|
|
35452
|
-
Local providers (OpenAI-compatible):
|
|
35453
|
-
OLLAMA_BASE_URL Ollama server (default: http://localhost:11434)
|
|
35454
|
-
OLLAMA_HOST Alias for OLLAMA_BASE_URL (same default)
|
|
35455
|
-
LMSTUDIO_BASE_URL LM Studio server (default: http://localhost:1234)
|
|
35456
|
-
VLLM_BASE_URL vLLM server (default: http://localhost:8000)
|
|
35457
35473
|
|
|
35458
35474
|
EXAMPLES:
|
|
35459
35475
|
# Interactive mode (default) - shows model selector
|
|
@@ -35463,26 +35479,28 @@ EXAMPLES:
|
|
|
35463
35479
|
# Interactive mode with only FREE models
|
|
35464
35480
|
claudish --free
|
|
35465
35481
|
|
|
35466
|
-
#
|
|
35467
|
-
claudish --model
|
|
35482
|
+
# OpenRouter models (default)
|
|
35483
|
+
claudish --model openai/gpt-5.2 "implement user authentication"
|
|
35484
|
+
claudish --model deepseek/deepseek-v3.2 "add tests for login"
|
|
35468
35485
|
|
|
35469
|
-
#
|
|
35470
|
-
claudish --model
|
|
35471
|
-
claudish --model
|
|
35486
|
+
# Direct Gemini API (lower latency)
|
|
35487
|
+
claudish --model g/gemini-2.0-flash "quick fix"
|
|
35488
|
+
claudish --model gemini/gemini-2.5-pro "complex analysis"
|
|
35472
35489
|
|
|
35473
|
-
#
|
|
35474
|
-
claudish --model
|
|
35490
|
+
# Direct OpenAI API
|
|
35491
|
+
claudish --model oai/gpt-4o "implement feature"
|
|
35492
|
+
claudish --model openai/o1 "complex reasoning"
|
|
35475
35493
|
|
|
35476
|
-
#
|
|
35477
|
-
claudish
|
|
35478
|
-
claudish --
|
|
35494
|
+
# Local models (free, private)
|
|
35495
|
+
claudish --model ollama/llama3.2 "code review"
|
|
35496
|
+
claudish --model lmstudio/qwen2.5-coder "refactor"
|
|
35479
35497
|
|
|
35480
|
-
#
|
|
35481
|
-
claudish --model-opus
|
|
35498
|
+
# Per-role model mapping
|
|
35499
|
+
claudish --model-opus openai/gpt-5.2 --model-sonnet deepseek/deepseek-v3.2 --model-haiku minimax/minimax-m2.1
|
|
35482
35500
|
|
|
35483
35501
|
# Use stdin for large prompts (e.g., git diffs, code review)
|
|
35484
|
-
echo "Review this code..." | claudish --stdin --model
|
|
35485
|
-
git diff | claudish --stdin --model openai/gpt-5
|
|
35502
|
+
echo "Review this code..." | claudish --stdin --model g/gemini-2.0-flash
|
|
35503
|
+
git diff | claudish --stdin --model openai/gpt-5.2 "Review these changes"
|
|
35486
35504
|
|
|
35487
35505
|
# Monitor mode - understand how Claude Code works (requires real Anthropic API key)
|
|
35488
35506
|
claudish --monitor --debug "analyze code structure"
|
|
@@ -61965,7 +61983,7 @@ var getRemoteProviders = () => [
|
|
|
61965
61983
|
baseUrl: process.env.GEMINI_BASE_URL || "https://generativelanguage.googleapis.com",
|
|
61966
61984
|
apiPath: "/v1beta/models/{model}:streamGenerateContent?alt=sse",
|
|
61967
61985
|
apiKeyEnvVar: "GEMINI_API_KEY",
|
|
61968
|
-
prefixes: ["g/", "gemini/"
|
|
61986
|
+
prefixes: ["g/", "gemini/"],
|
|
61969
61987
|
capabilities: {
|
|
61970
61988
|
supportsTools: true,
|
|
61971
61989
|
supportsVision: true,
|
package/package.json
CHANGED
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: claudish-usage
|
|
3
|
-
description: CRITICAL - Guide for using Claudish CLI ONLY through sub-agents to run Claude Code with
|
|
3
|
+
description: CRITICAL - Guide for using Claudish CLI ONLY through sub-agents to run Claude Code with any AI model (OpenRouter, Gemini, OpenAI, local models). NEVER run Claudish directly in main context unless user explicitly requests it. Use when user mentions external AI models, Claudish, OpenRouter, Gemini, OpenAI, Ollama, or alternative models. Includes mandatory sub-agent delegation patterns, agent selection guide, file-based instructions, and strict rules to prevent context window pollution.
|
|
4
4
|
---
|
|
5
5
|
|
|
6
6
|
# Claudish Usage Skill
|
|
7
7
|
|
|
8
|
-
**Version:**
|
|
9
|
-
**Purpose:** Guide AI agents on how to use Claudish CLI to run Claude Code with
|
|
8
|
+
**Version:** 2.0.0
|
|
9
|
+
**Purpose:** Guide AI agents on how to use Claudish CLI to run Claude Code with any AI model
|
|
10
10
|
**Status:** Production Ready
|
|
11
11
|
|
|
12
12
|
## ⚠️ CRITICAL RULES - READ FIRST
|
|
@@ -151,50 +151,71 @@ Decision:
|
|
|
151
151
|
|
|
152
152
|
## Overview
|
|
153
153
|
|
|
154
|
-
**Claudish** is a CLI tool that allows running Claude Code with any
|
|
154
|
+
**Claudish** is a CLI tool that allows running Claude Code with any AI model via prefix-based routing. Supports OpenRouter (100+ models), direct Google Gemini API, direct OpenAI API, and local models (Ollama, LM Studio, vLLM, MLX).
|
|
155
155
|
|
|
156
156
|
**Key Principle:** **ALWAYS** use Claudish through sub-agents with file-based instructions to avoid context window pollution.
|
|
157
157
|
|
|
158
158
|
## What is Claudish?
|
|
159
159
|
|
|
160
160
|
Claudish (Claude-ish) is a proxy tool that:
|
|
161
|
-
- ✅ Runs Claude Code with **any
|
|
161
|
+
- ✅ Runs Claude Code with **any AI model** via prefix-based routing
|
|
162
|
+
- ✅ Supports OpenRouter, Gemini, OpenAI, and local models
|
|
162
163
|
- ✅ Uses local API-compatible proxy server
|
|
163
164
|
- ✅ Supports 100% of Claude Code features
|
|
164
165
|
- ✅ Provides cost tracking and model selection
|
|
165
166
|
- ✅ Enables multi-model workflows
|
|
166
167
|
|
|
168
|
+
## Model Routing
|
|
169
|
+
|
|
170
|
+
| Prefix | Backend | Example |
|
|
171
|
+
|--------|---------|---------|
|
|
172
|
+
| _(none)_ | OpenRouter | `openai/gpt-5.2` |
|
|
173
|
+
| `g/` `gemini/` | Google Gemini | `g/gemini-2.0-flash` |
|
|
174
|
+
| `oai/` `openai/` | OpenAI | `oai/gpt-4o` |
|
|
175
|
+
| `ollama/` | Ollama | `ollama/llama3.2` |
|
|
176
|
+
| `lmstudio/` | LM Studio | `lmstudio/model` |
|
|
177
|
+
| `http://...` | Custom | `http://localhost:8000/model` |
|
|
178
|
+
|
|
167
179
|
**Use Cases:**
|
|
168
|
-
- Run tasks with different AI models (Grok for speed, GPT-5 for reasoning, Gemini for
|
|
180
|
+
- Run tasks with different AI models (Grok for speed, GPT-5 for reasoning, Gemini for large context)
|
|
181
|
+
- Use direct APIs for lower latency (Gemini, OpenAI)
|
|
182
|
+
- Use local models for free, private inference (Ollama, LM Studio)
|
|
169
183
|
- Compare model performance on same task
|
|
170
184
|
- Reduce costs with cheaper models for simple tasks
|
|
171
|
-
- Access models with specialized capabilities
|
|
172
185
|
|
|
173
186
|
## Requirements
|
|
174
187
|
|
|
175
188
|
### System Requirements
|
|
176
|
-
- **OpenRouter API Key** - Required (set as `OPENROUTER_API_KEY` environment variable)
|
|
177
189
|
- **Claudish CLI** - Install with: `npm install -g claudish` or `bun install -g claudish`
|
|
178
190
|
- **Claude Code** - Must be installed
|
|
191
|
+
- **At least one API key** (see below)
|
|
179
192
|
|
|
180
193
|
### Environment Variables
|
|
181
194
|
|
|
182
195
|
```bash
|
|
183
|
-
#
|
|
184
|
-
export OPENROUTER_API_KEY='sk-or-v1-...' #
|
|
196
|
+
# API Keys (at least one required)
|
|
197
|
+
export OPENROUTER_API_KEY='sk-or-v1-...' # OpenRouter (100+ models)
|
|
198
|
+
export GEMINI_API_KEY='...' # Direct Gemini API (g/ prefix)
|
|
199
|
+
export OPENAI_API_KEY='sk-...' # Direct OpenAI API (oai/ prefix)
|
|
200
|
+
|
|
201
|
+
# Placeholder (required to prevent Claude Code dialog)
|
|
202
|
+
export ANTHROPIC_API_KEY='sk-ant-api03-placeholder'
|
|
185
203
|
|
|
186
|
-
#
|
|
187
|
-
export
|
|
204
|
+
# Custom endpoints (optional)
|
|
205
|
+
export GEMINI_BASE_URL='https://...' # Custom Gemini endpoint
|
|
206
|
+
export OPENAI_BASE_URL='https://...' # Custom OpenAI/Azure endpoint
|
|
207
|
+
export OLLAMA_BASE_URL='http://...' # Custom Ollama server
|
|
208
|
+
export LMSTUDIO_BASE_URL='http://...' # Custom LM Studio server
|
|
188
209
|
|
|
189
|
-
#
|
|
190
|
-
export CLAUDISH_MODEL='
|
|
210
|
+
# Default model (optional)
|
|
211
|
+
export CLAUDISH_MODEL='openai/gpt-5.2' # Default model
|
|
191
212
|
```
|
|
192
213
|
|
|
193
|
-
**Get
|
|
194
|
-
|
|
195
|
-
|
|
196
|
-
|
|
197
|
-
|
|
214
|
+
**Get API Keys:**
|
|
215
|
+
- OpenRouter: https://openrouter.ai/keys (free tier available)
|
|
216
|
+
- Gemini: https://aistudio.google.com/apikey
|
|
217
|
+
- OpenAI: https://platform.openai.com/api-keys
|
|
218
|
+
- Local models: No API key needed
|
|
198
219
|
|
|
199
220
|
## Quick Start Guide
|
|
200
221
|
|
|
@@ -254,32 +275,25 @@ git diff | claudish --stdin --model openai/gpt-5-codex "Review these changes"
|
|
|
254
275
|
|
|
255
276
|
## Recommended Models
|
|
256
277
|
|
|
257
|
-
**Top Models for Development (
|
|
258
|
-
|
|
259
|
-
1. **x-ai/grok-code-fast-1** - xAI's Grok (fast coding, visible reasoning)
|
|
260
|
-
- Category: coding
|
|
261
|
-
- Context: 256K
|
|
262
|
-
- Best for: Quick iterations, agentic coding
|
|
263
|
-
|
|
264
|
-
2. **google/gemini-2.5-flash** - Google's Gemini (state-of-the-art reasoning)
|
|
265
|
-
- Category: reasoning
|
|
266
|
-
- Context: 1000K
|
|
267
|
-
- Best for: Complex analysis, multi-step reasoning
|
|
278
|
+
**Top Models for Development (v3.1.1):**
|
|
268
279
|
|
|
269
|
-
|
|
270
|
-
|
|
271
|
-
|
|
272
|
-
|
|
280
|
+
| Model | Provider | Best For |
|
|
281
|
+
|-------|----------|----------|
|
|
282
|
+
| `openai/gpt-5.2` | OpenAI | **Default** - Most advanced reasoning |
|
|
283
|
+
| `minimax/minimax-m2.1` | MiniMax | Budget-friendly, fast |
|
|
284
|
+
| `z-ai/glm-4.7` | Z.AI | Balanced performance |
|
|
285
|
+
| `google/gemini-3-pro-preview` | Google | 1M context window |
|
|
286
|
+
| `moonshotai/kimi-k2-thinking` | MoonShot | Extended thinking |
|
|
287
|
+
| `deepseek/deepseek-v3.2` | DeepSeek | Code specialist |
|
|
288
|
+
| `qwen/qwen3-vl-235b-a22b-thinking` | Alibaba | Vision + reasoning |
|
|
273
289
|
|
|
274
|
-
|
|
275
|
-
- Category: reasoning
|
|
276
|
-
- Context: 128K
|
|
277
|
-
- Best for: Complex implementations, architecture decisions
|
|
290
|
+
**Direct API Options (lower latency):**
|
|
278
291
|
|
|
279
|
-
|
|
280
|
-
|
|
281
|
-
|
|
282
|
-
|
|
292
|
+
| Model | Backend | Best For |
|
|
293
|
+
|-------|---------|----------|
|
|
294
|
+
| `g/gemini-2.0-flash` | Gemini | Fast tasks, large context |
|
|
295
|
+
| `oai/gpt-4o` | OpenAI | General purpose |
|
|
296
|
+
| `ollama/llama3.2` | Local | Free, private |
|
|
283
297
|
|
|
284
298
|
**Get Latest Models:**
|
|
285
299
|
```bash
|
|
@@ -1294,5 +1308,5 @@ claudish --help-ai # AI agent usage guide
|
|
|
1294
1308
|
---
|
|
1295
1309
|
|
|
1296
1310
|
**Maintained by:** MadAppGang
|
|
1297
|
-
**Last Updated:**
|
|
1298
|
-
**Skill Version:**
|
|
1311
|
+
**Last Updated:** January 5, 2026
|
|
1312
|
+
**Skill Version:** 2.0.0
|