claudish 3.1.1 → 3.1.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/AI_AGENT_GUIDE.md CHANGED
@@ -1,6 +1,6 @@
1
1
  # Claudish AI Agent Usage Guide
2
2
 
3
- **Version:** 1.0.0
3
+ **Version:** 2.0.0
4
4
  **Target Audience:** AI Agents running within Claude Code
5
5
  **Purpose:** Quick reference for using Claudish CLI in agentic workflows
6
6
 
@@ -10,24 +10,43 @@
10
10
 
11
11
  ```bash
12
12
  # 1. Get available models
13
- claudish --list-models --json
13
+ claudish --models --json
14
+
15
+ # 2. Run task with specific model (OpenRouter)
16
+ claudish --model openai/gpt-5.2 "your task here"
17
+
18
+ # 3. Run with direct Gemini API
19
+ claudish --model g/gemini-2.0-flash "your task here"
14
20
 
15
- # 2. Run task with specific model
16
- claudish --model x-ai/grok-code-fast-1 "your task here"
21
+ # 4. Run with local model
22
+ claudish --model ollama/llama3.2 "your task here"
17
23
 
18
- # 3. For large prompts, use stdin
19
- echo "your task" | claudish --stdin --model x-ai/grok-code-fast-1
24
+ # 5. For large prompts, use stdin
25
+ echo "your task" | claudish --stdin --model openai/gpt-5.2
20
26
  ```
21
27
 
22
28
  ## What is Claudish?
23
29
 
24
- Claudish = Claude Code + OpenRouter models
30
+ Claudish = Claude Code + Any AI Model
25
31
 
26
- - ✅ Run Claude Code with **any OpenRouter model** (Grok, GPT-5, Gemini, MiniMax, etc.)
32
+ - ✅ Run Claude Code with **any AI model** via prefix-based routing
33
+ - ✅ Supports OpenRouter (100+ models), direct Gemini API, direct OpenAI API
34
+ - ✅ Supports local models (Ollama, LM Studio, vLLM, MLX)
27
35
  - ✅ 100% Claude Code feature compatibility
28
36
  - ✅ Local proxy server (no data sent to Claudish servers)
29
37
  - ✅ Cost tracking and model selection
30
38
 
39
+ ## Model Routing
40
+
41
+ | Prefix | Backend | Example |
42
+ |--------|---------|---------|
43
+ | _(none)_ | OpenRouter | `openai/gpt-5.2` |
44
+ | `g/` `gemini/` | Google Gemini | `g/gemini-2.0-flash` |
45
+ | `oai/` `openai/` | OpenAI | `oai/gpt-4o` |
46
+ | `ollama/` | Ollama | `ollama/llama3.2` |
47
+ | `lmstudio/` | LM Studio | `lmstudio/model` |
48
+ | `http://...` | Custom | `http://localhost:8000/model` |
49
+
31
50
  ## Prerequisites
32
51
 
33
52
  1. **Install Claudish:**
@@ -49,15 +68,25 @@ Claudish = Claude Code + OpenRouter models
49
68
 
50
69
  | Model ID | Provider | Category | Best For |
51
70
  |----------|----------|----------|----------|
52
- | `x-ai/grok-code-fast-1` | xAI | Coding | Fast iterations, agentic coding |
53
- | `google/gemini-2.5-flash` | Google | Reasoning | Complex analysis, 1000K context |
54
- | `minimax/minimax-m2` | MiniMax | Coding | General coding tasks |
55
- | `openai/gpt-5` | OpenAI | Reasoning | Architecture decisions |
56
- | `qwen/qwen3-vl-235b-a22b-instruct` | Alibaba | Vision | UI/visual tasks |
71
+ | `openai/gpt-5.2` | OpenAI | Reasoning | **Default** - Most advanced reasoning |
72
+ | `minimax/minimax-m2.1` | MiniMax | Coding | Budget-friendly, fast |
73
+ | `z-ai/glm-4.7` | Z.AI | Coding | Balanced performance |
74
+ | `google/gemini-3-pro-preview` | Google | Reasoning | 1M context window |
75
+ | `moonshotai/kimi-k2-thinking` | MoonShot | Reasoning | Extended thinking |
76
+ | `deepseek/deepseek-v3.2` | DeepSeek | Coding | Code specialist |
77
+ | `qwen/qwen3-vl-235b-a22b-thinking` | Alibaba | Vision | Vision + reasoning |
78
+
79
+ **Direct API Options (lower latency):**
80
+
81
+ | Model ID | Backend | Best For |
82
+ |----------|---------|----------|
83
+ | `g/gemini-2.0-flash` | Gemini | Fast tasks, large context |
84
+ | `oai/gpt-4o` | OpenAI | General purpose |
85
+ | `ollama/llama3.2` | Local | Free, private |
57
86
 
58
87
  **Update models:**
59
88
  ```bash
60
- claudish --list-models --force-update
89
+ claudish --models --force-update
61
90
  ```
62
91
 
63
92
  ## Critical: File-Based Pattern for Sub-Agents
@@ -529,6 +558,6 @@ claudish --help-ai > claudish-agent-guide.md
529
558
 
530
559
  ---
531
560
 
532
- **Version:** 1.0.0
533
- **Last Updated:** November 19, 2025
561
+ **Version:** 2.0.0
562
+ **Last Updated:** January 5, 2026
534
563
  **Maintained by:** MadAppGang
package/README.md CHANGED
@@ -1,17 +1,19 @@
1
1
  # Claudish
2
2
 
3
- > Run Claude Code with OpenRouter models via local proxy
3
+ > Run Claude Code with any AI model - OpenRouter, Gemini, OpenAI, or local models
4
4
 
5
- **Claudish** (Claude-ish) is a CLI tool that allows you to run Claude Code with any OpenRouter model by proxying requests through a local Anthropic API-compatible server.
5
+ **Claudish** (Claude-ish) is a CLI tool that allows you to run Claude Code with any AI model by proxying requests through a local Anthropic API-compatible server. Supports OpenRouter (100+ models), direct Google Gemini API, direct OpenAI API, and local models (Ollama, LM Studio, vLLM, MLX).
6
6
 
7
7
  ## Features
8
8
 
9
+ - ✅ **Multi-provider support** - OpenRouter, Gemini, OpenAI, and local models via prefix routing
10
+ - ✅ **Direct API access** - Use `g/gemini-2.0-flash` or `oai/gpt-4o` for direct API calls
11
+ - ✅ **Local model support** - Ollama, LM Studio, vLLM, MLX with `ollama/`, `lmstudio/` prefixes
9
12
  - ✅ **Cross-platform** - Works with both Node.js and Bun (v1.3.0+)
10
13
  - ✅ **Universal compatibility** - Use with `npx` or `bunx` - no installation required
11
14
  - ✅ **Interactive setup** - Prompts for API key and model if not provided (zero config!)
12
15
  - ✅ **Monitor mode** - Proxy to real Anthropic API and log all traffic (for debugging)
13
16
  - ✅ **Protocol compliance** - 1:1 compatibility with Claude Code communication protocol
14
- - ✅ **Snapshot testing** - Comprehensive test suite with 13/13 passing tests
15
17
  - ✅ **Headless mode** - Automatic print mode for non-interactive execution
16
18
  - ✅ **Quiet mode** - Clean output by default (no log pollution)
17
19
  - ✅ **JSON output** - Structured data for tool integration
@@ -19,7 +21,6 @@
19
21
  - ✅ **Parallel runs** - Each instance gets isolated proxy
20
22
  - ✅ **Autonomous mode** - Bypass all prompts with flags
21
23
  - ✅ **Context inheritance** - Runs in current directory with same `.claude` settings
22
- - ✅ **Multiple models** - 10+ prioritized OpenRouter models
23
24
  - ✅ **Agent support** - Use Claude Code agents in headless mode with `--agent`
24
25
 
25
26
  ## Installation
@@ -43,7 +44,11 @@ bun install -g claudish
43
44
  ### Prerequisites
44
45
 
45
46
  - [Claude Code](https://claude.com/claude-code) - Claude CLI must be installed
46
- - [OpenRouter API Key](https://openrouter.ai/keys) - Free tier available
47
+ - At least one API key:
48
+ - [OpenRouter API Key](https://openrouter.ai/keys) - Access 100+ models (free tier available)
49
+ - [Google Gemini API Key](https://aistudio.google.com/apikey) - For direct Gemini access
50
+ - [OpenAI API Key](https://platform.openai.com/api-keys) - For direct OpenAI access
51
+ - Or local models (Ollama, LM Studio) - No API key needed
47
52
 
48
53
  ### Other Install Options
49
54
 
@@ -228,46 +233,94 @@ claudish [OPTIONS] <claude-args...>
228
233
 
229
234
  ### Environment Variables
230
235
 
231
- | Variable | Description | Required |
236
+ #### API Keys (at least one required)
237
+
238
+ | Variable | Description | Used For |
232
239
  |----------|-------------|----------|
233
- | `OPENROUTER_API_KEY` | Your OpenRouter API key | **Optional in interactive mode** (will prompt if not set)<br>✅ **Required in non-interactive mode** |
234
- | `ANTHROPIC_API_KEY` | Placeholder to prevent Claude Code dialog (not used for auth) | ✅ **Required** |
235
- | `CLAUDISH_MODEL` | Default model to use | No |
236
- | `CLAUDISH_PORT` | Default proxy port | No |
237
- | `CLAUDISH_ACTIVE_MODEL_NAME` | Automatically set by claudish to show active model in status line (read-only) | ❌ No |
240
+ | `OPENROUTER_API_KEY` | OpenRouter API key | Default backend (100+ models) |
241
+ | `GEMINI_API_KEY` | Google Gemini API key | Direct Gemini access (`g/` prefix) |
242
+ | `OPENAI_API_KEY` | OpenAI API key | Direct OpenAI access (`oai/` prefix) |
243
+ | `ANTHROPIC_API_KEY` | Placeholder (any value) | Prevents Claude Code dialog |
244
+
245
+ #### Custom Endpoints (optional)
246
+
247
+ | Variable | Description | Default |
248
+ |----------|-------------|---------|
249
+ | `GEMINI_BASE_URL` | Custom Gemini endpoint | `https://generativelanguage.googleapis.com` |
250
+ | `OPENAI_BASE_URL` | Custom OpenAI/Azure endpoint | `https://api.openai.com` |
251
+ | `OLLAMA_BASE_URL` | Ollama server URL | `http://localhost:11434` |
252
+ | `LMSTUDIO_BASE_URL` | LM Studio server URL | `http://localhost:1234` |
253
+ | `VLLM_BASE_URL` | vLLM server URL | `http://localhost:8000` |
254
+ | `MLX_BASE_URL` | MLX server URL | `http://127.0.0.1:8080` |
255
+
256
+ #### Other Settings
257
+
258
+ | Variable | Description | Default |
259
+ |----------|-------------|---------|
260
+ | `CLAUDISH_MODEL` | Default model to use | `openai/gpt-5.2` |
261
+ | `CLAUDISH_PORT` | Default proxy port | Random (3000-9000) |
262
+ | `CLAUDISH_CONTEXT_WINDOW` | Override context window size | Auto-detected |
238
263
 
239
264
  **Important Notes:**
240
- - **NEW in v1.3.0:** In interactive mode, if `OPENROUTER_API_KEY` is not set, you'll be prompted to enter it
241
265
  - You MUST set `ANTHROPIC_API_KEY=sk-ant-api03-placeholder` (or any value). Without it, Claude Code will show a dialog
266
+ - In interactive mode, if no API key is set, you'll be prompted to enter one
267
+
268
+ ## Model Routing (v3.1.0+)
269
+
270
+ Claudish uses **prefix-based routing** to determine which API backend to use:
271
+
272
+ | Prefix | Backend | API Key | Example |
273
+ |--------|---------|---------|---------|
274
+ | _(none)_ | OpenRouter | `OPENROUTER_API_KEY` | `openai/gpt-5.2` |
275
+ | `or/` | OpenRouter | `OPENROUTER_API_KEY` | `or/anthropic/claude-3.5-sonnet` |
276
+ | `g/` `gemini/` `google/` | Google Gemini | `GEMINI_API_KEY` | `g/gemini-2.0-flash` |
277
+ | `oai/` `openai/` | OpenAI | `OPENAI_API_KEY` | `oai/gpt-4o` |
278
+ | `ollama/` | Ollama | _(none)_ | `ollama/llama3.2` |
279
+ | `lmstudio/` | LM Studio | _(none)_ | `lmstudio/qwen2.5-coder` |
280
+ | `vllm/` | vLLM | _(none)_ | `vllm/mistral-7b` |
281
+ | `mlx/` | MLX | _(none)_ | `mlx/llama-3.2-3b` |
282
+ | `http://...` | Custom | _(none)_ | `http://localhost:8000/model` |
242
283
 
243
- ## Available Models
284
+ ### Examples
244
285
 
245
- Claudish supports 5 OpenRouter models in priority order:
286
+ ```bash
287
+ # OpenRouter (default) - 100+ models via unified API
288
+ claudish --model openai/gpt-5.2 "implement feature"
289
+ claudish --model anthropic/claude-3.5-sonnet "review code"
290
+
291
+ # Direct Gemini API - lower latency, direct billing
292
+ claudish --model g/gemini-2.0-flash "quick task"
293
+ claudish --model gemini/gemini-2.5-pro "complex analysis"
246
294
 
247
- 1. **x-ai/grok-code-fast-1** (Default)
248
- - Fast coding-focused model from xAI
249
- - Best for quick iterations
295
+ # Direct OpenAI API - lower latency, direct billing
296
+ claudish --model oai/gpt-4o "implement feature"
297
+ claudish --model openai/o1 "complex reasoning"
250
298
 
251
- 2. **openai/gpt-5-codex**
252
- - Advanced coding model from OpenAI
253
- - Best for complex implementations
299
+ # Local models - free, private, no API key needed
300
+ claudish --model ollama/llama3.2 "code review"
301
+ claudish --model lmstudio/qwen2.5-coder "refactor"
302
+ ```
254
303
 
255
- 3. **minimax/minimax-m2**
256
- - High-performance model from MiniMax
257
- - Good for general coding tasks
304
+ ## Curated Models
258
305
 
259
- 4. **zhipu-ai/glm-4.6**
260
- - Advanced model from Zhipu AI
261
- - Good for multilingual code
306
+ Top recommended models for development (v3.1.1):
262
307
 
263
- 5. **qwen/qwen3-vl-235b-a22b-instruct**
264
- - Vision-language model from Alibaba
265
- - Best for UI/visual tasks
308
+ | Model | Provider | Best For |
309
+ |-------|----------|----------|
310
+ | `openai/gpt-5.2` | OpenAI | **Default** - Most advanced reasoning |
311
+ | `minimax/minimax-m2.1` | MiniMax | Budget-friendly, fast |
312
+ | `z-ai/glm-4.7` | Z.AI | Balanced performance |
313
+ | `google/gemini-3-pro-preview` | Google | 1M context window |
314
+ | `moonshotai/kimi-k2-thinking` | MoonShot | Extended reasoning |
315
+ | `deepseek/deepseek-v3.2` | DeepSeek | Code specialist |
316
+ | `qwen/qwen3-vl-235b-a22b-thinking` | Alibaba | Vision + reasoning |
266
317
 
267
- List models anytime with:
318
+ List all models:
268
319
 
269
320
  ```bash
270
- claudish --models
321
+ claudish --models # List all OpenRouter models
322
+ claudish --models gemini # Search for specific models
323
+ claudish --top-models # Show curated recommendations
271
324
  ```
272
325
 
273
326
  ## Agent Support (NEW in v2.1.0)
@@ -363,14 +416,19 @@ claudish "implement user authentication with JWT tokens"
363
416
  ### With Specific Model
364
417
 
365
418
  ```bash
366
- # Use Grok for fast coding
367
- claudish --model x-ai/grok-code-fast-1 "add error handling"
419
+ # Use OpenRouter models (default)
420
+ claudish --model openai/gpt-5.2 "refactor entire API layer"
421
+ claudish --model deepseek/deepseek-v3.2 "add error handling"
422
+
423
+ # Use direct Gemini API (faster, direct billing)
424
+ claudish --model g/gemini-2.0-flash "quick fix"
368
425
 
369
- # Use GPT-5 Codex for complex tasks
370
- claudish --model openai/gpt-5-codex "refactor entire API layer"
426
+ # Use direct OpenAI API
427
+ claudish --model oai/gpt-4o "implement feature"
371
428
 
372
- # Use Qwen for UI tasks
373
- claudish --model qwen/qwen3-vl-235b-a22b-instruct "implement dashboard UI"
429
+ # Use local models (free, private)
430
+ claudish --model ollama/llama3.2 "code review"
431
+ claudish --model lmstudio/qwen2.5-coder "implement dashboard UI"
374
432
  ```
375
433
 
376
434
  ### Autonomous Mode
package/dist/index.js CHANGED
@@ -35367,12 +35367,22 @@ function printVersion() {
35367
35367
  }
35368
35368
  function printHelp() {
35369
35369
  console.log(`
35370
- claudish - Run Claude Code with OpenRouter models
35370
+ claudish - Run Claude Code with any AI model (OpenRouter, Gemini, OpenAI, Local)
35371
35371
 
35372
35372
  USAGE:
35373
35373
  claudish # Interactive mode (default, shows model selector)
35374
35374
  claudish [OPTIONS] <claude-args...> # Single-shot mode (requires --model)
35375
35375
 
35376
+ MODEL ROUTING (prefix-based):
35377
+ (no prefix) OpenRouter (default) claudish --model openai/gpt-5.2 "task"
35378
+ g/, gemini/ Google Gemini API claudish --model g/gemini-2.0-flash "task"
35379
+ oai/, openai/ OpenAI API claudish --model oai/gpt-4o "task"
35380
+ ollama/ Ollama (local) claudish --model ollama/llama3.2 "task"
35381
+ lmstudio/ LM Studio (local) claudish --model lmstudio/qwen "task"
35382
+ vllm/ vLLM (local) claudish --model vllm/model "task"
35383
+ mlx/ MLX (local) claudish --model mlx/model "task"
35384
+ http://... Custom endpoint claudish --model http://localhost:8000/model "task"
35385
+
35376
35386
  OPTIONS:
35377
35387
  -i, --interactive Run in interactive mode (default when no prompt given)
35378
35388
  -m, --model <model> OpenRouter model to use (required for single-shot mode)
@@ -35433,27 +35443,33 @@ NOTES:
35433
35443
  ENVIRONMENT VARIABLES:
35434
35444
  Claudish automatically loads .env file from current directory.
35435
35445
 
35436
- OPENROUTER_API_KEY Required: Your OpenRouter API key (for OpenRouter models)
35437
- CLAUDISH_MODEL Default model to use (takes priority)
35438
- ANTHROPIC_MODEL Claude Code standard: model to use (fallback)
35446
+ API Keys (at least one required for cloud models):
35447
+ OPENROUTER_API_KEY OpenRouter API key (default backend)
35448
+ GEMINI_API_KEY Google Gemini API key (for g/ prefix)
35449
+ OPENAI_API_KEY OpenAI API key (for oai/ prefix)
35450
+ ANTHROPIC_API_KEY Placeholder (prevents Claude Code dialog)
35451
+
35452
+ Custom endpoints:
35453
+ GEMINI_BASE_URL Custom Gemini endpoint
35454
+ OPENAI_BASE_URL Custom OpenAI/Azure endpoint
35455
+
35456
+ Local providers:
35457
+ OLLAMA_BASE_URL Ollama server (default: http://localhost:11434)
35458
+ OLLAMA_HOST Alias for OLLAMA_BASE_URL
35459
+ LMSTUDIO_BASE_URL LM Studio server (default: http://localhost:1234)
35460
+ VLLM_BASE_URL vLLM server (default: http://localhost:8000)
35461
+ MLX_BASE_URL MLX server (default: http://127.0.0.1:8080)
35462
+
35463
+ Model settings:
35464
+ CLAUDISH_MODEL Default model to use (default: openai/gpt-5.2)
35439
35465
  CLAUDISH_PORT Default port for proxy
35440
- CLAUDISH_ACTIVE_MODEL_NAME Auto-set by claudish (read-only) - shows active model
35466
+ CLAUDISH_CONTEXT_WINDOW Override context window size
35441
35467
 
35442
- Model mapping (CLAUDISH_* takes priority over ANTHROPIC_DEFAULT_*):
35468
+ Model mapping (per-role):
35443
35469
  CLAUDISH_MODEL_OPUS Override model for Opus role
35444
35470
  CLAUDISH_MODEL_SONNET Override model for Sonnet role
35445
35471
  CLAUDISH_MODEL_HAIKU Override model for Haiku role
35446
35472
  CLAUDISH_MODEL_SUBAGENT Override model for sub-agents
35447
- ANTHROPIC_DEFAULT_OPUS_MODEL Claude Code standard: Opus model (fallback)
35448
- ANTHROPIC_DEFAULT_SONNET_MODEL Claude Code standard: Sonnet model (fallback)
35449
- ANTHROPIC_DEFAULT_HAIKU_MODEL Claude Code standard: Haiku model (fallback)
35450
- CLAUDE_CODE_SUBAGENT_MODEL Claude Code standard: sub-agent model (fallback)
35451
-
35452
- Local providers (OpenAI-compatible):
35453
- OLLAMA_BASE_URL Ollama server (default: http://localhost:11434)
35454
- OLLAMA_HOST Alias for OLLAMA_BASE_URL (same default)
35455
- LMSTUDIO_BASE_URL LM Studio server (default: http://localhost:1234)
35456
- VLLM_BASE_URL vLLM server (default: http://localhost:8000)
35457
35473
 
35458
35474
  EXAMPLES:
35459
35475
  # Interactive mode (default) - shows model selector
@@ -35463,26 +35479,28 @@ EXAMPLES:
35463
35479
  # Interactive mode with only FREE models
35464
35480
  claudish --free
35465
35481
 
35466
- # Interactive mode with pre-selected model
35467
- claudish --model x-ai/grok-code-fast-1
35482
+ # OpenRouter models (default)
35483
+ claudish --model openai/gpt-5.2 "implement user authentication"
35484
+ claudish --model deepseek/deepseek-v3.2 "add tests for login"
35468
35485
 
35469
- # Single-shot mode - one task and exit (requires --model or CLAUDISH_MODEL env var)
35470
- claudish --model openai/gpt-5-codex "implement user authentication"
35471
- claudish --model x-ai/grok-code-fast-1 "add tests for login"
35486
+ # Direct Gemini API (lower latency)
35487
+ claudish --model g/gemini-2.0-flash "quick fix"
35488
+ claudish --model gemini/gemini-2.5-pro "complex analysis"
35472
35489
 
35473
- # Per-role model mapping (use different models for different Claude Code roles)
35474
- claudish --model-opus openai/gpt-5 --model-sonnet x-ai/grok-code-fast-1 --model-haiku minimax/minimax-m2
35490
+ # Direct OpenAI API
35491
+ claudish --model oai/gpt-4o "implement feature"
35492
+ claudish --model openai/o1 "complex reasoning"
35475
35493
 
35476
- # Use named profiles for pre-configured model mappings
35477
- claudish -p frontend "implement component"
35478
- claudish --profile debug "investigate error"
35494
+ # Local models (free, private)
35495
+ claudish --model ollama/llama3.2 "code review"
35496
+ claudish --model lmstudio/qwen2.5-coder "refactor"
35479
35497
 
35480
- # Hybrid: Native Anthropic for Opus, OpenRouter for Sonnet/Haiku
35481
- claudish --model-opus claude-3-opus-20240229 --model-sonnet x-ai/grok-code-fast-1
35498
+ # Per-role model mapping
35499
+ claudish --model-opus openai/gpt-5.2 --model-sonnet deepseek/deepseek-v3.2 --model-haiku minimax/minimax-m2.1
35482
35500
 
35483
35501
  # Use stdin for large prompts (e.g., git diffs, code review)
35484
- echo "Review this code..." | claudish --stdin --model x-ai/grok-code-fast-1
35485
- git diff | claudish --stdin --model openai/gpt-5-codex "Review these changes"
35502
+ echo "Review this code..." | claudish --stdin --model g/gemini-2.0-flash
35503
+ git diff | claudish --stdin --model openai/gpt-5.2 "Review these changes"
35486
35504
 
35487
35505
  # Monitor mode - understand how Claude Code works (requires real Anthropic API key)
35488
35506
  claudish --monitor --debug "analyze code structure"
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "claudish",
3
- "version": "3.1.1",
3
+ "version": "3.1.2",
4
4
  "description": "Run Claude Code with any model - OpenRouter, Ollama, LM Studio & local models",
5
5
  "type": "module",
6
6
  "main": "./dist/index.js",
@@ -1,12 +1,12 @@
1
1
  ---
2
2
  name: claudish-usage
3
- description: CRITICAL - Guide for using Claudish CLI ONLY through sub-agents to run Claude Code with OpenRouter models (Grok, GPT-5, Gemini, MiniMax). NEVER run Claudish directly in main context unless user explicitly requests it. Use when user mentions external AI models, Claudish, OpenRouter, or alternative models. Includes mandatory sub-agent delegation patterns, agent selection guide, file-based instructions, and strict rules to prevent context window pollution.
3
+ description: CRITICAL - Guide for using Claudish CLI ONLY through sub-agents to run Claude Code with any AI model (OpenRouter, Gemini, OpenAI, local models). NEVER run Claudish directly in main context unless user explicitly requests it. Use when user mentions external AI models, Claudish, OpenRouter, Gemini, OpenAI, Ollama, or alternative models. Includes mandatory sub-agent delegation patterns, agent selection guide, file-based instructions, and strict rules to prevent context window pollution.
4
4
  ---
5
5
 
6
6
  # Claudish Usage Skill
7
7
 
8
- **Version:** 1.1.0
9
- **Purpose:** Guide AI agents on how to use Claudish CLI to run Claude Code with OpenRouter models
8
+ **Version:** 2.0.0
9
+ **Purpose:** Guide AI agents on how to use Claudish CLI to run Claude Code with any AI model
10
10
  **Status:** Production Ready
11
11
 
12
12
  ## ⚠️ CRITICAL RULES - READ FIRST
@@ -151,50 +151,71 @@ Decision:
151
151
 
152
152
  ## Overview
153
153
 
154
- **Claudish** is a CLI tool that allows running Claude Code with any OpenRouter model (Grok, GPT-5, MiniMax, Gemini, etc.) by proxying requests through a local Anthropic API-compatible server.
154
+ **Claudish** is a CLI tool that allows running Claude Code with any AI model via prefix-based routing. Supports OpenRouter (100+ models), direct Google Gemini API, direct OpenAI API, and local models (Ollama, LM Studio, vLLM, MLX).
155
155
 
156
156
  **Key Principle:** **ALWAYS** use Claudish through sub-agents with file-based instructions to avoid context window pollution.
157
157
 
158
158
  ## What is Claudish?
159
159
 
160
160
  Claudish (Claude-ish) is a proxy tool that:
161
- - ✅ Runs Claude Code with **any OpenRouter model** (not just Anthropic models)
161
+ - ✅ Runs Claude Code with **any AI model** via prefix-based routing
162
+ - ✅ Supports OpenRouter, Gemini, OpenAI, and local models
162
163
  - ✅ Uses local API-compatible proxy server
163
164
  - ✅ Supports 100% of Claude Code features
164
165
  - ✅ Provides cost tracking and model selection
165
166
  - ✅ Enables multi-model workflows
166
167
 
168
+ ## Model Routing
169
+
170
+ | Prefix | Backend | Example |
171
+ |--------|---------|---------|
172
+ | _(none)_ | OpenRouter | `openai/gpt-5.2` |
173
+ | `g/` `gemini/` | Google Gemini | `g/gemini-2.0-flash` |
174
+ | `oai/` `openai/` | OpenAI | `oai/gpt-4o` |
175
+ | `ollama/` | Ollama | `ollama/llama3.2` |
176
+ | `lmstudio/` | LM Studio | `lmstudio/model` |
177
+ | `http://...` | Custom | `http://localhost:8000/model` |
178
+
167
179
  **Use Cases:**
168
- - Run tasks with different AI models (Grok for speed, GPT-5 for reasoning, Gemini for vision)
180
+ - Run tasks with different AI models (Grok for speed, GPT-5 for reasoning, Gemini for large context)
181
+ - Use direct APIs for lower latency (Gemini, OpenAI)
182
+ - Use local models for free, private inference (Ollama, LM Studio)
169
183
  - Compare model performance on same task
170
184
  - Reduce costs with cheaper models for simple tasks
171
- - Access models with specialized capabilities
172
185
 
173
186
  ## Requirements
174
187
 
175
188
  ### System Requirements
176
- - **OpenRouter API Key** - Required (set as `OPENROUTER_API_KEY` environment variable)
177
189
  - **Claudish CLI** - Install with: `npm install -g claudish` or `bun install -g claudish`
178
190
  - **Claude Code** - Must be installed
191
+ - **At least one API key** (see below)
179
192
 
180
193
  ### Environment Variables
181
194
 
182
195
  ```bash
183
- # Required
184
- export OPENROUTER_API_KEY='sk-or-v1-...' # Your OpenRouter API key
196
+ # API Keys (at least one required)
197
+ export OPENROUTER_API_KEY='sk-or-v1-...' # OpenRouter (100+ models)
198
+ export GEMINI_API_KEY='...' # Direct Gemini API (g/ prefix)
199
+ export OPENAI_API_KEY='sk-...' # Direct OpenAI API (oai/ prefix)
200
+
201
+ # Placeholder (required to prevent Claude Code dialog)
202
+ export ANTHROPIC_API_KEY='sk-ant-api03-placeholder'
185
203
 
186
- # Optional (but recommended)
187
- export ANTHROPIC_API_KEY='sk-ant-api03-placeholder' # Prevents Claude Code dialog
204
+ # Custom endpoints (optional)
205
+ export GEMINI_BASE_URL='https://...' # Custom Gemini endpoint
206
+ export OPENAI_BASE_URL='https://...' # Custom OpenAI/Azure endpoint
207
+ export OLLAMA_BASE_URL='http://...' # Custom Ollama server
208
+ export LMSTUDIO_BASE_URL='http://...' # Custom LM Studio server
188
209
 
189
- # Optional - default model
190
- export CLAUDISH_MODEL='x-ai/grok-code-fast-1' # or ANTHROPIC_MODEL
210
+ # Default model (optional)
211
+ export CLAUDISH_MODEL='openai/gpt-5.2' # Default model
191
212
  ```
192
213
 
193
- **Get OpenRouter API Key:**
194
- 1. Visit https://openrouter.ai/keys
195
- 2. Sign up (free tier available)
196
- 3. Create API key
197
- 4. Set as environment variable
214
+ **Get API Keys:**
215
+ - OpenRouter: https://openrouter.ai/keys (free tier available)
216
+ - Gemini: https://aistudio.google.com/apikey
217
+ - OpenAI: https://platform.openai.com/api-keys
218
+ - Local models: No API key needed
198
219
 
199
220
  ## Quick Start Guide
200
221
 
@@ -254,32 +275,25 @@ git diff | claudish --stdin --model openai/gpt-5-codex "Review these changes"
254
275
 
255
276
  ## Recommended Models
256
277
 
257
- **Top Models for Development (verified from OpenRouter):**
258
-
259
- 1. **x-ai/grok-code-fast-1** - xAI's Grok (fast coding, visible reasoning)
260
- - Category: coding
261
- - Context: 256K
262
- - Best for: Quick iterations, agentic coding
263
-
264
- 2. **google/gemini-2.5-flash** - Google's Gemini (state-of-the-art reasoning)
265
- - Category: reasoning
266
- - Context: 1000K
267
- - Best for: Complex analysis, multi-step reasoning
278
+ **Top Models for Development (v3.1.1):**
268
279
 
269
- 3. **minimax/minimax-m2** - MiniMax M2 (high performance)
270
- - Category: coding
271
- - Context: 128K
272
- - Best for: General coding tasks
280
+ | Model | Provider | Best For |
281
+ |-------|----------|----------|
282
+ | `openai/gpt-5.2` | OpenAI | **Default** - Most advanced reasoning |
283
+ | `minimax/minimax-m2.1` | MiniMax | Budget-friendly, fast |
284
+ | `z-ai/glm-4.7` | Z.AI | Balanced performance |
285
+ | `google/gemini-3-pro-preview` | Google | 1M context window |
286
+ | `moonshotai/kimi-k2-thinking` | MoonShot | Extended thinking |
287
+ | `deepseek/deepseek-v3.2` | DeepSeek | Code specialist |
288
+ | `qwen/qwen3-vl-235b-a22b-thinking` | Alibaba | Vision + reasoning |
273
289
 
274
- 4. **openai/gpt-5** - OpenAI's GPT-5 (advanced reasoning)
275
- - Category: reasoning
276
- - Context: 128K
277
- - Best for: Complex implementations, architecture decisions
290
+ **Direct API Options (lower latency):**
278
291
 
279
- 5. **qwen/qwen3-vl-235b-a22b-instruct** - Alibaba's Qwen (vision-language)
280
- - Category: vision
281
- - Context: 32K
282
- - Best for: UI/visual tasks, design implementation
292
+ | Model | Backend | Best For |
293
+ |-------|---------|----------|
294
+ | `g/gemini-2.0-flash` | Gemini | Fast tasks, large context |
295
+ | `oai/gpt-4o` | OpenAI | General purpose |
296
+ | `ollama/llama3.2` | Local | Free, private |
283
297
 
284
298
  **Get Latest Models:**
285
299
  ```bash
@@ -1294,5 +1308,5 @@ claudish --help-ai # AI agent usage guide
1294
1308
  ---
1295
1309
 
1296
1310
  **Maintained by:** MadAppGang
1297
- **Last Updated:** November 25, 2025
1298
- **Skill Version:** 1.1.0
1311
+ **Last Updated:** January 5, 2026
1312
+ **Skill Version:** 2.0.0