claudish 2.2.1 → 2.5.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -135,7 +135,12 @@ claudish --help-ai > claudish-agent-guide.md
135
135
 
136
136
  1. **Get available models:**
137
137
  ```bash
138
- claudish --list-models --json
138
+ # List all models or search
139
+ claudish --models
140
+ claudish --models gemini
141
+
142
+ # Get top recommended models (JSON)
143
+ claudish --top-models --json
139
144
  ```
140
145
 
141
146
  2. **Run Claudish through sub-agent** (recommended pattern):
@@ -187,7 +192,7 @@ claudish --help-ai > claudish-agent-guide.md
187
192
  - ✅ Use file-based patterns to avoid context window pollution
188
193
  - ✅ Delegate to sub-agents instead of running directly
189
194
  - ✅ Return summaries only (not full conversation transcripts)
190
- - ✅ Choose appropriate model for task (see `--list-models`)
195
+ - ✅ Choose appropriate model for task (see `--models` or `--top-models`)
191
196
 
192
197
  **Resources:**
193
198
  - Full AI agent guide: `claudish --help-ai`
@@ -216,7 +221,8 @@ claudish [OPTIONS] <claude-args...>
216
221
  | `--no-auto-approve` | Disable auto-approve (require prompts) | Auto-approve **enabled** |
217
222
  | `--dangerous` | Pass `--dangerouslyDisableSandbox` | `false` |
218
223
  | `--agent <agent>` | Use specific agent (e.g., `frontend:developer`) | - |
219
- | `--list-models` | List available OpenRouter models | - |
224
+ | `--models` | List all models or search (e.g., `--models gemini`) | - |
225
+ | `--top-models` | Show top recommended programming models | - |
220
226
  | `--list-agents` | List available agents in current project | - |
221
227
  | `--force-update` | Force refresh model cache | - |
222
228
  | `--init` | Install Claudish skill in current project | - |
@@ -264,7 +270,7 @@ Claudish supports 5 OpenRouter models in priority order:
264
270
  List models anytime with:
265
271
 
266
272
  ```bash
267
- claudish --list-models
273
+ claudish --models
268
274
  ```
269
275
 
270
276
  ## Agent Support (NEW in v2.1.0)
@@ -600,6 +606,24 @@ claudish --model minimax/minimax-m2 "task C"
600
606
 
601
607
  **NEW in v1.1.0**: Claudish now fully supports models with extended thinking/reasoning capabilities (Grok, o1, etc.) with complete Anthropic Messages API protocol compliance.
602
608
 
609
+ ### Thinking Translation Model (v1.5.0)
610
+
611
+ Claudish includes a sophisticated **Thinking Translation Model** that aligns Claude Code's native thinking budget with the unique requirements of every major AI provider.
612
+
613
+ When you set a thinking budget in Claude (e.g., `budget: 16000`), Claudish automatically translates it:
614
+
615
+ | Provider | Model | Translation Logic |
616
+ | :--- | :--- | :--- |
617
+ | **OpenAI** | o1, o3 | Maps budget to `reasoning_effort` (minimal/low/medium/high) |
618
+ | **Google** | Gemini 3 | Maps to `thinking_level` (low/high) |
619
+ | **Google** | Gemini 2.x | Passes exact `thinking_budget` (capped at 24k) |
620
+ | **xAI** | Grok 3 Mini | Maps to `reasoning_effort` (low/high) |
621
+ | **Qwen** | Qwen 2.5 | Enables `enable_thinking` + exact budget |
622
+ | **MiniMax** | M2 | Enables `reasoning_split` (interleaved thinking) |
623
+ | **DeepSeek** | R1 | Automatically manages reasoning (params stripped for safety) |
624
+
625
+ This ensures you can use standard Claude Code thinking controls with **ANY** supported model, without worrying about API specificities.
626
+
603
627
  ### What is Extended Thinking?
604
628
 
605
629
  Some AI models (like Grok and OpenAI's o1) can show their internal reasoning process before providing the final answer. This "thinking" content helps you understand how the model arrived at its conclusion.
@@ -678,6 +702,61 @@ For complete protocol documentation, see:
678
702
  - [COMPREHENSIVE_UX_ISSUE_ANALYSIS.md](./COMPREHENSIVE_UX_ISSUE_ANALYSIS.md) - Technical analysis
679
703
  - [THINKING_BLOCKS_IMPLEMENTATION.md](./THINKING_BLOCKS_IMPLEMENTATION.md) - Implementation summary
680
704
 
705
+ ## Dynamic Reasoning Support (NEW in v1.4.0)
706
+
707
+ **Claudish now intelligently adapts to ANY reasoning model!**
708
+
709
+ No more hardcoded lists or manual flags. Claudish dynamically queries OpenRouter metadata to enable thinking capabilities for any model that supports them.
710
+
711
+ ### 🧠 Dynamic Thinking Features
712
+
713
+ 1. **Auto-Detection**:
714
+ - Automatically checks model capabilities at startup
715
+ - Enables Extended Thinking UI *only* when supported
716
+ - Future-proof: Works instantly with new models (e.g., `deepseek-r1` or `minimax-m2`)
717
+
718
+ 2. **Smart Parameter Mapping**:
719
+ - **Claude**: Passes token budget directly (e.g., 16k tokens)
720
+ - **OpenAI (o1/o3)**: Translates budget to `reasoning_effort`
721
+ - "ultrathink" (≥32k) → `high`
722
+ - "think hard" (16k-32k) → `medium`
723
+ - "think" (<16k) → `low`
724
+ - **Gemini & Grok**: Preserves thought signatures and XML traces automatically
725
+
726
+ 3. **Universal Compatibility**:
727
+ - Use "ultrathink" or "think hard" prompts with ANY supported model
728
+ - Claudish handles the translation layer for you
729
+
730
+ ## Context Scaling & Auto-Compaction
731
+
732
+ **NEW in v1.2.0**: Claudish now intelligently manages token counting to support ANY context window size (from 128k to 2M+) while preserving Claude Code's native auto-compaction behavior.
733
+
734
+ ### The Challenge
735
+
736
+ Claude Code naturally assumes a fixed context window (typically 200k tokens for Sonnet).
737
+ - **Small Models (e.g., Grok 128k)**: Claude might overuse context and crash.
738
+ - **Massive Models (e.g., Gemini 2M)**: Claude would compact way too early (at 10% usage), wasting the model's potential.
739
+
740
+ ### The Solution: Token Scaling
741
+
742
+ Claudish implements a "Dual-Accounting" system:
743
+
744
+ 1. **Internal Scaling (For Claude):**
745
+ - We fetch the *real* context limit from OpenRouter (e.g., 1M tokens).
746
+ - We scale reported token usage so Claude *thinks* 1M tokens is 200k.
747
+ - **Result:** Auto-compaction triggers at the correct *percentage* of usage (e.g., 90% full), regardless of the actual limit.
748
+
749
+ 2. **Accurate Reporting (For You):**
750
+ - The status line displays the **Real Unscaled Usage** and **Real Context %**.
751
+ - You see specific costs and limits, while Claude remains blissfully unaware and stable.
752
+
753
+ **Benefits:**
754
+ - ✅ **Works with ANY model** size (128k, 1M, 2M, etc.)
755
+ - ✅ **Unlocks massive context** windows (Claude Code becomes 10x more powerful with Gemini!)
756
+ - ✅ **Prevents crashes** on smaller models (Grok)
757
+ - ✅ **Native behavior** (compaction just works)
758
+
759
+
681
760
  ## Development
682
761
 
683
762
  ### Project Structure