claude-evolve 1.8.29 → 1.8.30

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/lib/ai-cli.sh CHANGED
@@ -180,6 +180,13 @@ $prompt"
180
180
  ai_output=$(timeout -k 30 600 kimi --print -c "$prompt" 2>&1)
181
181
  local ai_exit_code=$?
182
182
  ;;
183
+ kimi-coder)
184
+ local ai_output
185
+ # Kimi for Coding model via kimi CLI (fast coding-focused model)
186
+ # Use --print to see agent actions while still allowing file modifications
187
+ ai_output=$(timeout -k 30 600 kimi --print -y -m kimi-for-coding -c "$prompt" 2>&1)
188
+ local ai_exit_code=$?
189
+ ;;
183
190
  codex-oss-local)
184
191
  # Codex-OSS via Codex CLI with Ollama backend
185
192
  local ai_output
package/lib/config.sh CHANGED
@@ -58,7 +58,7 @@ DEFAULT_MEMORY_LIMIT_MB=12288
58
58
  DEFAULT_WORKER_MAX_CANDIDATES=3
59
59
 
60
60
  # Default LLM CLI configuration
61
- DEFAULT_LLM_RUN="glm-zai glm-zai glm-zai glm-zai glm-zai codex-oss-local haiku"
61
+ DEFAULT_LLM_RUN="glm-zai glm-zai glm-zai kimi-coder glm-zai glm-zai codex-oss-local haiku"
62
62
  # Ideate: Commercial models for idea generation + local fallback
63
63
  DEFAULT_LLM_IDEATE="opus-openrouter kimi-k2-openrouter gemini-3-pro-preview sonnet-think gpt5high grok-4-openrouter deepseek-openrouter glm-zai"
64
64
 
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "claude-evolve",
3
- "version": "1.8.29",
3
+ "version": "1.8.30",
4
4
  "bin": {
5
5
  "claude-evolve": "./bin/claude-evolve",
6
6
  "claude-evolve-main": "./bin/claude-evolve-main",
@@ -77,9 +77,9 @@ llm_cli:
77
77
  # Models are tried in order, with round-robin distribution across candidates
78
78
  # You can repeat models for weighted selection (e.g., "sonnet sonnet gemini" for 2:1 ratio)
79
79
 
80
- # Default configuration: 100% local code generation, commercial ideation + local fallback
80
+ # Default configuration: Mix of local and commercial code generation, commercial ideation
81
81
  # Commented out because these change over time; uncomment to override
82
- #run: codex-qwen3
82
+ #run: glm-zai glm-zai kimi-coder codex-oss-local
83
83
  #ideate: opus-openrouter kimi-k2-openrouter gemini-pro sonnet-think gpt5high grok-4-openrouter deepseek-openrouter glm-zai
84
84
 
85
85
  # Available models:
@@ -99,5 +99,6 @@ llm_cli:
99
99
  # - opus-openrouter: Claude Opus 4.1 via OpenRouter
100
100
  # - kimi-k2-openrouter: Kimi K2 Thinking via OpenRouter (RECOMMENDED - no separate auth)
101
101
  # - kimi-k2-think-moonshot: Kimi K2 Thinking via Moonshot CLI (requires separate kimi CLI setup)
102
+ # - kimi-coder: Kimi for Coding model via kimi CLI (fast, good for code generation)
102
103
  # - codex-qwen3: Qwen3-Coder via Codex + Ollama (local, free, RECOMMENDED)
103
104
  # - aider-qwen3: Qwen3-Coder via Aider + Ollama (local, free, experimental)