claude-evolve 1.8.29 → 1.8.31
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/ai-cli.sh +7 -0
- package/lib/config.sh +1 -1
- package/package.json +1 -1
- package/templates/config.yaml +3 -2
package/lib/ai-cli.sh
CHANGED
|
@@ -180,6 +180,13 @@ $prompt"
|
|
|
180
180
|
ai_output=$(timeout -k 30 600 kimi --print -c "$prompt" 2>&1)
|
|
181
181
|
local ai_exit_code=$?
|
|
182
182
|
;;
|
|
183
|
+
kimi-coder)
|
|
184
|
+
local ai_output
|
|
185
|
+
# Kimi for Coding model via kimi CLI (fast coding-focused model)
|
|
186
|
+
# Use --print to see agent actions while still allowing file modifications
|
|
187
|
+
ai_output=$(timeout -k 30 600 kimi --print -y -m kimi-for-coding -c "$prompt" 2>&1)
|
|
188
|
+
local ai_exit_code=$?
|
|
189
|
+
;;
|
|
183
190
|
codex-oss-local)
|
|
184
191
|
# Codex-OSS via Codex CLI with Ollama backend
|
|
185
192
|
local ai_output
|
package/lib/config.sh
CHANGED
|
@@ -58,7 +58,7 @@ DEFAULT_MEMORY_LIMIT_MB=12288
|
|
|
58
58
|
DEFAULT_WORKER_MAX_CANDIDATES=3
|
|
59
59
|
|
|
60
60
|
# Default LLM CLI configuration
|
|
61
|
-
DEFAULT_LLM_RUN="glm-zai glm-zai glm-zai glm-zai glm-zai codex-oss-local haiku"
|
|
61
|
+
DEFAULT_LLM_RUN="glm-zai glm-zai glm-zai kimi-coder glm-zai glm-zai codex-oss-local haiku"
|
|
62
62
|
# Ideate: Commercial models for idea generation + local fallback
|
|
63
63
|
DEFAULT_LLM_IDEATE="opus-openrouter kimi-k2-openrouter gemini-3-pro-preview sonnet-think gpt5high grok-4-openrouter deepseek-openrouter glm-zai"
|
|
64
64
|
|
package/package.json
CHANGED
package/templates/config.yaml
CHANGED
|
@@ -77,9 +77,9 @@ llm_cli:
|
|
|
77
77
|
# Models are tried in order, with round-robin distribution across candidates
|
|
78
78
|
# You can repeat models for weighted selection (e.g., "sonnet sonnet gemini" for 2:1 ratio)
|
|
79
79
|
|
|
80
|
-
# Default configuration:
|
|
80
|
+
# Default configuration: Mix of local and commercial code generation, commercial ideation
|
|
81
81
|
# Commented out because these change over time; uncomment to override
|
|
82
|
-
#run: codex-
|
|
82
|
+
#run: glm-zai glm-zai kimi-coder codex-oss-local
|
|
83
83
|
#ideate: opus-openrouter kimi-k2-openrouter gemini-pro sonnet-think gpt5high grok-4-openrouter deepseek-openrouter glm-zai
|
|
84
84
|
|
|
85
85
|
# Available models:
|
|
@@ -99,5 +99,6 @@ llm_cli:
|
|
|
99
99
|
# - opus-openrouter: Claude Opus 4.1 via OpenRouter
|
|
100
100
|
# - kimi-k2-openrouter: Kimi K2 Thinking via OpenRouter (RECOMMENDED - no separate auth)
|
|
101
101
|
# - kimi-k2-think-moonshot: Kimi K2 Thinking via Moonshot CLI (requires separate kimi CLI setup)
|
|
102
|
+
# - kimi-coder: Kimi for Coding model via kimi CLI (fast, good for code generation)
|
|
102
103
|
# - codex-qwen3: Qwen3-Coder via Codex + Ollama (local, free, RECOMMENDED)
|
|
103
104
|
# - aider-qwen3: Qwen3-Coder via Aider + Ollama (local, free, experimental)
|