claude-evolve 1.6.8 → 1.6.9
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/ai-cli.sh +8 -2
- package/lib/config.sh +2 -2
- package/package.json +1 -1
package/lib/ai-cli.sh
CHANGED
|
@@ -97,11 +97,17 @@ $prompt"
|
|
|
97
97
|
local ai_exit_code=$?
|
|
98
98
|
;;
|
|
99
99
|
codex-qwen3)
|
|
100
|
-
# Qwen3-Coder via Codex CLI with Ollama backend (
|
|
100
|
+
# Qwen3-Coder via Codex CLI with Ollama backend (only mildly agentic)
|
|
101
101
|
local ai_output
|
|
102
102
|
ai_output=$(timeout 600 codex exec --dangerously-bypass-approvals-and-sandbox --skip-git-repo-check --oss --model qwen3-coder:30b "$prompt" 2>&1)
|
|
103
103
|
local ai_exit_code=$?
|
|
104
104
|
;;
|
|
105
|
+
codex-oss)
|
|
106
|
+
# Codex-OSS via Codex CLI with Ollama backend
|
|
107
|
+
local ai_output
|
|
108
|
+
ai_output=$(timeout 600 codex exec --dangerously-bypass-approvals-and-sandbox --skip-git-repo-check --oss "$prompt" 2>&1)
|
|
109
|
+
local ai_exit_code=$?
|
|
110
|
+
;;
|
|
105
111
|
aider-qwen3|qwen3)
|
|
106
112
|
# Qwen3-Coder via Aider + Ollama
|
|
107
113
|
# Extract the target filename from the prompt (e.g., "Modify the algorithm in evolution_gen01-001.py...")
|
|
@@ -345,4 +351,4 @@ call_ai_with_round_robin() {
|
|
|
345
351
|
# Legacy function name for compatibility
|
|
346
352
|
call_ai_with_fallbacks() {
|
|
347
353
|
call_ai_with_round_robin "$@"
|
|
348
|
-
}
|
|
354
|
+
}
|
package/lib/config.sh
CHANGED
|
@@ -55,9 +55,9 @@ DEFAULT_MEMORY_LIMIT_MB=12288
|
|
|
55
55
|
|
|
56
56
|
# Default LLM CLI configuration - use simple variables instead of arrays
|
|
57
57
|
# Run: 100% local with qwen3 via Codex+Ollama (more reliable than aider)
|
|
58
|
-
DEFAULT_LLM_RUN="codex-qwen3"
|
|
58
|
+
DEFAULT_LLM_RUN="codex-qwen3 codex-oss"
|
|
59
59
|
# Ideate: Commercial models for idea generation + local fallback
|
|
60
|
-
DEFAULT_LLM_IDEATE="gemini sonnet-think gpt5high o3high glm grok-4 codex-qwen3"
|
|
60
|
+
DEFAULT_LLM_IDEATE="gemini sonnet-think gpt5high o3high glm grok-4 codex-qwen3 codex-oss"
|
|
61
61
|
|
|
62
62
|
# Load configuration from config file
|
|
63
63
|
load_config() {
|