@kody-ade/kody-engine-lite 0.1.56 → 0.1.58
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/bin/cli.js +699 -274
- package/kody.config.schema.json +2 -2
- package/package.json +1 -1
- package/prompts/autofix.md +27 -9
- package/prompts/review.md +83 -16
- package/templates/kody.yml +8 -8
- package/dist/agent-runner.d.ts +0 -4
- package/dist/agent-runner.js +0 -122
- package/dist/ci/parse-inputs.d.ts +0 -6
- package/dist/ci/parse-inputs.js +0 -76
- package/dist/ci/parse-safety.d.ts +0 -6
- package/dist/ci/parse-safety.js +0 -22
- package/dist/cli/args.d.ts +0 -13
- package/dist/cli/args.js +0 -42
- package/dist/cli/litellm.d.ts +0 -2
- package/dist/cli/litellm.js +0 -85
- package/dist/cli/task-resolution.d.ts +0 -2
- package/dist/cli/task-resolution.js +0 -41
- package/dist/config.d.ts +0 -49
- package/dist/config.js +0 -72
- package/dist/context.d.ts +0 -4
- package/dist/context.js +0 -83
- package/dist/definitions.d.ts +0 -3
- package/dist/definitions.js +0 -59
- package/dist/entry.d.ts +0 -1
- package/dist/entry.js +0 -236
- package/dist/git-utils.d.ts +0 -13
- package/dist/git-utils.js +0 -174
- package/dist/github-api.d.ts +0 -14
- package/dist/github-api.js +0 -114
- package/dist/kody-utils.d.ts +0 -1
- package/dist/kody-utils.js +0 -9
- package/dist/learning/auto-learn.d.ts +0 -2
- package/dist/learning/auto-learn.js +0 -169
- package/dist/logger.d.ts +0 -14
- package/dist/logger.js +0 -51
- package/dist/memory.d.ts +0 -1
- package/dist/memory.js +0 -20
- package/dist/observer.d.ts +0 -9
- package/dist/observer.js +0 -80
- package/dist/pipeline/complexity.d.ts +0 -3
- package/dist/pipeline/complexity.js +0 -12
- package/dist/pipeline/executor-registry.d.ts +0 -3
- package/dist/pipeline/executor-registry.js +0 -20
- package/dist/pipeline/hooks.d.ts +0 -17
- package/dist/pipeline/hooks.js +0 -110
- package/dist/pipeline/questions.d.ts +0 -2
- package/dist/pipeline/questions.js +0 -44
- package/dist/pipeline/runner-selection.d.ts +0 -2
- package/dist/pipeline/runner-selection.js +0 -13
- package/dist/pipeline/state.d.ts +0 -4
- package/dist/pipeline/state.js +0 -37
- package/dist/pipeline.d.ts +0 -3
- package/dist/pipeline.js +0 -213
- package/dist/preflight.d.ts +0 -1
- package/dist/preflight.js +0 -69
- package/dist/retrospective.d.ts +0 -26
- package/dist/retrospective.js +0 -211
- package/dist/stages/agent.d.ts +0 -2
- package/dist/stages/agent.js +0 -94
- package/dist/stages/gate.d.ts +0 -2
- package/dist/stages/gate.js +0 -32
- package/dist/stages/review.d.ts +0 -2
- package/dist/stages/review.js +0 -32
- package/dist/stages/ship.d.ts +0 -3
- package/dist/stages/ship.js +0 -154
- package/dist/stages/verify.d.ts +0 -2
- package/dist/stages/verify.js +0 -94
- package/dist/types.d.ts +0 -61
- package/dist/types.js +0 -1
- package/dist/validators.d.ts +0 -8
- package/dist/validators.js +0 -42
- package/dist/verify-runner.d.ts +0 -11
- package/dist/verify-runner.js +0 -110
package/kody.config.schema.json
CHANGED
|
@@ -82,8 +82,8 @@
|
|
|
82
82
|
"properties": {
|
|
83
83
|
"taskDir": {
|
|
84
84
|
"type": "string",
|
|
85
|
-
"description": "Directory for pipeline artifacts and state (
|
|
86
|
-
"default": ".tasks"
|
|
85
|
+
"description": "Directory for pipeline artifacts and state (committed to branch)",
|
|
86
|
+
"default": ".kody/tasks"
|
|
87
87
|
}
|
|
88
88
|
},
|
|
89
89
|
"additionalProperties": false
|
package/package.json
CHANGED
package/prompts/autofix.md
CHANGED
|
@@ -1,21 +1,39 @@
|
|
|
1
1
|
---
|
|
2
2
|
name: autofix
|
|
3
|
-
description:
|
|
3
|
+
description: Investigate root cause then fix verification errors (typecheck, lint, test failures)
|
|
4
4
|
mode: primary
|
|
5
5
|
tools: [read, write, edit, bash, glob, grep]
|
|
6
6
|
---
|
|
7
7
|
|
|
8
8
|
You are an autofix agent. The verification stage failed. Fix the errors below.
|
|
9
9
|
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
10
|
+
IRON LAW: NO FIXES WITHOUT INVESTIGATION FIRST. Do not jump to changing code. Understand the failure first.
|
|
11
|
+
|
|
12
|
+
## Phase 1 — Investigate (do this BEFORE any edits)
|
|
13
|
+
1. Read the full error output — what exactly failed?
|
|
14
|
+
2. Identify the affected files — Read them to understand context
|
|
15
|
+
3. Check recent changes: run `git diff HEAD~1` to see what changed
|
|
16
|
+
4. Classify the failure pattern:
|
|
17
|
+
- **Type error**: mismatched types, missing properties, wrong generics
|
|
18
|
+
- **Test failure**: assertion mismatch, missing mock, changed behavior
|
|
19
|
+
- **Lint error**: style violation, unused import, naming convention
|
|
20
|
+
- **Runtime error**: null reference, missing dependency, config issue
|
|
21
|
+
- **Integration failure**: API contract mismatch, schema drift
|
|
22
|
+
5. Identify root cause — is this a direct error in new code, or a side effect of a change elsewhere?
|
|
23
|
+
|
|
24
|
+
## Phase 2 — Fix (only after root cause is clear)
|
|
25
|
+
1. Try quick wins first: run configured lintFix and formatFix commands via Bash
|
|
26
|
+
2. For type errors: fix the type mismatch at its source, not by adding type assertions
|
|
27
|
+
3. For test failures: fix the root cause (implementation or test), not both — determine which is correct
|
|
28
|
+
4. For lint errors: apply the specific fix the linter suggests
|
|
29
|
+
5. For integration failures: trace the contract back to its definition, fix the mismatch at source
|
|
16
30
|
6. After EACH fix, re-run the failing command to verify it passes
|
|
17
|
-
7.
|
|
31
|
+
7. If a fix introduces new failures, REVERT and try a different approach
|
|
32
|
+
8. Do NOT commit or push — the orchestrator handles git
|
|
18
33
|
|
|
19
|
-
|
|
34
|
+
## Rules
|
|
35
|
+
- Fix ONLY the reported errors. Do NOT make unrelated changes.
|
|
36
|
+
- Minimal diff — use Edit for surgical changes, not Write for rewrites
|
|
37
|
+
- If the failure is pre-existing (not caused by this PR's changes), document it and move on
|
|
20
38
|
|
|
21
39
|
{{TASK_CONTEXT}}
|
package/prompts/review.md
CHANGED
|
@@ -5,9 +5,10 @@ mode: primary
|
|
|
5
5
|
tools: [read, glob, grep, bash]
|
|
6
6
|
---
|
|
7
7
|
|
|
8
|
-
You are a code review agent
|
|
8
|
+
You are a code review agent following the Superpowers Structured Review methodology.
|
|
9
9
|
|
|
10
10
|
Use Bash to run `git diff` to see what changed. Use Read to examine modified files in full context.
|
|
11
|
+
When the diff introduces new enum values, status strings, or type constants — use Grep to trace ALL consumers outside the diff.
|
|
11
12
|
|
|
12
13
|
CRITICAL: You MUST output a structured review in the EXACT format below. Do NOT output conversational text, status updates, or summaries. Your entire output must be the structured review markdown.
|
|
13
14
|
|
|
@@ -21,28 +22,94 @@ Output markdown with this EXACT structure:
|
|
|
21
22
|
## Findings
|
|
22
23
|
|
|
23
24
|
### Critical
|
|
24
|
-
<Security vulnerabilities, data loss risks, crashes, broken authentication>
|
|
25
25
|
<If none: "None.">
|
|
26
26
|
|
|
27
27
|
### Major
|
|
28
|
-
<Logic errors, missing edge cases, broken tests, significant performance issues, missing error handling>
|
|
29
28
|
<If none: "None.">
|
|
30
29
|
|
|
31
30
|
### Minor
|
|
32
|
-
<Style issues, naming improvements, readability, trivial performance, minor refactoring opportunities>
|
|
33
31
|
<If none: "None.">
|
|
34
32
|
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
-
|
|
45
|
-
-
|
|
46
|
-
-
|
|
33
|
+
For each finding use: `file:line` — problem description. Suggested fix.
|
|
34
|
+
|
|
35
|
+
---
|
|
36
|
+
|
|
37
|
+
## Two-Pass Review
|
|
38
|
+
|
|
39
|
+
**Pass 1 — CRITICAL (must fix before merge):**
|
|
40
|
+
|
|
41
|
+
### SQL & Data Safety
|
|
42
|
+
- String interpolation in SQL — use parameterized queries even for `.to_i`/`.to_f` values
|
|
43
|
+
- TOCTOU races: check-then-set patterns that should be atomic `WHERE` + update
|
|
44
|
+
- Bypassing model validations via direct DB writes (e.g., `update_column`, raw queries)
|
|
45
|
+
- N+1 queries: missing eager loading for associations used in loops/views
|
|
46
|
+
|
|
47
|
+
### Race Conditions & Concurrency
|
|
48
|
+
- Read-check-write without uniqueness constraint or duplicate key handling
|
|
49
|
+
- find-or-create without unique DB index — concurrent calls create duplicates
|
|
50
|
+
- Status transitions without atomic `WHERE old_status = ? UPDATE SET new_status`
|
|
51
|
+
- Unsafe HTML rendering (`dangerouslySetInnerHTML`, `v-html`, `.html_safe`) on user-controlled data (XSS)
|
|
52
|
+
|
|
53
|
+
### LLM Output Trust Boundary
|
|
54
|
+
- LLM-generated values (emails, URLs, names) written to DB without format validation
|
|
55
|
+
- Structured tool output accepted without type/shape checks before DB writes
|
|
56
|
+
- LLM-generated URLs fetched without allowlist — SSRF risk
|
|
57
|
+
- LLM output stored in vector DBs without sanitization — stored prompt injection risk
|
|
58
|
+
|
|
59
|
+
### Shell Injection
|
|
60
|
+
- `subprocess.run()` / `os.system()` with `shell=True` AND string interpolation — use argument arrays
|
|
61
|
+
- `eval()` / `exec()` on LLM-generated code without sandboxing
|
|
62
|
+
|
|
63
|
+
### Enum & Value Completeness
|
|
64
|
+
When the diff introduces a new enum value, status string, tier name, or type constant:
|
|
65
|
+
- Trace it through every consumer (READ each file that switches/filters on that value)
|
|
66
|
+
- Check allowlists/filter arrays containing sibling values
|
|
67
|
+
- Check `case`/`if-elsif` chains — does the new value fall through to a wrong default?
|
|
68
|
+
|
|
69
|
+
**Pass 2 — INFORMATIONAL (should review, may auto-fix):**
|
|
70
|
+
|
|
71
|
+
### Conditional Side Effects
|
|
72
|
+
- Code paths that branch but forget a side effect on one branch (e.g., promoted but URL only attached conditionally)
|
|
73
|
+
- Log messages claiming an action happened when it was conditionally skipped
|
|
74
|
+
|
|
75
|
+
### Test Gaps
|
|
76
|
+
- Negative-path tests asserting type/status but not side effects
|
|
77
|
+
- Security enforcement features (blocking, rate limiting, auth) without integration tests
|
|
78
|
+
- Missing `.expects(:something).never` when a path should NOT call an external service
|
|
79
|
+
|
|
80
|
+
### Dead Code & Consistency
|
|
81
|
+
- Variables assigned but never read
|
|
82
|
+
- Comments/docstrings describing old behavior after code changed
|
|
83
|
+
- Version mismatch between PR title and VERSION/CHANGELOG
|
|
84
|
+
|
|
85
|
+
### Crypto & Entropy
|
|
86
|
+
- Truncation instead of hashing — less entropy, easier collisions
|
|
87
|
+
- `rand()` / `Math.random()` for security-sensitive values — use crypto-secure alternatives
|
|
88
|
+
- Non-constant-time comparisons (`==`) on secrets or tokens — timing attack risk
|
|
89
|
+
|
|
90
|
+
### Performance & Bundle Impact
|
|
91
|
+
- Known-heavy dependencies added: moment.js (→ date-fns), full lodash (→ lodash-es), jquery
|
|
92
|
+
- Images without `loading="lazy"` or explicit dimensions (CLS)
|
|
93
|
+
- `useEffect` fetch waterfalls — combine or parallelize
|
|
94
|
+
- Synchronous `<script>` without async/defer
|
|
95
|
+
|
|
96
|
+
### Type Coercion at Boundaries
|
|
97
|
+
- Values crossing language/serialization boundaries where type could change (numeric vs string)
|
|
98
|
+
- Hash/digest inputs without `.toString()` normalization before serialization
|
|
99
|
+
|
|
100
|
+
---
|
|
101
|
+
|
|
102
|
+
## Severity Definitions
|
|
103
|
+
|
|
104
|
+
- **Critical**: Security vulnerability, data loss, application crash, broken authentication, injection risk, race condition. MUST fix before merge.
|
|
105
|
+
- **Major**: Logic error, missing edge case, broken test, significant performance issue, missing input validation, enum completeness gap. SHOULD fix before merge.
|
|
106
|
+
- **Minor**: Style issue, naming improvement, readability, micro-optimization, stale comments. NICE to fix, not blocking.
|
|
107
|
+
|
|
108
|
+
## Suppressions — do NOT flag these:
|
|
109
|
+
- Redundancy that aids readability
|
|
110
|
+
- "Add a comment explaining this threshold" — thresholds change, comments rot
|
|
111
|
+
- Consistency-only changes with no behavioral impact
|
|
112
|
+
- Issues already addressed in the diff you are reviewing — read the FULL diff first
|
|
113
|
+
- devDependencies additions (no production impact)
|
|
47
114
|
|
|
48
115
|
{{TASK_CONTEXT}}
|
package/templates/kody.yml
CHANGED
|
@@ -254,7 +254,7 @@ jobs:
|
|
|
254
254
|
[ -n "$TASK_ID" ] && ARGS="$ARGS --task-id $TASK_ID"
|
|
255
255
|
[ -n "$PR_NUMBER" ] && ARGS="$ARGS --pr-number $PR_NUMBER"
|
|
256
256
|
[ -n "$FROM_STAGE" ] && ARGS="$ARGS --from $FROM_STAGE"
|
|
257
|
-
|
|
257
|
+
# FEEDBACK is passed via env var, not CLI arg (avoids shell escaping issues)
|
|
258
258
|
[ "$DRY_RUN" = "true" ] && ARGS="$ARGS --dry-run"
|
|
259
259
|
kody-engine-lite $CMD $ARGS
|
|
260
260
|
fi
|
|
@@ -263,7 +263,7 @@ jobs:
|
|
|
263
263
|
if: always()
|
|
264
264
|
run: |
|
|
265
265
|
TASK_ID="${{ github.event.inputs.task_id || needs.parse.outputs.task_id }}"
|
|
266
|
-
STATUS_FILE=".tasks/${TASK_ID}/status.json"
|
|
266
|
+
STATUS_FILE=".kody/tasks/${TASK_ID}/status.json"
|
|
267
267
|
if [ -f "$STATUS_FILE" ]; then
|
|
268
268
|
STATE=$(jq -r '.state' "$STATUS_FILE")
|
|
269
269
|
ICON="❌"
|
|
@@ -292,7 +292,7 @@ jobs:
|
|
|
292
292
|
uses: actions/upload-artifact@v4
|
|
293
293
|
with:
|
|
294
294
|
name: kody-tasks-${{ github.event.inputs.task_id || needs.parse.outputs.task_id }}
|
|
295
|
-
path: .tasks/
|
|
295
|
+
path: .kody/tasks/
|
|
296
296
|
retention-days: 7
|
|
297
297
|
|
|
298
298
|
# ─── Error Notifications ─────────────────────────────────────────────────────
|
|
@@ -355,11 +355,11 @@ jobs:
|
|
|
355
355
|
run: kody-engine-lite --help
|
|
356
356
|
- name: Dry run
|
|
357
357
|
run: |
|
|
358
|
-
mkdir -p .tasks/smoke-test
|
|
359
|
-
echo "Smoke test task" > .tasks/smoke-test/task.md
|
|
358
|
+
mkdir -p .kody/tasks/smoke-test
|
|
359
|
+
echo "Smoke test task" > .kody/tasks/smoke-test/task.md
|
|
360
360
|
kody-engine-lite run --task-id smoke-test --dry-run || true
|
|
361
|
-
if [ -f ".tasks/smoke-test/status.json" ]; then
|
|
361
|
+
if [ -f ".kody/tasks/smoke-test/status.json" ]; then
|
|
362
362
|
echo "✓ status.json created"
|
|
363
|
-
cat .tasks/smoke-test/status.json
|
|
363
|
+
cat .kody/tasks/smoke-test/status.json
|
|
364
364
|
fi
|
|
365
|
-
rm -rf .tasks/smoke-test
|
|
365
|
+
rm -rf .kody/tasks/smoke-test
|
package/dist/agent-runner.d.ts
DELETED
package/dist/agent-runner.js
DELETED
|
@@ -1,122 +0,0 @@
|
|
|
1
|
-
import { spawn, execFileSync } from "child_process";
|
|
2
|
-
const SIGKILL_GRACE_MS = 5000;
|
|
3
|
-
const STDERR_TAIL_CHARS = 500;
|
|
4
|
-
function writeStdin(child, prompt) {
|
|
5
|
-
return new Promise((resolve, reject) => {
|
|
6
|
-
if (!child.stdin) {
|
|
7
|
-
resolve();
|
|
8
|
-
return;
|
|
9
|
-
}
|
|
10
|
-
child.stdin.write(prompt, (err) => {
|
|
11
|
-
if (err)
|
|
12
|
-
reject(err);
|
|
13
|
-
else {
|
|
14
|
-
child.stdin.end();
|
|
15
|
-
resolve();
|
|
16
|
-
}
|
|
17
|
-
});
|
|
18
|
-
});
|
|
19
|
-
}
|
|
20
|
-
function waitForProcess(child, timeout) {
|
|
21
|
-
return new Promise((resolve) => {
|
|
22
|
-
const stdoutChunks = [];
|
|
23
|
-
const stderrChunks = [];
|
|
24
|
-
child.stdout?.on("data", (chunk) => stdoutChunks.push(chunk));
|
|
25
|
-
child.stderr?.on("data", (chunk) => stderrChunks.push(chunk));
|
|
26
|
-
const timer = setTimeout(() => {
|
|
27
|
-
child.kill("SIGTERM");
|
|
28
|
-
setTimeout(() => {
|
|
29
|
-
if (!child.killed)
|
|
30
|
-
child.kill("SIGKILL");
|
|
31
|
-
}, SIGKILL_GRACE_MS);
|
|
32
|
-
}, timeout);
|
|
33
|
-
child.on("exit", (code) => {
|
|
34
|
-
clearTimeout(timer);
|
|
35
|
-
resolve({
|
|
36
|
-
code,
|
|
37
|
-
stdout: Buffer.concat(stdoutChunks).toString(),
|
|
38
|
-
stderr: Buffer.concat(stderrChunks).toString(),
|
|
39
|
-
});
|
|
40
|
-
});
|
|
41
|
-
child.on("error", (err) => {
|
|
42
|
-
clearTimeout(timer);
|
|
43
|
-
resolve({ code: -1, stdout: "", stderr: err.message });
|
|
44
|
-
});
|
|
45
|
-
});
|
|
46
|
-
}
|
|
47
|
-
async function runSubprocess(command, args, prompt, timeout, options) {
|
|
48
|
-
const child = spawn(command, args, {
|
|
49
|
-
cwd: options?.cwd ?? process.cwd(),
|
|
50
|
-
env: {
|
|
51
|
-
...process.env,
|
|
52
|
-
SKIP_BUILD: "1",
|
|
53
|
-
SKIP_HOOKS: "1",
|
|
54
|
-
...options?.env,
|
|
55
|
-
},
|
|
56
|
-
stdio: ["pipe", "pipe", "pipe"],
|
|
57
|
-
});
|
|
58
|
-
try {
|
|
59
|
-
await writeStdin(child, prompt);
|
|
60
|
-
}
|
|
61
|
-
catch (err) {
|
|
62
|
-
return {
|
|
63
|
-
outcome: "failed",
|
|
64
|
-
error: `Failed to send prompt: ${err instanceof Error ? err.message : String(err)}`,
|
|
65
|
-
};
|
|
66
|
-
}
|
|
67
|
-
const { code, stdout, stderr } = await waitForProcess(child, timeout);
|
|
68
|
-
if (code === 0) {
|
|
69
|
-
return { outcome: "completed", output: stdout };
|
|
70
|
-
}
|
|
71
|
-
return {
|
|
72
|
-
outcome: code === null ? "timed_out" : "failed",
|
|
73
|
-
error: `Exit code ${code}\n${stderr.slice(-STDERR_TAIL_CHARS)}`,
|
|
74
|
-
};
|
|
75
|
-
}
|
|
76
|
-
function checkCommand(command, args) {
|
|
77
|
-
try {
|
|
78
|
-
execFileSync(command, args, { timeout: 10_000, stdio: "pipe" });
|
|
79
|
-
return true;
|
|
80
|
-
}
|
|
81
|
-
catch {
|
|
82
|
-
return false;
|
|
83
|
-
}
|
|
84
|
-
}
|
|
85
|
-
// ─── Claude Code Runner ──────────────────────────────────────────────────────
|
|
86
|
-
export function createClaudeCodeRunner() {
|
|
87
|
-
return {
|
|
88
|
-
async run(_stageName, prompt, model, timeout, _taskDir, options) {
|
|
89
|
-
return runSubprocess("claude", [
|
|
90
|
-
"--print",
|
|
91
|
-
"--model", model,
|
|
92
|
-
"--dangerously-skip-permissions",
|
|
93
|
-
"--allowedTools", "Bash,Edit,Read,Write,Glob,Grep",
|
|
94
|
-
], prompt, timeout, options);
|
|
95
|
-
},
|
|
96
|
-
async healthCheck() {
|
|
97
|
-
return checkCommand("claude", ["--version"]);
|
|
98
|
-
},
|
|
99
|
-
};
|
|
100
|
-
}
|
|
101
|
-
// ─── Runner Factory ──────────────────────────────────────────────────────────
|
|
102
|
-
const RUNNER_FACTORIES = {
|
|
103
|
-
"claude-code": createClaudeCodeRunner,
|
|
104
|
-
};
|
|
105
|
-
export function createRunners(config) {
|
|
106
|
-
// New multi-runner config
|
|
107
|
-
if (config.agent.runners && Object.keys(config.agent.runners).length > 0) {
|
|
108
|
-
const runners = {};
|
|
109
|
-
for (const [name, runnerConfig] of Object.entries(config.agent.runners)) {
|
|
110
|
-
const factory = RUNNER_FACTORIES[runnerConfig.type];
|
|
111
|
-
if (factory) {
|
|
112
|
-
runners[name] = factory();
|
|
113
|
-
}
|
|
114
|
-
}
|
|
115
|
-
return runners;
|
|
116
|
-
}
|
|
117
|
-
// Legacy single-runner fallback
|
|
118
|
-
const runnerType = config.agent.runner ?? "claude-code";
|
|
119
|
-
const factory = RUNNER_FACTORIES[runnerType];
|
|
120
|
-
const defaultName = config.agent.defaultRunner ?? "claude";
|
|
121
|
-
return { [defaultName]: factory ? factory() : createClaudeCodeRunner() };
|
|
122
|
-
}
|
package/dist/ci/parse-inputs.js
DELETED
|
@@ -1,76 +0,0 @@
|
|
|
1
|
-
/**
|
|
2
|
-
* Parses @kody / /kody comment body into structured inputs.
|
|
3
|
-
* Run by the parse job in GitHub Actions.
|
|
4
|
-
* Reads from env, writes to $GITHUB_OUTPUT.
|
|
5
|
-
*/
|
|
6
|
-
import * as fs from "fs";
|
|
7
|
-
const outputFile = process.env.GITHUB_OUTPUT;
|
|
8
|
-
const triggerType = process.env.TRIGGER_TYPE ?? "dispatch";
|
|
9
|
-
function output(key, value) {
|
|
10
|
-
if (outputFile) {
|
|
11
|
-
fs.appendFileSync(outputFile, `${key}=${value}\n`);
|
|
12
|
-
}
|
|
13
|
-
console.log(`${key}=${value}`);
|
|
14
|
-
}
|
|
15
|
-
// For workflow_dispatch, pass through inputs
|
|
16
|
-
if (triggerType === "dispatch") {
|
|
17
|
-
output("task_id", process.env.INPUT_TASK_ID ?? "");
|
|
18
|
-
output("mode", process.env.INPUT_MODE ?? "full");
|
|
19
|
-
output("from_stage", process.env.INPUT_FROM_STAGE ?? "");
|
|
20
|
-
output("issue_number", process.env.INPUT_ISSUE_NUMBER ?? "");
|
|
21
|
-
output("feedback", process.env.INPUT_FEEDBACK ?? "");
|
|
22
|
-
output("valid", process.env.INPUT_TASK_ID ? "true" : "false");
|
|
23
|
-
output("trigger_type", "dispatch");
|
|
24
|
-
process.exit(0);
|
|
25
|
-
}
|
|
26
|
-
// For issue_comment, parse the comment body
|
|
27
|
-
const commentBody = process.env.COMMENT_BODY ?? "";
|
|
28
|
-
const issueNumber = process.env.ISSUE_NUMBER ?? "";
|
|
29
|
-
// Match: @kody [mode] [task-id] [--from stage] [--feedback "text"]
|
|
30
|
-
const kodyMatch = commentBody.match(/(?:@kody|\/kody)\s*(.*)/i);
|
|
31
|
-
if (!kodyMatch) {
|
|
32
|
-
output("valid", "false");
|
|
33
|
-
output("trigger_type", "comment");
|
|
34
|
-
process.exit(0);
|
|
35
|
-
}
|
|
36
|
-
const parts = kodyMatch[1].trim().split(/\s+/);
|
|
37
|
-
const validModes = ["full", "rerun", "status"];
|
|
38
|
-
let mode = "full";
|
|
39
|
-
let taskId = "";
|
|
40
|
-
let fromStage = "";
|
|
41
|
-
let feedback = "";
|
|
42
|
-
let i = 0;
|
|
43
|
-
// First arg: mode or task-id
|
|
44
|
-
if (parts[i] && validModes.includes(parts[i])) {
|
|
45
|
-
mode = parts[i];
|
|
46
|
-
i++;
|
|
47
|
-
}
|
|
48
|
-
// Second arg: task-id
|
|
49
|
-
if (parts[i] && !parts[i].startsWith("--")) {
|
|
50
|
-
taskId = parts[i];
|
|
51
|
-
i++;
|
|
52
|
-
}
|
|
53
|
-
// Named args
|
|
54
|
-
while (i < parts.length) {
|
|
55
|
-
if (parts[i] === "--from" && parts[i + 1]) {
|
|
56
|
-
fromStage = parts[i + 1];
|
|
57
|
-
i += 2;
|
|
58
|
-
}
|
|
59
|
-
else if (parts[i] === "--feedback" && parts[i + 1]) {
|
|
60
|
-
// Collect quoted feedback
|
|
61
|
-
const rest = parts.slice(i + 1).join(" ");
|
|
62
|
-
const quoted = rest.match(/^"([^"]*)"/);
|
|
63
|
-
feedback = quoted ? quoted[1] : parts[i + 1];
|
|
64
|
-
break;
|
|
65
|
-
}
|
|
66
|
-
else {
|
|
67
|
-
i++;
|
|
68
|
-
}
|
|
69
|
-
}
|
|
70
|
-
output("task_id", taskId);
|
|
71
|
-
output("mode", mode);
|
|
72
|
-
output("from_stage", fromStage);
|
|
73
|
-
output("issue_number", issueNumber);
|
|
74
|
-
output("feedback", feedback);
|
|
75
|
-
output("valid", taskId ? "true" : "false");
|
|
76
|
-
output("trigger_type", "comment");
|
package/dist/ci/parse-safety.js
DELETED
|
@@ -1,22 +0,0 @@
|
|
|
1
|
-
/**
|
|
2
|
-
* Validates that a comment trigger is safe to execute.
|
|
3
|
-
* Run by the parse job in GitHub Actions.
|
|
4
|
-
* Reads from env, writes to $GITHUB_OUTPUT.
|
|
5
|
-
*/
|
|
6
|
-
import * as fs from "fs";
|
|
7
|
-
const ALLOWED_ASSOCIATIONS = ["COLLABORATOR", "MEMBER", "OWNER"];
|
|
8
|
-
const association = process.env.COMMENT_AUTHOR_ASSOCIATION ?? "";
|
|
9
|
-
const outputFile = process.env.GITHUB_OUTPUT;
|
|
10
|
-
function output(key, value) {
|
|
11
|
-
if (outputFile) {
|
|
12
|
-
fs.appendFileSync(outputFile, `${key}=${value}\n`);
|
|
13
|
-
}
|
|
14
|
-
console.log(`${key}=${value}`);
|
|
15
|
-
}
|
|
16
|
-
if (!ALLOWED_ASSOCIATIONS.includes(association)) {
|
|
17
|
-
output("valid", "false");
|
|
18
|
-
output("reason", `Author association '${association}' not in allowlist: ${ALLOWED_ASSOCIATIONS.join(", ")}`);
|
|
19
|
-
process.exit(0);
|
|
20
|
-
}
|
|
21
|
-
output("valid", "true");
|
|
22
|
-
output("reason", "");
|
package/dist/cli/args.d.ts
DELETED
|
@@ -1,13 +0,0 @@
|
|
|
1
|
-
export interface CliInput {
|
|
2
|
-
command: "run" | "rerun" | "fix" | "status";
|
|
3
|
-
taskId?: string;
|
|
4
|
-
task?: string;
|
|
5
|
-
fromStage?: string;
|
|
6
|
-
dryRun?: boolean;
|
|
7
|
-
cwd?: string;
|
|
8
|
-
issueNumber?: number;
|
|
9
|
-
feedback?: string;
|
|
10
|
-
local?: boolean;
|
|
11
|
-
complexity?: "low" | "medium" | "high";
|
|
12
|
-
}
|
|
13
|
-
export declare function parseArgs(): CliInput;
|
package/dist/cli/args.js
DELETED
|
@@ -1,42 +0,0 @@
|
|
|
1
|
-
const isCI = !!process.env.GITHUB_ACTIONS;
|
|
2
|
-
function getArg(args, flag) {
|
|
3
|
-
const idx = args.indexOf(flag);
|
|
4
|
-
if (idx !== -1 && args[idx + 1] && !args[idx + 1].startsWith("--")) {
|
|
5
|
-
return args[idx + 1];
|
|
6
|
-
}
|
|
7
|
-
return undefined;
|
|
8
|
-
}
|
|
9
|
-
function hasFlag(args, flag) {
|
|
10
|
-
return args.includes(flag);
|
|
11
|
-
}
|
|
12
|
-
export function parseArgs() {
|
|
13
|
-
const args = process.argv.slice(2);
|
|
14
|
-
if (hasFlag(args, "--help") || hasFlag(args, "-h") || args.length === 0) {
|
|
15
|
-
console.log(`Usage:
|
|
16
|
-
kody run --task-id <id> [--task "<desc>"] [--cwd <path>] [--issue-number <n>] [--complexity low|medium|high] [--feedback "<text>"] [--local] [--dry-run]
|
|
17
|
-
kody rerun --task-id <id> --from <stage> [--cwd <path>] [--issue-number <n>]
|
|
18
|
-
kody fix --task-id <id> [--cwd <path>] [--issue-number <n>] [--feedback "<text>"]
|
|
19
|
-
kody status --task-id <id> [--cwd <path>]
|
|
20
|
-
kody --help`);
|
|
21
|
-
process.exit(0);
|
|
22
|
-
}
|
|
23
|
-
const command = args[0];
|
|
24
|
-
if (!["run", "rerun", "fix", "status"].includes(command)) {
|
|
25
|
-
console.error(`Unknown command: ${command}`);
|
|
26
|
-
process.exit(1);
|
|
27
|
-
}
|
|
28
|
-
const issueStr = getArg(args, "--issue-number") ?? process.env.ISSUE_NUMBER;
|
|
29
|
-
const localFlag = hasFlag(args, "--local");
|
|
30
|
-
return {
|
|
31
|
-
command,
|
|
32
|
-
taskId: getArg(args, "--task-id") ?? process.env.TASK_ID,
|
|
33
|
-
task: getArg(args, "--task"),
|
|
34
|
-
fromStage: getArg(args, "--from") ?? process.env.FROM_STAGE,
|
|
35
|
-
dryRun: hasFlag(args, "--dry-run") || process.env.DRY_RUN === "true",
|
|
36
|
-
cwd: getArg(args, "--cwd"),
|
|
37
|
-
issueNumber: issueStr ? parseInt(issueStr, 10) : undefined,
|
|
38
|
-
feedback: getArg(args, "--feedback") ?? process.env.FEEDBACK,
|
|
39
|
-
local: localFlag || (!isCI && !hasFlag(args, "--no-local")),
|
|
40
|
-
complexity: (getArg(args, "--complexity") ?? process.env.COMPLEXITY),
|
|
41
|
-
};
|
|
42
|
-
}
|
package/dist/cli/litellm.d.ts
DELETED
package/dist/cli/litellm.js
DELETED
|
@@ -1,85 +0,0 @@
|
|
|
1
|
-
import * as fs from "fs";
|
|
2
|
-
import * as path from "path";
|
|
3
|
-
import { execFileSync } from "child_process";
|
|
4
|
-
import { logger } from "../logger.js";
|
|
5
|
-
export async function checkLitellmHealth(url) {
|
|
6
|
-
try {
|
|
7
|
-
const response = await fetch(`${url}/health`, { signal: AbortSignal.timeout(3000) });
|
|
8
|
-
return response.ok;
|
|
9
|
-
}
|
|
10
|
-
catch {
|
|
11
|
-
return false;
|
|
12
|
-
}
|
|
13
|
-
}
|
|
14
|
-
export async function tryStartLitellm(url, projectDir) {
|
|
15
|
-
const configPath = path.join(projectDir, "litellm-config.yaml");
|
|
16
|
-
if (!fs.existsSync(configPath)) {
|
|
17
|
-
logger.warn("litellm-config.yaml not found — cannot start proxy");
|
|
18
|
-
return null;
|
|
19
|
-
}
|
|
20
|
-
// Extract port from URL
|
|
21
|
-
const portMatch = url.match(/:(\d+)/);
|
|
22
|
-
const port = portMatch ? portMatch[1] : "4000";
|
|
23
|
-
// Check if litellm is installed
|
|
24
|
-
try {
|
|
25
|
-
execFileSync("litellm", ["--version"], { timeout: 5000, stdio: "pipe" });
|
|
26
|
-
}
|
|
27
|
-
catch {
|
|
28
|
-
try {
|
|
29
|
-
execFileSync("python3", ["-m", "litellm", "--version"], { timeout: 5000, stdio: "pipe" });
|
|
30
|
-
}
|
|
31
|
-
catch {
|
|
32
|
-
logger.warn("litellm not installed (pip install 'litellm[proxy]')");
|
|
33
|
-
return null;
|
|
34
|
-
}
|
|
35
|
-
}
|
|
36
|
-
logger.info(`Starting LiteLLM proxy on port ${port}...`);
|
|
37
|
-
// Determine command
|
|
38
|
-
let cmd;
|
|
39
|
-
let args;
|
|
40
|
-
try {
|
|
41
|
-
execFileSync("litellm", ["--version"], { timeout: 5000, stdio: "pipe" });
|
|
42
|
-
cmd = "litellm";
|
|
43
|
-
args = ["--config", configPath, "--port", port];
|
|
44
|
-
}
|
|
45
|
-
catch {
|
|
46
|
-
cmd = "python3";
|
|
47
|
-
args = ["-m", "litellm", "--config", configPath, "--port", port];
|
|
48
|
-
}
|
|
49
|
-
// Load API key env vars from project .env (only *_API_KEY patterns)
|
|
50
|
-
const dotenvPath = path.join(projectDir, ".env");
|
|
51
|
-
const dotenvVars = {};
|
|
52
|
-
if (fs.existsSync(dotenvPath)) {
|
|
53
|
-
for (const line of fs.readFileSync(dotenvPath, "utf-8").split("\n")) {
|
|
54
|
-
const match = line.match(/^([A-Z_][A-Z0-9_]*_API_KEY)=(.*)$/);
|
|
55
|
-
if (match)
|
|
56
|
-
dotenvVars[match[1]] = match[2];
|
|
57
|
-
}
|
|
58
|
-
if (Object.keys(dotenvVars).length > 0) {
|
|
59
|
-
logger.info(` Loaded API keys: ${Object.keys(dotenvVars).join(", ")}`);
|
|
60
|
-
}
|
|
61
|
-
}
|
|
62
|
-
const { spawn } = await import("child_process");
|
|
63
|
-
const child = spawn(cmd, args, {
|
|
64
|
-
stdio: ["ignore", "pipe", "pipe"],
|
|
65
|
-
detached: true,
|
|
66
|
-
env: { ...process.env, ...dotenvVars },
|
|
67
|
-
});
|
|
68
|
-
// Capture stderr for debugging
|
|
69
|
-
let proxyStderr = "";
|
|
70
|
-
child.stderr?.on("data", (chunk) => { proxyStderr += chunk.toString(); });
|
|
71
|
-
// Wait for health
|
|
72
|
-
for (let i = 0; i < 30; i++) {
|
|
73
|
-
await new Promise((r) => setTimeout(r, 2000));
|
|
74
|
-
if (await checkLitellmHealth(url)) {
|
|
75
|
-
logger.info(`LiteLLM proxy ready at ${url}`);
|
|
76
|
-
return child;
|
|
77
|
-
}
|
|
78
|
-
}
|
|
79
|
-
if (proxyStderr) {
|
|
80
|
-
logger.warn(`LiteLLM stderr: ${proxyStderr.slice(-1000)}`);
|
|
81
|
-
}
|
|
82
|
-
logger.warn("LiteLLM proxy failed to start within 60s");
|
|
83
|
-
child.kill();
|
|
84
|
-
return null;
|
|
85
|
-
}
|