gsd-trae 1.0.0 → 1.0.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +40 -0
- package/README.md +7 -76
- package/assets/screenshot.png +0 -0
- package/package.json +12 -3
- package/.claude/settings.local.json +0 -8
- package/.gitmodules +0 -6
- package/.trae/project_rules.md +0 -56
- package/.trae/rules/project_rules.md +0 -56
- package/.vscode/code-counter/code-counter.db +0 -0
- package/.vscode/settings.json +0 -5
- package/refs/gsd/.github/CODEOWNERS +0 -2
- package/refs/gsd/.github/FUNDING.yml +0 -1
- package/refs/gsd/.github/ISSUE_TEMPLATE/bug_report.yml +0 -59
- package/refs/gsd/.github/ISSUE_TEMPLATE/feature_request.yml +0 -37
- package/refs/gsd/.github/pull_request_template.md +0 -24
- package/refs/gsd/.github/workflows/auto-label-issues.yml +0 -21
- package/refs/gsd/CHANGELOG.md +0 -1520
- package/refs/gsd/LICENSE +0 -21
- package/refs/gsd/README.md +0 -704
- package/refs/gsd/SECURITY.md +0 -33
- package/refs/gsd/agents/gsd-codebase-mapper.md +0 -764
- package/refs/gsd/agents/gsd-debugger.md +0 -1246
- package/refs/gsd/agents/gsd-executor.md +0 -469
- package/refs/gsd/agents/gsd-integration-checker.md +0 -443
- package/refs/gsd/agents/gsd-phase-researcher.md +0 -546
- package/refs/gsd/agents/gsd-plan-checker.md +0 -690
- package/refs/gsd/agents/gsd-planner.md +0 -1275
- package/refs/gsd/agents/gsd-project-researcher.md +0 -621
- package/refs/gsd/agents/gsd-research-synthesizer.md +0 -239
- package/refs/gsd/agents/gsd-roadmapper.md +0 -642
- package/refs/gsd/agents/gsd-verifier.md +0 -573
- package/refs/gsd/assets/gsd-logo-2000-transparent.png +0 -0
- package/refs/gsd/assets/gsd-logo-2000-transparent.svg +0 -17
- package/refs/gsd/assets/gsd-logo-2000.png +0 -0
- package/refs/gsd/assets/gsd-logo-2000.svg +0 -21
- package/refs/gsd/assets/terminal.svg +0 -68
- package/refs/gsd/bin/install.js +0 -2090
- package/refs/gsd/commands/gsd/add-phase.md +0 -43
- package/refs/gsd/commands/gsd/add-tests.md +0 -41
- package/refs/gsd/commands/gsd/add-todo.md +0 -47
- package/refs/gsd/commands/gsd/audit-milestone.md +0 -36
- package/refs/gsd/commands/gsd/check-todos.md +0 -45
- package/refs/gsd/commands/gsd/cleanup.md +0 -18
- package/refs/gsd/commands/gsd/complete-milestone.md +0 -136
- package/refs/gsd/commands/gsd/debug.md +0 -167
- package/refs/gsd/commands/gsd/discuss-phase.md +0 -83
- package/refs/gsd/commands/gsd/execute-phase.md +0 -41
- package/refs/gsd/commands/gsd/health.md +0 -22
- package/refs/gsd/commands/gsd/help.md +0 -22
- package/refs/gsd/commands/gsd/insert-phase.md +0 -32
- package/refs/gsd/commands/gsd/join-discord.md +0 -18
- package/refs/gsd/commands/gsd/list-phase-assumptions.md +0 -46
- package/refs/gsd/commands/gsd/map-codebase.md +0 -71
- package/refs/gsd/commands/gsd/new-milestone.md +0 -44
- package/refs/gsd/commands/gsd/new-project.md +0 -42
- package/refs/gsd/commands/gsd/new-project.md.bak +0 -1041
- package/refs/gsd/commands/gsd/pause-work.md +0 -38
- package/refs/gsd/commands/gsd/plan-milestone-gaps.md +0 -34
- package/refs/gsd/commands/gsd/plan-phase.md +0 -45
- package/refs/gsd/commands/gsd/progress.md +0 -24
- package/refs/gsd/commands/gsd/quick.md +0 -41
- package/refs/gsd/commands/gsd/reapply-patches.md +0 -110
- package/refs/gsd/commands/gsd/remove-phase.md +0 -31
- package/refs/gsd/commands/gsd/research-phase.md +0 -189
- package/refs/gsd/commands/gsd/resume-work.md +0 -40
- package/refs/gsd/commands/gsd/set-profile.md +0 -34
- package/refs/gsd/commands/gsd/settings.md +0 -36
- package/refs/gsd/commands/gsd/update.md +0 -37
- package/refs/gsd/commands/gsd/verify-work.md +0 -38
- package/refs/gsd/docs/USER-GUIDE.md +0 -471
- package/refs/gsd/docs/context-monitor.md +0 -96
- package/refs/gsd/get-shit-done/bin/gsd-tools.cjs +0 -585
- package/refs/gsd/get-shit-done/bin/lib/commands.cjs +0 -553
- package/refs/gsd/get-shit-done/bin/lib/config.cjs +0 -162
- package/refs/gsd/get-shit-done/bin/lib/core.cjs +0 -411
- package/refs/gsd/get-shit-done/bin/lib/frontmatter.cjs +0 -299
- package/refs/gsd/get-shit-done/bin/lib/init.cjs +0 -710
- package/refs/gsd/get-shit-done/bin/lib/milestone.cjs +0 -215
- package/refs/gsd/get-shit-done/bin/lib/phase.cjs +0 -870
- package/refs/gsd/get-shit-done/bin/lib/roadmap.cjs +0 -298
- package/refs/gsd/get-shit-done/bin/lib/state.cjs +0 -521
- package/refs/gsd/get-shit-done/bin/lib/template.cjs +0 -222
- package/refs/gsd/get-shit-done/bin/lib/verify.cjs +0 -772
- package/refs/gsd/get-shit-done/references/checkpoints.md +0 -776
- package/refs/gsd/get-shit-done/references/continuation-format.md +0 -249
- package/refs/gsd/get-shit-done/references/decimal-phase-calculation.md +0 -65
- package/refs/gsd/get-shit-done/references/git-integration.md +0 -248
- package/refs/gsd/get-shit-done/references/git-planning-commit.md +0 -38
- package/refs/gsd/get-shit-done/references/model-profile-resolution.md +0 -34
- package/refs/gsd/get-shit-done/references/model-profiles.md +0 -92
- package/refs/gsd/get-shit-done/references/phase-argument-parsing.md +0 -61
- package/refs/gsd/get-shit-done/references/planning-config.md +0 -196
- package/refs/gsd/get-shit-done/references/questioning.md +0 -145
- package/refs/gsd/get-shit-done/references/tdd.md +0 -263
- package/refs/gsd/get-shit-done/references/ui-brand.md +0 -160
- package/refs/gsd/get-shit-done/references/verification-patterns.md +0 -612
- package/refs/gsd/get-shit-done/templates/DEBUG.md +0 -164
- package/refs/gsd/get-shit-done/templates/UAT.md +0 -247
- package/refs/gsd/get-shit-done/templates/VALIDATION.md +0 -76
- package/refs/gsd/get-shit-done/templates/codebase/architecture.md +0 -255
- package/refs/gsd/get-shit-done/templates/codebase/concerns.md +0 -310
- package/refs/gsd/get-shit-done/templates/codebase/conventions.md +0 -307
- package/refs/gsd/get-shit-done/templates/codebase/integrations.md +0 -280
- package/refs/gsd/get-shit-done/templates/codebase/stack.md +0 -186
- package/refs/gsd/get-shit-done/templates/codebase/structure.md +0 -285
- package/refs/gsd/get-shit-done/templates/codebase/testing.md +0 -480
- package/refs/gsd/get-shit-done/templates/config.json +0 -37
- package/refs/gsd/get-shit-done/templates/context.md +0 -283
- package/refs/gsd/get-shit-done/templates/continue-here.md +0 -78
- package/refs/gsd/get-shit-done/templates/debug-subagent-prompt.md +0 -91
- package/refs/gsd/get-shit-done/templates/discovery.md +0 -146
- package/refs/gsd/get-shit-done/templates/milestone-archive.md +0 -123
- package/refs/gsd/get-shit-done/templates/milestone.md +0 -115
- package/refs/gsd/get-shit-done/templates/phase-prompt.md +0 -569
- package/refs/gsd/get-shit-done/templates/planner-subagent-prompt.md +0 -117
- package/refs/gsd/get-shit-done/templates/project.md +0 -184
- package/refs/gsd/get-shit-done/templates/requirements.md +0 -231
- package/refs/gsd/get-shit-done/templates/research-project/ARCHITECTURE.md +0 -204
- package/refs/gsd/get-shit-done/templates/research-project/FEATURES.md +0 -147
- package/refs/gsd/get-shit-done/templates/research-project/PITFALLS.md +0 -200
- package/refs/gsd/get-shit-done/templates/research-project/STACK.md +0 -120
- package/refs/gsd/get-shit-done/templates/research-project/SUMMARY.md +0 -170
- package/refs/gsd/get-shit-done/templates/research.md +0 -552
- package/refs/gsd/get-shit-done/templates/retrospective.md +0 -54
- package/refs/gsd/get-shit-done/templates/roadmap.md +0 -202
- package/refs/gsd/get-shit-done/templates/state.md +0 -176
- package/refs/gsd/get-shit-done/templates/summary-complex.md +0 -59
- package/refs/gsd/get-shit-done/templates/summary-minimal.md +0 -41
- package/refs/gsd/get-shit-done/templates/summary-standard.md +0 -48
- package/refs/gsd/get-shit-done/templates/summary.md +0 -248
- package/refs/gsd/get-shit-done/templates/user-setup.md +0 -311
- package/refs/gsd/get-shit-done/templates/verification-report.md +0 -322
- package/refs/gsd/get-shit-done/workflows/add-phase.md +0 -111
- package/refs/gsd/get-shit-done/workflows/add-tests.md +0 -350
- package/refs/gsd/get-shit-done/workflows/add-todo.md +0 -157
- package/refs/gsd/get-shit-done/workflows/audit-milestone.md +0 -297
- package/refs/gsd/get-shit-done/workflows/check-todos.md +0 -176
- package/refs/gsd/get-shit-done/workflows/cleanup.md +0 -152
- package/refs/gsd/get-shit-done/workflows/complete-milestone.md +0 -763
- package/refs/gsd/get-shit-done/workflows/diagnose-issues.md +0 -219
- package/refs/gsd/get-shit-done/workflows/discovery-phase.md +0 -289
- package/refs/gsd/get-shit-done/workflows/discuss-phase.md +0 -542
- package/refs/gsd/get-shit-done/workflows/execute-phase.md +0 -449
- package/refs/gsd/get-shit-done/workflows/execute-plan.md +0 -448
- package/refs/gsd/get-shit-done/workflows/health.md +0 -156
- package/refs/gsd/get-shit-done/workflows/help.md +0 -489
- package/refs/gsd/get-shit-done/workflows/insert-phase.md +0 -129
- package/refs/gsd/get-shit-done/workflows/list-phase-assumptions.md +0 -178
- package/refs/gsd/get-shit-done/workflows/map-codebase.md +0 -315
- package/refs/gsd/get-shit-done/workflows/new-milestone.md +0 -382
- package/refs/gsd/get-shit-done/workflows/new-project.md +0 -1116
- package/refs/gsd/get-shit-done/workflows/pause-work.md +0 -122
- package/refs/gsd/get-shit-done/workflows/plan-milestone-gaps.md +0 -274
- package/refs/gsd/get-shit-done/workflows/plan-phase.md +0 -569
- package/refs/gsd/get-shit-done/workflows/progress.md +0 -381
- package/refs/gsd/get-shit-done/workflows/quick.md +0 -453
- package/refs/gsd/get-shit-done/workflows/remove-phase.md +0 -154
- package/refs/gsd/get-shit-done/workflows/research-phase.md +0 -73
- package/refs/gsd/get-shit-done/workflows/resume-project.md +0 -306
- package/refs/gsd/get-shit-done/workflows/set-profile.md +0 -80
- package/refs/gsd/get-shit-done/workflows/settings.md +0 -213
- package/refs/gsd/get-shit-done/workflows/transition.md +0 -544
- package/refs/gsd/get-shit-done/workflows/update.md +0 -219
- package/refs/gsd/get-shit-done/workflows/verify-phase.md +0 -242
- package/refs/gsd/get-shit-done/workflows/verify-work.md +0 -569
- package/refs/gsd/hooks/gsd-check-update.js +0 -62
- package/refs/gsd/hooks/gsd-context-monitor.js +0 -122
- package/refs/gsd/hooks/gsd-statusline.js +0 -108
- package/refs/gsd/package.json +0 -50
- package/refs/gsd/scripts/build-hooks.js +0 -43
- package/refs/gsd/tests/commands.test.cjs +0 -661
- package/refs/gsd/tests/helpers.cjs +0 -40
- package/refs/gsd/tests/init.test.cjs +0 -205
- package/refs/gsd/tests/milestone.test.cjs +0 -98
- package/refs/gsd/tests/phase.test.cjs +0 -1241
- package/refs/gsd/tests/roadmap.test.cjs +0 -265
- package/refs/gsd/tests/state.test.cjs +0 -302
- package/refs/gsd/tests/verify.test.cjs +0 -80
- package/refs/vbenchmark/.agent/agents/codebase-explorer.md +0 -224
- package/refs/vbenchmark/.agent/agents/debugger.md +0 -180
- package/refs/vbenchmark/.agent/agents/documenter.md +0 -166
- package/refs/vbenchmark/.agent/agents/implementer.md +0 -70
- package/refs/vbenchmark/.agent/agents/orchestrator.md +0 -212
- package/refs/vbenchmark/.agent/agents/researcher.md +0 -80
- package/refs/vbenchmark/.agent/agents/reviewer.md +0 -184
- package/refs/vbenchmark/.agent/agents/tester.md +0 -170
- package/refs/vbenchmark/.agent/commands/commit.md +0 -29
- package/refs/vbenchmark/.agent/commands/debug.md +0 -59
- package/refs/vbenchmark/.agent/commands/document.md +0 -52
- package/refs/vbenchmark/.agent/commands/gather-context.md +0 -58
- package/refs/vbenchmark/.agent/commands/init.md +0 -56
- package/refs/vbenchmark/.agent/commands/preset-help.md +0 -50
- package/refs/vbenchmark/.agent/commands/refactor.md +0 -71
- package/refs/vbenchmark/.agent/commands/research.md +0 -37
- package/refs/vbenchmark/.agent/commands/review.md +0 -38
- package/refs/vbenchmark/.agent/commands/test.md +0 -61
- package/refs/vbenchmark/.agent/rules/01-code-quality.md +0 -33
- package/refs/vbenchmark/.agent/rules/02-typescript-go.md +0 -46
- package/refs/vbenchmark/.agent/rules/03-security-git.md +0 -34
- package/refs/vbenchmark/.agent/rules/04-architecture.md +0 -40
- package/refs/vbenchmark/.agent/sync.js +0 -536
- package/refs/vbenchmark/.agent/workflows/commit.md +0 -29
- package/refs/vbenchmark/.agent/workflows/debug.md +0 -59
- package/refs/vbenchmark/.agent/workflows/document.md +0 -52
- package/refs/vbenchmark/.agent/workflows/gather-context.md +0 -58
- package/refs/vbenchmark/.agent/workflows/init.md +0 -56
- package/refs/vbenchmark/.agent/workflows/preset-help.md +0 -50
- package/refs/vbenchmark/.agent/workflows/refactor.md +0 -71
- package/refs/vbenchmark/.agent/workflows/research.md +0 -37
- package/refs/vbenchmark/.agent/workflows/review.md +0 -38
- package/refs/vbenchmark/.agent/workflows/test.md +0 -61
- package/refs/vbenchmark/.claude/commands/agentic-dev/apply.md +0 -222
- package/refs/vbenchmark/.claude/commands/agentic-dev/done.md +0 -166
- package/refs/vbenchmark/.claude/commands/agentic-dev/proposal.md +0 -220
- package/refs/vbenchmark/.claude/commands/openspec/apply.md +0 -23
- package/refs/vbenchmark/.claude/commands/openspec/archive.md +0 -27
- package/refs/vbenchmark/.claude/commands/openspec/proposal.md +0 -28
- package/refs/vbenchmark/.clinerules/01-rules.md +0 -73
- package/refs/vbenchmark/.clinerules/02-agents.md +0 -34
- package/refs/vbenchmark/.cursor/commands/commit.md +0 -29
- package/refs/vbenchmark/.cursor/commands/debug.md +0 -59
- package/refs/vbenchmark/.cursor/commands/document.md +0 -52
- package/refs/vbenchmark/.cursor/commands/gather-context.md +0 -58
- package/refs/vbenchmark/.cursor/commands/init.md +0 -56
- package/refs/vbenchmark/.cursor/commands/preset-help.md +0 -50
- package/refs/vbenchmark/.cursor/commands/refactor.md +0 -71
- package/refs/vbenchmark/.cursor/commands/research.md +0 -37
- package/refs/vbenchmark/.cursor/commands/review.md +0 -38
- package/refs/vbenchmark/.cursor/commands/test.md +0 -61
- package/refs/vbenchmark/.cursor/rules/agents.mdc +0 -1357
- package/refs/vbenchmark/.factory/droids/codebase-explorer.md +0 -224
- package/refs/vbenchmark/.factory/droids/debugger.md +0 -180
- package/refs/vbenchmark/.factory/droids/documenter.md +0 -166
- package/refs/vbenchmark/.factory/droids/implementer.md +0 -70
- package/refs/vbenchmark/.factory/droids/orchestrator.md +0 -212
- package/refs/vbenchmark/.factory/droids/researcher.md +0 -80
- package/refs/vbenchmark/.factory/droids/reviewer.md +0 -184
- package/refs/vbenchmark/.factory/droids/tester.md +0 -170
- package/refs/vbenchmark/.gemini/workflows/commit.md +0 -29
- package/refs/vbenchmark/.gemini/workflows/debug.md +0 -59
- package/refs/vbenchmark/.gemini/workflows/document.md +0 -52
- package/refs/vbenchmark/.gemini/workflows/gather-context.md +0 -58
- package/refs/vbenchmark/.gemini/workflows/init.md +0 -56
- package/refs/vbenchmark/.gemini/workflows/preset-help.md +0 -50
- package/refs/vbenchmark/.gemini/workflows/refactor.md +0 -71
- package/refs/vbenchmark/.gemini/workflows/research.md +0 -37
- package/refs/vbenchmark/.gemini/workflows/review.md +0 -38
- package/refs/vbenchmark/.gemini/workflows/test.md +0 -61
- package/refs/vbenchmark/.github/CODEOWNERS +0 -20
- package/refs/vbenchmark/.github/FUNDING.yml +0 -4
- package/refs/vbenchmark/.github/ISSUE_TEMPLATE/bug-report.yml +0 -76
- package/refs/vbenchmark/.github/ISSUE_TEMPLATE/new-task.yml +0 -106
- package/refs/vbenchmark/.github/PULL_REQUEST_TEMPLATE.md +0 -38
- package/refs/vbenchmark/.github/copilot-instructions.md +0 -73
- package/refs/vbenchmark/.github/workflows/ci.yaml +0 -33
- package/refs/vbenchmark/.github/workflows/vercel-auto-pr.yml +0 -478
- package/refs/vbenchmark/.github/workflows/vercel-deploy.yaml +0 -487
- package/refs/vbenchmark/.github/workflows/vercel-pr-command.yaml +0 -337
- package/refs/vbenchmark/.github/workflows/vercel-project-init.yaml +0 -208
- package/refs/vbenchmark/.opencode/agent/codebase-explorer.md +0 -224
- package/refs/vbenchmark/.opencode/agent/debugger.md +0 -180
- package/refs/vbenchmark/.opencode/agent/documenter.md +0 -166
- package/refs/vbenchmark/.opencode/agent/implementer.md +0 -70
- package/refs/vbenchmark/.opencode/agent/orchestrator.md +0 -212
- package/refs/vbenchmark/.opencode/agent/researcher.md +0 -80
- package/refs/vbenchmark/.opencode/agent/reviewer.md +0 -184
- package/refs/vbenchmark/.opencode/agent/tester.md +0 -170
- package/refs/vbenchmark/.opencode/command/commit.md +0 -29
- package/refs/vbenchmark/.opencode/command/debug.md +0 -59
- package/refs/vbenchmark/.opencode/command/document.md +0 -52
- package/refs/vbenchmark/.opencode/command/gather-context.md +0 -58
- package/refs/vbenchmark/.opencode/command/init.md +0 -56
- package/refs/vbenchmark/.opencode/command/preset-help.md +0 -50
- package/refs/vbenchmark/.opencode/command/refactor.md +0 -71
- package/refs/vbenchmark/.opencode/command/research.md +0 -37
- package/refs/vbenchmark/.opencode/command/review.md +0 -38
- package/refs/vbenchmark/.opencode/command/test.md +0 -61
- package/refs/vbenchmark/.trae/project_rules.md +0 -73
- package/refs/vbenchmark/.windsurf/rules/rules.md +0 -85
- package/refs/vbenchmark/AGENTS.md +0 -73
- package/refs/vbenchmark/CONTRIBUTING.md +0 -332
- package/refs/vbenchmark/Caddyfile +0 -3
- package/refs/vbenchmark/LICENSE +0 -47
- package/refs/vbenchmark/README.md +0 -354
- package/refs/vbenchmark/docker-compose.prod.yaml +0 -35
- package/refs/vbenchmark/docker-compose.yaml +0 -53
- package/refs/vbenchmark/docs/TASK_EXPANSION_PLAN.md +0 -211
- package/refs/vbenchmark/docs/THESIS.md +0 -441
- package/refs/vbenchmark/docs/categories/code-evolution.md +0 -138
- package/refs/vbenchmark/openspec/changes/init-vibecodingbench/design.md +0 -111
- package/refs/vbenchmark/openspec/changes/init-vibecodingbench/proposal.md +0 -15
- package/refs/vbenchmark/openspec/changes/init-vibecodingbench/specs/evaluation/spec.md +0 -105
- package/refs/vbenchmark/openspec/changes/init-vibecodingbench/specs/leaderboard/spec.md +0 -68
- package/refs/vbenchmark/openspec/changes/init-vibecodingbench/specs/task-definition/spec.md +0 -45
- package/refs/vbenchmark/openspec/changes/init-vibecodingbench/specs/task-runner/spec.md +0 -49
- package/refs/vbenchmark/openspec/changes/init-vibecodingbench/tasks.md +0 -413
- package/refs/vbenchmark/package.json +0 -51
- package/refs/vbenchmark/packages/cli/eslint.config.js +0 -16
- package/refs/vbenchmark/packages/cli/package.json +0 -35
- package/refs/vbenchmark/packages/cli/src/agents/index.ts +0 -655
- package/refs/vbenchmark/packages/cli/src/commands/eval.ts +0 -197
- package/refs/vbenchmark/packages/cli/src/commands/list.ts +0 -63
- package/refs/vbenchmark/packages/cli/src/commands/run.ts +0 -147
- package/refs/vbenchmark/packages/cli/src/evaluator.ts +0 -125
- package/refs/vbenchmark/packages/cli/src/index.ts +0 -21
- package/refs/vbenchmark/packages/cli/src/lib/task-variation.ts +0 -153
- package/refs/vbenchmark/packages/cli/src/loader.ts +0 -258
- package/refs/vbenchmark/packages/cli/src/reporter.ts +0 -222
- package/refs/vbenchmark/packages/cli/src/runtime/docker.ts +0 -385
- package/refs/vbenchmark/packages/cli/tsconfig.json +0 -8
- package/refs/vbenchmark/packages/dashboard/Dockerfile +0 -42
- package/refs/vbenchmark/packages/dashboard/index.html +0 -21
- package/refs/vbenchmark/packages/dashboard/package.json +0 -29
- package/refs/vbenchmark/packages/dashboard/postcss.config.js +0 -6
- package/refs/vbenchmark/packages/dashboard/public/favicon.svg +0 -24
- package/refs/vbenchmark/packages/dashboard/public/logo.png +0 -0
- package/refs/vbenchmark/packages/dashboard/public/logo.svg +0 -39
- package/refs/vbenchmark/packages/dashboard/src/App.tsx +0 -1468
- package/refs/vbenchmark/packages/dashboard/src/data/category-performance.json +0 -1
- package/refs/vbenchmark/packages/dashboard/src/data/leaderboard.json +0 -1
- package/refs/vbenchmark/packages/dashboard/src/data/task-results.json +0 -1
- package/refs/vbenchmark/packages/dashboard/src/data/tasks.json +0 -1
- package/refs/vbenchmark/packages/dashboard/src/index.css +0 -3
- package/refs/vbenchmark/packages/dashboard/src/main.tsx +0 -13
- package/refs/vbenchmark/packages/dashboard/src/vite-env.d.ts +0 -9
- package/refs/vbenchmark/packages/dashboard/tailwind.config.js +0 -11
- package/refs/vbenchmark/packages/dashboard/tsconfig.json +0 -21
- package/refs/vbenchmark/packages/dashboard/tsconfig.node.json +0 -11
- package/refs/vbenchmark/packages/dashboard/vercel.json +0 -6
- package/refs/vbenchmark/packages/dashboard/vite.config.ts +0 -28
- package/refs/vbenchmark/packages/evaluator/eslint.config.js +0 -16
- package/refs/vbenchmark/packages/evaluator/package.json +0 -24
- package/refs/vbenchmark/packages/evaluator/src/index.ts +0 -15
- package/refs/vbenchmark/packages/evaluator/src/runners/functional.ts +0 -88
- package/refs/vbenchmark/packages/evaluator/src/runners/quality.ts +0 -140
- package/refs/vbenchmark/packages/evaluator/src/runners/security.ts +0 -94
- package/refs/vbenchmark/packages/evaluator/src/runners/visual.ts +0 -108
- package/refs/vbenchmark/packages/evaluator/src/types.d.ts +0 -19
- package/refs/vbenchmark/packages/evaluator/tsconfig.json +0 -8
- package/refs/vbenchmark/packages/leaderboard/Dockerfile +0 -38
- package/refs/vbenchmark/packages/leaderboard/drizzle.config.ts +0 -10
- package/refs/vbenchmark/packages/leaderboard/eslint.config.js +0 -16
- package/refs/vbenchmark/packages/leaderboard/fly.toml +0 -29
- package/refs/vbenchmark/packages/leaderboard/package.json +0 -36
- package/refs/vbenchmark/packages/leaderboard/src/app.ts +0 -29
- package/refs/vbenchmark/packages/leaderboard/src/components/BrowserPreview.tsx +0 -190
- package/refs/vbenchmark/packages/leaderboard/src/components/ComparisonView.tsx +0 -205
- package/refs/vbenchmark/packages/leaderboard/src/components/LeaderboardTable.tsx +0 -150
- package/refs/vbenchmark/packages/leaderboard/src/components/LiveRunCard.tsx +0 -133
- package/refs/vbenchmark/packages/leaderboard/src/components/SubmissionForm.tsx +0 -406
- package/refs/vbenchmark/packages/leaderboard/src/components/SubmitForm.tsx +0 -293
- package/refs/vbenchmark/packages/leaderboard/src/components/TerminalStream.tsx +0 -111
- package/refs/vbenchmark/packages/leaderboard/src/config/pricing.ts +0 -206
- package/refs/vbenchmark/packages/leaderboard/src/db/index.ts +0 -31
- package/refs/vbenchmark/packages/leaderboard/src/db/schema.ts +0 -125
- package/refs/vbenchmark/packages/leaderboard/src/index.ts +0 -13
- package/refs/vbenchmark/packages/leaderboard/src/lib/websocket.ts +0 -124
- package/refs/vbenchmark/packages/leaderboard/src/routes/leaderboard.ts +0 -698
- package/refs/vbenchmark/packages/leaderboard/src/routes/live.ts +0 -175
- package/refs/vbenchmark/packages/leaderboard/src/routes/submissions.ts +0 -183
- package/refs/vbenchmark/packages/leaderboard/src/routes/tasks.ts +0 -215
- package/refs/vbenchmark/packages/leaderboard/tests/api.test.ts +0 -228
- package/refs/vbenchmark/packages/leaderboard/tsconfig.json +0 -9
- package/refs/vbenchmark/scripts/deploy.sh +0 -70
- package/refs/vbenchmark/tasks/ai-integration/advanced/context-management/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/ai-integration/advanced/context-management/task.yaml +0 -16
- package/refs/vbenchmark/tasks/ai-integration/advanced/evaluation-framework/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/ai-integration/advanced/evaluation-framework/task.yaml +0 -16
- package/refs/vbenchmark/tasks/ai-integration/advanced/guardrails-safety/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/ai-integration/advanced/guardrails-safety/task.yaml +0 -16
- package/refs/vbenchmark/tasks/ai-integration/advanced/memory-system/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/ai-integration/advanced/memory-system/task.yaml +0 -16
- package/refs/vbenchmark/tasks/ai-integration/advanced/model-routing/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/ai-integration/advanced/model-routing/task.yaml +0 -16
- package/refs/vbenchmark/tasks/ai-integration/advanced/multi-agent-system/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/ai-integration/advanced/multi-agent-system/task.yaml +0 -16
- package/refs/vbenchmark/tasks/ai-integration/advanced/prompt-optimization/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/ai-integration/advanced/prompt-optimization/task.yaml +0 -16
- package/refs/vbenchmark/tasks/ai-integration/advanced/reasoning-chain/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/ai-integration/advanced/reasoning-chain/task.yaml +0 -16
- package/refs/vbenchmark/tasks/ai-integration/advanced/streaming-pipeline/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/ai-integration/advanced/streaming-pipeline/task.yaml +0 -16
- package/refs/vbenchmark/tasks/ai-integration/advanced/tool-use-orchestration/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/ai-integration/advanced/tool-use-orchestration/task.yaml +0 -16
- package/refs/vbenchmark/tasks/ai-integration/agents/code-review-agent/PROMPT.md +0 -64
- package/refs/vbenchmark/tasks/ai-integration/agents/code-review-agent/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/agents/research-agent/PROMPT.md +0 -61
- package/refs/vbenchmark/tasks/ai-integration/agents/research-agent/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/agents/web-scraper-agent/PROMPT.md +0 -57
- package/refs/vbenchmark/tasks/ai-integration/agents/web-scraper-agent/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/embeddings/duplicate-detection/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/ai-integration/embeddings/duplicate-detection/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/embeddings/recommendation-engine/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/ai-integration/embeddings/recommendation-engine/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/embeddings/semantic-search/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/ai-integration/embeddings/semantic-search/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/fine-tuning/classification-model/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/ai-integration/fine-tuning/classification-model/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/function-calling/api-orchestrator/PROMPT.md +0 -60
- package/refs/vbenchmark/tasks/ai-integration/function-calling/api-orchestrator/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/function-calling/calendar-assistant/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/ai-integration/function-calling/calendar-assistant/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/function-calling/database-query/PROMPT.md +0 -62
- package/refs/vbenchmark/tasks/ai-integration/function-calling/database-query/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/multimodal/chart-interpreter/PROMPT.md +0 -60
- package/refs/vbenchmark/tasks/ai-integration/multimodal/chart-interpreter/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/multimodal/image-captioning/PROMPT.md +0 -49
- package/refs/vbenchmark/tasks/ai-integration/multimodal/image-captioning/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/rag-chatbot/code-assistant/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/ai-integration/rag-chatbot/code-assistant/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/rag-chatbot/doc-search/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/ai-integration/rag-chatbot/doc-search/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/rag-chatbot/pdf-qa/PROMPT.md +0 -76
- package/refs/vbenchmark/tasks/ai-integration/rag-chatbot/pdf-qa/docker-compose.yaml +0 -30
- package/refs/vbenchmark/tasks/ai-integration/rag-chatbot/pdf-qa/task.yaml +0 -30
- package/refs/vbenchmark/tasks/ai-integration/rag-chatbot/pdf-qa/tests/functional/qa.test.py +0 -146
- package/refs/vbenchmark/tasks/ai-integration/rag-chatbot/support-bot/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/ai-integration/rag-chatbot/support-bot/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/structured-output/contract-analyzer/PROMPT.md +0 -67
- package/refs/vbenchmark/tasks/ai-integration/structured-output/contract-analyzer/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/structured-output/invoice-parser/PROMPT.md +0 -61
- package/refs/vbenchmark/tasks/ai-integration/structured-output/invoice-parser/task.yaml +0 -27
- package/refs/vbenchmark/tasks/ai-integration/structured-output/receipt-scanner/PROMPT.md +0 -65
- package/refs/vbenchmark/tasks/ai-integration/structured-output/receipt-scanner/task.yaml +0 -24
- package/refs/vbenchmark/tasks/ai-integration/structured-output/resume-parser/PROMPT.md +0 -70
- package/refs/vbenchmark/tasks/ai-integration/structured-output/resume-parser/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/advanced/api-analytics/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/api-integrations/advanced/api-analytics/task.yaml +0 -16
- package/refs/vbenchmark/tasks/api-integrations/advanced/api-gateway/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/api-integrations/advanced/api-gateway/task.yaml +0 -16
- package/refs/vbenchmark/tasks/api-integrations/advanced/api-mocking/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/api-integrations/advanced/api-mocking/task.yaml +0 -16
- package/refs/vbenchmark/tasks/api-integrations/advanced/contract-testing/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/api-integrations/advanced/contract-testing/task.yaml +0 -16
- package/refs/vbenchmark/tasks/api-integrations/advanced/graphql-federation/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/api-integrations/advanced/graphql-federation/task.yaml +0 -16
- package/refs/vbenchmark/tasks/api-integrations/advanced/grpc-gateway/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/api-integrations/advanced/grpc-gateway/task.yaml +0 -16
- package/refs/vbenchmark/tasks/api-integrations/advanced/rate-limiter/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/api-integrations/advanced/rate-limiter/task.yaml +0 -16
- package/refs/vbenchmark/tasks/api-integrations/advanced/request-validator/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/api-integrations/advanced/request-validator/task.yaml +0 -16
- package/refs/vbenchmark/tasks/api-integrations/advanced/sdk-generator/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/api-integrations/advanced/sdk-generator/task.yaml +0 -16
- package/refs/vbenchmark/tasks/api-integrations/advanced/webhook-processor/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/api-integrations/advanced/webhook-processor/task.yaml +0 -16
- package/refs/vbenchmark/tasks/api-integrations/analytics/mixpanel-events/PROMPT.md +0 -42
- package/refs/vbenchmark/tasks/api-integrations/analytics/mixpanel-events/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/analytics/segment-tracking/PROMPT.md +0 -42
- package/refs/vbenchmark/tasks/api-integrations/analytics/segment-tracking/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/auth-provider/oauth2-github/PROMPT.md +0 -42
- package/refs/vbenchmark/tasks/api-integrations/auth-provider/oauth2-github/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/auth-provider/okta-integration/PROMPT.md +0 -44
- package/refs/vbenchmark/tasks/api-integrations/auth-provider/okta-integration/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/auth-provider/saml-sso/PROMPT.md +0 -42
- package/refs/vbenchmark/tasks/api-integrations/auth-provider/saml-sso/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/communication/discord-webhook/PROMPT.md +0 -44
- package/refs/vbenchmark/tasks/api-integrations/communication/discord-webhook/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/communication/slack-bot/PROMPT.md +0 -42
- package/refs/vbenchmark/tasks/api-integrations/communication/slack-bot/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/communication/twilio-sms/PROMPT.md +0 -42
- package/refs/vbenchmark/tasks/api-integrations/communication/twilio-sms/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/email/transactional/PROMPT.md +0 -82
- package/refs/vbenchmark/tasks/api-integrations/email/transactional/task.yaml +0 -27
- package/refs/vbenchmark/tasks/api-integrations/maps/google-maps-geocoding/PROMPT.md +0 -41
- package/refs/vbenchmark/tasks/api-integrations/maps/google-maps-geocoding/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/maps/mapbox-directions/PROMPT.md +0 -41
- package/refs/vbenchmark/tasks/api-integrations/maps/mapbox-directions/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/payment/crypto-payments/PROMPT.md +0 -43
- package/refs/vbenchmark/tasks/api-integrations/payment/crypto-payments/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/payment/paypal-integration/PROMPT.md +0 -41
- package/refs/vbenchmark/tasks/api-integrations/payment/paypal-integration/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/social/twitter-api/PROMPT.md +0 -41
- package/refs/vbenchmark/tasks/api-integrations/social/twitter-api/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/storage/cloudinary-upload/PROMPT.md +0 -43
- package/refs/vbenchmark/tasks/api-integrations/storage/cloudinary-upload/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/storage/gcs-streaming/PROMPT.md +0 -43
- package/refs/vbenchmark/tasks/api-integrations/storage/gcs-streaming/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/storage/s3-presigned-urls/PROMPT.md +0 -41
- package/refs/vbenchmark/tasks/api-integrations/storage/s3-presigned-urls/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/stripe/checkout-session/PROMPT.md +0 -41
- package/refs/vbenchmark/tasks/api-integrations/stripe/checkout-session/task.yaml +0 -24
- package/refs/vbenchmark/tasks/api-integrations/stripe/payment-webhook/PROMPT.md +0 -60
- package/refs/vbenchmark/tasks/api-integrations/stripe/payment-webhook/docker-compose.yaml +0 -38
- package/refs/vbenchmark/tasks/api-integrations/stripe/payment-webhook/task.yaml +0 -31
- package/refs/vbenchmark/tasks/api-integrations/stripe/payment-webhook/tests/webhook.test.ts +0 -193
- package/refs/vbenchmark/tasks/api-integrations/stripe/subscription-portal/PROMPT.md +0 -41
- package/refs/vbenchmark/tasks/api-integrations/stripe/subscription-portal/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/advanced/api-deprecation/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/code-evolution/advanced/api-deprecation/task.yaml +0 -16
- package/refs/vbenchmark/tasks/code-evolution/advanced/ast-refactoring/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/code-evolution/advanced/ast-refactoring/task.yaml +0 -16
- package/refs/vbenchmark/tasks/code-evolution/advanced/concurrency-fix/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/code-evolution/advanced/concurrency-fix/task.yaml +0 -16
- package/refs/vbenchmark/tasks/code-evolution/advanced/database-schema-migration/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/code-evolution/advanced/database-schema-migration/task.yaml +0 -16
- package/refs/vbenchmark/tasks/code-evolution/advanced/dead-code-elimination/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/code-evolution/advanced/dead-code-elimination/task.yaml +0 -16
- package/refs/vbenchmark/tasks/code-evolution/advanced/dependency-upgrade/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/code-evolution/advanced/dependency-upgrade/task.yaml +0 -16
- package/refs/vbenchmark/tasks/code-evolution/advanced/memory-optimization/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/code-evolution/advanced/memory-optimization/task.yaml +0 -16
- package/refs/vbenchmark/tasks/code-evolution/advanced/monorepo-extraction/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/code-evolution/advanced/monorepo-extraction/task.yaml +0 -16
- package/refs/vbenchmark/tasks/code-evolution/advanced/performance-profiling/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/code-evolution/advanced/performance-profiling/task.yaml +0 -16
- package/refs/vbenchmark/tasks/code-evolution/advanced/type-migration/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/code-evolution/advanced/type-migration/task.yaml +0 -16
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/callback-to-async/PROMPT.md +0 -47
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/callback-to-async/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/express-to-fastify/PROMPT.md +0 -49
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/express-to-fastify/base-code/src/app.ts +0 -22
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/express-to-fastify/task.yaml +0 -37
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/express-to-fastify/tests/api.test.ts +0 -70
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/flask-to-fastapi/PROMPT.md +0 -46
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/flask-to-fastapi/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/java-to-kotlin/PROMPT.md +0 -45
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/java-to-kotlin/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/jquery-to-react/PROMPT.md +0 -47
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/jquery-to-react/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/rest-to-grpc/PROMPT.md +0 -47
- package/refs/vbenchmark/tasks/code-evolution/legacy-migration/rest-to-grpc/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/performance/async-refactor/PROMPT.md +0 -47
- package/refs/vbenchmark/tasks/code-evolution/performance/async-refactor/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/performance/memory-leak-fix/PROMPT.md +0 -47
- package/refs/vbenchmark/tasks/code-evolution/performance/memory-leak-fix/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/performance/query-optimization/PROMPT.md +0 -49
- package/refs/vbenchmark/tasks/code-evolution/performance/query-optimization/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/refactoring/class-to-hooks/PROMPT.md +0 -96
- package/refs/vbenchmark/tasks/code-evolution/refactoring/class-to-hooks/task.yaml +0 -27
- package/refs/vbenchmark/tasks/code-evolution/refactoring/dependency-injection/PROMPT.md +0 -47
- package/refs/vbenchmark/tasks/code-evolution/refactoring/dependency-injection/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/refactoring/error-handling/PROMPT.md +0 -48
- package/refs/vbenchmark/tasks/code-evolution/refactoring/error-handling/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/refactoring/monolith-to-modules/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/code-evolution/refactoring/monolith-to-modules/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/refactoring/orm-migration/PROMPT.md +0 -47
- package/refs/vbenchmark/tasks/code-evolution/refactoring/orm-migration/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/security/secrets-rotation/PROMPT.md +0 -49
- package/refs/vbenchmark/tasks/code-evolution/security/secrets-rotation/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/security/sql-injection-fix/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/code-evolution/security/sql-injection-fix/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/security/xss-prevention/PROMPT.md +0 -47
- package/refs/vbenchmark/tasks/code-evolution/security/xss-prevention/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/testing/add-unit-tests/PROMPT.md +0 -48
- package/refs/vbenchmark/tasks/code-evolution/testing/add-unit-tests/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/testing/e2e-playwright/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/code-evolution/testing/e2e-playwright/task.yaml +0 -24
- package/refs/vbenchmark/tasks/code-evolution/testing/pytest-fixtures/PROMPT.md +0 -47
- package/refs/vbenchmark/tasks/code-evolution/testing/pytest-fixtures/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/accessibility/keyboard-shortcuts/PROMPT.md +0 -44
- package/refs/vbenchmark/tasks/frontend/accessibility/keyboard-shortcuts/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/accessibility/screen-reader-nav/PROMPT.md +0 -44
- package/refs/vbenchmark/tasks/frontend/accessibility/screen-reader-nav/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/advanced/canvas-editor/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/frontend/advanced/canvas-editor/task.yaml +0 -16
- package/refs/vbenchmark/tasks/frontend/advanced/micro-frontend/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/frontend/advanced/micro-frontend/task.yaml +0 -16
- package/refs/vbenchmark/tasks/frontend/advanced/offline-first/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/frontend/advanced/offline-first/task.yaml +0 -16
- package/refs/vbenchmark/tasks/frontend/advanced/realtime-collab/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/frontend/advanced/realtime-collab/task.yaml +0 -16
- package/refs/vbenchmark/tasks/frontend/advanced/service-worker/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/frontend/advanced/service-worker/task.yaml +0 -16
- package/refs/vbenchmark/tasks/frontend/advanced/state-machine/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/frontend/advanced/state-machine/task.yaml +0 -16
- package/refs/vbenchmark/tasks/frontend/advanced/virtual-list/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/frontend/advanced/virtual-list/task.yaml +0 -16
- package/refs/vbenchmark/tasks/frontend/advanced/wasm-integration/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/frontend/advanced/wasm-integration/task.yaml +0 -16
- package/refs/vbenchmark/tasks/frontend/advanced/web-worker/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/frontend/advanced/web-worker/task.yaml +0 -16
- package/refs/vbenchmark/tasks/frontend/advanced/webgl-visualization/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/frontend/advanced/webgl-visualization/task.yaml +0 -16
- package/refs/vbenchmark/tasks/frontend/animation/page-transitions/PROMPT.md +0 -44
- package/refs/vbenchmark/tasks/frontend/animation/page-transitions/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/components/data-grid/PROMPT.md +0 -59
- package/refs/vbenchmark/tasks/frontend/components/data-grid/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/components/date-range-picker/PROMPT.md +0 -57
- package/refs/vbenchmark/tasks/frontend/components/date-range-picker/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/components/file-uploader/PROMPT.md +0 -55
- package/refs/vbenchmark/tasks/frontend/components/file-uploader/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/components/form-builder/PROMPT.md +0 -96
- package/refs/vbenchmark/tasks/frontend/components/form-builder/task.yaml +0 -28
- package/refs/vbenchmark/tasks/frontend/components/rich-text-editor/PROMPT.md +0 -45
- package/refs/vbenchmark/tasks/frontend/components/rich-text-editor/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/figma-to-code/dashboard-layout/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/frontend/figma-to-code/dashboard-layout/task.yaml +0 -25
- package/refs/vbenchmark/tasks/frontend/figma-to-code/landing-page/PROMPT.md +0 -49
- package/refs/vbenchmark/tasks/frontend/figma-to-code/landing-page/task.yaml +0 -25
- package/refs/vbenchmark/tasks/frontend/figma-to-code/mobile-app-screen/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/frontend/figma-to-code/mobile-app-screen/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/figma-to-code/pricing-card/PROMPT.md +0 -93
- package/refs/vbenchmark/tasks/frontend/figma-to-code/pricing-card/docker-compose.yaml +0 -23
- package/refs/vbenchmark/tasks/frontend/figma-to-code/pricing-card/task.yaml +0 -30
- package/refs/vbenchmark/tasks/frontend/figma-to-code/pricing-card/tests/visual/diff.test.ts +0 -107
- package/refs/vbenchmark/tasks/frontend/figma-to-code/pricing-card/tests/visual/interaction.test.ts +0 -88
- package/refs/vbenchmark/tasks/frontend/performance/image-lazy-load/PROMPT.md +0 -43
- package/refs/vbenchmark/tasks/frontend/performance/image-lazy-load/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/performance/infinite-scroll/PROMPT.md +0 -44
- package/refs/vbenchmark/tasks/frontend/performance/infinite-scroll/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/state-management/collaborative-editor/PROMPT.md +0 -44
- package/refs/vbenchmark/tasks/frontend/state-management/collaborative-editor/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/state-management/shopping-cart/PROMPT.md +0 -53
- package/refs/vbenchmark/tasks/frontend/state-management/shopping-cart/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/visualization/chart-dashboard/PROMPT.md +0 -83
- package/refs/vbenchmark/tasks/frontend/visualization/chart-dashboard/task.yaml +0 -28
- package/refs/vbenchmark/tasks/frontend/visualization/gantt-chart/PROMPT.md +0 -57
- package/refs/vbenchmark/tasks/frontend/visualization/gantt-chart/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/visualization/map-dashboard/PROMPT.md +0 -44
- package/refs/vbenchmark/tasks/frontend/visualization/map-dashboard/task.yaml +0 -24
- package/refs/vbenchmark/tasks/frontend/visualization/realtime-charts/PROMPT.md +0 -43
- package/refs/vbenchmark/tasks/frontend/visualization/realtime-charts/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/advanced/blue-green-deploy/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/glue-code/advanced/blue-green-deploy/task.yaml +0 -16
- package/refs/vbenchmark/tasks/glue-code/advanced/canary-release/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/glue-code/advanced/canary-release/task.yaml +0 -16
- package/refs/vbenchmark/tasks/glue-code/advanced/change-data-capture/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/glue-code/advanced/change-data-capture/task.yaml +0 -16
- package/refs/vbenchmark/tasks/glue-code/advanced/config-management/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/glue-code/advanced/config-management/task.yaml +0 -16
- package/refs/vbenchmark/tasks/glue-code/advanced/data-pipeline/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/glue-code/advanced/data-pipeline/task.yaml +0 -16
- package/refs/vbenchmark/tasks/glue-code/advanced/distributed-tracing/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/glue-code/advanced/distributed-tracing/task.yaml +0 -16
- package/refs/vbenchmark/tasks/glue-code/advanced/log-aggregation/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/glue-code/advanced/log-aggregation/task.yaml +0 -16
- package/refs/vbenchmark/tasks/glue-code/advanced/schema-registry/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/glue-code/advanced/schema-registry/task.yaml +0 -16
- package/refs/vbenchmark/tasks/glue-code/advanced/secret-rotation/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/glue-code/advanced/secret-rotation/task.yaml +0 -16
- package/refs/vbenchmark/tasks/glue-code/advanced/stream-processing/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/glue-code/advanced/stream-processing/task.yaml +0 -16
- package/refs/vbenchmark/tasks/glue-code/api-sync/rest-to-graphql/PROMPT.md +0 -66
- package/refs/vbenchmark/tasks/glue-code/api-sync/rest-to-graphql/task.yaml +0 -27
- package/refs/vbenchmark/tasks/glue-code/caching/redis-cache/PROMPT.md +0 -82
- package/refs/vbenchmark/tasks/glue-code/caching/redis-cache/task.yaml +0 -27
- package/refs/vbenchmark/tasks/glue-code/data-transform/avro-schema-evolution/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/glue-code/data-transform/avro-schema-evolution/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/data-transform/csv-normalizer/PROMPT.md +0 -49
- package/refs/vbenchmark/tasks/glue-code/data-transform/csv-normalizer/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/data-transform/excel-to-json/PROMPT.md +0 -67
- package/refs/vbenchmark/tasks/glue-code/data-transform/excel-to-json/task.yaml +0 -28
- package/refs/vbenchmark/tasks/glue-code/data-transform/excel-to-json/tests/transform.test.py +0 -137
- package/refs/vbenchmark/tasks/glue-code/data-transform/json-to-xml/PROMPT.md +0 -45
- package/refs/vbenchmark/tasks/glue-code/data-transform/json-to-xml/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/data-transform/protobuf-converter/PROMPT.md +0 -44
- package/refs/vbenchmark/tasks/glue-code/data-transform/protobuf-converter/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/etl/cdc-pipeline/PROMPT.md +0 -52
- package/refs/vbenchmark/tasks/glue-code/etl/cdc-pipeline/task.yaml +0 -27
- package/refs/vbenchmark/tasks/glue-code/etl/database-sync/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/glue-code/etl/database-sync/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/etl/s3-to-warehouse/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/glue-code/etl/s3-to-warehouse/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/file-processing/image-resizer/PROMPT.md +0 -52
- package/refs/vbenchmark/tasks/glue-code/file-processing/image-resizer/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/file-processing/pdf-merger/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/glue-code/file-processing/pdf-merger/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/file-processing/video-transcoder/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/glue-code/file-processing/video-transcoder/task.yaml +0 -27
- package/refs/vbenchmark/tasks/glue-code/migration/data-backfill/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/glue-code/migration/data-backfill/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/migration/database-versioning/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/glue-code/migration/database-versioning/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/queue/kafka-producer/PROMPT.md +0 -49
- package/refs/vbenchmark/tasks/glue-code/queue/kafka-producer/task.yaml +0 -27
- package/refs/vbenchmark/tasks/glue-code/queue/rabbitmq-consumer/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/glue-code/queue/rabbitmq-consumer/task.yaml +0 -27
- package/refs/vbenchmark/tasks/glue-code/queue/sqs-batch-processor/PROMPT.md +0 -47
- package/refs/vbenchmark/tasks/glue-code/queue/sqs-batch-processor/task.yaml +0 -24
- package/refs/vbenchmark/tasks/glue-code/scheduler/cron-job-manager/PROMPT.md +0 -52
- package/refs/vbenchmark/tasks/glue-code/scheduler/cron-job-manager/task.yaml +0 -27
- package/refs/vbenchmark/tasks/glue-code/scheduler/delayed-tasks/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/glue-code/scheduler/delayed-tasks/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/advanced/api-versioning/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/saas-core/advanced/api-versioning/task.yaml +0 -16
- package/refs/vbenchmark/tasks/saas-core/advanced/circuit-breaker/PROMPT.md +0 -13
- package/refs/vbenchmark/tasks/saas-core/advanced/circuit-breaker/task.yaml +0 -16
- package/refs/vbenchmark/tasks/saas-core/advanced/compliance-gdpr/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/saas-core/advanced/compliance-gdpr/task.yaml +0 -16
- package/refs/vbenchmark/tasks/saas-core/advanced/cqrs-pattern/PROMPT.md +0 -13
- package/refs/vbenchmark/tasks/saas-core/advanced/cqrs-pattern/task.yaml +0 -16
- package/refs/vbenchmark/tasks/saas-core/advanced/data-encryption/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/saas-core/advanced/data-encryption/task.yaml +0 -16
- package/refs/vbenchmark/tasks/saas-core/advanced/distributed-locking/PROMPT.md +0 -46
- package/refs/vbenchmark/tasks/saas-core/advanced/distributed-locking/task.yaml +0 -24
- package/refs/vbenchmark/tasks/saas-core/advanced/event-sourcing/PROMPT.md +0 -23
- package/refs/vbenchmark/tasks/saas-core/advanced/event-sourcing/task.yaml +0 -16
- package/refs/vbenchmark/tasks/saas-core/advanced/feature-flags-ab/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/saas-core/advanced/feature-flags-ab/task.yaml +0 -16
- package/refs/vbenchmark/tasks/saas-core/advanced/saga-orchestration/PROMPT.md +0 -13
- package/refs/vbenchmark/tasks/saas-core/advanced/saga-orchestration/task.yaml +0 -16
- package/refs/vbenchmark/tasks/saas-core/advanced/webhook-delivery/PROMPT.md +0 -15
- package/refs/vbenchmark/tasks/saas-core/advanced/webhook-delivery/task.yaml +0 -16
- package/refs/vbenchmark/tasks/saas-core/audit/activity-logging/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/saas-core/audit/activity-logging/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/auth/jwt-refresh-tokens/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/saas-core/auth/jwt-refresh-tokens/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/auth/magic-link-email/PROMPT.md +0 -53
- package/refs/vbenchmark/tasks/saas-core/auth/magic-link-email/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/auth/mfa-totp/PROMPT.md +0 -79
- package/refs/vbenchmark/tasks/saas-core/auth/mfa-totp/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/auth/rbac-permissions/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/saas-core/auth/rbac-permissions/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/auth/session-management/PROMPT.md +0 -52
- package/refs/vbenchmark/tasks/saas-core/auth/session-management/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/auth/supabase-oauth/PROMPT.md +0 -45
- package/refs/vbenchmark/tasks/saas-core/auth/supabase-oauth/docker-compose.yaml +0 -47
- package/refs/vbenchmark/tasks/saas-core/auth/supabase-oauth/task.yaml +0 -32
- package/refs/vbenchmark/tasks/saas-core/auth/supabase-oauth/tests/auth.test.ts +0 -59
- package/refs/vbenchmark/tasks/saas-core/billing/invoice-generation/PROMPT.md +0 -53
- package/refs/vbenchmark/tasks/saas-core/billing/invoice-generation/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/billing/stripe-subscriptions/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/saas-core/billing/stripe-subscriptions/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/billing/usage-metering/PROMPT.md +0 -52
- package/refs/vbenchmark/tasks/saas-core/billing/usage-metering/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/crud/dashboard-table/PROMPT.md +0 -48
- package/refs/vbenchmark/tasks/saas-core/crud/dashboard-table/task.yaml +0 -28
- package/refs/vbenchmark/tasks/saas-core/multi-tenant/org-isolation/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/saas-core/multi-tenant/org-isolation/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/multi-tenant/subdomain-routing/PROMPT.md +0 -50
- package/refs/vbenchmark/tasks/saas-core/multi-tenant/subdomain-routing/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/notifications/email-queue/PROMPT.md +0 -53
- package/refs/vbenchmark/tasks/saas-core/notifications/email-queue/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/notifications/in-app-alerts/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/saas-core/notifications/in-app-alerts/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/notifications/push-notifications/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/saas-core/notifications/push-notifications/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/realtime/websocket-chat/PROMPT.md +0 -80
- package/refs/vbenchmark/tasks/saas-core/realtime/websocket-chat/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/search/full-text-search/PROMPT.md +0 -51
- package/refs/vbenchmark/tasks/saas-core/search/full-text-search/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/security/rate-limiter/PROMPT.md +0 -99
- package/refs/vbenchmark/tasks/saas-core/security/rate-limiter/task.yaml +0 -27
- package/refs/vbenchmark/tasks/saas-core/settings/user-preferences/PROMPT.md +0 -78
- package/refs/vbenchmark/tasks/saas-core/settings/user-preferences/task.yaml +0 -27
- package/refs/vbenchmark/templates/fastapi-postgres/docker-compose.yaml +0 -36
- package/refs/vbenchmark/templates/fastapi-postgres/pyproject.toml +0 -34
- package/refs/vbenchmark/templates/fastapi-postgres/src/__init__.py +0 -0
- package/refs/vbenchmark/templates/fastapi-postgres/src/config.py +0 -12
- package/refs/vbenchmark/templates/fastapi-postgres/src/database.py +0 -15
- package/refs/vbenchmark/templates/fastapi-postgres/src/main.py +0 -51
- package/refs/vbenchmark/templates/fastapi-postgres/src/models.py +0 -12
- package/refs/vbenchmark/templates/fastapi-postgres/src/schemas.py +0 -20
- package/refs/vbenchmark/templates/go-fiber/docker-compose.yaml +0 -34
- package/refs/vbenchmark/templates/go-fiber/go.mod +0 -33
- package/refs/vbenchmark/templates/go-fiber/go.sum +0 -68
- package/refs/vbenchmark/templates/go-fiber/main.go +0 -98
- package/refs/vbenchmark/templates/nextjs-supabase/.env.example +0 -3
- package/refs/vbenchmark/templates/nextjs-supabase/docker-compose.yaml +0 -68
- package/refs/vbenchmark/templates/nextjs-supabase/src/app/globals.css +0 -13
- package/refs/vbenchmark/templates/nextjs-supabase/src/app/layout.tsx +0 -19
- package/refs/vbenchmark/templates/nextjs-supabase/src/app/page.tsx +0 -38
- package/refs/vbenchmark/templates/nextjs-supabase/src/lib/supabase/client.ts +0 -8
- package/refs/vbenchmark/templates/nextjs-supabase/src/lib/supabase/server.ts +0 -32
- package/refs/vbenchmark/templates/rust-axum/Cargo.lock +0 -2371
- package/refs/vbenchmark/templates/rust-axum/Cargo.toml +0 -16
- package/refs/vbenchmark/templates/rust-axum/docker-compose.yaml +0 -34
- package/refs/vbenchmark/templates/rust-axum/migrations/20240101000000_init.sql +0 -20
- package/refs/vbenchmark/templates/rust-axum/src/main.rs +0 -121
- package/refs/vbenchmark/tsconfig.base.json +0 -18
- package/refs/vbenchmark/turbo.json +0 -23
- package/refs/vbenchmark/vercel.json +0 -10
|
@@ -1,441 +0,0 @@
|
|
|
1
|
-
# VibeCodingBench: Benchmark Vibe Coding Models for Fun
|
|
2
|
-
|
|
3
|
-
**A Position Paper on Benchmark Methodology**
|
|
4
|
-
|
|
5
|
-
---
|
|
6
|
-
|
|
7
|
-
## Abstract
|
|
8
|
-
|
|
9
|
-
Current benchmarks for evaluating AI coding agents suffer from a fundamental ecological validity crisis. While models achieve impressive scores on SWE-bench (70%+) and HumanEval (90%+), these metrics fail to predict real-world utility. We present VibeCodingBench, a benchmark grounded in empirical data about actual developer workflows, designed to measure what matters: the ability to accelerate production software development. Our approach addresses three critical gaps in existing evaluation: (1) task distribution mismatch, (2) single-dimension scoring, and (3) isolation from production concerns.
|
|
10
|
-
|
|
11
|
-
---
|
|
12
|
-
|
|
13
|
-
## 1. The Evaluation Crisis
|
|
14
|
-
|
|
15
|
-
### 1.1 The Disconnect Between Benchmarks and Reality
|
|
16
|
-
|
|
17
|
-
The AI coding agent landscape exhibits a troubling paradox: benchmark performance continues to climb while practical utility remains uncertain. Consider the evidence:
|
|
18
|
-
|
|
19
|
-
- **SWE-bench Verified**: Top models score 70%+ on bug fixes, yet these "represent a substantial proportion of relatively trivial problems (161 out of 500) that require only one- to two-line modifications" ([Runloop AI, 2025](https://runloop.ai/blog/swe-bench-deep-dive-unmasking-the-limitations-of-a-popular-benchmark))
|
|
20
|
-
|
|
21
|
-
- **SWE-bench Pro Performance Gap**: When evaluated on multi-file, long-horizon tasks, the same models drop to ~23%—a 47-point decline ([Scale AI Leaderboard](https://scale.com/leaderboard/swe_bench_pro_public))
|
|
22
|
-
|
|
23
|
-
- **SWE-EVO Findings**: "Even the best-performing model (GPT-5) resolves only 21% of SWE-EVO tasks compared to 65% on SWE-Bench Verified" ([arxiv:2512.18470](https://arxiv.org/abs/2512.18470))
|
|
24
|
-
|
|
25
|
-
- **Developer Trust**: Stack Overflow 2025 reports that 66% of developers cite "AI solutions that are almost right, but not quite" as their biggest frustration ([Stack Overflow Developer Survey 2025](https://survey.stackoverflow.co/2025/))
|
|
26
|
-
|
|
27
|
-
### 1.2 Why Current Benchmarks Fail
|
|
28
|
-
|
|
29
|
-
We identify four structural limitations:
|
|
30
|
-
|
|
31
|
-
| Limitation | Evidence | Impact |
|
|
32
|
-
|------------|----------|--------|
|
|
33
|
-
| **Task Triviality** | 63.75% of SWE-Agent patches are "suspicious" due to weak tests | Overstates capability |
|
|
34
|
-
| **Data Contamination** | Training corpora include benchmark repositories | Inflates scores |
|
|
35
|
-
| **Narrow Scope** | Focus on Python bug fixes in 12 repositories | Misses language/domain breadth |
|
|
36
|
-
| **Single Metric** | Pass rate alone ignores quality, security, cost | Hides production risks |
|
|
37
|
-
|
|
38
|
-
As one analysis notes: "Even when LLM-generated code passes functional performance benchmarks, it is not free of underlying quality defects—for example, Claude Sonnet 4 (a top performer on pass rate) averaged 2.11 issues per passing task" ([Qodo State of AI Code Quality 2025](https://www.qodo.ai/reports/state-of-ai-code-quality/)).
|
|
39
|
-
|
|
40
|
-
---
|
|
41
|
-
|
|
42
|
-
## 2. What Developers Actually Do
|
|
43
|
-
|
|
44
|
-
### 2.1 Time Distribution in Software Engineering
|
|
45
|
-
|
|
46
|
-
To build a valid benchmark, we must first understand developer workflows. Empirical research reveals:
|
|
47
|
-
|
|
48
|
-
**Daily Activity Breakdown** ([Sonar Research](https://www.sonarsource.com/blog/how-much-time-do-developers-spend-actually-writing-code/)):
|
|
49
|
-
- Writing new code: **32%**
|
|
50
|
-
- Code maintenance: **19%**
|
|
51
|
-
- Testing: **12%**
|
|
52
|
-
- Meetings/operations: **23%**
|
|
53
|
-
- Security issues: **4%**
|
|
54
|
-
|
|
55
|
-
**Code Time Statistics** ([Software.com Report](https://www.software.com/reports/code-time-report)):
|
|
56
|
-
- Median active coding: **52 minutes/day**
|
|
57
|
-
- Top quartile focus time: **5.8 hours/day**
|
|
58
|
-
- Reading vs. writing ratio: **7:1 to 200:1**
|
|
59
|
-
|
|
60
|
-
**Microsoft Research** found that developers consider workdays "good" when they achieve balance between coding and collaboration, with "efficient use of time" cited by 54% as key ([Microsoft TSE 2019](https://www.microsoft.com/en-us/research/wp-content/uploads/2019/04/devtime-preprint-TSE19.pdf)).
|
|
61
|
-
|
|
62
|
-
**Microsoft Time Warp Study (2024)** surveyed 484 developers and found "significant deviations between a developer's ideal workweek and their actual workweek, with a clear correlation: as the gap between these two workweeks widens, there is a decline in both productivity and satisfaction" ([Microsoft Research 2024](https://www.microsoft.com/en-us/research/wp-content/uploads/2024/11/Time-Warp-Developer-Productivity-Study.pdf)).
|
|
63
|
-
|
|
64
|
-
**Atlassian 2025 DevEx Survey** of 3,500 developers found: "Developers only spend **16% of their time coding**, and coding is not a friction point for developers." Additionally, "63% of developers now say leaders don't understand their pain points, up sharply from 44% last year" ([Atlassian 2025](https://www.atlassian.com/blog/developer/developer-experience-report-2025)).
|
|
65
|
-
|
|
66
|
-
**JetBrains 2025** reports that 85% of developers regularly use AI tools, with 62% relying on at least one AI coding assistant. However, "developers prefer to stay in charge of creative and complex tasks, like debugging or designing application logic" ([JetBrains DevEco 2025](https://devecosystem-2025.jetbrains.com/)).
|
|
67
|
-
|
|
68
|
-
### 2.2 The Boilerplate Burden
|
|
69
|
-
|
|
70
|
-
A critical insight emerges: developers spend enormous time on repetitive, well-understood patterns:
|
|
71
|
-
|
|
72
|
-
> "As time spent on boilerplate code is reduced, developers focus more on improving the software's essential features." ([Iterators HQ](https://www.iteratorshq.com/blog/boilerplate-code-productivity-and-consistency-in-software-development/))
|
|
73
|
-
|
|
74
|
-
GitHub Octoverse 2025 confirms this pattern:
|
|
75
|
-
- **2.4M repositories** now use Notebooks (+75% YoY)
|
|
76
|
-
- **1.9M repositories** use Dockerfiles (+120% YoY)
|
|
77
|
-
- This growth is "fueled by the need to sandbox agents and LLMs" ([GitHub Octoverse 2025](https://github.blog/news-insights/octoverse/))
|
|
78
|
-
|
|
79
|
-
The implication is clear: **an agent that excels at boilerplate tasks (auth, CRUD, integrations) would deliver outsized value**.
|
|
80
|
-
|
|
81
|
-
### 2.3 Common Task Categories
|
|
82
|
-
|
|
83
|
-
Cross-referencing GitHub activity, Stack Overflow surveys, and developer productivity research, we identify six high-impact task categories:
|
|
84
|
-
|
|
85
|
-
| Category | % of Dev Work | Examples | Current Benchmark Coverage |
|
|
86
|
-
|----------|---------------|----------|---------------------------|
|
|
87
|
-
| **SaaS Features** | 25% | Auth, billing, settings | ❌ Minimal |
|
|
88
|
-
| **Glue Code** | 20% | Data transforms, API adapters | ❌ Minimal |
|
|
89
|
-
| **AI Integration** | 15% | RAG, function calling, embeddings | ❌ Minimal |
|
|
90
|
-
| **Frontend** | 15% | Components, forms, dashboards | ⚠️ HumanEval UI only |
|
|
91
|
-
| **API Integrations** | 15% | Webhooks, OAuth, third-party SDKs | ❌ Minimal |
|
|
92
|
-
| **Code Evolution** | 10% | Migrations, refactoring, upgrades | ⚠️ SWE-EVO partial |
|
|
93
|
-
|
|
94
|
-
---
|
|
95
|
-
|
|
96
|
-
## 3. VibeCodingBench Methodology
|
|
97
|
-
|
|
98
|
-
### 3.1 Design Principles
|
|
99
|
-
|
|
100
|
-
We propose five principles for ecological validity:
|
|
101
|
-
|
|
102
|
-
1. **Representative Tasks**: Mirror actual developer work distribution, not algorithmic puzzles
|
|
103
|
-
2. **Multi-Dimensional Scoring**: Evaluate quality, security, cost, and speed alongside correctness
|
|
104
|
-
3. **Production Realism**: Include Docker environments, real APIs, and security constraints
|
|
105
|
-
4. **Language Diversity**: Cover TypeScript, Python, Go, and Rust—not just Python
|
|
106
|
-
5. **Hot-Reload Extensibility**: Allow community task contributions without benchmark rebuilds
|
|
107
|
-
|
|
108
|
-
### 3.2 Evaluation Framework
|
|
109
|
-
|
|
110
|
-
Unlike pass-rate-only benchmarks, VibeCodingBench computes a weighted composite score:
|
|
111
|
-
|
|
112
|
-
```
|
|
113
|
-
Final Score = (Functional × 0.40) + (Visual × 0.20) + (Quality × 0.20)
|
|
114
|
-
- Cost_Penalty - Speed_Penalty
|
|
115
|
-
|
|
116
|
-
If Security_Critical_Fail: Final Score = 0
|
|
117
|
-
```
|
|
118
|
-
|
|
119
|
-
**Dimension Details**:
|
|
120
|
-
|
|
121
|
-
| Dimension | Weight | Measurement | Rationale |
|
|
122
|
-
|-----------|--------|-------------|-----------|
|
|
123
|
-
| Functional | 40% | Playwright E2E tests, Pass@k | Core correctness |
|
|
124
|
-
| Visual | 20% | Pixel diff vs reference design | UI fidelity matters |
|
|
125
|
-
| Quality | 20% | ESLint + Semgrep + complexity | Maintainability |
|
|
126
|
-
| Cost | 10% | Tokens used, API calls | Efficiency |
|
|
127
|
-
| Speed | 10% | Wall-clock time | Practical latency |
|
|
128
|
-
|
|
129
|
-
**Security Gate**: Any OWASP Top 10 vulnerability triggers automatic failure. As [Walturn research](https://www.walturn.com/insights/quantitative-evaluation-of-ai-code-generation-tools) notes: "Organizations should analyze security incident rates for AI-generated versus human code using CWE vulnerability classification."
|
|
130
|
-
|
|
131
|
-
### 3.3 Evaluation Methodology: Real Test Execution
|
|
132
|
-
|
|
133
|
-
VibeCodingBench uses **real test execution**, not pattern matching or fixed expected outputs. This approach reflects production reality where there are multiple valid implementations for any given requirement.
|
|
134
|
-
|
|
135
|
-
#### How It Works
|
|
136
|
-
|
|
137
|
-
1. **Task Submission**: Each task provides the agent with:
|
|
138
|
-
- `PROMPT.md`: Natural language requirements (what a human developer would receive)
|
|
139
|
-
- `manifest.yaml`: Task metadata, dependencies, and evaluation criteria
|
|
140
|
-
- Starter template code (one of 4 stacks: React/Vite, Next.js, Express, FastAPI)
|
|
141
|
-
|
|
142
|
-
2. **Agent Execution**: The agent writes code to satisfy the requirements. Unlike academic benchmarks with single correct answers, agents can choose their own:
|
|
143
|
-
- Implementation patterns (functional vs class-based, hooks vs context)
|
|
144
|
-
- Library choices (within template constraints)
|
|
145
|
-
- Code organization and file structure
|
|
146
|
-
|
|
147
|
-
3. **Real Test Execution**: The evaluator runs actual tests against the generated code:
|
|
148
|
-
|
|
149
|
-
```
|
|
150
|
-
┌─────────────────────────────────────────────────────────────────────┐
|
|
151
|
-
│ EVALUATION PIPELINE │
|
|
152
|
-
├─────────────────────────────────────────────────────────────────────┤
|
|
153
|
-
│ │
|
|
154
|
-
│ Agent Output ──► Build ──► Test Runners ──► Score Aggregation │
|
|
155
|
-
│ │ │
|
|
156
|
-
│ ▼ │
|
|
157
|
-
│ ┌─────────────────────────────────────────────────┐ │
|
|
158
|
-
│ │ PARALLEL TEST RUNNERS │ │
|
|
159
|
-
│ ├─────────────────────────────────────────────────┤ │
|
|
160
|
-
│ │ FunctionalRunner │ Playwright E2E tests │ │
|
|
161
|
-
│ │ QualityRunner │ ESLint + complexity │ │
|
|
162
|
-
│ │ VisualRunner │ Pixel diff vs reference │ │
|
|
163
|
-
│ │ SecurityRunner │ Semgrep OWASP Top 10 │ │
|
|
164
|
-
│ │ CostRunner │ Token usage tracking │ │
|
|
165
|
-
│ │ SpeedRunner │ Wall-clock timing │ │
|
|
166
|
-
│ └─────────────────────────────────────────────────┘ │
|
|
167
|
-
│ │
|
|
168
|
-
└─────────────────────────────────────────────────────────────────────┘
|
|
169
|
-
```
|
|
170
|
-
|
|
171
|
-
#### Test Execution Details
|
|
172
|
-
|
|
173
|
-
**Functional Tests (40% weight)**
|
|
174
|
-
- Real Playwright browser automation
|
|
175
|
-
- Tests verify observable behavior, not implementation details
|
|
176
|
-
- Example: "User can log in" tests that login works, not that you used a specific auth library
|
|
177
|
-
- Pass rate calculated as: (passed tests / total tests) × 100
|
|
178
|
-
|
|
179
|
-
**Quality Analysis (20% weight)**
|
|
180
|
-
- ESLint with strict configuration
|
|
181
|
-
- Cyclomatic complexity analysis
|
|
182
|
-
- Code duplication detection
|
|
183
|
-
- TypeScript strict mode compliance (for TS tasks)
|
|
184
|
-
|
|
185
|
-
**Visual Regression (20% weight)**
|
|
186
|
-
- Screenshot comparison against reference designs
|
|
187
|
-
- Pixel-diff with configurable tolerance (default 5%)
|
|
188
|
-
- Layout validation for responsive breakpoints
|
|
189
|
-
|
|
190
|
-
**Security Gate (Binary)**
|
|
191
|
-
- Semgrep with OWASP Top 10 ruleset
|
|
192
|
-
- Critical vulnerabilities (SQL injection, XSS, command injection) → automatic 0 score
|
|
193
|
-
- Warnings logged but don't fail the task
|
|
194
|
-
|
|
195
|
-
#### Flexible vs Deterministic
|
|
196
|
-
|
|
197
|
-
VibeCodingBench is **intentionally flexible** on implementation while **deterministic on outcomes**:
|
|
198
|
-
|
|
199
|
-
| Aspect | Flexible | Deterministic |
|
|
200
|
-
|--------|----------|---------------|
|
|
201
|
-
| Code style | ✅ Agent's choice | - |
|
|
202
|
-
| Library selection | ✅ Within constraints | - |
|
|
203
|
-
| File organization | ✅ Agent's choice | - |
|
|
204
|
-
| Functional behavior | - | ✅ Must pass tests |
|
|
205
|
-
| Security compliance | - | ✅ Must pass scan |
|
|
206
|
-
| Visual fidelity | - | ✅ Must match reference |
|
|
207
|
-
|
|
208
|
-
This mirrors real development: multiple valid solutions exist, but acceptance criteria are fixed.
|
|
209
|
-
|
|
210
|
-
#### Task Structure Example
|
|
211
|
-
|
|
212
|
-
```yaml
|
|
213
|
-
# manifest.yaml
|
|
214
|
-
name: stripe-webhook-handler
|
|
215
|
-
category: api-integrations
|
|
216
|
-
difficulty: medium
|
|
217
|
-
template: express-ts
|
|
218
|
-
timeout: 300s
|
|
219
|
-
|
|
220
|
-
evaluation:
|
|
221
|
-
functional:
|
|
222
|
-
test_file: tests/webhook.spec.ts
|
|
223
|
-
pass_threshold: 0.8
|
|
224
|
-
quality:
|
|
225
|
-
eslint_config: .eslintrc.strict.json
|
|
226
|
-
max_complexity: 10
|
|
227
|
-
security:
|
|
228
|
-
ruleset: owasp-top-10
|
|
229
|
-
fail_on: [critical, high]
|
|
230
|
-
visual: null # No visual tests for API tasks
|
|
231
|
-
|
|
232
|
-
dependencies:
|
|
233
|
-
- stripe: ^14.0.0
|
|
234
|
-
- express: ^4.18.0
|
|
235
|
-
```
|
|
236
|
-
|
|
237
|
-
The evaluator reads this manifest, spins up the appropriate test environment, executes all runners, and aggregates scores using the weighted formula.
|
|
238
|
-
|
|
239
|
-
### 3.4 Task Taxonomy
|
|
240
|
-
|
|
241
|
-
VibeCodingBench organizes tasks into six categories aligned with developer workflow data:
|
|
242
|
-
|
|
243
|
-
**SaaS Core (25% weight)**
|
|
244
|
-
- OAuth integration (Google, GitHub, Microsoft)
|
|
245
|
-
- Multi-factor authentication (TOTP)
|
|
246
|
-
- CRUD dashboards with filtering/pagination
|
|
247
|
-
- User preference systems
|
|
248
|
-
|
|
249
|
-
**Glue Code (20% weight)**
|
|
250
|
-
- Data format transformations (Excel → JSON, CSV → SQL)
|
|
251
|
-
- API adapters (REST → GraphQL)
|
|
252
|
-
- Message queue consumers
|
|
253
|
-
- Cron job implementations
|
|
254
|
-
|
|
255
|
-
**AI Integration (20% weight)**
|
|
256
|
-
- RAG chatbot with PDF ingestion
|
|
257
|
-
- Structured output extraction (invoices, receipts)
|
|
258
|
-
- Function calling implementations
|
|
259
|
-
- Embedding-based search
|
|
260
|
-
|
|
261
|
-
**Frontend (15% weight)**
|
|
262
|
-
- Figma-to-code translation
|
|
263
|
-
- Dynamic form builders
|
|
264
|
-
- Analytics dashboards with charts
|
|
265
|
-
- Responsive component libraries
|
|
266
|
-
|
|
267
|
-
**API Integrations (10% weight)**
|
|
268
|
-
- Payment webhooks (Stripe, Paddle)
|
|
269
|
-
- Email service integration (SendGrid, Resend)
|
|
270
|
-
- Cloud storage (S3, GCS)
|
|
271
|
-
- Third-party OAuth consumers
|
|
272
|
-
|
|
273
|
-
**Code Evolution (10% weight)**
|
|
274
|
-
- Framework migrations (Express → Fastify)
|
|
275
|
-
- Class components → Hooks refactoring
|
|
276
|
-
- Monolith decomposition
|
|
277
|
-
- Dependency upgrades
|
|
278
|
-
|
|
279
|
-
---
|
|
280
|
-
|
|
281
|
-
## 4. Addressing Benchmark Contamination
|
|
282
|
-
|
|
283
|
-
A persistent concern with AI benchmarks is data contamination—models may have seen solutions during training. VibeCodingBench mitigates this through:
|
|
284
|
-
|
|
285
|
-
1. **Novel Task Design**: Tasks are synthesized from common patterns, not extracted from existing repositories
|
|
286
|
-
2. **Template Variation**: Each task can be instantiated with different naming, structure, and requirements
|
|
287
|
-
3. **Temporal Freshness**: New tasks can be added post-training cutoff
|
|
288
|
-
4. **Private Test Suites**: Full test implementations are withheld from public repository
|
|
289
|
-
|
|
290
|
-
As OpenAI acknowledged regarding SWE-bench: "existing benchmarks are often contaminated by LLM pretraining data" ([OpenAI SWE-bench Verified](https://openai.com/index/introducing-swe-bench-verified/)).
|
|
291
|
-
|
|
292
|
-
---
|
|
293
|
-
|
|
294
|
-
## 5. Why This Matters
|
|
295
|
-
|
|
296
|
-
### 5.1 The AI Productivity Paradox
|
|
297
|
-
|
|
298
|
-
Multiple studies reveal a striking paradox between AI promise and reality:
|
|
299
|
-
|
|
300
|
-
**GitHub Octoverse 2025**:
|
|
301
|
-
> "Developers expect 24% productivity gains but experience 19% slowdowns in controlled conditions. Less than 44% of AI-generated code is accepted without modification."
|
|
302
|
-
|
|
303
|
-
**METR Randomized Controlled Trial (2025)**: A rigorous study of 16 experienced developers completing 246 tasks found that "allowing AI actually increased completion time by 19%"—despite developers forecasting a 24% reduction ([METR 2025](https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/)).
|
|
304
|
-
|
|
305
|
-
**McKinsey Research (2025)**: While optimistic about AI potential, McKinsey found that "true impact comes when adoption is deep and organization-wide. Companies with 80-100% developer adoption saw gains of more than 110%." However, partial adoption shows minimal gains ([McKinsey 2025](https://www.mckinsey.com/capabilities/tech-and-ai/our-insights/unleashing-developer-productivity-with-generative-ai)).
|
|
306
|
-
|
|
307
|
-
**Stack Overflow 2025** corroborates the frustration:
|
|
308
|
-
> "70% of agent users agree that agents have reduced time spent on specific development tasks"—but 66% cite "AI solutions that are almost right, but not quite" as their biggest frustration, and 45% say "debugging AI code is more time-consuming."
|
|
309
|
-
|
|
310
|
-
The gap between expectation and reality stems from **evaluation-deployment mismatch**: we optimize for the wrong signals.
|
|
311
|
-
|
|
312
|
-
### 5.2 The Path Forward
|
|
313
|
-
|
|
314
|
-
VibeCodingBench aims to close this gap by:
|
|
315
|
-
|
|
316
|
-
1. **Measuring what matters**: Tasks that occupy developer time, not academic exercises
|
|
317
|
-
2. **Penalizing hidden costs**: A fast solution with security flaws scores zero
|
|
318
|
-
3. **Enabling comparison**: Standardized multi-agent evaluation with transparent methodology
|
|
319
|
-
4. **Supporting iteration**: Hot-reload task addition for continuous benchmark evolution
|
|
320
|
-
|
|
321
|
-
---
|
|
322
|
-
|
|
323
|
-
## 6. Initial Results (January 2026)
|
|
324
|
-
|
|
325
|
-
We evaluated 15 leading AI coding models across all 180 VibeCodingBench tasks. The results reveal important insights about the current state of AI coding agents.
|
|
326
|
-
|
|
327
|
-
### 6.1 Leaderboard Summary
|
|
328
|
-
|
|
329
|
-
| Rank | Model | Final Score | Pass Rate | Cost | Avg Time |
|
|
330
|
-
|------|-------|-------------|-----------|------|----------|
|
|
331
|
-
| #1 | Claude Opus 4.5 | 89.2% | 100.0% (180/180) | $12.31 | 44s |
|
|
332
|
-
| #2 | Claude Haiku 4.5 | 89.0% | 99.4% (179/180) | $3.03 | 22s |
|
|
333
|
-
| #3 | Grok 4 Fast | 88.8% | 98.9% (178/180) | $0.21 | 70s |
|
|
334
|
-
| #4 | OpenAI GPT-5.2 | 88.8% | 98.3% (177/180) | $5.01 | 28s |
|
|
335
|
-
| #5 | Qwen3 Max | 88.6% | 100.0% (180/180) | $5.42 | 45s |
|
|
336
|
-
| #6 | Claude Sonnet 4.5 | 88.6% | 98.3% (177/180) | $6.98 | 42s |
|
|
337
|
-
| #7 | GLM 4-Plus | 88.2% | 98.9% (178/180) | $0.93 | 96s |
|
|
338
|
-
| #8 | DeepSeek v3.2 | 88.2% | 98.3% (177/180) | $0.50 | 90s |
|
|
339
|
-
| #9 | Grok 4 | 88.0% | 97.8% (176/180) | $5.47 | 75s |
|
|
340
|
-
| #10 | MiniMax M2.1 | 87.4% | 99.4% (179/180) | $2.40 | 165s |
|
|
341
|
-
| #11 | Grok 4.1 Fast | 86.8% | 97.2% (175/180) | $0.24 | 89s |
|
|
342
|
-
| #12 | Gemini 3 Pro Preview | 85.8% | 95.0% (171/180) | $10.34 | 32s |
|
|
343
|
-
| #13 | GLM-4.7 | 83.9% | 85.6% (154/180) | $0.73 | 57s |
|
|
344
|
-
| #14 | GLM 4.7 Flash | 83.8% | 92.8% (167/180) | $1.11 | 45s |
|
|
345
|
-
| #15 | Gemini 3 Flash | 83.4% | 92.2% (166/180) | $0.86 | 28s |
|
|
346
|
-
|
|
347
|
-
### 6.2 Key Findings
|
|
348
|
-
|
|
349
|
-
**1. Tight Competition at the Top**
|
|
350
|
-
The top 10 models are clustered within 2 percentage points (87.4%–89.2%). This narrow spread suggests frontier models have converged on similar capabilities for production coding tasks.
|
|
351
|
-
|
|
352
|
-
**2. Cost-Performance Tradeoffs**
|
|
353
|
-
Cost varies by 60x between models at similar performance levels:
|
|
354
|
-
- **Best value**: Grok 4 Fast ($0.21 total, 88.8% score) and DeepSeek v3.2 ($0.50, 88.2%)
|
|
355
|
-
- **Premium tier**: Claude Opus 4.5 ($12.31) and Gemini 3 Pro ($10.34)
|
|
356
|
-
- **Middle ground**: Claude Haiku 4.5 ($3.03, 89.0%) offers excellent cost-performance balance
|
|
357
|
-
|
|
358
|
-
**3. Speed vs Quality**
|
|
359
|
-
Faster models don't necessarily score lower:
|
|
360
|
-
- Claude Haiku 4.5: 22s average, 89.0% score
|
|
361
|
-
- Gemini 3 Flash: 28s average, 83.4% score
|
|
362
|
-
- MiniMax M2.1: 165s average, 87.4% score
|
|
363
|
-
|
|
364
|
-
**4. Pass Rate ≠ Final Score**
|
|
365
|
-
Our multi-dimensional scoring reveals quality differences hidden by pass rate:
|
|
366
|
-
- Qwen3 Max: 100% pass rate but 88.6% final score (quality/cost penalties)
|
|
367
|
-
- Claude Opus 4.5: 100% pass rate and 89.2% final score (better quality metrics)
|
|
368
|
-
|
|
369
|
-
**5. All Models Pass Security Gate**
|
|
370
|
-
Every evaluated model achieved 100% on security checks—a positive signal that modern LLMs avoid obvious OWASP vulnerabilities in generated code.
|
|
371
|
-
|
|
372
|
-
### 6.3 Dimensional Breakdown
|
|
373
|
-
|
|
374
|
-
| Model | Functional | Quality | Visual | Cost Score | Speed Score |
|
|
375
|
-
|-------|------------|---------|--------|------------|-------------|
|
|
376
|
-
| Claude Opus 4.5 | 85.0% | 80.0% | 80.0% | 70% | 80% |
|
|
377
|
-
| Claude Haiku 4.5 | 84.5% | 79.6% | 80.0% | 88% | 95% |
|
|
378
|
-
| Grok 4 Fast | 84.1% | 80.0% | 80.0% | 94% | 72% |
|
|
379
|
-
| Qwen3 Max | 85.0% | 80.0% | 80.0% | 90% | 75% |
|
|
380
|
-
| DeepSeek v3.2 | 83.6% | 80.0% | 80.0% | 96% | 65% |
|
|
381
|
-
| GLM-4.7 | 72.7% | 79.6% | 80.0% | 94% | 82% |
|
|
382
|
-
| Gemini 3 Flash | 78.4% | 75.1% | 80.0% | 95% | 90% |
|
|
383
|
-
|
|
384
|
-
**Observations**:
|
|
385
|
-
- Functional scores range from 72.7% to 85.0%—a 12-point spread revealing real capability differences
|
|
386
|
-
- Quality scores cluster around 79-80%, suggesting similar code hygiene across models
|
|
387
|
-
- Visual compliance is uniformly high (80%), indicating models handle UI requirements well
|
|
388
|
-
- Cost efficiency varies dramatically based on token usage and pricing
|
|
389
|
-
|
|
390
|
-
---
|
|
391
|
-
|
|
392
|
-
## 7. Conclusion
|
|
393
|
-
|
|
394
|
-
The current generation of AI coding agent benchmarks optimizes for the wrong objective. High SWE-bench scores do not translate to production readiness. Pass rates ignore quality debt. Single-repository evaluations miss the polyglot reality of modern development.
|
|
395
|
-
|
|
396
|
-
VibeCodingBench proposes a return to first principles: measure what developers actually do, evaluate what production actually requires, and score what businesses actually value. By grounding our benchmark in empirical workflow data and multi-dimensional evaluation, we aim to create a signal that predicts real-world utility—not just benchmark leaderboard position.
|
|
397
|
-
|
|
398
|
-
The ultimate test of an AI coding agent is not whether it can fix a bug in scikit-learn. It is whether it can help a developer ship a feature their users need, safely and efficiently, before the sprint ends.
|
|
399
|
-
|
|
400
|
-
---
|
|
401
|
-
|
|
402
|
-
## References
|
|
403
|
-
|
|
404
|
-
### Industry Reports
|
|
405
|
-
|
|
406
|
-
1. GitHub. (2025). *Octoverse 2025: The State of Open Source*. https://octoverse.github.com/
|
|
407
|
-
2. GitHub Blog. (2025). *What 986 million code pushes say about the developer workflow in 2025*. https://github.blog/news-insights/octoverse/what-986-million-code-pushes-say-about-the-developer-workflow-in-2025/
|
|
408
|
-
3. Stack Overflow. (2025). *2025 Developer Survey*. https://survey.stackoverflow.co/2025/
|
|
409
|
-
4. Stack Overflow Blog. (2025). *Developers remain willing but reluctant to use AI*. https://stackoverflow.blog/2025/12/29/developers-remain-willing-but-reluctant-to-use-ai-the-2025-developer-survey-results-are-here
|
|
410
|
-
5. Qodo. (2025). *State of AI Code Quality in 2025*. https://www.qodo.ai/reports/state-of-ai-code-quality/
|
|
411
|
-
6. Worklytics. (2025). *Software Engineering Productivity Benchmarks 2025*. https://www.worklytics.co/resources/software-engineering-productivity-benchmarks-2025-good-scores
|
|
412
|
-
|
|
413
|
-
### Academic Research
|
|
414
|
-
|
|
415
|
-
7. Meyer, A., et al. (2019). *Today was a Good Day: The Daily Life of Software Developers*. IEEE Transactions on Software Engineering. https://www.microsoft.com/en-us/research/wp-content/uploads/2019/04/devtime-preprint-TSE19.pdf
|
|
416
|
-
8. SWE-EVO Authors. (2025). *SWE-EVO: Benchmarking Coding Agents in Long-Horizon Software Evolution Scenarios*. arXiv:2512.18470. https://arxiv.org/abs/2512.18470
|
|
417
|
-
9. SWE-bench Pro Authors. (2025). *SWE-Bench Pro: Can AI Agents Solve Long-Horizon Software Engineering Tasks?*. arXiv:2509.16941. https://arxiv.org/abs/2509.16941
|
|
418
|
-
10. Security Analysis. (2025). *Assessing the Quality and Security of AI-Generated Code: A Quantitative Analysis*. arXiv:2508.14727. https://arxiv.org/abs/2508.14727
|
|
419
|
-
|
|
420
|
-
### Benchmark Analysis
|
|
421
|
-
|
|
422
|
-
11. OpenAI. (2024). *Introducing SWE-bench Verified*. https://openai.com/index/introducing-swe-bench-verified/
|
|
423
|
-
12. Runloop AI. (2025). *SWE-Bench Deep Dive: Unmasking the Limitations of a Popular Benchmark*. https://runloop.ai/blog/swe-bench-deep-dive-unmasking-the-limitations-of-a-popular-benchmark
|
|
424
|
-
13. Scale AI. (2025). *SWE-Bench Pro Leaderboard*. https://scale.com/leaderboard/swe_bench_pro_public
|
|
425
|
-
14. Amazon AWS. (2025). *Amazon introduces SWE-PolyBench*. https://aws.amazon.com/blogs/devops/amazon-introduces-swe-polybench-a-multi-lingual-benchmark-for-ai-coding-agents/
|
|
426
|
-
|
|
427
|
-
### Developer Productivity
|
|
428
|
-
|
|
429
|
-
15. Sonar. (2025). *How much time do developers spend actually writing code?*. https://www.sonarsource.com/blog/how-much-time-do-developers-spend-actually-writing-code/
|
|
430
|
-
16. Software.com. (2025). *Code Time Report*. https://www.software.com/reports/code-time-report
|
|
431
|
-
17. DX. (2025). *Measuring AI code assistants and agents*. https://getdx.com/research/measuring-ai-code-assistants-and-agents/
|
|
432
|
-
18. Augment Code. (2025). *How to Test AI Coding Assistants: 7 Enterprise Benchmarks*. https://www.augmentcode.com/guides/how-to-test-ai-coding-assistants-7-enterprise-benchmarks
|
|
433
|
-
19. Microsoft Research. (2024). *Time Warp: The Gap Between Developers' Ideal vs Actual Workweek*. https://www.microsoft.com/en-us/research/wp-content/uploads/2024/11/Time-Warp-Developer-Productivity-Study.pdf
|
|
434
|
-
20. JetBrains. (2025). *State of Developer Ecosystem 2025*. https://devecosystem-2025.jetbrains.com/
|
|
435
|
-
21. Atlassian. (2025). *2025 State of Developer Experience Report*. https://www.atlassian.com/blog/developer/developer-experience-report-2025
|
|
436
|
-
22. METR. (2025). *Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity*. https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/
|
|
437
|
-
23. McKinsey. (2025). *Unleashing Developer Productivity with Generative AI*. https://www.mckinsey.com/capabilities/tech-and-ai/our-insights/unleashing-developer-productivity-with-generative-ai
|
|
438
|
-
|
|
439
|
-
---
|
|
440
|
-
|
|
441
|
-
*VibeCodingBench is a project by Alt Research. Contributions welcome at https://github.com/alt-research/vibe-coding-benchmark-public*
|
|
@@ -1,138 +0,0 @@
|
|
|
1
|
-
# Code Evolution Category
|
|
2
|
-
|
|
3
|
-
Research shows 80% of real engineering work is maintenance/evolution, not greenfield development. The SWE-EVO benchmark demonstrates that agents score only 21% on evolution tasks vs 65% on simple fixes, highlighting a critical gap in current benchmarks.
|
|
4
|
-
|
|
5
|
-
## Motivation
|
|
6
|
-
|
|
7
|
-
Most existing benchmarks focus on:
|
|
8
|
-
- Bug fixes with clear error messages
|
|
9
|
-
- Greenfield implementations with no constraints
|
|
10
|
-
- Isolated changes without broader context
|
|
11
|
-
|
|
12
|
-
Real-world developers primarily:
|
|
13
|
-
- Migrate legacy codebases to modern frameworks
|
|
14
|
-
- Extend features in existing, complex systems
|
|
15
|
-
- Refactor code while maintaining backwards compatibility
|
|
16
|
-
- Review PRs and suggest improvements
|
|
17
|
-
|
|
18
|
-
## Subcategories
|
|
19
|
-
|
|
20
|
-
### 1. Legacy Migration (`legacy-migration/`)
|
|
21
|
-
|
|
22
|
-
Tasks involving upgrading dependencies or migrating between frameworks while preserving functionality.
|
|
23
|
-
|
|
24
|
-
**Examples:**
|
|
25
|
-
- jQuery → React migration
|
|
26
|
-
- Express → Fastify migration
|
|
27
|
-
- Class components → Hooks
|
|
28
|
-
- CommonJS → ESM
|
|
29
|
-
- Upgrade from deprecated APIs
|
|
30
|
-
|
|
31
|
-
**Evaluation Criteria:**
|
|
32
|
-
- All existing tests pass
|
|
33
|
-
- No functional regressions
|
|
34
|
-
- Migration completeness (no legacy remnants)
|
|
35
|
-
- Code quality improvement
|
|
36
|
-
|
|
37
|
-
### 2. Feature Extension (`feature-extension/`)
|
|
38
|
-
|
|
39
|
-
Adding new features to existing codebases, not greenfield. The agent must understand and integrate with existing architecture.
|
|
40
|
-
|
|
41
|
-
**Examples:**
|
|
42
|
-
- Add pagination to existing list endpoint
|
|
43
|
-
- Add caching layer to existing service
|
|
44
|
-
- Add audit logging to existing operations
|
|
45
|
-
- Add rate limiting to existing API
|
|
46
|
-
|
|
47
|
-
**Evaluation Criteria:**
|
|
48
|
-
- Feature works correctly
|
|
49
|
-
- Integrates with existing patterns
|
|
50
|
-
- Doesn't break existing functionality
|
|
51
|
-
- Follows codebase conventions
|
|
52
|
-
|
|
53
|
-
### 3. Refactor & Preserve (`refactor-preserve/`)
|
|
54
|
-
|
|
55
|
-
Restructuring code while maintaining all existing tests and behavior. Emphasizes understanding invariants.
|
|
56
|
-
|
|
57
|
-
**Examples:**
|
|
58
|
-
- Extract service from monolith
|
|
59
|
-
- Convert sync to async operations
|
|
60
|
-
- Improve error handling patterns
|
|
61
|
-
- Reduce cyclomatic complexity
|
|
62
|
-
|
|
63
|
-
**Evaluation Criteria:**
|
|
64
|
-
- All tests still pass
|
|
65
|
-
- Same API surface (unless explicitly changing)
|
|
66
|
-
- Measurable improvement (complexity, coverage, etc.)
|
|
67
|
-
- No behavioral changes
|
|
68
|
-
|
|
69
|
-
### 4. Code Review (`code-review/`)
|
|
70
|
-
|
|
71
|
-
Given a PR diff, identify issues, suggest fixes, and potentially implement corrections.
|
|
72
|
-
|
|
73
|
-
**Examples:**
|
|
74
|
-
- Security vulnerability detection
|
|
75
|
-
- Performance issue identification
|
|
76
|
-
- Logic bug detection
|
|
77
|
-
- Style/pattern consistency review
|
|
78
|
-
|
|
79
|
-
**Evaluation Criteria:**
|
|
80
|
-
- Issues correctly identified
|
|
81
|
-
- False positive rate
|
|
82
|
-
- Fix suggestions are correct
|
|
83
|
-
- Explanation quality
|
|
84
|
-
|
|
85
|
-
## Task Structure
|
|
86
|
-
|
|
87
|
-
Each task in code-evolution provides:
|
|
88
|
-
|
|
89
|
-
```yaml
|
|
90
|
-
name: "Migration/Feature Name"
|
|
91
|
-
category: code-evolution
|
|
92
|
-
subcategory: legacy-migration | feature-extension | refactor-preserve | code-review
|
|
93
|
-
difficulty: easy | medium | hard
|
|
94
|
-
baseCode: path/to/existing/codebase
|
|
95
|
-
tests:
|
|
96
|
-
existing: tests/ # Must continue to pass
|
|
97
|
-
new: tests/new/ # New tests for added functionality
|
|
98
|
-
constraints:
|
|
99
|
-
- "Must not modify public API"
|
|
100
|
-
- "All existing tests must pass"
|
|
101
|
-
- "No new dependencies unless approved"
|
|
102
|
-
```
|
|
103
|
-
|
|
104
|
-
## Scoring
|
|
105
|
-
|
|
106
|
-
Evolution tasks use weighted scoring:
|
|
107
|
-
|
|
108
|
-
| Dimension | Weight | Description |
|
|
109
|
-
|-----------|--------|-------------|
|
|
110
|
-
| Existing Tests | 40% | All pre-existing tests must pass |
|
|
111
|
-
| New Functionality | 30% | New feature/migration works correctly |
|
|
112
|
-
| Code Quality | 15% | Maintainability, complexity reduction |
|
|
113
|
-
| Integration | 15% | Follows existing patterns, conventions |
|
|
114
|
-
|
|
115
|
-
## Example Task: Express to Fastify Migration
|
|
116
|
-
|
|
117
|
-
```yaml
|
|
118
|
-
name: Express to Fastify Migration
|
|
119
|
-
category: code-evolution
|
|
120
|
-
subcategory: legacy-migration
|
|
121
|
-
difficulty: medium
|
|
122
|
-
baseCode: examples/express-api/
|
|
123
|
-
stack: nodejs
|
|
124
|
-
|
|
125
|
-
description: |
|
|
126
|
-
Migrate a REST API from Express.js to Fastify while
|
|
127
|
-
maintaining all existing functionality and API contracts.
|
|
128
|
-
|
|
129
|
-
constraints:
|
|
130
|
-
- All existing integration tests must pass
|
|
131
|
-
- API endpoints must remain identical
|
|
132
|
-
- Response formats must match exactly
|
|
133
|
-
- Error handling patterns must be preserved
|
|
134
|
-
|
|
135
|
-
tests:
|
|
136
|
-
existing: tests/integration/
|
|
137
|
-
new: tests/fastify-specific/
|
|
138
|
-
```
|