pumuki 6.3.137 → 6.3.139

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -6,6 +6,19 @@ This project follows [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
6
6
 
7
7
  ## [Unreleased]
8
8
 
9
+ ## [6.3.139] - 2026-05-05
10
+
11
+ ### Fixed
12
+
13
+ - **PUMUKI-INC-060 baseline test gate:** la evidencia TDD/BDD exige ahora un baseline test `passed` por slice antes del evento `red`; si falta o falla, el gate bloquea con `TDD_BASELINE_TEST_REQUIRED` o `TDD_BASELINE_TEST_MUST_PASS`.
14
+ - **Alineación all-severities en evidencia:** los tests de evidencia quedan sincronizados con el contrato publicado de bloqueo por cualquier severidad (`BLOCK` / `BLOCKED`), incluyendo findings `WARN`.
15
+
16
+ ## [6.3.138] - 2026-05-05
17
+
18
+ ### Fixed
19
+
20
+ - **Doc-only evidence hygiene:** en commits documentales, `.ai_evidence.json` trackeado queda restaurado a `HEAD` y no deja modificaciones de hook que hagan fallar integraciones `pre-commit` con `files were modified by this hook`.
21
+
9
22
  ## [6.3.137] - 2026-05-05
10
23
 
11
24
  ### Fixed
@@ -4,6 +4,17 @@ This file tracks the active deterministic framework line used in this repository
4
4
  Canonical release chronology lives in `CHANGELOG.md`.
5
5
  This file keeps only the operational highlights and rollout notes that matter while running the framework.
6
6
 
7
+ ### 2026-05-05 (v6.3.139)
8
+
9
+ - **Baseline tests antes de editar:** Pumuki bloquea cambios TDD/BDD in-scope si cada slice no registra un baseline test pasado antes de RED.
10
+ - **RuralGo INC060:** cierra el gap por el que una regresión de test preexistente en un componente relacionado podía pasar desapercibida antes de iterar.
11
+ - **Rollout:** publicar `pumuki@6.3.139`, repinear RuralGo primero y revalidar `PRE_WRITE`, `PRE_COMMIT`, `PRE_PUSH` y `CI`.
12
+
13
+ ### 2026-05-05 (v6.3.138)
14
+
15
+ - **Doc-only sin worktree sucio:** cuando el índice solo contiene `*.md` / `*.mdx`, Pumuki restaura `.ai_evidence.json` trackeado a `HEAD` tras refrescar el gate, evitando el fallo del framework `pre-commit` por “files were modified by this hook”.
16
+ - **RuralGo follow-up:** corrige el bloqueo observado al commitear el feedback `docs/technical/08-validation/refactor/pumuki-integration-feedback.md` después del repin.
17
+
7
18
  ### 2026-05-05 (v6.3.137)
8
19
 
9
20
  - **Atomicidad compatible con evidencia gestionada:** `.ai_evidence.json` / `.AI_EVIDENCE.json` ya no cuentan como scope o fichero funcional en `git-atomicity`, de modo que un commit de repin no queda bloqueado solo porque el hook refresque evidencia trackeada.
@@ -346,6 +346,7 @@ test('writeEvidence preserva snapshot.tdd_bdd cuando viene en evidencia', async
346
346
  slices_invalid: 0,
347
347
  integrity_ok: true,
348
348
  errors: [],
349
+ baseline: { required: true, passed: 0, missing: 0, failed: 0 },
349
350
  },
350
351
  waiver: {
351
352
  applied: true,
@@ -272,6 +272,12 @@ const normalizeTddBddSnapshot = (snapshot: TddBddSnapshot | undefined): TddBddSn
272
272
  slices_invalid: snapshot.evidence.slices_invalid,
273
273
  integrity_ok: snapshot.evidence.integrity_ok,
274
274
  errors: [...snapshot.evidence.errors],
275
+ baseline: {
276
+ required: snapshot.evidence.baseline.required,
277
+ passed: snapshot.evidence.baseline.passed,
278
+ missing: snapshot.evidence.baseline.missing,
279
+ failed: snapshot.evidence.baseline.failed,
280
+ },
275
281
  },
276
282
  waiver: {
277
283
  applied: snapshot.waiver.applied,
@@ -20,9 +20,7 @@ import {
20
20
  } from '../notifications/emitAuditSummaryNotification';
21
21
  import { existsSync, readFileSync, unlinkSync, writeFileSync } from 'node:fs';
22
22
  import { join } from 'node:path';
23
- import { buildEvidenceOperationalHints } from '../evidence/operationalHints';
24
23
  import { readEvidence, readEvidenceResult } from '../evidence/readEvidence';
25
- import { writeEvidence } from '../evidence/writeEvidence';
26
24
  import type { EvidenceReadResult } from '../evidence/readEvidence';
27
25
  import type { SnapshotFinding } from '../evidence/schema';
28
26
  import { ensureRuntimeArtifactsIgnored } from '../lifecycle/artifacts';
@@ -135,6 +133,7 @@ type StageRunnerDependencies = {
135
133
  isPathTracked: (repoRoot: string, relativePath: string) => boolean;
136
134
  listStagedIndexPaths: (repoRoot: string) => ReadonlyArray<string>;
137
135
  stagePath: (repoRoot: string, relativePath: string) => void;
136
+ restorePathFromHead: (repoRoot: string, relativePath: string) => void;
138
137
  resolveHeadOid: (repoRoot: string) => string | null;
139
138
  resolveGitAtomicityEnforcement: () => GitAtomicityEnforcementResolution;
140
139
  };
@@ -223,6 +222,9 @@ const defaultDependencies: StageRunnerDependencies = {
223
222
  stagePath: (repoRoot, relativePath) => {
224
223
  new GitService().runGit(['add', '--', relativePath], repoRoot);
225
224
  },
225
+ restorePathFromHead: (repoRoot, relativePath) => {
226
+ new GitService().runGit(['checkout', '--', relativePath], repoRoot);
227
+ },
226
228
  resolveHeadOid: (repoRoot) => {
227
229
  try {
228
230
  return new GitService().runGit(['rev-parse', 'HEAD'], repoRoot).trim();
@@ -507,26 +509,6 @@ const runHookGateWithPolicyRetry = async (params: {
507
509
  }
508
510
  };
509
511
 
510
- const patchOperationalHintsAfterDocumentationOnlyEvidenceSync = (repoRoot: string): void => {
511
- const evidenceRead = readEvidenceResult(repoRoot);
512
- if (evidenceRead.kind !== 'valid') {
513
- return;
514
- }
515
- const evidence = evidenceRead.evidence;
516
- const hints = buildEvidenceOperationalHints({
517
- stage: evidence.snapshot.stage,
518
- outcome: evidence.snapshot.outcome,
519
- findings: evidence.snapshot.findings,
520
- rulesCoverage: evidence.snapshot.rules_coverage,
521
- evaluationMetrics: evidence.snapshot.evaluation_metrics,
522
- extra: {
523
- requires_second_pass: true,
524
- second_pass_reason: 'tracked_evidence_refreshed_on_disk_not_staged_documentation_only_commit',
525
- },
526
- });
527
- writeEvidence({ ...evidence, operational_hints: hints }, { repoRoot });
528
- };
529
-
530
512
  const syncTrackedEvidenceAfterSuccessfulPreCommit = (params: {
531
513
  dependencies: StageRunnerDependencies;
532
514
  repoRoot: string;
@@ -547,12 +529,11 @@ const syncTrackedEvidenceAfterSuccessfulPreCommit = (params: {
547
529
  ) {
548
530
  if (!params.dependencies.isQuietMode()) {
549
531
  process.stderr.write(
550
- `[pumuki][evidence-sync] tracked ${EVIDENCE_FILE_PATH} updated on disk but not auto-staged (documentation-only staged paths: *.md / *.mdx). ` +
551
- `Include in this commit if needed: git add -- ${EVIDENCE_FILE_PATH}. ` +
532
+ `[pumuki][evidence-sync] tracked ${EVIDENCE_FILE_PATH} left unchanged for documentation-only staged paths (*.md / *.mdx). ` +
552
533
  `Force previous behavior: PUMUKI_PRE_COMMIT_ALWAYS_RESTAGE_TRACKED_EVIDENCE=1\n`
553
534
  );
554
535
  }
555
- patchOperationalHintsAfterDocumentationOnlyEvidenceSync(params.repoRoot);
536
+ params.dependencies.restorePathFromHead(params.repoRoot, EVIDENCE_FILE_PATH);
556
537
  return false;
557
538
  }
558
539
  try {
@@ -14,6 +14,7 @@ const tddEventSchema = z.object({
14
14
  const tddSliceSchema = z.object({
15
15
  id: z.string().min(1),
16
16
  scenario_ref: z.string().min(1),
17
+ baseline: tddEventSchema.optional(),
17
18
  red: tddEventSchema,
18
19
  green: tddEventSchema,
19
20
  refactor: tddEventSchema,
@@ -104,6 +104,12 @@ export const enforceTddBddPolicy = (params: {
104
104
  slices_invalid: 0,
105
105
  integrity_ok: true,
106
106
  errors: [],
107
+ baseline: {
108
+ required: scope.inScope,
109
+ passed: 0,
110
+ missing: 0,
111
+ failed: 0,
112
+ },
107
113
  },
108
114
  waiver: {
109
115
  applied: false,
@@ -207,6 +213,9 @@ export const enforceTddBddPolicy = (params: {
207
213
  const sliceFindings: Finding[] = [];
208
214
  const seenSliceIds = new Set<string>();
209
215
  let validSlices = 0;
216
+ let baselinePassed = 0;
217
+ let baselineMissing = 0;
218
+ let baselineFailed = 0;
210
219
 
211
220
  if (evidenceRead.evidence.slices.length === 0) {
212
221
  sliceFindings.push(
@@ -257,6 +266,30 @@ export const enforceTddBddPolicy = (params: {
257
266
  );
258
267
  }
259
268
 
269
+ if (!slice.baseline) {
270
+ baselineMissing += 1;
271
+ sliceFindings.push(
272
+ buildFinding({
273
+ ruleId: 'generic_tdd_baseline_required',
274
+ code: 'TDD_BASELINE_TEST_REQUIRED',
275
+ message: `Slice ${slice.id} must include passing baseline test evidence before RED.`,
276
+ filePath: evidenceRead.path,
277
+ })
278
+ );
279
+ } else if (slice.baseline.status !== 'passed') {
280
+ baselineFailed += 1;
281
+ sliceFindings.push(
282
+ buildFinding({
283
+ ruleId: 'generic_tdd_baseline_required',
284
+ code: 'TDD_BASELINE_TEST_MUST_PASS',
285
+ message: `Slice ${slice.id} baseline test evidence must pass before editing related code.`,
286
+ filePath: evidenceRead.path,
287
+ })
288
+ );
289
+ } else {
290
+ baselinePassed += 1;
291
+ }
292
+
260
293
  if (slice.red.status !== 'failed') {
261
294
  sliceFindings.push(
262
295
  buildFinding({
@@ -281,6 +314,7 @@ export const enforceTddBddPolicy = (params: {
281
314
 
282
315
  if (
283
316
  !isTimelineOrdered([
317
+ slice.baseline?.timestamp,
284
318
  slice.red.timestamp,
285
319
  slice.green.timestamp,
286
320
  slice.refactor.timestamp,
@@ -320,6 +354,12 @@ export const enforceTddBddPolicy = (params: {
320
354
  slices_invalid: invalidSlices,
321
355
  integrity_ok: evidenceRead.integrity.valid,
322
356
  errors: sliceFindings.map((finding) => finding.code),
357
+ baseline: {
358
+ required: true,
359
+ passed: baselinePassed,
360
+ missing: baselineMissing,
361
+ failed: baselineFailed,
362
+ },
323
363
  },
324
364
  waiver: {
325
365
  applied: false,
@@ -21,6 +21,12 @@ export type TddBddSnapshot = {
21
21
  slices_invalid: number;
22
22
  integrity_ok: boolean;
23
23
  errors: string[];
24
+ baseline: {
25
+ required: boolean;
26
+ passed: number;
27
+ missing: number;
28
+ failed: number;
29
+ };
24
30
  };
25
31
  waiver: {
26
32
  applied: boolean;
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "pumuki",
3
- "version": "6.3.137",
3
+ "version": "6.3.139",
4
4
  "description": "Enterprise-grade AST Intelligence System with multi-platform support (iOS, Android, Backend, Frontend) and Feature-First + DDD + Clean Architecture enforcement. Includes dynamic violations API for intelligent querying.",
5
5
  "main": "index.js",
6
6
  "bin": {