pumuki 6.3.143 → 6.3.145
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -532,6 +532,23 @@ let text = "XCTAssertEqual(value, expected)"
|
|
|
532
532
|
assert.equal(hasSwiftXCTestAssertionUsage(ignored), false);
|
|
533
533
|
});
|
|
534
534
|
|
|
535
|
+
test('hasSwiftXCTestAssertionUsage excluye XCTest compatible con UI automation', () => {
|
|
536
|
+
const uiSource = `
|
|
537
|
+
import XCTest
|
|
538
|
+
|
|
539
|
+
final class BuyerCommerceUISmokeTests: XCTestCase {
|
|
540
|
+
func test_buyer_flow() {
|
|
541
|
+
let app = XCUIApplication()
|
|
542
|
+
app.launch()
|
|
543
|
+
XCTAssertTrue(app.buttons["Comprar"].exists)
|
|
544
|
+
}
|
|
545
|
+
}
|
|
546
|
+
`;
|
|
547
|
+
|
|
548
|
+
assert.equal(hasSwiftXCTestAssertionUsage(uiSource), false);
|
|
549
|
+
assert.equal(hasSwiftXCTUnwrapUsage(`${uiSource}\nlet value = try XCTUnwrap(optional)`), false);
|
|
550
|
+
});
|
|
551
|
+
|
|
535
552
|
test('hasSwiftXCTUnwrapUsage detecta XCTUnwrap real y evita strings', () => {
|
|
536
553
|
const source = `
|
|
537
554
|
let value = try XCTUnwrap(optionalValue)
|
|
@@ -767,6 +767,10 @@ export const hasSwiftMixedTestingFrameworksUsage = (source: string): boolean =>
|
|
|
767
767
|
};
|
|
768
768
|
|
|
769
769
|
export const hasSwiftXCTestAssertionUsage = (source: string): boolean => {
|
|
770
|
+
if (hasSwiftLegacyXCTestUiOrPerformanceUsage(source)) {
|
|
771
|
+
return false;
|
|
772
|
+
}
|
|
773
|
+
|
|
770
774
|
return (
|
|
771
775
|
collectSwiftRegexLines(source, /\bXCTAssert[A-Za-z0-9_]*\s*\(/).length > 0 ||
|
|
772
776
|
collectSwiftRegexLines(source, /\bXCTFail\s*\(/).length > 0
|
|
@@ -774,6 +778,10 @@ export const hasSwiftXCTestAssertionUsage = (source: string): boolean => {
|
|
|
774
778
|
};
|
|
775
779
|
|
|
776
780
|
export const hasSwiftXCTUnwrapUsage = (source: string): boolean => {
|
|
781
|
+
if (hasSwiftLegacyXCTestUiOrPerformanceUsage(source)) {
|
|
782
|
+
return false;
|
|
783
|
+
}
|
|
784
|
+
|
|
777
785
|
return collectSwiftRegexLines(source, /\bXCTUnwrap\s*\(/).length > 0;
|
|
778
786
|
};
|
|
779
787
|
|
|
@@ -4,6 +4,18 @@ This file tracks the active deterministic framework line used in this repository
|
|
|
4
4
|
Canonical release chronology lives in `CHANGELOG.md`.
|
|
5
5
|
This file keeps only the operational highlights and rollout notes that matter while running the framework.
|
|
6
6
|
|
|
7
|
+
### 2026-05-05 (v6.3.145)
|
|
8
|
+
|
|
9
|
+
- **RuralGo PUMUKI-INC-124:** `skills.ios.critical-test-quality` deja de bloquear tests XCTest de UI automation/performance cuando usan `XCUIApplication`, `XCTMetric` o `measure`.
|
|
10
|
+
- **Swift Testing sin regresión:** los unit/integration tests XCTest modernizables siguen bloqueados si incumplen el contrato; solo se respeta la compatibilidad explícita que `swift-testing-expert` permite para UI/performance.
|
|
11
|
+
- **Rollout:** publicar `pumuki@6.3.145`, repinear primero RuralGo y revalidar el smoke UI XCTest que estaba bloqueado.
|
|
12
|
+
|
|
13
|
+
### 2026-05-05 (v6.3.144)
|
|
14
|
+
|
|
15
|
+
- **RuralGo PUMUKI-INC-122:** `pumuki sdd evidence` serializa escrituras concurrentes del artefacto `.pumuki/artifacts/pumuki-evidence-v1.json` con lock local y rename atómico.
|
|
16
|
+
- **Sin pérdida de slices:** dos ejecuciones paralelas fusionan slices existentes/nuevas en lugar de pisar el JSON compartido; además el scaffold genera baseline TDD/BDD válido desde la evidencia PRE_WRITE.
|
|
17
|
+
- **Rollout:** publicar `pumuki@6.3.144`, repinear primero RuralGo y revalidar `status`, `doctor` y una ráfaga concurrente de `pumuki sdd evidence`.
|
|
18
|
+
|
|
7
19
|
### 2026-05-05 (v6.3.143)
|
|
8
20
|
|
|
9
21
|
- **RuralGo PUMUKI-INC-060:** PRE_WRITE deja de aceptar evidencia TDD/BDD de baseline caducada para cambios in-scope.
|
|
@@ -442,6 +442,10 @@ const isXCTestSource = (content: string): boolean => {
|
|
|
442
442
|
return /\bimport\s+XCTest\b/.test(content) || /\bXCTestCase\b/.test(content);
|
|
443
443
|
};
|
|
444
444
|
|
|
445
|
+
const isXCTestUiOrPerformanceCompatibilitySource = (content: string): boolean => {
|
|
446
|
+
return /\bXCUIApplication\b|\bXCTMetric\b|\bmeasure\s*(?:\(|\{)/.test(content);
|
|
447
|
+
};
|
|
448
|
+
|
|
445
449
|
const hasMakeSUTPattern = (content: string): boolean => /\bmakeSUT\s*\(/.test(content);
|
|
446
450
|
|
|
447
451
|
const hasTrackForMemoryLeaksPattern = (content: string): boolean =>
|
|
@@ -461,6 +465,9 @@ const toIosTestsQualityBlockingFinding = (params: {
|
|
|
461
465
|
if (!isXCTestSource(testFile.content)) {
|
|
462
466
|
continue;
|
|
463
467
|
}
|
|
468
|
+
if (isXCTestUiOrPerformanceCompatibilitySource(testFile.content)) {
|
|
469
|
+
continue;
|
|
470
|
+
}
|
|
464
471
|
const missingMarkers: string[] = [];
|
|
465
472
|
if (!hasMakeSUTPattern(testFile.content)) {
|
|
466
473
|
missingMarkers.push('makeSUT()');
|
|
@@ -1,5 +1,14 @@
|
|
|
1
|
-
import { createHash } from 'node:crypto';
|
|
2
|
-
import {
|
|
1
|
+
import { createHash, randomUUID } from 'node:crypto';
|
|
2
|
+
import {
|
|
3
|
+
closeSync,
|
|
4
|
+
existsSync,
|
|
5
|
+
mkdirSync,
|
|
6
|
+
openSync,
|
|
7
|
+
readFileSync,
|
|
8
|
+
renameSync,
|
|
9
|
+
rmSync,
|
|
10
|
+
writeFileSync,
|
|
11
|
+
} from 'node:fs';
|
|
3
12
|
import { basename, dirname, isAbsolute, relative, resolve } from 'node:path';
|
|
4
13
|
import { readEvidenceResult, type EvidenceReadResult } from '../evidence/readEvidence';
|
|
5
14
|
|
|
@@ -27,6 +36,11 @@ export type SddEvidenceScaffoldResult = {
|
|
|
27
36
|
slices: Array<{
|
|
28
37
|
id: string;
|
|
29
38
|
scenario_ref: string;
|
|
39
|
+
baseline: {
|
|
40
|
+
status: 'passed';
|
|
41
|
+
timestamp: string;
|
|
42
|
+
test_ref: string;
|
|
43
|
+
};
|
|
30
44
|
red: {
|
|
31
45
|
status: 'failed';
|
|
32
46
|
timestamp: string;
|
|
@@ -62,9 +76,100 @@ export type SddEvidenceScaffoldResult = {
|
|
|
62
76
|
};
|
|
63
77
|
};
|
|
64
78
|
|
|
79
|
+
type SddEvidenceArtifact = SddEvidenceScaffoldResult['artifact'];
|
|
80
|
+
|
|
81
|
+
const LOCK_WAIT_BUFFER_BYTES = 4;
|
|
82
|
+
const EVIDENCE_ARTIFACT_LOCK_TIMEOUT_MS = 5000;
|
|
83
|
+
const EVIDENCE_ARTIFACT_LOCK_RETRY_DELAY_MS = 25;
|
|
84
|
+
const EVIDENCE_ARTIFACT_JSON_INDENT_SPACES = 2;
|
|
85
|
+
|
|
65
86
|
const computeDigest = (value: string): string =>
|
|
66
87
|
`sha256:${createHash('sha256').update(value, 'utf8').digest('hex')}`;
|
|
67
88
|
|
|
89
|
+
const sleepSync = (milliseconds: number): void => {
|
|
90
|
+
Atomics.wait(new Int32Array(new SharedArrayBuffer(LOCK_WAIT_BUFFER_BYTES)), 0, 0, milliseconds);
|
|
91
|
+
};
|
|
92
|
+
|
|
93
|
+
const acquireFileLock = (lockPath: string): (() => void) => {
|
|
94
|
+
const startedAt = Date.now();
|
|
95
|
+
mkdirSync(dirname(lockPath), { recursive: true });
|
|
96
|
+
while (true) {
|
|
97
|
+
try {
|
|
98
|
+
const fd = openSync(lockPath, 'wx');
|
|
99
|
+
writeFileSync(fd, `${process.pid}:${new Date().toISOString()}\n`, 'utf8');
|
|
100
|
+
return () => {
|
|
101
|
+
closeSync(fd);
|
|
102
|
+
rmSync(lockPath, { force: true });
|
|
103
|
+
};
|
|
104
|
+
} catch (error) {
|
|
105
|
+
const code = error instanceof Error && 'code' in error
|
|
106
|
+
? String((error as NodeJS.ErrnoException).code)
|
|
107
|
+
: '';
|
|
108
|
+
if (code !== 'EEXIST') {
|
|
109
|
+
throw error;
|
|
110
|
+
}
|
|
111
|
+
if (Date.now() - startedAt > EVIDENCE_ARTIFACT_LOCK_TIMEOUT_MS) {
|
|
112
|
+
throw new Error(
|
|
113
|
+
`[pumuki][sdd] evidence artifact is locked by another process: ${lockPath}. Retry when the current evidence write finishes.`
|
|
114
|
+
);
|
|
115
|
+
}
|
|
116
|
+
sleepSync(EVIDENCE_ARTIFACT_LOCK_RETRY_DELAY_MS);
|
|
117
|
+
}
|
|
118
|
+
}
|
|
119
|
+
};
|
|
120
|
+
|
|
121
|
+
const readExistingSddEvidenceArtifact = (path: string): SddEvidenceArtifact | null => {
|
|
122
|
+
if (!existsSync(path)) {
|
|
123
|
+
return null;
|
|
124
|
+
}
|
|
125
|
+
const parsed = JSON.parse(readFileSync(path, 'utf8')) as Partial<SddEvidenceArtifact>;
|
|
126
|
+
if (parsed.version !== '1' || !Array.isArray(parsed.slices)) {
|
|
127
|
+
return null;
|
|
128
|
+
}
|
|
129
|
+
return parsed as SddEvidenceArtifact;
|
|
130
|
+
};
|
|
131
|
+
|
|
132
|
+
const mergeSddEvidenceArtifacts = (params: {
|
|
133
|
+
existing: SddEvidenceArtifact | null;
|
|
134
|
+
next: SddEvidenceArtifact;
|
|
135
|
+
}): SddEvidenceArtifact => {
|
|
136
|
+
if (!params.existing) {
|
|
137
|
+
return params.next;
|
|
138
|
+
}
|
|
139
|
+
const slices = params.existing.slices.filter(
|
|
140
|
+
(slice) => !params.next.slices.some((nextSlice) => nextSlice.id === slice.id)
|
|
141
|
+
);
|
|
142
|
+
return {
|
|
143
|
+
...params.next,
|
|
144
|
+
slices: [...slices, ...params.next.slices],
|
|
145
|
+
};
|
|
146
|
+
};
|
|
147
|
+
|
|
148
|
+
const writeSddEvidenceArtifactAtomically = (params: {
|
|
149
|
+
outputPath: string;
|
|
150
|
+
artifact: SddEvidenceArtifact;
|
|
151
|
+
}): {
|
|
152
|
+
artifact: SddEvidenceArtifact;
|
|
153
|
+
serialized: string;
|
|
154
|
+
} => {
|
|
155
|
+
const lockPath = `${params.outputPath}.lock`;
|
|
156
|
+
const release = acquireFileLock(lockPath);
|
|
157
|
+
const tempPath = `${params.outputPath}.${process.pid}.${randomUUID()}.tmp`;
|
|
158
|
+
try {
|
|
159
|
+
const artifact = mergeSddEvidenceArtifacts({
|
|
160
|
+
existing: readExistingSddEvidenceArtifact(params.outputPath),
|
|
161
|
+
next: params.artifact,
|
|
162
|
+
});
|
|
163
|
+
const serialized = `${JSON.stringify(artifact, null, EVIDENCE_ARTIFACT_JSON_INDENT_SPACES)}\n`;
|
|
164
|
+
writeFileSync(tempPath, serialized, 'utf8');
|
|
165
|
+
renameSync(tempPath, params.outputPath);
|
|
166
|
+
return { artifact, serialized };
|
|
167
|
+
} finally {
|
|
168
|
+
rmSync(tempPath, { force: true });
|
|
169
|
+
release();
|
|
170
|
+
}
|
|
171
|
+
};
|
|
172
|
+
|
|
68
173
|
const resolveRepoBoundPath = (params: {
|
|
69
174
|
repoRoot: string;
|
|
70
175
|
candidatePath: string;
|
|
@@ -91,7 +196,7 @@ const resolveRepoBoundPath = (params: {
|
|
|
91
196
|
return resolved;
|
|
92
197
|
};
|
|
93
198
|
|
|
94
|
-
const
|
|
199
|
+
const isReservedInputValue = (value: string): boolean => {
|
|
95
200
|
const normalized = value.trim().toLowerCase();
|
|
96
201
|
return (
|
|
97
202
|
normalized === 'todo' ||
|
|
@@ -110,7 +215,7 @@ const normalizeRequired = (value: string | undefined, flagName: string): string
|
|
|
110
215
|
if (normalized.length === 0) {
|
|
111
216
|
throw new Error(`[pumuki][sdd] evidence requires ${flagName}.`);
|
|
112
217
|
}
|
|
113
|
-
if (
|
|
218
|
+
if (isReservedInputValue(normalized)) {
|
|
114
219
|
throw new Error(`[pumuki][sdd] evidence ${flagName} must not be a placeholder value.`);
|
|
115
220
|
}
|
|
116
221
|
return normalized;
|
|
@@ -214,6 +319,11 @@ export const runSddEvidenceScaffold = (params?: {
|
|
|
214
319
|
{
|
|
215
320
|
id: scenarioId,
|
|
216
321
|
scenario_ref: resolveScenarioReference(scenarioId),
|
|
322
|
+
baseline: {
|
|
323
|
+
status: 'passed',
|
|
324
|
+
timestamp: validEvidence.source_descriptor.generated_at ?? generatedAt,
|
|
325
|
+
test_ref: 'pumuki sdd validate --stage=PRE_WRITE --json',
|
|
326
|
+
},
|
|
217
327
|
red: {
|
|
218
328
|
status: 'failed',
|
|
219
329
|
timestamp: generatedAt,
|
|
@@ -249,13 +359,19 @@ export const runSddEvidenceScaffold = (params?: {
|
|
|
249
359
|
status: 'valid',
|
|
250
360
|
},
|
|
251
361
|
};
|
|
252
|
-
|
|
253
|
-
|
|
362
|
+
let finalArtifact = artifact;
|
|
363
|
+
let serialized = `${JSON.stringify(finalArtifact, null, 2)}\n`;
|
|
254
364
|
|
|
255
365
|
if (!dryRun) {
|
|
256
366
|
mkdirSync(dirname(outputAbsolutePath), { recursive: true });
|
|
257
|
-
|
|
367
|
+
const writeResult = writeSddEvidenceArtifactAtomically({
|
|
368
|
+
outputPath: outputAbsolutePath,
|
|
369
|
+
artifact,
|
|
370
|
+
});
|
|
371
|
+
finalArtifact = writeResult.artifact;
|
|
372
|
+
serialized = writeResult.serialized;
|
|
258
373
|
}
|
|
374
|
+
const digest = computeDigest(serialized);
|
|
259
375
|
|
|
260
376
|
return {
|
|
261
377
|
command: 'pumuki sdd evidence',
|
|
@@ -273,6 +389,6 @@ export const runSddEvidenceScaffold = (params?: {
|
|
|
273
389
|
written: !dryRun,
|
|
274
390
|
digest,
|
|
275
391
|
},
|
|
276
|
-
artifact,
|
|
392
|
+
artifact: finalArtifact,
|
|
277
393
|
};
|
|
278
394
|
};
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "pumuki",
|
|
3
|
-
"version": "6.3.
|
|
3
|
+
"version": "6.3.145",
|
|
4
4
|
"description": "Enterprise-grade AST Intelligence System with multi-platform support (iOS, Android, Backend, Frontend) and Feature-First + DDD + Clean Architecture enforcement. Includes dynamic violations API for intelligent querying.",
|
|
5
5
|
"main": "index.js",
|
|
6
6
|
"bin": {
|