docrev 0.9.14 → 0.9.16
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +66 -51
- package/dist/lib/wordcomments.d.ts.map +1 -1
- package/dist/lib/wordcomments.js +25 -2
- package/dist/lib/wordcomments.js.map +1 -1
- package/docs-src/build.py +113 -0
- package/docs-src/extra.css +208 -0
- package/docs-src/md-to-html.lua +6 -0
- package/docs-src/template.html +116 -0
- package/lib/wordcomments.ts +25 -2
- package/mkdocs.yml +64 -0
- package/package.json +1 -1
- package/site/assets/extra.css +208 -0
- package/site/commands.html +926 -0
- package/site/configuration.html +469 -0
- package/site/index.html +288 -0
- package/site/troubleshooting.html +461 -0
- package/site/workflow.html +518 -0
- package/skill/REFERENCE.md +44 -2
- package/skill/SKILL.md +20 -4
- package/dev_notes/bug_repro_comment_parser.md +0 -71
- package/dev_notes/stress2/adversarial.docx +0 -0
- package/dev_notes/stress2/build_adversarial.ts +0 -186
- package/dev_notes/stress2/drift_matcher.ts +0 -62
- package/dev_notes/stress2/probe_anchors.ts +0 -35
- package/dev_notes/stress2/project/adversarial.docx +0 -0
- package/dev_notes/stress2/project/discussion.before.md +0 -3
- package/dev_notes/stress2/project/discussion.md +0 -3
- package/dev_notes/stress2/project/methods.before.md +0 -20
- package/dev_notes/stress2/project/methods.md +0 -20
- package/dev_notes/stress2/project/rev.yaml +0 -5
- package/dev_notes/stress2/project/sections.yaml +0 -4
- package/dev_notes/stress2/sections.yaml +0 -5
- package/dev_notes/stress2/trace_placement.ts +0 -50
- package/dev_notes/stresstest_boundaries.ts +0 -27
- package/dev_notes/stresstest_drift_apply.ts +0 -43
- package/dev_notes/stresstest_drift_compare.ts +0 -43
- package/dev_notes/stresstest_drift_v2.ts +0 -54
- package/dev_notes/stresstest_inspect.ts +0 -54
- package/dev_notes/stresstest_pstyle.ts +0 -55
- package/dev_notes/stresstest_section_debug.ts +0 -23
- package/dev_notes/stresstest_split.ts +0 -70
- package/dev_notes/stresstest_trace.ts +0 -19
- package/dev_notes/stresstest_verify_no_overwrite.ts +0 -40
|
@@ -1,54 +0,0 @@
|
|
|
1
|
-
// Apply drift edits that actually intersect real comment anchors
|
|
2
|
-
import * as fs from 'fs';
|
|
3
|
-
|
|
4
|
-
const dir = 'C:/GillesC/tmp/docrev-stress/project-drifted';
|
|
5
|
-
|
|
6
|
-
function edit(file: string, edits: Array<[string, string]>) {
|
|
7
|
-
const p = `${dir}/${file}`;
|
|
8
|
-
let t = fs.readFileSync(p, 'utf-8');
|
|
9
|
-
for (const [from, to] of edits) {
|
|
10
|
-
if (!t.includes(from)) {
|
|
11
|
-
console.error(`!! ${file}: pattern not found: "${from.slice(0, 60)}"`);
|
|
12
|
-
continue;
|
|
13
|
-
}
|
|
14
|
-
t = t.replace(from, to);
|
|
15
|
-
console.log(`OK ${file}: replaced "${from.slice(0, 60)}..."`);
|
|
16
|
-
}
|
|
17
|
-
fs.writeFileSync(p, t);
|
|
18
|
-
}
|
|
19
|
-
|
|
20
|
-
// Each of these intersects real anchors; expected effect noted.
|
|
21
|
-
edit('abstract.md', [
|
|
22
|
-
// Title: 3 anchors are 'Patterns of habitat niche expansion' (full title) — but title isn't in abstract.md, it's in YAML frontmatter, so probably already failing
|
|
23
|
-
// 'overrepresented' is the most common anchor (6 comments). Replace it everywhere.
|
|
24
|
-
// After this, the word 'overrepresented' is gone → all those comments should go to context-only or unmatched
|
|
25
|
-
['overrepresented in more than one habitat type', 'preferentially distributed across habitat types'],
|
|
26
|
-
// 'undisturbed' appears in 'undisturbed habitat types' → 2 comments
|
|
27
|
-
['undisturbed habitat types', 'pristine vegetation communities'],
|
|
28
|
-
// 'human economies' (3 anchors)
|
|
29
|
-
['human economies', 'economic activity'],
|
|
30
|
-
// 'nutrient-rich'
|
|
31
|
-
['nutrient-rich, disturbed habitats', 'anthropic, modified environments'],
|
|
32
|
-
// 'when residence time was prolonged.' — long anchor, should go to drift / partial-start
|
|
33
|
-
['when residence time was prolonged.', 'as residence time increased.'],
|
|
34
|
-
// 'pervasive' (1 anchor)
|
|
35
|
-
['Invasion debt in terms of habitat niche breadth is pervasive', 'Invasion debt in terms of habitat niche breadth is widespread'],
|
|
36
|
-
// 'alpine habitats' (1 anchor)
|
|
37
|
-
['alpine habitats', 'high-altitude environments'],
|
|
38
|
-
// 'semi-natural vegetation' (2 anchors)
|
|
39
|
-
['semi-natural vegetation', 'natural plant communities'],
|
|
40
|
-
// 'Man-made habitats function as gateways' (1 anchor)
|
|
41
|
-
['Man-made habitats function as gateways', 'Anthropogenic environments serve as entry routes'],
|
|
42
|
-
// 'Invasion debt in terms of habitat niche breadth' (2 anchors)
|
|
43
|
-
// Already mutated above (pervasive→widespread keeps anchor matchable). Add another small change to test partial-start
|
|
44
|
-
// 'habitat niche expansion' — anchor for #27
|
|
45
|
-
['habitat niche expansion', 'ecological niche broadening'],
|
|
46
|
-
// ', yet' anchor for #19
|
|
47
|
-
[', yet', ', however'],
|
|
48
|
-
// '44' — short numeric anchor
|
|
49
|
-
['44%', '53%'],
|
|
50
|
-
// '17 habitats'
|
|
51
|
-
['17 habitats', '18 habitats'],
|
|
52
|
-
]);
|
|
53
|
-
|
|
54
|
-
console.log('---done');
|
|
@@ -1,54 +0,0 @@
|
|
|
1
|
-
import { extractWordComments, extractCommentAnchors } from '../lib/import.js';
|
|
2
|
-
|
|
3
|
-
const docx = 'C:/GillesC/tmp/docrev-stress/reviewed.docx';
|
|
4
|
-
|
|
5
|
-
const comments = await extractWordComments(docx);
|
|
6
|
-
const { anchors, fullDocText } = await extractCommentAnchors(docx);
|
|
7
|
-
|
|
8
|
-
console.log(`comments: ${comments.length}`);
|
|
9
|
-
console.log(`anchors: ${anchors.size}`);
|
|
10
|
-
console.log(`docText: ${fullDocText.length} chars`);
|
|
11
|
-
console.log();
|
|
12
|
-
|
|
13
|
-
const byAuthor: Record<string, number> = {};
|
|
14
|
-
for (const c of comments) byAuthor[c.author] = (byAuthor[c.author] || 0) + 1;
|
|
15
|
-
console.log('by author:', byAuthor);
|
|
16
|
-
console.log();
|
|
17
|
-
|
|
18
|
-
console.log('first 5 comments:');
|
|
19
|
-
for (const c of comments.slice(0, 5)) {
|
|
20
|
-
const a = anchors.get(c.id);
|
|
21
|
-
const pos = a?.docPosition ?? -1;
|
|
22
|
-
const anchor = (a?.anchor || '').slice(0, 60);
|
|
23
|
-
const text = c.text.replace(/\s+/g, ' ').slice(0, 80);
|
|
24
|
-
console.log(` #${c.id} [${c.author}] pos=${pos} anchor="${anchor}" text="${text}"`);
|
|
25
|
-
}
|
|
26
|
-
console.log();
|
|
27
|
-
console.log('last 3 comments:');
|
|
28
|
-
for (const c of comments.slice(-3)) {
|
|
29
|
-
const a = anchors.get(c.id);
|
|
30
|
-
const pos = a?.docPosition ?? -1;
|
|
31
|
-
const anchor = (a?.anchor || '').slice(0, 60);
|
|
32
|
-
const text = c.text.replace(/\s+/g, ' ').slice(0, 80);
|
|
33
|
-
console.log(` #${c.id} [${c.author}] pos=${pos} anchor="${anchor}" text="${text}"`);
|
|
34
|
-
}
|
|
35
|
-
console.log();
|
|
36
|
-
|
|
37
|
-
const sectionKeywords = ['Abstract', 'Introduction', 'Methods', 'Results', 'Discussion', 'Conclusion', 'References', 'Acknowledgements', 'Data Availability', 'Author Contributions'];
|
|
38
|
-
console.log('candidate headings:');
|
|
39
|
-
for (const kw of sectionKeywords) {
|
|
40
|
-
const idx = fullDocText.indexOf(kw);
|
|
41
|
-
if (idx >= 0) {
|
|
42
|
-
const context = fullDocText.slice(idx, idx + 80).replace(/\s+/g, ' ');
|
|
43
|
-
console.log(` ${kw} @ ${idx}: "${context}"`);
|
|
44
|
-
}
|
|
45
|
-
}
|
|
46
|
-
|
|
47
|
-
// Check anchor distribution to understand section spans
|
|
48
|
-
const positions = [...anchors.values()].map(a => a.docPosition).sort((a, b) => a - b);
|
|
49
|
-
console.log();
|
|
50
|
-
console.log(`anchor positions: min=${positions[0]} max=${positions[positions.length-1]} median=${positions[Math.floor(positions.length/2)]}`);
|
|
51
|
-
|
|
52
|
-
// Check for empty anchors
|
|
53
|
-
const empty = [...anchors.values()].filter(a => a.isEmpty).length;
|
|
54
|
-
console.log(`empty anchors: ${empty}`);
|
|
@@ -1,55 +0,0 @@
|
|
|
1
|
-
// Inspect the document.xml paragraph styles to find true headings.
|
|
2
|
-
import AdmZip from 'adm-zip';
|
|
3
|
-
import * as fs from 'fs';
|
|
4
|
-
|
|
5
|
-
const docx = 'C:/GillesC/tmp/docrev-stress/reviewed.docx';
|
|
6
|
-
const zip = new AdmZip(docx);
|
|
7
|
-
const xml = zip.getEntry('word/document.xml')!.getData().toString('utf8');
|
|
8
|
-
|
|
9
|
-
// Build text-position map from <w:t> runs the same way extractCommentAnchors does
|
|
10
|
-
const textNodePattern = /<w:t[^>]*>([^<]*)<\/w:t>/g;
|
|
11
|
-
const textNodes: Array<{ xmlStart: number; xmlEnd: number; textStart: number; textEnd: number; text: string }> = [];
|
|
12
|
-
let textPosition = 0;
|
|
13
|
-
let m;
|
|
14
|
-
function decode(s: string): string {
|
|
15
|
-
return s.replace(/&/g, '&').replace(/</g, '<').replace(/>/g, '>').replace(/"/g, '"').replace(/'/g, "'");
|
|
16
|
-
}
|
|
17
|
-
while ((m = textNodePattern.exec(xml)) !== null) {
|
|
18
|
-
const decoded = decode(m[1]);
|
|
19
|
-
textNodes.push({ xmlStart: m.index, xmlEnd: m.index + m[0].length, textStart: textPosition, textEnd: textPosition + decoded.length, text: decoded });
|
|
20
|
-
textPosition += decoded.length;
|
|
21
|
-
}
|
|
22
|
-
|
|
23
|
-
function xmlToText(xmlPos: number): number {
|
|
24
|
-
for (const n of textNodes) {
|
|
25
|
-
if (xmlPos >= n.xmlStart && xmlPos < n.xmlEnd) return n.textStart;
|
|
26
|
-
if (xmlPos < n.xmlStart) return n.textStart;
|
|
27
|
-
}
|
|
28
|
-
return textNodes.length ? textNodes[textNodes.length - 1].textEnd : 0;
|
|
29
|
-
}
|
|
30
|
-
|
|
31
|
-
// Now walk paragraphs; each <w:p ...>...</w:p>. Inside, find <w:pStyle w:val="..."/> if present, and concatenate text runs.
|
|
32
|
-
const paraPattern = /<w:p\b[^>]*>([\s\S]*?)<\/w:p>/g;
|
|
33
|
-
const headings: Array<{ style: string; text: string; xmlStart: number; textPos: number }> = [];
|
|
34
|
-
let pm;
|
|
35
|
-
while ((pm = paraPattern.exec(xml)) !== null) {
|
|
36
|
-
const inner = pm[1];
|
|
37
|
-
const styleMatch = inner.match(/<w:pStyle[^>]*w:val="([^"]+)"/);
|
|
38
|
-
if (!styleMatch) continue;
|
|
39
|
-
const style = styleMatch[1];
|
|
40
|
-
if (!/heading/i.test(style)) continue;
|
|
41
|
-
// Extract text from runs
|
|
42
|
-
const textInRange = /<w:t[^>]*>([^<]*)<\/w:t>/g;
|
|
43
|
-
let txt = '';
|
|
44
|
-
let tm;
|
|
45
|
-
while ((tm = textInRange.exec(inner)) !== null) txt += decode(tm[1]);
|
|
46
|
-
if (!txt.trim()) continue;
|
|
47
|
-
const xmlStart = pm.index;
|
|
48
|
-
const textPos = xmlToText(xmlStart);
|
|
49
|
-
headings.push({ style, text: txt.trim(), xmlStart, textPos });
|
|
50
|
-
}
|
|
51
|
-
|
|
52
|
-
console.log(`Found ${headings.length} heading paragraphs:`);
|
|
53
|
-
for (const h of headings) {
|
|
54
|
-
console.log(` [${h.style.padEnd(8)}] textPos=${String(h.textPos).padStart(6)} "${h.text.slice(0, 60)}"`);
|
|
55
|
-
}
|
|
@@ -1,23 +0,0 @@
|
|
|
1
|
-
import { extractCommentAnchors } from '../lib/import.js';
|
|
2
|
-
|
|
3
|
-
const docx = 'C:/GillesC/tmp/docrev-stress/reviewed.docx';
|
|
4
|
-
const { fullDocText } = await extractCommentAnchors(docx);
|
|
5
|
-
|
|
6
|
-
const keywords = ['Abstract', 'Introduction', 'Methods', 'Results', 'Discussion', 'Conclusion', 'References'];
|
|
7
|
-
for (const kw of keywords) {
|
|
8
|
-
const lower = fullDocText.toLowerCase();
|
|
9
|
-
const needle = kw.toLowerCase();
|
|
10
|
-
const occ: Array<{ idx: number; ctx: string; afterChar: string }> = [];
|
|
11
|
-
let idx = 0;
|
|
12
|
-
while ((idx = lower.indexOf(needle, idx)) !== -1) {
|
|
13
|
-
const ctx = fullDocText.slice(Math.max(0, idx - 20), idx + needle.length + 20).replace(/\s+/g, ' ');
|
|
14
|
-
const afterChar = fullDocText.slice(idx + needle.length, idx + needle.length + 1);
|
|
15
|
-
occ.push({ idx, ctx, afterChar });
|
|
16
|
-
idx++;
|
|
17
|
-
}
|
|
18
|
-
console.log(`\n--- ${kw} (${occ.length} occurrences) ---`);
|
|
19
|
-
for (const o of occ.slice(0, 6)) {
|
|
20
|
-
console.log(` @${o.idx} after="${o.afterChar}" ctx: ...${o.ctx}...`);
|
|
21
|
-
}
|
|
22
|
-
if (occ.length > 6) console.log(` ... and ${occ.length - 6} more`);
|
|
23
|
-
}
|
|
@@ -1,70 +0,0 @@
|
|
|
1
|
-
import * as fs from 'fs';
|
|
2
|
-
import * as path from 'path';
|
|
3
|
-
|
|
4
|
-
const ROOT = 'C:/GillesC/tmp/docrev-stress/project';
|
|
5
|
-
fs.mkdirSync(ROOT, { recursive: true });
|
|
6
|
-
|
|
7
|
-
const full = fs.readFileSync('C:/GillesC/tmp/docrev-stress/full.md', 'utf-8');
|
|
8
|
-
const lines = full.split(/\r?\n/);
|
|
9
|
-
|
|
10
|
-
interface Section { file: string; header: string; lines: string[]; }
|
|
11
|
-
|
|
12
|
-
const sections: Section[] = [];
|
|
13
|
-
|
|
14
|
-
// Map H1 heading text to section file
|
|
15
|
-
const headerToFile: Record<string, string> = {
|
|
16
|
-
'Abstract': 'abstract.md',
|
|
17
|
-
'Introduction': 'introduction.md',
|
|
18
|
-
'Methods': 'methods.md',
|
|
19
|
-
'Results': 'results.md',
|
|
20
|
-
'Discussion': 'discussion.md',
|
|
21
|
-
'Conclusion': 'conclusion.md',
|
|
22
|
-
'References': 'references.md',
|
|
23
|
-
'Supplementary Materials': 'supplement.md',
|
|
24
|
-
};
|
|
25
|
-
|
|
26
|
-
let current: Section | null = null;
|
|
27
|
-
const preambleLines: string[] = [];
|
|
28
|
-
|
|
29
|
-
for (const line of lines) {
|
|
30
|
-
const m = line.match(/^# (.+)$/);
|
|
31
|
-
if (m && headerToFile[m[1].trim()]) {
|
|
32
|
-
if (current) sections.push(current);
|
|
33
|
-
current = { file: headerToFile[m[1].trim()], header: m[1].trim(), lines: [line] };
|
|
34
|
-
} else if (current) {
|
|
35
|
-
current.lines.push(line);
|
|
36
|
-
} else {
|
|
37
|
-
preambleLines.push(line);
|
|
38
|
-
}
|
|
39
|
-
}
|
|
40
|
-
if (current) sections.push(current);
|
|
41
|
-
|
|
42
|
-
// Write sections
|
|
43
|
-
for (const s of sections) {
|
|
44
|
-
const out = path.join(ROOT, s.file);
|
|
45
|
-
fs.writeFileSync(out, s.lines.join('\n'), 'utf-8');
|
|
46
|
-
console.log(`wrote ${out} (${s.lines.length} lines)`);
|
|
47
|
-
}
|
|
48
|
-
|
|
49
|
-
// Write sections.yaml
|
|
50
|
-
const yaml: string[] = ['version: 1', 'sections:'];
|
|
51
|
-
for (const s of sections) {
|
|
52
|
-
yaml.push(` ${s.file}:`);
|
|
53
|
-
yaml.push(` header: "${s.header}"`);
|
|
54
|
-
yaml.push(` aliases: []`);
|
|
55
|
-
}
|
|
56
|
-
fs.writeFileSync(path.join(ROOT, 'sections.yaml'), yaml.join('\n') + '\n', 'utf-8');
|
|
57
|
-
console.log(`wrote sections.yaml`);
|
|
58
|
-
|
|
59
|
-
// Minimal rev.yaml so commands that look for project root don't choke
|
|
60
|
-
const revYaml = `title: "Stress Test"
|
|
61
|
-
authors:
|
|
62
|
-
- name: Gilles Colling
|
|
63
|
-
output:
|
|
64
|
-
docx: {}
|
|
65
|
-
pdf: {}
|
|
66
|
-
`;
|
|
67
|
-
fs.writeFileSync(path.join(ROOT, 'rev.yaml'), revYaml, 'utf-8');
|
|
68
|
-
|
|
69
|
-
console.log();
|
|
70
|
-
console.log(`section files: ${sections.length}`);
|
|
@@ -1,19 +0,0 @@
|
|
|
1
|
-
import { extractCommentAnchors } from '../lib/import.js';
|
|
2
|
-
|
|
3
|
-
const docx = 'C:/GillesC/tmp/docrev-stress/reviewed.docx';
|
|
4
|
-
const { fullDocText } = await extractCommentAnchors(docx);
|
|
5
|
-
|
|
6
|
-
const needle = 'methods';
|
|
7
|
-
const lower = fullDocText.toLowerCase();
|
|
8
|
-
let idx = 0;
|
|
9
|
-
const trace: Array<{idx:number, after:string, skipped:boolean}> = [];
|
|
10
|
-
while ((idx = lower.indexOf(needle, idx)) !== -1 && trace.length < 10) {
|
|
11
|
-
const after = fullDocText.slice(idx + needle.length, idx + needle.length + 5);
|
|
12
|
-
const skip = after.startsWith(':') || after.startsWith(' :');
|
|
13
|
-
trace.push({ idx, after, skipped: skip });
|
|
14
|
-
if (!skip) break;
|
|
15
|
-
idx++;
|
|
16
|
-
}
|
|
17
|
-
for (const t of trace) console.log(`idx=${t.idx} after="${t.after}" skipped=${t.skipped}`);
|
|
18
|
-
console.log();
|
|
19
|
-
console.log(`first non-:-prefixed Methods idx = ${trace.find(t => !t.skipped)?.idx ?? -1}`);
|
|
@@ -1,40 +0,0 @@
|
|
|
1
|
-
// Verify sync --comments-only didn't touch existing prose, only inserted comments.
|
|
2
|
-
import * as fs from 'fs';
|
|
3
|
-
import { stripCriticMarkup } from '../lib/anchor-match.js';
|
|
4
|
-
|
|
5
|
-
const original = 'C:/GillesC/tmp/docrev-stress/project';
|
|
6
|
-
const synced = 'C:/GillesC/tmp/docrev-stress/project-sync';
|
|
7
|
-
const files = ['abstract.md', 'introduction.md', 'methods.md', 'results.md', 'discussion.md', 'conclusion.md', 'references.md', 'supplement.md'];
|
|
8
|
-
|
|
9
|
-
let totalComments = 0;
|
|
10
|
-
let mismatches = 0;
|
|
11
|
-
for (const f of files) {
|
|
12
|
-
const orig = fs.readFileSync(`${original}/${f}`, 'utf-8');
|
|
13
|
-
const synd = fs.readFileSync(`${synced}/${f}`, 'utf-8');
|
|
14
|
-
|
|
15
|
-
// Count comments inserted
|
|
16
|
-
const commentMatches = synd.match(/\{>>.*?<<\}/gs) || [];
|
|
17
|
-
totalComments += commentMatches.length;
|
|
18
|
-
|
|
19
|
-
// The originals may already contain CriticMarkup (Word highlights -> [X]{.mark}),
|
|
20
|
-
// so strip both sides to compare just the underlying prose.
|
|
21
|
-
const synStripped = stripCriticMarkup(synd);
|
|
22
|
-
const origNorm = stripCriticMarkup(orig);
|
|
23
|
-
|
|
24
|
-
if (synStripped !== origNorm) {
|
|
25
|
-
// Find first difference
|
|
26
|
-
let i = 0;
|
|
27
|
-
while (i < origNorm.length && i < synStripped.length && origNorm[i] === synStripped[i]) i++;
|
|
28
|
-
const ctxOrig = origNorm.slice(Math.max(0, i - 30), i + 50).replace(/\n/g, '\\n');
|
|
29
|
-
const ctxSyn = synStripped.slice(Math.max(0, i - 30), i + 50).replace(/\n/g, '\\n');
|
|
30
|
-
console.log(`✗ ${f}: prose differs at byte ${i} (lengths orig=${origNorm.length} stripped=${synStripped.length})`);
|
|
31
|
-
console.log(` orig: ${ctxOrig}`);
|
|
32
|
-
console.log(` syn: ${ctxSyn}`);
|
|
33
|
-
mismatches++;
|
|
34
|
-
} else {
|
|
35
|
-
console.log(`✓ ${f}: prose preserved (${commentMatches.length} comments inserted)`);
|
|
36
|
-
}
|
|
37
|
-
}
|
|
38
|
-
console.log();
|
|
39
|
-
console.log(`total comments inserted: ${totalComments}`);
|
|
40
|
-
console.log(`prose mismatches: ${mismatches}/${files.length}`);
|