docrev 0.9.13 → 0.9.14
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.claude/settings.local.json +9 -9
- package/.gitattributes +1 -1
- package/CHANGELOG.md +149 -149
- package/PLAN-tables-and-postprocess.md +850 -850
- package/README.md +391 -391
- package/bin/rev.js +11 -11
- package/bin/rev.ts +145 -145
- package/completions/rev.bash +127 -127
- package/completions/rev.ps1 +210 -210
- package/completions/rev.zsh +207 -207
- package/dev_notes/stress2/build_adversarial.ts +186 -186
- package/dev_notes/stress2/drift_matcher.ts +62 -62
- package/dev_notes/stress2/probe_anchors.ts +35 -35
- package/dev_notes/stress2/project/discussion.before.md +3 -3
- package/dev_notes/stress2/project/discussion.md +3 -3
- package/dev_notes/stress2/project/methods.before.md +20 -20
- package/dev_notes/stress2/project/methods.md +20 -20
- package/dev_notes/stress2/project/rev.yaml +5 -5
- package/dev_notes/stress2/project/sections.yaml +4 -4
- package/dev_notes/stress2/sections.yaml +5 -5
- package/dev_notes/stress2/trace_placement.ts +50 -50
- package/dev_notes/stresstest_boundaries.ts +27 -27
- package/dev_notes/stresstest_drift_apply.ts +43 -43
- package/dev_notes/stresstest_drift_compare.ts +43 -43
- package/dev_notes/stresstest_drift_v2.ts +54 -54
- package/dev_notes/stresstest_inspect.ts +54 -54
- package/dev_notes/stresstest_pstyle.ts +55 -55
- package/dev_notes/stresstest_section_debug.ts +23 -23
- package/dev_notes/stresstest_split.ts +70 -70
- package/dev_notes/stresstest_trace.ts +19 -19
- package/dev_notes/stresstest_verify_no_overwrite.ts +40 -40
- package/dist/lib/build.d.ts +38 -1
- package/dist/lib/build.d.ts.map +1 -1
- package/dist/lib/build.js +68 -30
- package/dist/lib/build.js.map +1 -1
- package/dist/lib/commands/build.d.ts.map +1 -1
- package/dist/lib/commands/build.js +38 -5
- package/dist/lib/commands/build.js.map +1 -1
- package/dist/lib/commands/utilities.js +164 -164
- package/dist/lib/commands/word-tools.js +8 -8
- package/dist/lib/grammar.js +3 -3
- package/dist/lib/pdf-comments.js +44 -44
- package/dist/lib/plugins.js +57 -57
- package/dist/lib/pptx-themes.js +115 -115
- package/dist/lib/spelling.js +2 -2
- package/dist/lib/templates.js +387 -387
- package/dist/lib/themes.js +51 -51
- package/eslint.config.js +27 -27
- package/lib/anchor-match.ts +276 -276
- package/lib/annotations.ts +644 -644
- package/lib/build.ts +1300 -1251
- package/lib/citations.ts +160 -160
- package/lib/commands/build.ts +833 -801
- package/lib/commands/citations.ts +515 -515
- package/lib/commands/comments.ts +1050 -1050
- package/lib/commands/context.ts +174 -174
- package/lib/commands/core.ts +309 -309
- package/lib/commands/doi.ts +435 -435
- package/lib/commands/file-ops.ts +372 -372
- package/lib/commands/history.ts +320 -320
- package/lib/commands/index.ts +87 -87
- package/lib/commands/init.ts +259 -259
- package/lib/commands/merge-resolve.ts +378 -378
- package/lib/commands/preview.ts +178 -178
- package/lib/commands/project-info.ts +244 -244
- package/lib/commands/quality.ts +517 -517
- package/lib/commands/response.ts +454 -454
- package/lib/commands/section-boundaries.ts +82 -82
- package/lib/commands/sections.ts +451 -451
- package/lib/commands/sync.ts +706 -706
- package/lib/commands/text-ops.ts +449 -449
- package/lib/commands/utilities.ts +448 -448
- package/lib/commands/verify-anchors.ts +272 -272
- package/lib/commands/word-tools.ts +340 -340
- package/lib/comment-realign.ts +517 -517
- package/lib/config.ts +84 -84
- package/lib/crossref.ts +781 -781
- package/lib/csl.ts +191 -191
- package/lib/dependencies.ts +98 -98
- package/lib/diff-engine.ts +465 -465
- package/lib/doi-cache.ts +115 -115
- package/lib/doi.ts +897 -897
- package/lib/equations.ts +506 -506
- package/lib/errors.ts +346 -346
- package/lib/format.ts +541 -541
- package/lib/git.ts +326 -326
- package/lib/grammar.ts +303 -303
- package/lib/image-registry.ts +180 -180
- package/lib/import.ts +911 -911
- package/lib/journals.ts +543 -543
- package/lib/merge.ts +633 -633
- package/lib/orcid.ts +144 -144
- package/lib/pdf-comments.ts +263 -263
- package/lib/pdf-import.ts +524 -524
- package/lib/plugins.ts +362 -362
- package/lib/postprocess.ts +188 -188
- package/lib/pptx-color-filter.lua +37 -37
- package/lib/pptx-template.ts +469 -469
- package/lib/pptx-themes.ts +483 -483
- package/lib/protect-restore.ts +520 -520
- package/lib/rate-limiter.ts +94 -94
- package/lib/response.ts +197 -197
- package/lib/restore-references.ts +240 -240
- package/lib/review.ts +327 -327
- package/lib/schema.ts +417 -417
- package/lib/scientific-words.ts +73 -73
- package/lib/sections.ts +335 -335
- package/lib/slides.ts +756 -756
- package/lib/spelling.ts +334 -334
- package/lib/templates.ts +526 -526
- package/lib/themes.ts +742 -742
- package/lib/trackchanges.ts +247 -247
- package/lib/tui.ts +450 -450
- package/lib/types.ts +550 -550
- package/lib/undo.ts +250 -250
- package/lib/utils.ts +69 -69
- package/lib/variables.ts +179 -179
- package/lib/word-extraction.ts +806 -806
- package/lib/word.ts +643 -643
- package/lib/wordcomments.ts +817 -817
- package/package.json +137 -137
- package/scripts/postbuild.js +28 -28
- package/skill/REFERENCE.md +431 -431
- package/skill/SKILL.md +258 -258
- package/tsconfig.json +26 -26
- package/types/index.d.ts +525 -525
package/lib/grammar.ts
CHANGED
|
@@ -1,303 +1,303 @@
|
|
|
1
|
-
/**
|
|
2
|
-
* Grammar checker module with custom dictionary support
|
|
3
|
-
*
|
|
4
|
-
* Features:
|
|
5
|
-
* - Common grammar/style issues detection
|
|
6
|
-
* - Custom dictionary for project-specific terms
|
|
7
|
-
* - Learn mode to add words to dictionary
|
|
8
|
-
*/
|
|
9
|
-
|
|
10
|
-
import * as fs from 'fs';
|
|
11
|
-
import * as path from 'path';
|
|
12
|
-
|
|
13
|
-
// Default dictionary location
|
|
14
|
-
const DEFAULT_DICT_NAME = '.rev-dictionary';
|
|
15
|
-
|
|
16
|
-
interface GrammarRule {
|
|
17
|
-
id: string;
|
|
18
|
-
pattern: RegExp;
|
|
19
|
-
message: string;
|
|
20
|
-
severity: 'error' | 'warning' | 'info';
|
|
21
|
-
check?: (match: string) => boolean;
|
|
22
|
-
}
|
|
23
|
-
|
|
24
|
-
interface GrammarIssue {
|
|
25
|
-
rule: string;
|
|
26
|
-
message: string;
|
|
27
|
-
severity: 'error' | 'warning' | 'info';
|
|
28
|
-
line: number;
|
|
29
|
-
column: number;
|
|
30
|
-
match: string;
|
|
31
|
-
context: string;
|
|
32
|
-
}
|
|
33
|
-
|
|
34
|
-
interface GrammarSummary {
|
|
35
|
-
total: number;
|
|
36
|
-
errors: number;
|
|
37
|
-
warnings: number;
|
|
38
|
-
info: number;
|
|
39
|
-
byRule: Record<string, number>;
|
|
40
|
-
}
|
|
41
|
-
|
|
42
|
-
interface CheckGrammarOptions {
|
|
43
|
-
scientific?: boolean;
|
|
44
|
-
directory?: string;
|
|
45
|
-
}
|
|
46
|
-
|
|
47
|
-
/**
|
|
48
|
-
* Common grammar/style rules
|
|
49
|
-
*/
|
|
50
|
-
const GRAMMAR_RULES: GrammarRule[] = [
|
|
51
|
-
{
|
|
52
|
-
id: 'passive-voice',
|
|
53
|
-
pattern: /\b(is|are|was|were|be|been|being)\s+(being\s+)?\w+ed\b/gi,
|
|
54
|
-
message: 'Possible passive voice',
|
|
55
|
-
severity: 'info',
|
|
56
|
-
},
|
|
57
|
-
{
|
|
58
|
-
id: 'weasel-words',
|
|
59
|
-
pattern: /\b(very|really|quite|extremely|fairly|rather|somewhat|just)\b/gi,
|
|
60
|
-
message: 'Weasel word - consider removing or being more specific',
|
|
61
|
-
severity: 'warning',
|
|
62
|
-
},
|
|
63
|
-
{
|
|
64
|
-
id: 'weak-start',
|
|
65
|
-
pattern: /^\s*(There (is|are|was|were)|It is)\b/gmi,
|
|
66
|
-
message: 'Weak sentence start - consider restructuring',
|
|
67
|
-
severity: 'info',
|
|
68
|
-
},
|
|
69
|
-
{
|
|
70
|
-
id: 'duplicate-words',
|
|
71
|
-
pattern: /\b(\w+)\s+\1\b/gi,
|
|
72
|
-
message: 'Duplicate word',
|
|
73
|
-
severity: 'error',
|
|
74
|
-
},
|
|
75
|
-
{
|
|
76
|
-
id: 'split-infinitive',
|
|
77
|
-
pattern: /\bto\s+(\w+ly)\s+\w+\b/gi,
|
|
78
|
-
message: 'Split infinitive',
|
|
79
|
-
severity: 'info',
|
|
80
|
-
},
|
|
81
|
-
{
|
|
82
|
-
id: 'sentence-length',
|
|
83
|
-
pattern: /[^.!?]*[.!?]/g,
|
|
84
|
-
check: (match: string): boolean => {
|
|
85
|
-
const words = match.trim().split(/\s+/).length;
|
|
86
|
-
return words > 40;
|
|
87
|
-
},
|
|
88
|
-
message: 'Long sentence (>40 words) - consider breaking up',
|
|
89
|
-
severity: 'warning',
|
|
90
|
-
},
|
|
91
|
-
{
|
|
92
|
-
id: 'cliches',
|
|
93
|
-
pattern: /\b(at the end of the day|in terms of|it goes without saying|needless to say|as a matter of fact|first and foremost|last but not least)\b/gi,
|
|
94
|
-
message: 'Cliche - consider rephrasing',
|
|
95
|
-
severity: 'warning',
|
|
96
|
-
},
|
|
97
|
-
{
|
|
98
|
-
id: 'hedging',
|
|
99
|
-
pattern: /\b(seems to|appears to|tends to|might|may|could possibly|would seem)\b/gi,
|
|
100
|
-
message: 'Hedging language - be more direct if appropriate',
|
|
101
|
-
severity: 'info',
|
|
102
|
-
},
|
|
103
|
-
{
|
|
104
|
-
id: 'redundancy',
|
|
105
|
-
pattern: /\b(basic fundamentals|end result|free gift|future plans|past history|completely unique|absolutely essential|close proximity|each and every|first began|true fact|advance planning|final outcome)\b/gi,
|
|
106
|
-
message: 'Redundant phrase',
|
|
107
|
-
severity: 'warning',
|
|
108
|
-
},
|
|
109
|
-
];
|
|
110
|
-
|
|
111
|
-
/**
|
|
112
|
-
* Scientific writing specific rules
|
|
113
|
-
*/
|
|
114
|
-
const SCIENTIFIC_RULES: GrammarRule[] = [
|
|
115
|
-
{
|
|
116
|
-
id: 'first-person',
|
|
117
|
-
pattern: /\b(I|we|my|our)\b/gi,
|
|
118
|
-
message: 'First person pronoun - check if appropriate for your journal',
|
|
119
|
-
severity: 'info',
|
|
120
|
-
},
|
|
121
|
-
{
|
|
122
|
-
id: 'significant',
|
|
123
|
-
pattern: /\bsignificant(ly)?\b(?!\s+(P|p|α|difference|effect|increase|decrease|correlation))/gi,
|
|
124
|
-
message: '"Significant" without statistical context - clarify or use different word',
|
|
125
|
-
severity: 'warning',
|
|
126
|
-
},
|
|
127
|
-
{
|
|
128
|
-
id: 'prove',
|
|
129
|
-
pattern: /\b(prove[ds]?|proof)\b/gi,
|
|
130
|
-
message: 'Avoid "prove" in science - use "demonstrate", "show", "suggest"',
|
|
131
|
-
severity: 'warning',
|
|
132
|
-
},
|
|
133
|
-
{
|
|
134
|
-
id: 'obviously',
|
|
135
|
-
pattern: /\b(obviously|clearly|of course)\b/gi,
|
|
136
|
-
message: 'If obvious, no need to say so; if not obvious, this doesn\'t help',
|
|
137
|
-
severity: 'warning',
|
|
138
|
-
},
|
|
139
|
-
];
|
|
140
|
-
|
|
141
|
-
/**
|
|
142
|
-
* Load custom dictionary from file
|
|
143
|
-
*/
|
|
144
|
-
export function loadDictionary(directory: string = '.'): Set<string> {
|
|
145
|
-
const dictPath = path.join(directory, DEFAULT_DICT_NAME);
|
|
146
|
-
const words = new Set<string>();
|
|
147
|
-
|
|
148
|
-
if (fs.existsSync(dictPath)) {
|
|
149
|
-
const content = fs.readFileSync(dictPath, 'utf-8');
|
|
150
|
-
const lines = content.split('\n');
|
|
151
|
-
|
|
152
|
-
for (const line of lines) {
|
|
153
|
-
const word = line.trim().toLowerCase();
|
|
154
|
-
if (word && !word.startsWith('#')) {
|
|
155
|
-
words.add(word);
|
|
156
|
-
}
|
|
157
|
-
}
|
|
158
|
-
}
|
|
159
|
-
|
|
160
|
-
return words;
|
|
161
|
-
}
|
|
162
|
-
|
|
163
|
-
/**
|
|
164
|
-
* Save custom dictionary to file
|
|
165
|
-
*/
|
|
166
|
-
export function saveDictionary(words: Set<string>, directory: string = '.'): void {
|
|
167
|
-
const dictPath = path.join(directory, DEFAULT_DICT_NAME);
|
|
168
|
-
|
|
169
|
-
const header = `# Custom dictionary for docrev
|
|
170
|
-
# Add one word per line
|
|
171
|
-
# Lines starting with # are comments
|
|
172
|
-
`;
|
|
173
|
-
|
|
174
|
-
const content = header + [...words].sort().join('\n') + '\n';
|
|
175
|
-
fs.writeFileSync(dictPath, content, 'utf-8');
|
|
176
|
-
}
|
|
177
|
-
|
|
178
|
-
/**
|
|
179
|
-
* Add word to custom dictionary
|
|
180
|
-
*/
|
|
181
|
-
export function addToDictionary(word: string, directory: string = '.'): boolean {
|
|
182
|
-
const words = loadDictionary(directory);
|
|
183
|
-
const normalizedWord = word.trim().toLowerCase();
|
|
184
|
-
|
|
185
|
-
if (words.has(normalizedWord)) {
|
|
186
|
-
return false;
|
|
187
|
-
}
|
|
188
|
-
|
|
189
|
-
words.add(normalizedWord);
|
|
190
|
-
saveDictionary(words, directory);
|
|
191
|
-
return true;
|
|
192
|
-
}
|
|
193
|
-
|
|
194
|
-
/**
|
|
195
|
-
* Remove word from custom dictionary
|
|
196
|
-
*/
|
|
197
|
-
export function removeFromDictionary(word: string, directory: string = '.'): boolean {
|
|
198
|
-
const words = loadDictionary(directory);
|
|
199
|
-
const normalizedWord = word.trim().toLowerCase();
|
|
200
|
-
|
|
201
|
-
if (!words.has(normalizedWord)) {
|
|
202
|
-
return false;
|
|
203
|
-
}
|
|
204
|
-
|
|
205
|
-
words.delete(normalizedWord);
|
|
206
|
-
saveDictionary(words, directory);
|
|
207
|
-
return true;
|
|
208
|
-
}
|
|
209
|
-
|
|
210
|
-
/**
|
|
211
|
-
* Check text for grammar/style issues
|
|
212
|
-
*/
|
|
213
|
-
export function checkGrammar(text: string, options: CheckGrammarOptions = {}): GrammarIssue[] {
|
|
214
|
-
const { scientific = true, directory = '.' } = options;
|
|
215
|
-
const customDict = loadDictionary(directory);
|
|
216
|
-
const issues: GrammarIssue[] = [];
|
|
217
|
-
|
|
218
|
-
// Get all rules
|
|
219
|
-
const rules = scientific ? [...GRAMMAR_RULES, ...SCIENTIFIC_RULES] : GRAMMAR_RULES;
|
|
220
|
-
|
|
221
|
-
// Split into lines for line number tracking
|
|
222
|
-
const lines = text.split('\n');
|
|
223
|
-
|
|
224
|
-
for (let lineNum = 0; lineNum < lines.length; lineNum++) {
|
|
225
|
-
const line = lines[lineNum];
|
|
226
|
-
if (!line) continue;
|
|
227
|
-
|
|
228
|
-
// Skip code blocks and YAML frontmatter
|
|
229
|
-
if (line.trim().startsWith('```') || line.trim().startsWith('---')) {
|
|
230
|
-
continue;
|
|
231
|
-
}
|
|
232
|
-
|
|
233
|
-
// Skip lines that are just markdown syntax
|
|
234
|
-
if (/^[#\-*>|]/.test(line.trim()) && line.trim().length < 5) {
|
|
235
|
-
continue;
|
|
236
|
-
}
|
|
237
|
-
|
|
238
|
-
for (const rule of rules) {
|
|
239
|
-
const pattern = new RegExp(rule.pattern.source, rule.pattern.flags);
|
|
240
|
-
let match;
|
|
241
|
-
|
|
242
|
-
while ((match = pattern.exec(line)) !== null) {
|
|
243
|
-
// Check if rule has additional check function
|
|
244
|
-
if (rule.check && !rule.check(match[0])) {
|
|
245
|
-
continue;
|
|
246
|
-
}
|
|
247
|
-
|
|
248
|
-
// Skip if word is in custom dictionary
|
|
249
|
-
const word = match[0].toLowerCase();
|
|
250
|
-
if (customDict.has(word)) {
|
|
251
|
-
continue;
|
|
252
|
-
}
|
|
253
|
-
|
|
254
|
-
issues.push({
|
|
255
|
-
rule: rule.id,
|
|
256
|
-
message: rule.message,
|
|
257
|
-
severity: rule.severity,
|
|
258
|
-
line: lineNum + 1,
|
|
259
|
-
column: match.index + 1,
|
|
260
|
-
match: match[0],
|
|
261
|
-
context: line.trim(),
|
|
262
|
-
});
|
|
263
|
-
}
|
|
264
|
-
}
|
|
265
|
-
}
|
|
266
|
-
|
|
267
|
-
return issues;
|
|
268
|
-
}
|
|
269
|
-
|
|
270
|
-
/**
|
|
271
|
-
* Get grammar check summary
|
|
272
|
-
*/
|
|
273
|
-
export function getGrammarSummary(issues: GrammarIssue[]): GrammarSummary {
|
|
274
|
-
const summary: GrammarSummary = {
|
|
275
|
-
total: issues.length,
|
|
276
|
-
errors: 0,
|
|
277
|
-
warnings: 0,
|
|
278
|
-
info: 0,
|
|
279
|
-
byRule: {},
|
|
280
|
-
};
|
|
281
|
-
|
|
282
|
-
for (const issue of issues) {
|
|
283
|
-
if (issue.severity === 'error') summary.errors++;
|
|
284
|
-
else if (issue.severity === 'warning') summary.warnings++;
|
|
285
|
-
else summary.info++;
|
|
286
|
-
|
|
287
|
-
summary.byRule[issue.rule] = (summary.byRule[issue.rule] || 0) + 1;
|
|
288
|
-
}
|
|
289
|
-
|
|
290
|
-
return summary;
|
|
291
|
-
}
|
|
292
|
-
|
|
293
|
-
/**
|
|
294
|
-
* List available grammar rules
|
|
295
|
-
*/
|
|
296
|
-
export function listRules(scientific: boolean = true): Array<{ id: string; message: string; severity: string }> {
|
|
297
|
-
const rules = scientific ? [...GRAMMAR_RULES, ...SCIENTIFIC_RULES] : GRAMMAR_RULES;
|
|
298
|
-
return rules.map(r => ({
|
|
299
|
-
id: r.id,
|
|
300
|
-
message: r.message,
|
|
301
|
-
severity: r.severity,
|
|
302
|
-
}));
|
|
303
|
-
}
|
|
1
|
+
/**
|
|
2
|
+
* Grammar checker module with custom dictionary support
|
|
3
|
+
*
|
|
4
|
+
* Features:
|
|
5
|
+
* - Common grammar/style issues detection
|
|
6
|
+
* - Custom dictionary for project-specific terms
|
|
7
|
+
* - Learn mode to add words to dictionary
|
|
8
|
+
*/
|
|
9
|
+
|
|
10
|
+
import * as fs from 'fs';
|
|
11
|
+
import * as path from 'path';
|
|
12
|
+
|
|
13
|
+
// Default dictionary location
|
|
14
|
+
const DEFAULT_DICT_NAME = '.rev-dictionary';
|
|
15
|
+
|
|
16
|
+
interface GrammarRule {
|
|
17
|
+
id: string;
|
|
18
|
+
pattern: RegExp;
|
|
19
|
+
message: string;
|
|
20
|
+
severity: 'error' | 'warning' | 'info';
|
|
21
|
+
check?: (match: string) => boolean;
|
|
22
|
+
}
|
|
23
|
+
|
|
24
|
+
interface GrammarIssue {
|
|
25
|
+
rule: string;
|
|
26
|
+
message: string;
|
|
27
|
+
severity: 'error' | 'warning' | 'info';
|
|
28
|
+
line: number;
|
|
29
|
+
column: number;
|
|
30
|
+
match: string;
|
|
31
|
+
context: string;
|
|
32
|
+
}
|
|
33
|
+
|
|
34
|
+
interface GrammarSummary {
|
|
35
|
+
total: number;
|
|
36
|
+
errors: number;
|
|
37
|
+
warnings: number;
|
|
38
|
+
info: number;
|
|
39
|
+
byRule: Record<string, number>;
|
|
40
|
+
}
|
|
41
|
+
|
|
42
|
+
interface CheckGrammarOptions {
|
|
43
|
+
scientific?: boolean;
|
|
44
|
+
directory?: string;
|
|
45
|
+
}
|
|
46
|
+
|
|
47
|
+
/**
|
|
48
|
+
* Common grammar/style rules
|
|
49
|
+
*/
|
|
50
|
+
const GRAMMAR_RULES: GrammarRule[] = [
|
|
51
|
+
{
|
|
52
|
+
id: 'passive-voice',
|
|
53
|
+
pattern: /\b(is|are|was|were|be|been|being)\s+(being\s+)?\w+ed\b/gi,
|
|
54
|
+
message: 'Possible passive voice',
|
|
55
|
+
severity: 'info',
|
|
56
|
+
},
|
|
57
|
+
{
|
|
58
|
+
id: 'weasel-words',
|
|
59
|
+
pattern: /\b(very|really|quite|extremely|fairly|rather|somewhat|just)\b/gi,
|
|
60
|
+
message: 'Weasel word - consider removing or being more specific',
|
|
61
|
+
severity: 'warning',
|
|
62
|
+
},
|
|
63
|
+
{
|
|
64
|
+
id: 'weak-start',
|
|
65
|
+
pattern: /^\s*(There (is|are|was|were)|It is)\b/gmi,
|
|
66
|
+
message: 'Weak sentence start - consider restructuring',
|
|
67
|
+
severity: 'info',
|
|
68
|
+
},
|
|
69
|
+
{
|
|
70
|
+
id: 'duplicate-words',
|
|
71
|
+
pattern: /\b(\w+)\s+\1\b/gi,
|
|
72
|
+
message: 'Duplicate word',
|
|
73
|
+
severity: 'error',
|
|
74
|
+
},
|
|
75
|
+
{
|
|
76
|
+
id: 'split-infinitive',
|
|
77
|
+
pattern: /\bto\s+(\w+ly)\s+\w+\b/gi,
|
|
78
|
+
message: 'Split infinitive',
|
|
79
|
+
severity: 'info',
|
|
80
|
+
},
|
|
81
|
+
{
|
|
82
|
+
id: 'sentence-length',
|
|
83
|
+
pattern: /[^.!?]*[.!?]/g,
|
|
84
|
+
check: (match: string): boolean => {
|
|
85
|
+
const words = match.trim().split(/\s+/).length;
|
|
86
|
+
return words > 40;
|
|
87
|
+
},
|
|
88
|
+
message: 'Long sentence (>40 words) - consider breaking up',
|
|
89
|
+
severity: 'warning',
|
|
90
|
+
},
|
|
91
|
+
{
|
|
92
|
+
id: 'cliches',
|
|
93
|
+
pattern: /\b(at the end of the day|in terms of|it goes without saying|needless to say|as a matter of fact|first and foremost|last but not least)\b/gi,
|
|
94
|
+
message: 'Cliche - consider rephrasing',
|
|
95
|
+
severity: 'warning',
|
|
96
|
+
},
|
|
97
|
+
{
|
|
98
|
+
id: 'hedging',
|
|
99
|
+
pattern: /\b(seems to|appears to|tends to|might|may|could possibly|would seem)\b/gi,
|
|
100
|
+
message: 'Hedging language - be more direct if appropriate',
|
|
101
|
+
severity: 'info',
|
|
102
|
+
},
|
|
103
|
+
{
|
|
104
|
+
id: 'redundancy',
|
|
105
|
+
pattern: /\b(basic fundamentals|end result|free gift|future plans|past history|completely unique|absolutely essential|close proximity|each and every|first began|true fact|advance planning|final outcome)\b/gi,
|
|
106
|
+
message: 'Redundant phrase',
|
|
107
|
+
severity: 'warning',
|
|
108
|
+
},
|
|
109
|
+
];
|
|
110
|
+
|
|
111
|
+
/**
|
|
112
|
+
* Scientific writing specific rules
|
|
113
|
+
*/
|
|
114
|
+
const SCIENTIFIC_RULES: GrammarRule[] = [
|
|
115
|
+
{
|
|
116
|
+
id: 'first-person',
|
|
117
|
+
pattern: /\b(I|we|my|our)\b/gi,
|
|
118
|
+
message: 'First person pronoun - check if appropriate for your journal',
|
|
119
|
+
severity: 'info',
|
|
120
|
+
},
|
|
121
|
+
{
|
|
122
|
+
id: 'significant',
|
|
123
|
+
pattern: /\bsignificant(ly)?\b(?!\s+(P|p|α|difference|effect|increase|decrease|correlation))/gi,
|
|
124
|
+
message: '"Significant" without statistical context - clarify or use different word',
|
|
125
|
+
severity: 'warning',
|
|
126
|
+
},
|
|
127
|
+
{
|
|
128
|
+
id: 'prove',
|
|
129
|
+
pattern: /\b(prove[ds]?|proof)\b/gi,
|
|
130
|
+
message: 'Avoid "prove" in science - use "demonstrate", "show", "suggest"',
|
|
131
|
+
severity: 'warning',
|
|
132
|
+
},
|
|
133
|
+
{
|
|
134
|
+
id: 'obviously',
|
|
135
|
+
pattern: /\b(obviously|clearly|of course)\b/gi,
|
|
136
|
+
message: 'If obvious, no need to say so; if not obvious, this doesn\'t help',
|
|
137
|
+
severity: 'warning',
|
|
138
|
+
},
|
|
139
|
+
];
|
|
140
|
+
|
|
141
|
+
/**
|
|
142
|
+
* Load custom dictionary from file
|
|
143
|
+
*/
|
|
144
|
+
export function loadDictionary(directory: string = '.'): Set<string> {
|
|
145
|
+
const dictPath = path.join(directory, DEFAULT_DICT_NAME);
|
|
146
|
+
const words = new Set<string>();
|
|
147
|
+
|
|
148
|
+
if (fs.existsSync(dictPath)) {
|
|
149
|
+
const content = fs.readFileSync(dictPath, 'utf-8');
|
|
150
|
+
const lines = content.split('\n');
|
|
151
|
+
|
|
152
|
+
for (const line of lines) {
|
|
153
|
+
const word = line.trim().toLowerCase();
|
|
154
|
+
if (word && !word.startsWith('#')) {
|
|
155
|
+
words.add(word);
|
|
156
|
+
}
|
|
157
|
+
}
|
|
158
|
+
}
|
|
159
|
+
|
|
160
|
+
return words;
|
|
161
|
+
}
|
|
162
|
+
|
|
163
|
+
/**
|
|
164
|
+
* Save custom dictionary to file
|
|
165
|
+
*/
|
|
166
|
+
export function saveDictionary(words: Set<string>, directory: string = '.'): void {
|
|
167
|
+
const dictPath = path.join(directory, DEFAULT_DICT_NAME);
|
|
168
|
+
|
|
169
|
+
const header = `# Custom dictionary for docrev
|
|
170
|
+
# Add one word per line
|
|
171
|
+
# Lines starting with # are comments
|
|
172
|
+
`;
|
|
173
|
+
|
|
174
|
+
const content = header + [...words].sort().join('\n') + '\n';
|
|
175
|
+
fs.writeFileSync(dictPath, content, 'utf-8');
|
|
176
|
+
}
|
|
177
|
+
|
|
178
|
+
/**
|
|
179
|
+
* Add word to custom dictionary
|
|
180
|
+
*/
|
|
181
|
+
export function addToDictionary(word: string, directory: string = '.'): boolean {
|
|
182
|
+
const words = loadDictionary(directory);
|
|
183
|
+
const normalizedWord = word.trim().toLowerCase();
|
|
184
|
+
|
|
185
|
+
if (words.has(normalizedWord)) {
|
|
186
|
+
return false;
|
|
187
|
+
}
|
|
188
|
+
|
|
189
|
+
words.add(normalizedWord);
|
|
190
|
+
saveDictionary(words, directory);
|
|
191
|
+
return true;
|
|
192
|
+
}
|
|
193
|
+
|
|
194
|
+
/**
|
|
195
|
+
* Remove word from custom dictionary
|
|
196
|
+
*/
|
|
197
|
+
export function removeFromDictionary(word: string, directory: string = '.'): boolean {
|
|
198
|
+
const words = loadDictionary(directory);
|
|
199
|
+
const normalizedWord = word.trim().toLowerCase();
|
|
200
|
+
|
|
201
|
+
if (!words.has(normalizedWord)) {
|
|
202
|
+
return false;
|
|
203
|
+
}
|
|
204
|
+
|
|
205
|
+
words.delete(normalizedWord);
|
|
206
|
+
saveDictionary(words, directory);
|
|
207
|
+
return true;
|
|
208
|
+
}
|
|
209
|
+
|
|
210
|
+
/**
|
|
211
|
+
* Check text for grammar/style issues
|
|
212
|
+
*/
|
|
213
|
+
export function checkGrammar(text: string, options: CheckGrammarOptions = {}): GrammarIssue[] {
|
|
214
|
+
const { scientific = true, directory = '.' } = options;
|
|
215
|
+
const customDict = loadDictionary(directory);
|
|
216
|
+
const issues: GrammarIssue[] = [];
|
|
217
|
+
|
|
218
|
+
// Get all rules
|
|
219
|
+
const rules = scientific ? [...GRAMMAR_RULES, ...SCIENTIFIC_RULES] : GRAMMAR_RULES;
|
|
220
|
+
|
|
221
|
+
// Split into lines for line number tracking
|
|
222
|
+
const lines = text.split('\n');
|
|
223
|
+
|
|
224
|
+
for (let lineNum = 0; lineNum < lines.length; lineNum++) {
|
|
225
|
+
const line = lines[lineNum];
|
|
226
|
+
if (!line) continue;
|
|
227
|
+
|
|
228
|
+
// Skip code blocks and YAML frontmatter
|
|
229
|
+
if (line.trim().startsWith('```') || line.trim().startsWith('---')) {
|
|
230
|
+
continue;
|
|
231
|
+
}
|
|
232
|
+
|
|
233
|
+
// Skip lines that are just markdown syntax
|
|
234
|
+
if (/^[#\-*>|]/.test(line.trim()) && line.trim().length < 5) {
|
|
235
|
+
continue;
|
|
236
|
+
}
|
|
237
|
+
|
|
238
|
+
for (const rule of rules) {
|
|
239
|
+
const pattern = new RegExp(rule.pattern.source, rule.pattern.flags);
|
|
240
|
+
let match;
|
|
241
|
+
|
|
242
|
+
while ((match = pattern.exec(line)) !== null) {
|
|
243
|
+
// Check if rule has additional check function
|
|
244
|
+
if (rule.check && !rule.check(match[0])) {
|
|
245
|
+
continue;
|
|
246
|
+
}
|
|
247
|
+
|
|
248
|
+
// Skip if word is in custom dictionary
|
|
249
|
+
const word = match[0].toLowerCase();
|
|
250
|
+
if (customDict.has(word)) {
|
|
251
|
+
continue;
|
|
252
|
+
}
|
|
253
|
+
|
|
254
|
+
issues.push({
|
|
255
|
+
rule: rule.id,
|
|
256
|
+
message: rule.message,
|
|
257
|
+
severity: rule.severity,
|
|
258
|
+
line: lineNum + 1,
|
|
259
|
+
column: match.index + 1,
|
|
260
|
+
match: match[0],
|
|
261
|
+
context: line.trim(),
|
|
262
|
+
});
|
|
263
|
+
}
|
|
264
|
+
}
|
|
265
|
+
}
|
|
266
|
+
|
|
267
|
+
return issues;
|
|
268
|
+
}
|
|
269
|
+
|
|
270
|
+
/**
|
|
271
|
+
* Get grammar check summary
|
|
272
|
+
*/
|
|
273
|
+
export function getGrammarSummary(issues: GrammarIssue[]): GrammarSummary {
|
|
274
|
+
const summary: GrammarSummary = {
|
|
275
|
+
total: issues.length,
|
|
276
|
+
errors: 0,
|
|
277
|
+
warnings: 0,
|
|
278
|
+
info: 0,
|
|
279
|
+
byRule: {},
|
|
280
|
+
};
|
|
281
|
+
|
|
282
|
+
for (const issue of issues) {
|
|
283
|
+
if (issue.severity === 'error') summary.errors++;
|
|
284
|
+
else if (issue.severity === 'warning') summary.warnings++;
|
|
285
|
+
else summary.info++;
|
|
286
|
+
|
|
287
|
+
summary.byRule[issue.rule] = (summary.byRule[issue.rule] || 0) + 1;
|
|
288
|
+
}
|
|
289
|
+
|
|
290
|
+
return summary;
|
|
291
|
+
}
|
|
292
|
+
|
|
293
|
+
/**
|
|
294
|
+
* List available grammar rules
|
|
295
|
+
*/
|
|
296
|
+
export function listRules(scientific: boolean = true): Array<{ id: string; message: string; severity: string }> {
|
|
297
|
+
const rules = scientific ? [...GRAMMAR_RULES, ...SCIENTIFIC_RULES] : GRAMMAR_RULES;
|
|
298
|
+
return rules.map(r => ({
|
|
299
|
+
id: r.id,
|
|
300
|
+
message: r.message,
|
|
301
|
+
severity: r.severity,
|
|
302
|
+
}));
|
|
303
|
+
}
|