sentinel-scanner 1.1.0-alpha.1 → 1.1.0-alpha.2
Sign up to get free protection for your applications and to get access to all the features.
- package/.cspell.json +19 -5
- package/.github/workflows/pr.yaml +86 -0
- package/.github/workflows/welcome.yaml +66 -0
- package/CHANGELOG.md +2 -3
- package/DISCLAIMER.md +64 -0
- package/LICENSE +2 -2
- package/README.md +20 -1
- package/build/bin.js +376 -0
- package/build/bin.js.map +7 -0
- package/build/index.d.ts +24 -0
- package/build/index.js +115 -158
- package/build/index.js.map +3 -3
- package/package.json +6 -8
- package/scripts/build.ts +4 -1
- package/src/bin.ts +20 -0
- package/src/commands/spider.ts +193 -0
- package/src/index.ts +2 -85
- package/src/modules/spider/index.ts +104 -67
- package/src/utils/index.ts +29 -0
- package/tsconfig.json +5 -10
- package/src/__tests__/spider.test.ts +0 -1
- package/src/lib/logger.ts +0 -43
- package/src/modules/index.ts +0 -3
@@ -0,0 +1,193 @@
|
|
1
|
+
import fs from "node:fs";
|
2
|
+
import path from "node:path";
|
3
|
+
import type { ArgumentsCamelCase, CommandModule } from "yargs";
|
4
|
+
import SpiderScanner from "../modules/spider";
|
5
|
+
import { createLogger } from "../utils";
|
6
|
+
|
7
|
+
export type SpiderScannerCLIOptions = {
|
8
|
+
url: string;
|
9
|
+
depth?: number;
|
10
|
+
output?: string;
|
11
|
+
concurrency?: number;
|
12
|
+
timeout?: number;
|
13
|
+
retries?: number;
|
14
|
+
};
|
15
|
+
|
16
|
+
const cliLogger = createLogger("CLI");
|
17
|
+
|
18
|
+
export const spiderCommand: CommandModule = {
|
19
|
+
command: "spider",
|
20
|
+
describe:
|
21
|
+
"Crawl a website and get an array of URLs which are internal to the website",
|
22
|
+
builder: (yargs) => {
|
23
|
+
return yargs
|
24
|
+
.option("url", {
|
25
|
+
alias: "u",
|
26
|
+
type: "string",
|
27
|
+
description: "The URL of the website to scan",
|
28
|
+
demandOption: true,
|
29
|
+
coerce: (url) => {
|
30
|
+
try {
|
31
|
+
new URL(url);
|
32
|
+
|
33
|
+
return url;
|
34
|
+
} catch (error) {
|
35
|
+
throw new Error(`Invalid URL: ${url}`);
|
36
|
+
}
|
37
|
+
},
|
38
|
+
})
|
39
|
+
.option("depth", {
|
40
|
+
alias: "d",
|
41
|
+
type: "number",
|
42
|
+
description: "The maximum depth to crawl",
|
43
|
+
default: 250,
|
44
|
+
coerce: (depth) => {
|
45
|
+
if (depth < 0) {
|
46
|
+
throw new Error("Depth must be a positive number");
|
47
|
+
}
|
48
|
+
|
49
|
+
if (depth > 250) {
|
50
|
+
throw new Error("Depth must be less than 250");
|
51
|
+
}
|
52
|
+
|
53
|
+
return depth;
|
54
|
+
},
|
55
|
+
})
|
56
|
+
.option("output", {
|
57
|
+
alias: "o",
|
58
|
+
type: "string",
|
59
|
+
description:
|
60
|
+
"The output file to write the results to. Must be a JSON file",
|
61
|
+
coerce: (output) => {
|
62
|
+
try {
|
63
|
+
// Should throw an error if the path is invalid
|
64
|
+
// Should Be A JSON File
|
65
|
+
const resolvedPath = path.resolve(output);
|
66
|
+
const parsedPath = path.parse(resolvedPath);
|
67
|
+
|
68
|
+
if (parsedPath.ext !== ".json") {
|
69
|
+
throw new Error("Output file must be a JSON file");
|
70
|
+
}
|
71
|
+
|
72
|
+
if (fs.existsSync(resolvedPath)) {
|
73
|
+
throw new Error("Output file already exists");
|
74
|
+
}
|
75
|
+
|
76
|
+
return resolvedPath;
|
77
|
+
} catch (error) {
|
78
|
+
throw new Error(`Invalid output file: ${output}`);
|
79
|
+
}
|
80
|
+
},
|
81
|
+
default: getDefaultFilePath(),
|
82
|
+
})
|
83
|
+
.option("concurrency", {
|
84
|
+
alias: "c",
|
85
|
+
type: "number",
|
86
|
+
description: "The number of concurrent requests to make",
|
87
|
+
default: 10,
|
88
|
+
coerce: (concurrency) => {
|
89
|
+
if (concurrency < 1) {
|
90
|
+
throw new Error("Concurrency must be a positive number");
|
91
|
+
}
|
92
|
+
|
93
|
+
if (concurrency > 20) {
|
94
|
+
throw new Error("Concurrency must be less than 20");
|
95
|
+
}
|
96
|
+
|
97
|
+
return concurrency;
|
98
|
+
},
|
99
|
+
})
|
100
|
+
.option("timeout", {
|
101
|
+
alias: "t",
|
102
|
+
type: "number",
|
103
|
+
description: "The timeout for each request in milliseconds",
|
104
|
+
default: 5000,
|
105
|
+
coerce: (timeout) => {
|
106
|
+
if (timeout < 0) {
|
107
|
+
throw new Error("Timeout must be a positive number");
|
108
|
+
}
|
109
|
+
|
110
|
+
if (timeout > 25_000) {
|
111
|
+
throw new Error("Timeout must be less than 25,000");
|
112
|
+
}
|
113
|
+
|
114
|
+
return timeout;
|
115
|
+
},
|
116
|
+
})
|
117
|
+
.option("retries", {
|
118
|
+
alias: "r",
|
119
|
+
type: "number",
|
120
|
+
description: "The number of retries for each request",
|
121
|
+
default: 3,
|
122
|
+
coerce: (retries) => {
|
123
|
+
if (retries < 0) {
|
124
|
+
throw new Error("Retries must be a positive number");
|
125
|
+
}
|
126
|
+
|
127
|
+
if (retries > 10) {
|
128
|
+
throw new Error("Retries must be less than 10");
|
129
|
+
}
|
130
|
+
|
131
|
+
return retries;
|
132
|
+
},
|
133
|
+
});
|
134
|
+
},
|
135
|
+
handler: async (args) => {
|
136
|
+
try {
|
137
|
+
const argData = args as ArgumentsCamelCase<SpiderScannerCLIOptions>;
|
138
|
+
|
139
|
+
const scanner = new SpiderScanner(argData.url, {
|
140
|
+
depth: argData.depth ?? 250,
|
141
|
+
concurrency: argData.concurrency ?? 10,
|
142
|
+
timeout: argData.timeout ?? 5000,
|
143
|
+
retries: argData.retries ?? 3,
|
144
|
+
});
|
145
|
+
|
146
|
+
cliLogger.info("Starting to crawl website");
|
147
|
+
|
148
|
+
const results = await scanner.crawl();
|
149
|
+
|
150
|
+
if (argData.output) {
|
151
|
+
fs.writeFileSync(argData.output, JSON.stringify(results, null, 2));
|
152
|
+
cliLogger.info(`Results written to ${argData.output}`);
|
153
|
+
} else {
|
154
|
+
const resolvedPath = getDefaultFilePath();
|
155
|
+
fs.writeFileSync(resolvedPath, JSON.stringify(results, null, 2));
|
156
|
+
cliLogger.info(`Results written to ${resolvedPath}`);
|
157
|
+
}
|
158
|
+
} catch (error) {
|
159
|
+
if (error instanceof Error) {
|
160
|
+
cliLogger.error(error.message);
|
161
|
+
}
|
162
|
+
cliLogger.error("Failed to run spider command");
|
163
|
+
process.exit(1);
|
164
|
+
}
|
165
|
+
},
|
166
|
+
};
|
167
|
+
|
168
|
+
const getDefaultFilePath = () => {
|
169
|
+
try {
|
170
|
+
const resolvedDir = path.resolve("sentinel_output");
|
171
|
+
// Check If Directory Exists
|
172
|
+
if (!fs.existsSync(resolvedDir)) {
|
173
|
+
fs.mkdirSync(resolvedDir);
|
174
|
+
}
|
175
|
+
|
176
|
+
const resolvedPath = path.resolve(
|
177
|
+
`sentinel_output/spider_${Date.now()}.json`,
|
178
|
+
);
|
179
|
+
// Check If File Exists
|
180
|
+
if (fs.existsSync(resolvedPath)) {
|
181
|
+
throw new Error("Output file already exists");
|
182
|
+
}
|
183
|
+
const parsedPath = path.parse(resolvedPath);
|
184
|
+
|
185
|
+
if (parsedPath.ext !== ".json") {
|
186
|
+
throw new Error("Output file must be a JSON file");
|
187
|
+
}
|
188
|
+
|
189
|
+
return resolvedPath;
|
190
|
+
} catch (error) {
|
191
|
+
throw new Error("Invalid output file");
|
192
|
+
}
|
193
|
+
};
|
package/src/index.ts
CHANGED
@@ -1,86 +1,3 @@
|
|
1
|
-
|
1
|
+
import SpiderScanner, { type SpiderScannerOptions } from "./modules/spider";
|
2
2
|
|
3
|
-
|
4
|
-
import { hideBin } from "yargs/helpers";
|
5
|
-
import { SpiderScanner } from "./modules";
|
6
|
-
|
7
|
-
const commandHandler = yargs(hideBin(process.argv));
|
8
|
-
|
9
|
-
/**
|
10
|
-
* Command to scan for XSS vulnerabilities
|
11
|
-
*
|
12
|
-
* @param {string} url - URL to scan
|
13
|
-
* @param {string} wordlist - Path to wordlist file
|
14
|
-
* @returns {void}
|
15
|
-
*
|
16
|
-
* @example
|
17
|
-
* npx sentinel-scanner xss --url https://example.com
|
18
|
-
*/
|
19
|
-
commandHandler.command(
|
20
|
-
"xss",
|
21
|
-
"Scan for XSS vulnerabilities",
|
22
|
-
{
|
23
|
-
url: {
|
24
|
-
describe: "URL to scan",
|
25
|
-
demandOption: true,
|
26
|
-
type: "string",
|
27
|
-
coerce: (value) => {
|
28
|
-
try {
|
29
|
-
new URL(value);
|
30
|
-
return value;
|
31
|
-
} catch (err) {
|
32
|
-
throw new Error("Invalid URL format");
|
33
|
-
}
|
34
|
-
},
|
35
|
-
},
|
36
|
-
wordlist: {
|
37
|
-
describe: "Path to wordlist file",
|
38
|
-
type: "string",
|
39
|
-
},
|
40
|
-
},
|
41
|
-
(argv) => {
|
42
|
-
console.log("Scanning for XSS vulnerabilities...");
|
43
|
-
console.log(`URL: ${argv.url}`);
|
44
|
-
console.log(`Wordlist: ${argv.wordlist || "Default"}`);
|
45
|
-
},
|
46
|
-
);
|
47
|
-
|
48
|
-
// Command to Spider a website
|
49
|
-
commandHandler.command(
|
50
|
-
"spider",
|
51
|
-
"Scan a website for vulnerabilities",
|
52
|
-
{
|
53
|
-
url: {
|
54
|
-
describe: "URL to scan",
|
55
|
-
demandOption: true,
|
56
|
-
type: "string",
|
57
|
-
coerce: (value) => {
|
58
|
-
try {
|
59
|
-
new URL(value);
|
60
|
-
return value;
|
61
|
-
} catch (err) {
|
62
|
-
throw new Error("Invalid URL format");
|
63
|
-
}
|
64
|
-
},
|
65
|
-
},
|
66
|
-
},
|
67
|
-
(argv) => {
|
68
|
-
const spider = new SpiderScanner(argv.url);
|
69
|
-
|
70
|
-
spider.crawl().then((output) => {
|
71
|
-
console.log(
|
72
|
-
JSON.stringify(
|
73
|
-
{
|
74
|
-
forms: output.forms,
|
75
|
-
links: output.links,
|
76
|
-
},
|
77
|
-
null,
|
78
|
-
2,
|
79
|
-
),
|
80
|
-
);
|
81
|
-
});
|
82
|
-
},
|
83
|
-
);
|
84
|
-
|
85
|
-
// Parse arguments and handle commands
|
86
|
-
commandHandler.parse();
|
3
|
+
export { SpiderScanner, type SpiderScannerOptions };
|
@@ -1,31 +1,43 @@
|
|
1
1
|
import fetch from "isomorphic-fetch";
|
2
2
|
import jsdom from "jsdom";
|
3
3
|
import UserAgent from "user-agents";
|
4
|
-
import
|
4
|
+
import { createLogger } from "../../utils";
|
5
5
|
|
6
|
-
export
|
7
|
-
|
8
|
-
|
9
|
-
|
10
|
-
|
11
|
-
|
12
|
-
export type CrawlOutput = {
|
13
|
-
links: string[];
|
14
|
-
forms: FormOutput[];
|
15
|
-
};
|
6
|
+
export interface SpiderScannerOptions {
|
7
|
+
depth?: number;
|
8
|
+
concurrency?: number;
|
9
|
+
retries?: number;
|
10
|
+
timeout?: number;
|
11
|
+
}
|
16
12
|
|
17
13
|
export default class SpiderScanner {
|
18
14
|
private header: Record<string, string> = {
|
19
15
|
"User-Agent": new UserAgent().toString(),
|
20
16
|
};
|
21
17
|
private url: URL;
|
22
|
-
private logger =
|
18
|
+
private logger = createLogger("SpiderScanner");
|
19
|
+
|
20
|
+
private depth: number;
|
21
|
+
private concurrency: number;
|
22
|
+
private retries: number;
|
23
|
+
private timeout: number;
|
24
|
+
|
25
|
+
constructor(url: string, options: SpiderScannerOptions = {}) {
|
26
|
+
const {
|
27
|
+
depth = 250,
|
28
|
+
concurrency = 5,
|
29
|
+
retries = 3,
|
30
|
+
timeout = 5000,
|
31
|
+
} = options;
|
32
|
+
this.depth = depth;
|
33
|
+
this.concurrency = concurrency;
|
34
|
+
this.retries = retries;
|
35
|
+
this.timeout = timeout;
|
23
36
|
|
24
|
-
constructor(url: string) {
|
25
37
|
try {
|
26
38
|
this.url = new URL(url);
|
27
39
|
this.logger.info(
|
28
|
-
`Initialized with URL: ${url}
|
40
|
+
`Initialized with URL: ${url}, User-Agent: ${this.header["User-Agent"]}`,
|
29
41
|
);
|
30
42
|
} catch (error) {
|
31
43
|
if (error instanceof TypeError) {
|
@@ -37,7 +49,6 @@ export default class SpiderScanner {
|
|
37
49
|
}
|
38
50
|
}
|
39
51
|
|
40
|
-
// Normalize domains (removes 'www.')
|
41
52
|
private normalizeDomain(domain: string): string {
|
42
53
|
return domain.startsWith("www.") ? domain.slice(4) : domain;
|
43
54
|
}
|
@@ -61,20 +72,42 @@ export default class SpiderScanner {
|
|
61
72
|
}
|
62
73
|
}
|
63
74
|
|
64
|
-
private async
|
65
|
-
|
66
|
-
|
67
|
-
|
68
|
-
|
75
|
+
private async fetchWithRetries(
|
76
|
+
url: string,
|
77
|
+
retries: number,
|
78
|
+
): Promise<string | null> {
|
79
|
+
for (let attempt = 1; attempt <= retries; attempt++) {
|
80
|
+
const controller = new AbortController();
|
81
|
+
const timeoutId = setTimeout(() => controller.abort(), this.timeout);
|
82
|
+
|
83
|
+
try {
|
84
|
+
this.logger.debug(`Fetching URL (Attempt ${attempt}): ${url}`);
|
85
|
+
const randomUserAgent = new UserAgent().toString();
|
86
|
+
this.logger.info(`Changing User-Agent to: ${randomUserAgent}`);
|
87
|
+
this.header["User-Agent"] = randomUserAgent;
|
88
|
+
const response = await fetch(url, {
|
89
|
+
headers: this.header,
|
90
|
+
signal: controller.signal,
|
91
|
+
redirect: "follow",
|
92
|
+
});
|
93
|
+
|
94
|
+
clearTimeout(timeoutId);
|
95
|
+
|
96
|
+
if (response.ok) {
|
97
|
+
this.logger.info(`Successfully fetched URL: ${url}`);
|
98
|
+
return await response.text();
|
99
|
+
}
|
100
|
+
|
69
101
|
this.logger.warn(`Failed to fetch URL (${response.status}): ${url}`);
|
70
|
-
|
102
|
+
} catch (error) {
|
103
|
+
if ((error as Error).name === "AbortError") {
|
104
|
+
this.logger.warn(`Fetch timed out: ${url}`);
|
105
|
+
} else {
|
106
|
+
this.logger.error(`Error fetching URL: ${url} - ${error}`);
|
107
|
+
}
|
71
108
|
}
|
72
|
-
this.logger.info(`Successfully fetched URL: ${url}`);
|
73
|
-
return await response.text();
|
74
|
-
} catch (error) {
|
75
|
-
this.logger.error(`Error fetching URL: ${url} - ${error}`);
|
76
|
-
return null;
|
77
109
|
}
|
110
|
+
return null;
|
78
111
|
}
|
79
112
|
|
80
113
|
private extractLinks(html: string): string[] {
|
@@ -89,36 +122,40 @@ export default class SpiderScanner {
|
|
89
122
|
return internalLinks.map((link) => this.convertRelativeUrlToAbsolute(link));
|
90
123
|
}
|
91
124
|
|
92
|
-
|
93
|
-
const { JSDOM } = jsdom;
|
94
|
-
const dom = new JSDOM(html);
|
95
|
-
const forms = Array.from(dom.window.document.querySelectorAll("form"));
|
96
|
-
this.logger.debug(`Extracted ${forms.length} forms from HTML content`);
|
97
|
-
|
98
|
-
return forms.map((form, index) => {
|
99
|
-
const fields = Array.from(form.querySelectorAll("input")).map(
|
100
|
-
(input) => ({
|
101
|
-
name: input.name,
|
102
|
-
id: input.id,
|
103
|
-
class: input.className,
|
104
|
-
type: input.type,
|
105
|
-
}),
|
106
|
-
);
|
107
|
-
|
108
|
-
return {
|
109
|
-
id: index,
|
110
|
-
url: this.url.href,
|
111
|
-
fields,
|
112
|
-
};
|
113
|
-
});
|
114
|
-
}
|
115
|
-
|
116
|
-
// Main function to scan the website with concurrency support and return both links and forms
|
117
|
-
public async crawl(depth = 250, concurrency = 5): Promise<CrawlOutput> {
|
125
|
+
public async crawl(): Promise<Array<string>> {
|
118
126
|
const visited = new Set<string>();
|
119
127
|
const queue = new Set<string>([this.url.href]);
|
120
128
|
const resultLinks = new Set<string>();
|
121
|
-
|
129
|
+
|
130
|
+
// Assets to ignore
|
131
|
+
const assetExtensions = [
|
132
|
+
".css",
|
133
|
+
".js",
|
134
|
+
".png",
|
135
|
+
".jpg",
|
136
|
+
".jpeg",
|
137
|
+
".gif",
|
138
|
+
".svg",
|
139
|
+
".ico",
|
140
|
+
".webp",
|
141
|
+
".mp4",
|
142
|
+
".mp3",
|
143
|
+
".wav",
|
144
|
+
".avi",
|
145
|
+
".mov",
|
146
|
+
".webm",
|
147
|
+
".pdf",
|
148
|
+
".doc",
|
149
|
+
".docx",
|
150
|
+
".xls",
|
151
|
+
".xlsx",
|
152
|
+
".ppt",
|
153
|
+
".pptx",
|
154
|
+
".zip",
|
155
|
+
".rar",
|
156
|
+
".tar",
|
157
|
+
".gz",
|
158
|
+
];
|
122
159
|
|
123
160
|
const fetchAndExtract = async (currentUrl: string) => {
|
124
161
|
if (visited.has(currentUrl)) {
|
@@ -128,19 +165,22 @@ export default class SpiderScanner {
|
|
128
165
|
visited.add(currentUrl);
|
129
166
|
this.logger.info(`Visiting URL: ${currentUrl}`);
|
130
167
|
|
131
|
-
const html = await this.
|
168
|
+
const html = await this.fetchWithRetries(currentUrl, this.retries);
|
132
169
|
if (!html) return;
|
133
170
|
|
134
|
-
// Extract links and forms
|
135
171
|
const links = this.extractLinks(html);
|
136
|
-
const forms = this.extractForms(html);
|
137
172
|
|
138
|
-
|
139
|
-
|
173
|
+
// Filter out asset links
|
174
|
+
for (const link of links) {
|
175
|
+
if (assetExtensions.some((ext) => link.endsWith(ext))) {
|
176
|
+
this.logger.debug(`Ignoring asset link: ${link}`);
|
177
|
+
continue;
|
178
|
+
}
|
179
|
+
this.logger.debug(`Found link: ${link}`);
|
140
180
|
}
|
141
181
|
|
142
182
|
for (const link of links) {
|
143
|
-
if (!visited.has(link) && queue.size < depth) {
|
183
|
+
if (!visited.has(link) && queue.size < this.depth) {
|
144
184
|
queue.add(link);
|
145
185
|
this.logger.debug(`Added to queue: ${link}`);
|
146
186
|
}
|
@@ -149,7 +189,7 @@ export default class SpiderScanner {
|
|
149
189
|
};
|
150
190
|
|
151
191
|
const processBatch = async () => {
|
152
|
-
const batch = Array.from(queue).slice(0, concurrency);
|
192
|
+
const batch = Array.from(queue).slice(0, this.concurrency);
|
153
193
|
for (const url of batch) {
|
154
194
|
queue.delete(url);
|
155
195
|
}
|
@@ -157,19 +197,16 @@ export default class SpiderScanner {
|
|
157
197
|
};
|
158
198
|
|
159
199
|
this.logger.info(
|
160
|
-
`Starting crawl with depth: ${depth}, concurrency: ${concurrency}`,
|
200
|
+
`Starting crawl with depth: ${this.depth}, concurrency: ${this.concurrency}`,
|
161
201
|
);
|
162
|
-
while (queue.size > 0 && visited.size < depth) {
|
202
|
+
while (queue.size > 0 && visited.size < this.depth) {
|
163
203
|
await processBatch();
|
164
204
|
}
|
165
205
|
|
166
206
|
this.logger.info(
|
167
|
-
`Crawling completed. Total pages visited: ${resultLinks.size}
|
207
|
+
`Crawling completed. Total pages visited: ${resultLinks.size}`,
|
168
208
|
);
|
169
209
|
|
170
|
-
return
|
171
|
-
links: Array.from(resultLinks),
|
172
|
-
forms: Array.from(resultForms),
|
173
|
-
};
|
210
|
+
return Array.from(resultLinks);
|
174
211
|
}
|
175
212
|
}
|
@@ -0,0 +1,29 @@
|
|
1
|
+
import winston from "winston";
|
2
|
+
|
3
|
+
export const createLogger = (label: string) =>
|
4
|
+
winston.createLogger({
|
5
|
+
levels: {
|
6
|
+
error: 0,
|
7
|
+
warn: 1,
|
8
|
+
info: 2,
|
9
|
+
http: 3,
|
10
|
+
verbose: 4,
|
11
|
+
debug: 5,
|
12
|
+
silly: 6,
|
13
|
+
},
|
14
|
+
format: winston.format.combine(
|
15
|
+
winston.format.label({ label }),
|
16
|
+
winston.format.colorize(),
|
17
|
+
winston.format.timestamp({
|
18
|
+
format: () => {
|
19
|
+
return new Date().toLocaleString("en-US");
|
20
|
+
},
|
21
|
+
}),
|
22
|
+
winston.format.align(),
|
23
|
+
winston.format.printf(
|
24
|
+
(info) =>
|
25
|
+
`\x1b[34m(${info.label})\x1b[0m \x1b[33m${info.timestamp}\x1b[0m [${info.level}]: ${info.message}`,
|
26
|
+
),
|
27
|
+
),
|
28
|
+
transports: [new winston.transports.Console()],
|
29
|
+
});
|
package/tsconfig.json
CHANGED
@@ -2,9 +2,10 @@
|
|
2
2
|
"include": ["./src/**/*.ts"],
|
3
3
|
"compilerOptions": {
|
4
4
|
"lib": ["es2023"],
|
5
|
-
"module": "
|
5
|
+
"module": "CommonJS",
|
6
6
|
"target": "es2022",
|
7
7
|
"moduleResolution": "node",
|
8
|
+
"allowSyntheticDefaultImports": true,
|
8
9
|
|
9
10
|
"rootDir": "./src",
|
10
11
|
"outDir": "build",
|
@@ -17,13 +18,7 @@
|
|
17
18
|
"forceConsistentCasingInFileNames": true,
|
18
19
|
"declaration": true,
|
19
20
|
"resolveJsonModule": true,
|
20
|
-
"
|
21
|
-
|
22
|
-
|
23
|
-
"allowSyntheticDefaultImports": true,
|
24
|
-
"isolatedModules": true,
|
25
|
-
|
26
|
-
// Ensure TypeScript recognizes .ts file extensions in ESM
|
27
|
-
"allowImportingTsExtensions": true
|
28
|
-
}
|
21
|
+
"allowImportingTsExtensions": false
|
22
|
+
},
|
23
|
+
"exclude": ["src/__tests__/**/*"]
|
29
24
|
}
|
@@ -1 +0,0 @@
|
|
1
|
-
import { describe, it, test } from "node:test";
|
package/src/lib/logger.ts
DELETED
@@ -1,43 +0,0 @@
|
|
1
|
-
export default class Logger {
|
2
|
-
private moduleName: string;
|
3
|
-
private colors = {
|
4
|
-
error: "\x1b[31m",
|
5
|
-
info: "\x1b[32m",
|
6
|
-
warn: "\x1b[33m",
|
7
|
-
debug: "\x1b[35m",
|
8
|
-
reset: "\x1b[0m",
|
9
|
-
module: "\x1b[46m",
|
10
|
-
};
|
11
|
-
|
12
|
-
constructor(moduleName: string) {
|
13
|
-
this.moduleName = moduleName;
|
14
|
-
}
|
15
|
-
|
16
|
-
private formatMessage(
|
17
|
-
level: keyof typeof this.colors,
|
18
|
-
...message: string[]
|
19
|
-
): string {
|
20
|
-
const timestamp = new Date().toTimeString().split(" ")[0];
|
21
|
-
return `[${level}] ${this.colors[level]}${this.colors.reset}${this.colors[level]}[${timestamp}]${this.colors.reset} ${this.colors.module}[${this.moduleName}]${this.colors.reset} ${this.colors[level]}${message}${this.colors.reset}`;
|
22
|
-
}
|
23
|
-
|
24
|
-
public error(...message: string[]): void {
|
25
|
-
console.error(this.formatMessage("error", ...message));
|
26
|
-
}
|
27
|
-
|
28
|
-
public info(...message: string[]): void {
|
29
|
-
console.info(this.formatMessage("info", ...message));
|
30
|
-
}
|
31
|
-
|
32
|
-
public warn(...message: string[]): void {
|
33
|
-
console.warn(this.formatMessage("warn", ...message));
|
34
|
-
}
|
35
|
-
|
36
|
-
public log(...message: string[]): void {
|
37
|
-
console.log(this.formatMessage("info", ...message));
|
38
|
-
}
|
39
|
-
|
40
|
-
public debug(...message: string[]): void {
|
41
|
-
console.debug(this.formatMessage("debug", ...message));
|
42
|
-
}
|
43
|
-
}
|
package/src/modules/index.ts
DELETED