dbgate-api-premium 6.6.0 → 6.6.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +6 -6
- package/src/auth/authProvider.js +14 -2
- package/src/auth/storageAuthProvider.js +89 -22
- package/src/controllers/archive.js +1 -1
- package/src/controllers/auth.js +3 -2
- package/src/controllers/cloud.js +1 -1
- package/src/controllers/config.js +8 -5
- package/src/controllers/connections.js +12 -11
- package/src/controllers/databaseConnections.js +148 -83
- package/src/controllers/files.js +49 -19
- package/src/controllers/plugins.js +7 -4
- package/src/controllers/runners.js +10 -6
- package/src/controllers/scheduler.js +4 -3
- package/src/controllers/serverConnections.js +69 -14
- package/src/controllers/sessions.js +8 -5
- package/src/controllers/storage.js +81 -51
- package/src/controllers/storageDb.js +118 -4
- package/src/controllers/uploads.js +2 -2
- package/src/currentVersion.js +2 -2
- package/src/index.js +36 -5
- package/src/main.js +59 -20
- package/src/proc/databaseConnectionProcess.js +45 -13
- package/src/proc/serverConnectionProcess.js +32 -6
- package/src/proc/sessionProcess.js +2 -2
- package/src/proc/sshForwardProcess.js +1 -1
- package/src/shell/archiveWriter.js +1 -1
- package/src/shell/copyStream.js +1 -1
- package/src/shell/executeQuery.js +3 -3
- package/src/shell/importDatabase.js +3 -3
- package/src/shell/jsonLinesReader.js +1 -1
- package/src/shell/jsonLinesWriter.js +1 -1
- package/src/shell/jsonReader.js +1 -1
- package/src/shell/jsonWriter.js +1 -1
- package/src/shell/loadDatabase.js +2 -2
- package/src/shell/modifyJsonLinesReader.js +1 -1
- package/src/shell/queryReader.js +1 -1
- package/src/shell/requirePlugin.js +6 -1
- package/src/shell/runScript.js +1 -1
- package/src/shell/sqlDataWriter.js +1 -1
- package/src/shell/tableReader.js +3 -3
- package/src/shell/tableWriter.js +1 -1
- package/src/shell/unzipDirectory.js +4 -4
- package/src/shell/zipDirectory.js +3 -3
- package/src/shell/zipJsonLinesData.js +3 -3
- package/src/storageModel.js +726 -105
- package/src/utility/DatastoreProxy.js +3 -3
- package/src/utility/JsonLinesDatastore.js +4 -2
- package/src/utility/appLogStore.js +119 -0
- package/src/utility/auditlog.js +1 -1
- package/src/utility/authProxy.js +4 -4
- package/src/utility/checkLicense.js +10 -4
- package/src/utility/childProcessChecker.js +1 -1
- package/src/utility/cloudIntf.js +5 -5
- package/src/utility/cloudUpgrade.js +4 -4
- package/src/utility/connectUtility.js +1 -1
- package/src/utility/directories.js +2 -2
- package/src/utility/extractSingleFileFromZip.js +3 -3
- package/src/utility/hasPermission.js +286 -71
- package/src/utility/loadModelTransform.js +1 -1
- package/src/utility/sshTunnel.js +7 -7
- package/src/utility/sshTunnelProxy.js +1 -1
- package/src/utility/useController.js +3 -3
|
@@ -61,7 +61,7 @@ class DatastoreProxy {
|
|
|
61
61
|
this.subprocess = null;
|
|
62
62
|
});
|
|
63
63
|
this.subprocess.on('error', err => {
|
|
64
|
-
logger.error(extractErrorLogData(err), 'Error in data store subprocess');
|
|
64
|
+
logger.error(extractErrorLogData(err), 'DBGM-00167 Error in data store subprocess');
|
|
65
65
|
this.subprocess = null;
|
|
66
66
|
});
|
|
67
67
|
this.subprocess.send({ msgtype: 'open', file: this.file });
|
|
@@ -77,7 +77,7 @@ class DatastoreProxy {
|
|
|
77
77
|
try {
|
|
78
78
|
this.subprocess.send({ msgtype: 'read', msgid, offset, limit });
|
|
79
79
|
} catch (err) {
|
|
80
|
-
logger.error(extractErrorLogData(err), 'Error getting rows');
|
|
80
|
+
logger.error(extractErrorLogData(err), 'DBGM-00168 Error getting rows');
|
|
81
81
|
this.subprocess = null;
|
|
82
82
|
}
|
|
83
83
|
});
|
|
@@ -91,7 +91,7 @@ class DatastoreProxy {
|
|
|
91
91
|
try {
|
|
92
92
|
this.subprocess.send({ msgtype: 'notify', msgid });
|
|
93
93
|
} catch (err) {
|
|
94
|
-
logger.error(extractErrorLogData(err), 'Error notifying subprocess');
|
|
94
|
+
logger.error(extractErrorLogData(err), 'DBGM-00169 Error notifying subprocess');
|
|
95
95
|
this.subprocess = null;
|
|
96
96
|
}
|
|
97
97
|
});
|
|
@@ -7,7 +7,6 @@ const AsyncLock = require('async-lock');
|
|
|
7
7
|
const lock = new AsyncLock();
|
|
8
8
|
const stableStringify = require('json-stable-stringify');
|
|
9
9
|
const { evaluateCondition } = require('dbgate-sqltree');
|
|
10
|
-
const requirePluginFunction = require('./requirePluginFunction');
|
|
11
10
|
const esort = require('external-sorting');
|
|
12
11
|
const { jsldir } = require('./directories');
|
|
13
12
|
const LineReader = require('./LineReader');
|
|
@@ -23,7 +22,10 @@ class JsonLinesDatastore {
|
|
|
23
22
|
this.notifyChangedCallback = null;
|
|
24
23
|
this.currentFilter = null;
|
|
25
24
|
this.currentSort = null;
|
|
26
|
-
|
|
25
|
+
if (formatterFunction) {
|
|
26
|
+
const requirePluginFunction = require('./requirePluginFunction');
|
|
27
|
+
this.rowFormatter = requirePluginFunction(formatterFunction);
|
|
28
|
+
}
|
|
27
29
|
this.sortedFiles = {};
|
|
28
30
|
}
|
|
29
31
|
|
|
@@ -0,0 +1,119 @@
|
|
|
1
|
+
const fs = require('fs-extra');
|
|
2
|
+
const path = require('path');
|
|
3
|
+
const { logsdir } = require('./directories');
|
|
4
|
+
const { format, addDays, startOfDay } = require('date-fns');
|
|
5
|
+
const LineReader = require('./LineReader');
|
|
6
|
+
const socket = require('./socket');
|
|
7
|
+
const _ = require('lodash');
|
|
8
|
+
|
|
9
|
+
async function getLogFiles(timeFrom, timeTo) {
|
|
10
|
+
const dir = logsdir();
|
|
11
|
+
const files = await fs.readdir(dir);
|
|
12
|
+
const startPrefix = format(timeFrom, 'yyyy-MM-dd');
|
|
13
|
+
const endPrefix = format(addDays(timeTo, 1), 'yyyy-MM-dd');
|
|
14
|
+
const logFiles = files
|
|
15
|
+
.filter(file => file.endsWith('.ndjson'))
|
|
16
|
+
.filter(file => file >= startPrefix && file < endPrefix);
|
|
17
|
+
return logFiles.sort().map(x => path.join(dir, x));
|
|
18
|
+
}
|
|
19
|
+
|
|
20
|
+
const RECENT_LOG_LIMIT = 1000;
|
|
21
|
+
|
|
22
|
+
let recentLogs = null;
|
|
23
|
+
const beforeRecentLogs = [];
|
|
24
|
+
|
|
25
|
+
function adjustRecentLogs() {
|
|
26
|
+
if (recentLogs.length > RECENT_LOG_LIMIT) {
|
|
27
|
+
recentLogs.splice(0, recentLogs.length - RECENT_LOG_LIMIT);
|
|
28
|
+
}
|
|
29
|
+
}
|
|
30
|
+
|
|
31
|
+
function prepareEntryForExport(entry, lastEntry) {
|
|
32
|
+
return {
|
|
33
|
+
date: format(new Date(entry.time), 'yyyy-MM-dd'),
|
|
34
|
+
time: format(new Date(entry.time), 'HH:mm:ss'),
|
|
35
|
+
dtime: lastEntry ? entry.time - lastEntry.time : 0,
|
|
36
|
+
msgcode: entry.msgcode || '',
|
|
37
|
+
message: entry.msg || '',
|
|
38
|
+
..._.omit(entry, ['time', 'msg', 'msgcode']),
|
|
39
|
+
conid: entry.conid || '',
|
|
40
|
+
database: entry.database || '',
|
|
41
|
+
engine: entry.engine || '',
|
|
42
|
+
ts: entry.time,
|
|
43
|
+
};
|
|
44
|
+
}
|
|
45
|
+
|
|
46
|
+
async function copyAppLogsIntoFile(timeFrom, timeTo, fileName, prepareForExport) {
|
|
47
|
+
const writeStream = fs.createWriteStream(fileName);
|
|
48
|
+
|
|
49
|
+
let lastEntry = null;
|
|
50
|
+
for (const file of await getLogFiles(timeFrom, timeTo)) {
|
|
51
|
+
const readStream = fs.createReadStream(file);
|
|
52
|
+
const reader = new LineReader(readStream);
|
|
53
|
+
do {
|
|
54
|
+
const line = await reader.readLine();
|
|
55
|
+
if (line == null) break;
|
|
56
|
+
try {
|
|
57
|
+
const logEntry = JSON.parse(line);
|
|
58
|
+
if (logEntry.time >= timeFrom && logEntry.time <= timeTo) {
|
|
59
|
+
writeStream.write(
|
|
60
|
+
JSON.stringify(prepareForExport ? prepareEntryForExport(logEntry, lastEntry) : logEntry) + '\n'
|
|
61
|
+
);
|
|
62
|
+
lastEntry = logEntry;
|
|
63
|
+
}
|
|
64
|
+
} catch (e) {
|
|
65
|
+
continue;
|
|
66
|
+
}
|
|
67
|
+
} while (true);
|
|
68
|
+
}
|
|
69
|
+
}
|
|
70
|
+
|
|
71
|
+
async function initializeRecentLogProvider() {
|
|
72
|
+
const logs = [];
|
|
73
|
+
for (const file of await getLogFiles(startOfDay(new Date()), new Date())) {
|
|
74
|
+
const fileStream = fs.createReadStream(file);
|
|
75
|
+
const reader = new LineReader(fileStream);
|
|
76
|
+
do {
|
|
77
|
+
const line = await reader.readLine();
|
|
78
|
+
if (line == null) break;
|
|
79
|
+
try {
|
|
80
|
+
const logEntry = JSON.parse(line);
|
|
81
|
+
logs.push(logEntry);
|
|
82
|
+
if (logs.length > RECENT_LOG_LIMIT) {
|
|
83
|
+
logs.shift();
|
|
84
|
+
}
|
|
85
|
+
} catch (e) {
|
|
86
|
+
continue;
|
|
87
|
+
}
|
|
88
|
+
} while (true);
|
|
89
|
+
}
|
|
90
|
+
recentLogs = logs;
|
|
91
|
+
recentLogs.push(...beforeRecentLogs);
|
|
92
|
+
}
|
|
93
|
+
|
|
94
|
+
let counter = 0;
|
|
95
|
+
function pushToRecentLogs(msg) {
|
|
96
|
+
const finalMsg = {
|
|
97
|
+
...msg,
|
|
98
|
+
counter,
|
|
99
|
+
};
|
|
100
|
+
counter += 1;
|
|
101
|
+
if (recentLogs) {
|
|
102
|
+
recentLogs.push(finalMsg);
|
|
103
|
+
adjustRecentLogs();
|
|
104
|
+
socket.emit('applog-event', finalMsg);
|
|
105
|
+
} else {
|
|
106
|
+
beforeRecentLogs.push(finalMsg);
|
|
107
|
+
}
|
|
108
|
+
}
|
|
109
|
+
|
|
110
|
+
function getRecentAppLogRecords() {
|
|
111
|
+
return recentLogs ?? beforeRecentLogs;
|
|
112
|
+
}
|
|
113
|
+
|
|
114
|
+
module.exports = {
|
|
115
|
+
initializeRecentLogProvider,
|
|
116
|
+
getRecentAppLogRecords,
|
|
117
|
+
pushToRecentLogs,
|
|
118
|
+
copyAppLogsIntoFile,
|
|
119
|
+
};
|
package/src/utility/auditlog.js
CHANGED
|
@@ -92,7 +92,7 @@ async function processAuditLogQueue() {
|
|
|
92
92
|
element.message || null
|
|
93
93
|
);
|
|
94
94
|
} catch (err) {
|
|
95
|
-
logger.error(extractErrorLogData(err), 'Error processing audit log entry');
|
|
95
|
+
logger.error(extractErrorLogData(err), 'DBGM-00159 Error processing audit log entry');
|
|
96
96
|
}
|
|
97
97
|
}
|
|
98
98
|
|
package/src/utility/authProxy.js
CHANGED
|
@@ -10,7 +10,7 @@ const processArgs = require('./processArgs');
|
|
|
10
10
|
const logger = getLogger('authProxy');
|
|
11
11
|
|
|
12
12
|
const AUTH_PROXY_URL = process.env.LOCAL_AUTH_PROXY
|
|
13
|
-
? 'http://localhost:
|
|
13
|
+
? 'http://localhost:3110'
|
|
14
14
|
: process.env.DEVWEB || process.env.DEVMODE
|
|
15
15
|
? 'https://auth-proxy.dbgate.udolni.net'
|
|
16
16
|
: 'https://auth.dbgate.eu';
|
|
@@ -120,7 +120,7 @@ function startTokenChecking(sid, callback) {
|
|
|
120
120
|
callback(resp.data.token);
|
|
121
121
|
}
|
|
122
122
|
} catch (err) {
|
|
123
|
-
logger.error(extractErrorLogData(err), 'Error checking token');
|
|
123
|
+
logger.error(extractErrorLogData(err), 'DBGM-00160 Error checking token');
|
|
124
124
|
}
|
|
125
125
|
}, 500);
|
|
126
126
|
}
|
|
@@ -157,12 +157,12 @@ async function obtainRefreshedLicense() {
|
|
|
157
157
|
|
|
158
158
|
const decoded = jwt.decode(licenseKey?.trim());
|
|
159
159
|
if (!decoded?.end) {
|
|
160
|
-
logger.info('Invalid license found');
|
|
160
|
+
logger.info('DBGM-00078 Invalid license found');
|
|
161
161
|
return null;
|
|
162
162
|
}
|
|
163
163
|
|
|
164
164
|
if (Date.now() > decoded.end * 1000) {
|
|
165
|
-
logger.info('License expired, trying to obtain fresh license');
|
|
165
|
+
logger.info('DBGM-00079 License expired, trying to obtain fresh license');
|
|
166
166
|
|
|
167
167
|
try {
|
|
168
168
|
const respToken = await axios.default.post(
|
|
@@ -52,6 +52,12 @@ const licenseTypeById = {
|
|
|
52
52
|
isForWeb: true,
|
|
53
53
|
isForApp: true,
|
|
54
54
|
},
|
|
55
|
+
'6c734e30-9b66-417d-91a2-0f9aeb739b32': {
|
|
56
|
+
name: 'Demo',
|
|
57
|
+
isPremium: true,
|
|
58
|
+
isForWeb: true,
|
|
59
|
+
isForApp: true,
|
|
60
|
+
},
|
|
55
61
|
};
|
|
56
62
|
|
|
57
63
|
function getLicenseByDecoded(decoded) {
|
|
@@ -105,10 +111,10 @@ async function getAwsMetadata() {
|
|
|
105
111
|
awsMetadata = { amiId, region };
|
|
106
112
|
awsMetadataLoaded = true;
|
|
107
113
|
|
|
108
|
-
logger.info(`Loaded AWS metadata, AMIID=${amiId}, region=${region}`);
|
|
114
|
+
logger.info(`DBGM-00080 Loaded AWS metadata, AMIID=${amiId}, region=${region}`);
|
|
109
115
|
return { amiId, region };
|
|
110
116
|
} catch (error) {
|
|
111
|
-
logger.error(extractErrorLogData(error), 'Error getting AWS metadata');
|
|
117
|
+
logger.error(extractErrorLogData(error), 'DBGM-00081 Error getting AWS metadata');
|
|
112
118
|
awsMetadataLoaded = true;
|
|
113
119
|
return null;
|
|
114
120
|
}
|
|
@@ -134,7 +140,7 @@ function checkLicenseKey(licenseKey) {
|
|
|
134
140
|
(platformInfo.isElectron && !licenseTypeObj.isForApp) ||
|
|
135
141
|
(!platformInfo.isElectron && !licenseTypeObj.isForWeb)
|
|
136
142
|
) {
|
|
137
|
-
logger.error(`Incorrect license type, found ${licenseTypeObj?.name ?? 'n/a'}`);
|
|
143
|
+
logger.error(`DBGM-00161 Incorrect license type, found ${licenseTypeObj?.name ?? 'n/a'}`);
|
|
138
144
|
return {
|
|
139
145
|
status: 'error',
|
|
140
146
|
error: `Incorrect license type, found ${licenseTypeObj?.name ?? 'n/a'}`,
|
|
@@ -167,7 +173,7 @@ function checkLicenseKey(licenseKey) {
|
|
|
167
173
|
}
|
|
168
174
|
} catch (err) {}
|
|
169
175
|
|
|
170
|
-
logger.error(extractErrorLogData(err), 'License token is invalid');
|
|
176
|
+
logger.error(extractErrorLogData(err), 'DBGM-00162 License token is invalid');
|
|
171
177
|
return {
|
|
172
178
|
status: 'error',
|
|
173
179
|
error: err.message ?? 'License token is invalid',
|
|
@@ -12,7 +12,7 @@ function childProcessChecker() {
|
|
|
12
12
|
// This will come once parent dies.
|
|
13
13
|
// One way can be to check for error code ERR_IPC_CHANNEL_CLOSED
|
|
14
14
|
// and call process.exit()
|
|
15
|
-
logger.error(extractErrorLogData(err), 'parent died');
|
|
15
|
+
logger.error(extractErrorLogData(err), 'DBGM-00163 parent died');
|
|
16
16
|
process.exit(1);
|
|
17
17
|
}
|
|
18
18
|
}, 1000);
|
package/src/utility/cloudIntf.js
CHANGED
|
@@ -77,7 +77,7 @@ function startCloudTokenChecking(sid, callback) {
|
|
|
77
77
|
callback(resp.data);
|
|
78
78
|
}
|
|
79
79
|
} catch (err) {
|
|
80
|
-
logger.error(extractErrorLogData(err), 'Error checking cloud token');
|
|
80
|
+
logger.error(extractErrorLogData(err), 'DBGM-00164 Error checking cloud token');
|
|
81
81
|
}
|
|
82
82
|
}, 500);
|
|
83
83
|
}
|
|
@@ -125,7 +125,7 @@ async function getCloudUsedEngines() {
|
|
|
125
125
|
const resp = await callCloudApiGet('content-engines');
|
|
126
126
|
return resp || [];
|
|
127
127
|
} catch (err) {
|
|
128
|
-
logger.error(extractErrorLogData(err), 'Error getting cloud content list');
|
|
128
|
+
logger.error(extractErrorLogData(err), 'DBGM-00165 Error getting cloud content list');
|
|
129
129
|
return [];
|
|
130
130
|
}
|
|
131
131
|
}
|
|
@@ -208,7 +208,7 @@ async function updateCloudFiles(isRefresh) {
|
|
|
208
208
|
lastCheckedTm = _.max(cloudFiles.map(x => parseInt(x.modifiedTm)));
|
|
209
209
|
}
|
|
210
210
|
|
|
211
|
-
logger.info({ tags, lastCheckedTm }, 'Downloading cloud files');
|
|
211
|
+
logger.info({ tags, lastCheckedTm }, 'DBGM-00082 Downloading cloud files');
|
|
212
212
|
|
|
213
213
|
const resp = await axios.default.get(
|
|
214
214
|
`${DBGATE_CLOUD_URL}/public-cloud-updates?lastCheckedTm=${lastCheckedTm}&tags=${tags}&isRefresh=${
|
|
@@ -223,7 +223,7 @@ async function updateCloudFiles(isRefresh) {
|
|
|
223
223
|
}
|
|
224
224
|
);
|
|
225
225
|
|
|
226
|
-
logger.info(`Downloaded ${resp.data.length} cloud files`);
|
|
226
|
+
logger.info(`DBGM-00083 Downloaded ${resp.data.length} cloud files`);
|
|
227
227
|
|
|
228
228
|
const filesByPath = lastCheckedTm == 0 ? {} : _.keyBy(cloudFiles, 'path');
|
|
229
229
|
for (const file of resp.data) {
|
|
@@ -269,7 +269,7 @@ async function refreshPublicFiles(isRefresh) {
|
|
|
269
269
|
try {
|
|
270
270
|
await updateCloudFiles(isRefresh);
|
|
271
271
|
} catch (err) {
|
|
272
|
-
logger.error(extractErrorLogData(err), 'Error updating cloud files');
|
|
272
|
+
logger.error(extractErrorLogData(err), 'DBGM-00166 Error updating cloud files');
|
|
273
273
|
}
|
|
274
274
|
}
|
|
275
275
|
|
|
@@ -31,7 +31,7 @@ async function checkCloudUpgrade() {
|
|
|
31
31
|
semver.gt(version, currentVersion.version) &&
|
|
32
32
|
(!cloudDownloadedVersion || semver.gt(version, cloudDownloadedVersion))
|
|
33
33
|
) {
|
|
34
|
-
logger.info(`New version available: ${version}`);
|
|
34
|
+
logger.info(`DBGM-00084 New version available: ${version}`);
|
|
35
35
|
const zipUrl = json.assets.find(x => x.name == 'cloud-build.zip').browser_download_url;
|
|
36
36
|
|
|
37
37
|
const writer = fs.createWriteStream(process.env.CLOUD_UPGRADE_FILE);
|
|
@@ -50,16 +50,16 @@ async function checkCloudUpgrade() {
|
|
|
50
50
|
});
|
|
51
51
|
await fsp.writeFile(process.env.CLOUD_UPGRADE_FILE + '.version', version);
|
|
52
52
|
|
|
53
|
-
logger.info(`Downloaded new version from ${zipUrl}`);
|
|
53
|
+
logger.info(`DBGM-00085 Downloaded new version from ${zipUrl}`);
|
|
54
54
|
} else {
|
|
55
55
|
logger.info(
|
|
56
|
-
`Checked version ${version} is not newer than ${
|
|
56
|
+
`DBGM-00086 Checked version ${version} is not newer than ${
|
|
57
57
|
cloudDownloadedVersion ?? currentVersion.version
|
|
58
58
|
}, upgrade skippped`
|
|
59
59
|
);
|
|
60
60
|
}
|
|
61
61
|
} catch (err) {
|
|
62
|
-
logger.error(extractErrorLogData(err), 'Error checking cloud upgrade');
|
|
62
|
+
logger.error(extractErrorLogData(err), 'DBGM-00087 Error checking cloud upgrade');
|
|
63
63
|
}
|
|
64
64
|
}
|
|
65
65
|
|
|
@@ -132,7 +132,7 @@ async function connectUtility(driver, storedConnection, connectionMode, addition
|
|
|
132
132
|
|
|
133
133
|
connection.ssl = await extractConnectionSslParams(connection);
|
|
134
134
|
|
|
135
|
-
const conn = await driver.connect({ ...connection, ...additionalOptions });
|
|
135
|
+
const conn = await driver.connect({ conid: connectionLoaded?._id, ...connection, ...additionalOptions });
|
|
136
136
|
return conn;
|
|
137
137
|
}
|
|
138
138
|
|
|
@@ -14,11 +14,11 @@ const createDirectories = {};
|
|
|
14
14
|
const ensureDirectory = (dir, clean) => {
|
|
15
15
|
if (!createDirectories[dir]) {
|
|
16
16
|
if (clean && fs.existsSync(dir) && !platformInfo.isForkedApi) {
|
|
17
|
-
getLogger('directories').info(`Cleaning directory ${dir}`);
|
|
17
|
+
getLogger('directories').info(`DBGM-00170 Cleaning directory ${dir}`);
|
|
18
18
|
cleanDirectory(dir, _.isNumber(clean) ? clean : null);
|
|
19
19
|
}
|
|
20
20
|
if (!fs.existsSync(dir)) {
|
|
21
|
-
getLogger('directories').info(`Creating directory ${dir}`);
|
|
21
|
+
getLogger('directories').info(`DBGM-00171 Creating directory ${dir}`);
|
|
22
22
|
fs.mkdirSync(dir);
|
|
23
23
|
}
|
|
24
24
|
createDirectories[dir] = true;
|
|
@@ -42,13 +42,13 @@ function extractSingleFileFromZip(zipPath, fileInZip, outputPath) {
|
|
|
42
42
|
|
|
43
43
|
// When the file is finished writing, resolve
|
|
44
44
|
writeStream.on('finish', () => {
|
|
45
|
-
logger.info(`File "${fileInZip}" extracted to "${outputPath}".`);
|
|
45
|
+
logger.info(`DBGM-00088 File "${fileInZip}" extracted to "${outputPath}".`);
|
|
46
46
|
resolve(true);
|
|
47
47
|
});
|
|
48
48
|
|
|
49
49
|
// Handle write errors
|
|
50
50
|
writeStream.on('error', writeErr => {
|
|
51
|
-
logger.error(extractErrorLogData(writeErr), `Error extracting "${fileInZip}" from "${zipPath}".`);
|
|
51
|
+
logger.error(extractErrorLogData(writeErr), `DBGM-00089 Error extracting "${fileInZip}" from "${zipPath}".`);
|
|
52
52
|
reject(writeErr);
|
|
53
53
|
});
|
|
54
54
|
});
|
|
@@ -67,7 +67,7 @@ function extractSingleFileFromZip(zipPath, fileInZip, outputPath) {
|
|
|
67
67
|
|
|
68
68
|
// Handle general errors
|
|
69
69
|
zipFile.on('error', err => {
|
|
70
|
-
logger.error(extractErrorLogData(err), `ZIP file error in ${zipPath}.`);
|
|
70
|
+
logger.error(extractErrorLogData(err), `DBGM-00172 ZIP file error in ${zipPath}.`);
|
|
71
71
|
reject(err);
|
|
72
72
|
});
|
|
73
73
|
});
|