@ezs/basics 2.3.0 → 2.4.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +22 -0
- package/README.md +17 -0
- package/lib/fetch.js +74 -0
- package/lib/index.js +3 -0
- package/lib/obj-columns.js +53 -0
- package/lib/skos-object.js +73 -0
- package/lib/tar-dump.js +65 -0
- package/lib/url-pager.js +179 -0
- package/lib/utils.js +18 -0
- package/package.json +5 -4
package/CHANGELOG.md
CHANGED
|
@@ -3,6 +3,28 @@
|
|
|
3
3
|
All notable changes to this project will be documented in this file.
|
|
4
4
|
See [Conventional Commits](https://conventionalcommits.org) for commit guidelines.
|
|
5
5
|
|
|
6
|
+
## [2.4.1](https://github.com/Inist-CNRS/ezs/compare/@ezs/basics@2.4.0...@ezs/basics@2.4.1) (2023-07-20)
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
### Bug Fixes
|
|
10
|
+
|
|
11
|
+
* 🐛 manifest.json ([559cda8](https://github.com/Inist-CNRS/ezs/commit/559cda8372c65c3c1cfb5a0dcfee404ece4a79d6))
|
|
12
|
+
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
|
|
16
|
+
|
|
17
|
+
# [2.4.0](https://github.com/Inist-CNRS/ezs/compare/@ezs/basics@2.3.0...@ezs/basics@2.4.0) (2023-07-17)
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
### Features
|
|
21
|
+
|
|
22
|
+
* 🎸 add [TARDump] ([d3ff2a9](https://github.com/Inist-CNRS/ezs/commit/d3ff2a99650fc7c9eacd2c1ad015d9907addeebb))
|
|
23
|
+
|
|
24
|
+
|
|
25
|
+
|
|
26
|
+
|
|
27
|
+
|
|
6
28
|
# [2.3.0](https://github.com/Inist-CNRS/ezs/compare/@ezs/basics@2.2.0...@ezs/basics@2.3.0) (2023-06-23)
|
|
7
29
|
|
|
8
30
|
|
package/README.md
CHANGED
|
@@ -27,6 +27,7 @@ npm install @ezs/basics
|
|
|
27
27
|
- [OBJFlatten](#objflatten)
|
|
28
28
|
- [OBJNamespaces](#objnamespaces)
|
|
29
29
|
- [OBJStandardize](#objstandardize)
|
|
30
|
+
- [TARDump](#tardump)
|
|
30
31
|
- [TARExtract](#tarextract)
|
|
31
32
|
- [TXTConcat](#txtconcat)
|
|
32
33
|
- [TXTInflection](#txtinflection)
|
|
@@ -537,6 +538,22 @@ Output:
|
|
|
537
538
|
|
|
538
539
|
Returns **[Object](https://developer.mozilla.org/docs/Web/JavaScript/Reference/Global_Objects/Object)**
|
|
539
540
|
|
|
541
|
+
### TARDump
|
|
542
|
+
|
|
543
|
+
Take all recevied objects and build a tar file
|
|
544
|
+
|
|
545
|
+
```json
|
|
546
|
+
{
|
|
547
|
+
}
|
|
548
|
+
```
|
|
549
|
+
|
|
550
|
+
#### Parameters
|
|
551
|
+
|
|
552
|
+
- `manifest` **[String](https://developer.mozilla.org/docs/Web/JavaScript/Reference/Global_Objects/String)?** Location path to store files in the tarball
|
|
553
|
+
- `location` **[String](https://developer.mozilla.org/docs/Web/JavaScript/Reference/Global_Objects/String)** Location path to store files in the tarball (optional, default `data`)
|
|
554
|
+
- `json` **[String](https://developer.mozilla.org/docs/Web/JavaScript/Reference/Global_Objects/String)** Parse as JSON the content of each file (optional, default `true`)
|
|
555
|
+
- `compress` **[Boolean](https://developer.mozilla.org/docs/Web/JavaScript/Reference/Global_Objects/Boolean)** Enable gzip compression (optional, default `false`)
|
|
556
|
+
|
|
540
557
|
### TARExtract
|
|
541
558
|
|
|
542
559
|
Take the content of a tar file, extract some files.
|
package/lib/fetch.js
ADDED
|
@@ -0,0 +1,74 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
|
|
3
|
+
Object.defineProperty(exports, "__esModule", {
|
|
4
|
+
value: true
|
|
5
|
+
});
|
|
6
|
+
exports.default = fetch;
|
|
7
|
+
|
|
8
|
+
var _crossFetch = _interopRequireDefault(require("cross-fetch"));
|
|
9
|
+
|
|
10
|
+
var _proxyFromEnv = require("proxy-from-env");
|
|
11
|
+
|
|
12
|
+
var _http = _interopRequireDefault(require("http"));
|
|
13
|
+
|
|
14
|
+
var _https = _interopRequireDefault(require("https"));
|
|
15
|
+
|
|
16
|
+
var _betterHttpsProxyAgent = require("better-https-proxy-agent");
|
|
17
|
+
|
|
18
|
+
function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { default: obj }; }
|
|
19
|
+
|
|
20
|
+
const DefaultOptions = {
|
|
21
|
+
keepAlive: true,
|
|
22
|
+
timeout: 1000,
|
|
23
|
+
keepAliveMsecs: 500,
|
|
24
|
+
maxSockets: 200,
|
|
25
|
+
maxFreeSockets: 5,
|
|
26
|
+
maxCachedSessions: 500
|
|
27
|
+
};
|
|
28
|
+
|
|
29
|
+
const chooseAgent = (parsedURL, options) => {
|
|
30
|
+
const proxyurl = (0, _proxyFromEnv.getProxyForUrl)(parsedURL.href);
|
|
31
|
+
|
|
32
|
+
if (proxyurl) {
|
|
33
|
+
const proxyRequestOptions = new URL(proxyurl);
|
|
34
|
+
return new _betterHttpsProxyAgent.Agent(options, proxyRequestOptions);
|
|
35
|
+
}
|
|
36
|
+
|
|
37
|
+
if (parsedURL.protocol === 'https:') {
|
|
38
|
+
return new _https.default.Agent(options);
|
|
39
|
+
}
|
|
40
|
+
|
|
41
|
+
return new _http.default.Agent(options);
|
|
42
|
+
};
|
|
43
|
+
|
|
44
|
+
function fetch(url, options) {
|
|
45
|
+
const opts = options || {};
|
|
46
|
+
const {
|
|
47
|
+
keepAlive,
|
|
48
|
+
timeout,
|
|
49
|
+
keepAliveMsecs,
|
|
50
|
+
maxSockets,
|
|
51
|
+
maxFreeSockets,
|
|
52
|
+
maxCachedSessions
|
|
53
|
+
} = { ...options,
|
|
54
|
+
...DefaultOptions
|
|
55
|
+
};
|
|
56
|
+
let agent = chooseAgent(new URL(url), {
|
|
57
|
+
keepAlive,
|
|
58
|
+
timeout,
|
|
59
|
+
keepAliveMsecs,
|
|
60
|
+
maxSockets,
|
|
61
|
+
maxFreeSockets,
|
|
62
|
+
maxCachedSessions
|
|
63
|
+
});
|
|
64
|
+
opts.agent = agent;
|
|
65
|
+
|
|
66
|
+
if (opts.signal) {
|
|
67
|
+
opts.signal.addEventListener('abort', () => {
|
|
68
|
+
agent.destroy();
|
|
69
|
+
agent = null;
|
|
70
|
+
});
|
|
71
|
+
}
|
|
72
|
+
|
|
73
|
+
return (0, _crossFetch.default)(url, options);
|
|
74
|
+
}
|
package/lib/index.js
CHANGED
|
@@ -61,6 +61,8 @@ var _zipExtract = _interopRequireDefault(require("./zip-extract"));
|
|
|
61
61
|
|
|
62
62
|
var _tarExtract = _interopRequireDefault(require("./tar-extract"));
|
|
63
63
|
|
|
64
|
+
var _tarDump = _interopRequireDefault(require("./tar-dump"));
|
|
65
|
+
|
|
64
66
|
var _iniString = _interopRequireDefault(require("./ini-string"));
|
|
65
67
|
|
|
66
68
|
var _fileSave = _interopRequireDefault(require("./file-save"));
|
|
@@ -98,6 +100,7 @@ const funcs = {
|
|
|
98
100
|
TXTZip: _txtZip.default,
|
|
99
101
|
ZIPExtract: _zipExtract.default,
|
|
100
102
|
TARExtract: _tarExtract.default,
|
|
103
|
+
TARDump: _tarDump.default,
|
|
101
104
|
INIString: _iniString.default,
|
|
102
105
|
FILESave: _fileSave.default,
|
|
103
106
|
FILELoad: _fileLoad.default,
|
|
@@ -0,0 +1,53 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
|
|
3
|
+
Object.defineProperty(exports, "__esModule", {
|
|
4
|
+
value: true
|
|
5
|
+
});
|
|
6
|
+
exports.default = OBJColumns;
|
|
7
|
+
|
|
8
|
+
/**
|
|
9
|
+
* Take an `Object` and flatten it to get only one level of keys.
|
|
10
|
+
*
|
|
11
|
+
* <caption>Input:</caption>
|
|
12
|
+
*
|
|
13
|
+
* ```json
|
|
14
|
+
* [{
|
|
15
|
+
* "foo": {
|
|
16
|
+
* "hello": "world"
|
|
17
|
+
* },
|
|
18
|
+
* "bar": "anything else",
|
|
19
|
+
* "baz": 1
|
|
20
|
+
* }]
|
|
21
|
+
* ```
|
|
22
|
+
*
|
|
23
|
+
* <caption>Output:</caption>
|
|
24
|
+
*
|
|
25
|
+
* ```json
|
|
26
|
+
* [{
|
|
27
|
+
* "foo": "{\"hello\":\"world\"}",
|
|
28
|
+
* "bar": "anything else",
|
|
29
|
+
* "baz": 1
|
|
30
|
+
* }]
|
|
31
|
+
* ```
|
|
32
|
+
*
|
|
33
|
+
* @name OBJColumns
|
|
34
|
+
* @alias flatten
|
|
35
|
+
* @param {undefined} none
|
|
36
|
+
* @returns {Object}
|
|
37
|
+
*/
|
|
38
|
+
function OBJColumns(data, feed) {
|
|
39
|
+
if (this.isLast()) {
|
|
40
|
+
feed.close();
|
|
41
|
+
return;
|
|
42
|
+
}
|
|
43
|
+
|
|
44
|
+
const obj = {};
|
|
45
|
+
Object.keys(data).sort((x, y) => x.localeCompare(y)).forEach(key => {
|
|
46
|
+
if (typeof data[key] === 'object') {
|
|
47
|
+
obj[key] = JSON.stringify(data[key]);
|
|
48
|
+
} else {
|
|
49
|
+
obj[key] = data[key];
|
|
50
|
+
}
|
|
51
|
+
});
|
|
52
|
+
feed.send(obj);
|
|
53
|
+
}
|
|
@@ -0,0 +1,73 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
|
|
3
|
+
Object.defineProperty(exports, "__esModule", {
|
|
4
|
+
value: true
|
|
5
|
+
});
|
|
6
|
+
exports.default = void 0;
|
|
7
|
+
|
|
8
|
+
function Concept(data, feed) {
|
|
9
|
+
const obj = {};
|
|
10
|
+
Object.keys(data).forEach(key => {
|
|
11
|
+
const newkey = key.replace('skos$', '');
|
|
12
|
+
|
|
13
|
+
if (Array.isArray(data[key])) {
|
|
14
|
+
data[key].filter(x => x.xml$lang).forEach(item => {
|
|
15
|
+
const localkey = newkey.concat('@').concat(item.xml$lang);
|
|
16
|
+
obj[localkey] = item.$t;
|
|
17
|
+
});
|
|
18
|
+
} else if (data[key].rdf$resource && !obj[newkey]) {
|
|
19
|
+
obj[newkey] = data[key].rdf$resource;
|
|
20
|
+
} else if (data[key].rdf$resource && obj[newkey]) {
|
|
21
|
+
obj[newkey] = [obj[newkey], data[key].rdf$resource];
|
|
22
|
+
} else if (data[key].$t && data[key].xml$lang) {
|
|
23
|
+
const localkey = newkey.concat('@').concat(data[key].xml$lang);
|
|
24
|
+
obj[localkey] = data[key].$t;
|
|
25
|
+
} else if (data[key].$t && Array.isArray(obj[newkey])) {
|
|
26
|
+
obj[newkey].push(data[key].$t);
|
|
27
|
+
} else if (data[key].$t && obj[newkey]) {
|
|
28
|
+
obj[newkey] = [obj[newkey], data[key].$t];
|
|
29
|
+
} else if (data[key].$t && !obj[newkey]) {
|
|
30
|
+
obj[newkey] = data[key].$t;
|
|
31
|
+
} else if (typeof data[key] === 'object') {
|
|
32
|
+
obj[newkey] = (this.getIndex().toString(36) + Math.random().toString(36).substr(2, 5)).toUpperCase();
|
|
33
|
+
let counter = 0;
|
|
34
|
+
Object.keys(data[key]).forEach(key2 => {
|
|
35
|
+
if (typeof data[key][key2] === 'object') {
|
|
36
|
+
data[key][key2].rdf$about = obj[newkey];
|
|
37
|
+
Concept.call(this, data[key][key2], feed);
|
|
38
|
+
counter += 1;
|
|
39
|
+
}
|
|
40
|
+
});
|
|
41
|
+
|
|
42
|
+
if (counter === 0) {
|
|
43
|
+
delete obj[newkey];
|
|
44
|
+
}
|
|
45
|
+
} else {
|
|
46
|
+
obj[newkey] = data[key];
|
|
47
|
+
}
|
|
48
|
+
});
|
|
49
|
+
feed.write(obj);
|
|
50
|
+
}
|
|
51
|
+
|
|
52
|
+
function SKOSObject(data, feed) {
|
|
53
|
+
if (this.isLast()) {
|
|
54
|
+
feed.close();
|
|
55
|
+
} else {
|
|
56
|
+
Concept.call(this, data, feed);
|
|
57
|
+
feed.end();
|
|
58
|
+
}
|
|
59
|
+
}
|
|
60
|
+
/**
|
|
61
|
+
* Take `Object` generated by XMLMapping & SKOS data and
|
|
62
|
+
* create a new basic object with only keys & values
|
|
63
|
+
*
|
|
64
|
+
* @name SKOSObject
|
|
65
|
+
* @param {undefined} none
|
|
66
|
+
* @returns {Object}
|
|
67
|
+
*/
|
|
68
|
+
|
|
69
|
+
|
|
70
|
+
var _default = {
|
|
71
|
+
SKOSObject
|
|
72
|
+
};
|
|
73
|
+
exports.default = _default;
|
package/lib/tar-dump.js
ADDED
|
@@ -0,0 +1,65 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
|
|
3
|
+
Object.defineProperty(exports, "__esModule", {
|
|
4
|
+
value: true
|
|
5
|
+
});
|
|
6
|
+
exports.default = TARDump;
|
|
7
|
+
|
|
8
|
+
var _tarStream = _interopRequireDefault(require("tar-stream"));
|
|
9
|
+
|
|
10
|
+
var _zlib = require("zlib");
|
|
11
|
+
|
|
12
|
+
var _lodash = _interopRequireDefault(require("lodash.merge"));
|
|
13
|
+
|
|
14
|
+
function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { default: obj }; }
|
|
15
|
+
|
|
16
|
+
const eol = '\n';
|
|
17
|
+
/**
|
|
18
|
+
* Take all recevied objects and build a tar file
|
|
19
|
+
*
|
|
20
|
+
* ```json
|
|
21
|
+
* {
|
|
22
|
+
* }
|
|
23
|
+
* ```
|
|
24
|
+
*
|
|
25
|
+
* @name TARDump
|
|
26
|
+
* @param {String} [manifest] Location path to store files in the tarball
|
|
27
|
+
* @param {String} [location=data] Location path to store files in the tarball
|
|
28
|
+
* @param {String} [json=true] Parse as JSON the content of each file
|
|
29
|
+
* @param {Boolean} [compress=false] Enable gzip compression
|
|
30
|
+
*/
|
|
31
|
+
|
|
32
|
+
function TARDump(data, feed) {
|
|
33
|
+
if (!this.pack) {
|
|
34
|
+
this.pack = _tarStream.default.pack();
|
|
35
|
+
const compress = this.getParam('compress', false);
|
|
36
|
+
this.created = new Date();
|
|
37
|
+
const stream = compress ? this.pack.pipe((0, _zlib.createGzip)()) : this.pack;
|
|
38
|
+
feed.flow(stream).finally(() => feed.close());
|
|
39
|
+
}
|
|
40
|
+
|
|
41
|
+
if (this.isLast()) {
|
|
42
|
+
const updated = new Date();
|
|
43
|
+
const metadata = {
|
|
44
|
+
'creationDate': this.created.toUTCString(),
|
|
45
|
+
'updateDate': updated.toUTCString(),
|
|
46
|
+
'itemsCounter': this.getIndex() - 1,
|
|
47
|
+
'processingMSTime': this.getCumulativeTimeMS()
|
|
48
|
+
};
|
|
49
|
+
const manifestArray = [metadata].concat(this.getParam('manifest', [])).filter(Boolean);
|
|
50
|
+
const manifest = (0, _lodash.default)(...manifestArray);
|
|
51
|
+
this.pack.entry({
|
|
52
|
+
name: 'manifest.json'
|
|
53
|
+
}, JSON.stringify(manifest, null, ' '));
|
|
54
|
+
this.pack.finalize();
|
|
55
|
+
return;
|
|
56
|
+
}
|
|
57
|
+
|
|
58
|
+
const id = this.getIndex().toString().padStart(10, '0');
|
|
59
|
+
const value = JSON.stringify(data).concat(eol);
|
|
60
|
+
const location = this.getParam('location', 'data');
|
|
61
|
+
this.pack.entry({
|
|
62
|
+
name: `${location}/f${id}.json`
|
|
63
|
+
}, value);
|
|
64
|
+
feed.end();
|
|
65
|
+
}
|
package/lib/url-pager.js
ADDED
|
@@ -0,0 +1,179 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
|
|
3
|
+
Object.defineProperty(exports, "__esModule", {
|
|
4
|
+
value: true
|
|
5
|
+
});
|
|
6
|
+
exports.default = URLPager;
|
|
7
|
+
|
|
8
|
+
var _debug = _interopRequireDefault(require("debug"));
|
|
9
|
+
|
|
10
|
+
var _url = require("url");
|
|
11
|
+
|
|
12
|
+
var _nodeAbortController = _interopRequireDefault(require("node-abort-controller"));
|
|
13
|
+
|
|
14
|
+
var _lodash = _interopRequireDefault(require("lodash.get"));
|
|
15
|
+
|
|
16
|
+
var _parseHeaders = _interopRequireDefault(require("parse-headers"));
|
|
17
|
+
|
|
18
|
+
var _asyncRetry = _interopRequireDefault(require("async-retry"));
|
|
19
|
+
|
|
20
|
+
var _request = _interopRequireDefault(require("./request"));
|
|
21
|
+
|
|
22
|
+
function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { default: obj }; }
|
|
23
|
+
|
|
24
|
+
/**
|
|
25
|
+
* Take `Object` as parameters of URL, throw each chunk from the result
|
|
26
|
+
*
|
|
27
|
+
*
|
|
28
|
+
* Input:
|
|
29
|
+
*
|
|
30
|
+
* ```json
|
|
31
|
+
* [{"q": "a"}]
|
|
32
|
+
* ```
|
|
33
|
+
*
|
|
34
|
+
* Script:
|
|
35
|
+
*
|
|
36
|
+
* ```ini
|
|
37
|
+
* [URLPager]
|
|
38
|
+
* url = https://api.search.net
|
|
39
|
+
* path = total
|
|
40
|
+
* ```
|
|
41
|
+
*
|
|
42
|
+
* Output:
|
|
43
|
+
*
|
|
44
|
+
* ```json
|
|
45
|
+
* [
|
|
46
|
+
* {
|
|
47
|
+
* "q": "a",
|
|
48
|
+
* "total": 22
|
|
49
|
+
* "offset": 0,
|
|
50
|
+
* "pageNumber": 1,
|
|
51
|
+
* "totalPages", 3,
|
|
52
|
+
* "maxPages": 1000,
|
|
53
|
+
* "limit": 10
|
|
54
|
+
* },
|
|
55
|
+
* {
|
|
56
|
+
* "q": "a",
|
|
57
|
+
* "total": 22
|
|
58
|
+
* "offset": 10,
|
|
59
|
+
* "pageNumber": 2,
|
|
60
|
+
* "totalPages", 3,
|
|
61
|
+
* "maxPages": 1000,
|
|
62
|
+
* "limit": 10
|
|
63
|
+
* },
|
|
64
|
+
* {
|
|
65
|
+
* "q": "a",
|
|
66
|
+
* "total": 22
|
|
67
|
+
* "offset": 20,
|
|
68
|
+
* "pageNumber": 3,
|
|
69
|
+
* "totalPages", 3,
|
|
70
|
+
* "maxPages": 1000,
|
|
71
|
+
* "limit": 10
|
|
72
|
+
* }
|
|
73
|
+
* ]
|
|
74
|
+
* ```
|
|
75
|
+
*
|
|
76
|
+
* #### Example with URLs
|
|
77
|
+
*
|
|
78
|
+
* Input:
|
|
79
|
+
*
|
|
80
|
+
* ```json
|
|
81
|
+
* [
|
|
82
|
+
* "https://httpbin.org/get?a=a",
|
|
83
|
+
* "https://httpbin.org/get?a=b",
|
|
84
|
+
* "https://httpbin.org/get?a=c"
|
|
85
|
+
* ]
|
|
86
|
+
* ```
|
|
87
|
+
*
|
|
88
|
+
* Script:
|
|
89
|
+
*
|
|
90
|
+
* ```ini
|
|
91
|
+
* [URLPager]
|
|
92
|
+
* path = .args
|
|
93
|
+
* ```
|
|
94
|
+
*
|
|
95
|
+
* Output:
|
|
96
|
+
*
|
|
97
|
+
* ```json
|
|
98
|
+
* [{"a": "a"}, {"a": "b"}, {"a": "c" }]
|
|
99
|
+
* ```
|
|
100
|
+
*
|
|
101
|
+
* @name URLPager
|
|
102
|
+
* @param {String} [url] URL to fetch (by default input string is taken)
|
|
103
|
+
* @param {String} [path=total] choose the path to find the number of result
|
|
104
|
+
* @param {Number} [timeout=1000] Timeout in milliseconds
|
|
105
|
+
* @param {Boolean} [noerror=false] Ignore all errors, the target field will remain undefined
|
|
106
|
+
* @param {Number} [retries=5] The maximum amount of times to retry the connection
|
|
107
|
+
* @returns {Object}
|
|
108
|
+
*/
|
|
109
|
+
async function URLPager(data, feed) {
|
|
110
|
+
if (this.isLast()) {
|
|
111
|
+
return feed.close();
|
|
112
|
+
}
|
|
113
|
+
|
|
114
|
+
const url = this.getParam('url');
|
|
115
|
+
const path = this.getParam('path', 'total');
|
|
116
|
+
const limit = Number(this.getParam('limit', 10));
|
|
117
|
+
const maxPages = Number(this.getParam('maxPages', 1000));
|
|
118
|
+
const retries = Number(this.getParam('retries', 5));
|
|
119
|
+
const noerror = Boolean(this.getParam('noerror', false));
|
|
120
|
+
const timeout = Number(this.getParam('timeout')) || 1000;
|
|
121
|
+
const headers = (0, _parseHeaders.default)([].concat(this.getParam('header')).filter(Boolean).join('\n'));
|
|
122
|
+
const cURL = new _url.URL(url || data);
|
|
123
|
+
const controller = new _nodeAbortController.default();
|
|
124
|
+
const parameters = {
|
|
125
|
+
timeout,
|
|
126
|
+
headers,
|
|
127
|
+
signal: controller.signal
|
|
128
|
+
};
|
|
129
|
+
const options = {
|
|
130
|
+
retries
|
|
131
|
+
};
|
|
132
|
+
cURL.search = new _url.URLSearchParams(data);
|
|
133
|
+
|
|
134
|
+
const onError = e => {
|
|
135
|
+
controller.abort();
|
|
136
|
+
|
|
137
|
+
if (noerror) {
|
|
138
|
+
(0, _debug.default)('ezs')(`Ignore item #${this.getIndex()} [URLPager] <${e}>`);
|
|
139
|
+
return feed.send(data);
|
|
140
|
+
}
|
|
141
|
+
|
|
142
|
+
(0, _debug.default)('ezs')(`Break item #${this.getIndex()} [URLPager] <${e}>`);
|
|
143
|
+
return feed.send(e);
|
|
144
|
+
};
|
|
145
|
+
|
|
146
|
+
try {
|
|
147
|
+
const response = await (0, _asyncRetry.default)((0, _request.default)(cURL.href, parameters), options);
|
|
148
|
+
const json = await response.json();
|
|
149
|
+
const total = (0, _lodash.default)(json, path);
|
|
150
|
+
|
|
151
|
+
if (total === 0) {
|
|
152
|
+
return onError(new Error('No result.'));
|
|
153
|
+
}
|
|
154
|
+
|
|
155
|
+
if (total === undefined) {
|
|
156
|
+
return onError(new Error('Unexpected response.'));
|
|
157
|
+
}
|
|
158
|
+
|
|
159
|
+
let totalPages = Math.ceil(json.total / limit);
|
|
160
|
+
|
|
161
|
+
if (totalPages > maxPages) {
|
|
162
|
+
totalPages = maxPages;
|
|
163
|
+
}
|
|
164
|
+
|
|
165
|
+
for (let pageNumber = 1; pageNumber <= totalPages; pageNumber += 1) {
|
|
166
|
+
feed.write({ ...data,
|
|
167
|
+
offset: (pageNumber - 1) * limit,
|
|
168
|
+
pageNumber,
|
|
169
|
+
totalPages,
|
|
170
|
+
maxPages,
|
|
171
|
+
limit
|
|
172
|
+
});
|
|
173
|
+
}
|
|
174
|
+
|
|
175
|
+
feed.end();
|
|
176
|
+
} catch (e) {
|
|
177
|
+
onError(e);
|
|
178
|
+
}
|
|
179
|
+
}
|
package/lib/utils.js
ADDED
|
@@ -0,0 +1,18 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
|
|
3
|
+
Object.defineProperty(exports, "__esModule", {
|
|
4
|
+
value: true
|
|
5
|
+
});
|
|
6
|
+
exports.default = exports.writeTo = void 0;
|
|
7
|
+
|
|
8
|
+
const writeTo = function writeTo(stream, data, cb) {
|
|
9
|
+
if (!stream.write(data)) {
|
|
10
|
+
stream.once('drain', cb);
|
|
11
|
+
} else {
|
|
12
|
+
process.nextTick(cb);
|
|
13
|
+
}
|
|
14
|
+
};
|
|
15
|
+
|
|
16
|
+
exports.writeTo = writeTo;
|
|
17
|
+
var _default = writeTo;
|
|
18
|
+
exports.default = _default;
|
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@ezs/basics",
|
|
3
3
|
"description": "Basics statements for EZS",
|
|
4
|
-
"version": "2.
|
|
4
|
+
"version": "2.4.1",
|
|
5
5
|
"author": "Nicolas Thouvenin <nthouvenin@gmail.com>",
|
|
6
6
|
"bugs": "https://github.com/Inist-CNRS/ezs/issues",
|
|
7
7
|
"dependencies": {
|
|
@@ -20,15 +20,16 @@
|
|
|
20
20
|
"lodash.get": "4.4.2",
|
|
21
21
|
"lodash.mapkeys": "4.6.0",
|
|
22
22
|
"lodash.mapvalues": "4.6.0",
|
|
23
|
+
"lodash.merge": "4.6.2",
|
|
23
24
|
"lodash.set": "4.3.2",
|
|
24
25
|
"lodash.zipobject": "4.1.3",
|
|
25
|
-
"make-dir": "
|
|
26
|
+
"make-dir": "4.0.0",
|
|
26
27
|
"micromatch": "4.0.4",
|
|
27
28
|
"node-abort-controller": "1.1.0",
|
|
28
29
|
"parse-headers": "2.0.4",
|
|
29
30
|
"path-exists": "4.0.0",
|
|
30
31
|
"stream-write": "1.0.1",
|
|
31
|
-
"tar-stream": "3.1.
|
|
32
|
+
"tar-stream": "3.1.6",
|
|
32
33
|
"tmp-filepath": "2.0.0",
|
|
33
34
|
"unzipper": "0.10.11",
|
|
34
35
|
"xml-mapping": "1.7.2",
|
|
@@ -41,7 +42,7 @@
|
|
|
41
42
|
"directories": {
|
|
42
43
|
"test": "test"
|
|
43
44
|
},
|
|
44
|
-
"gitHead": "
|
|
45
|
+
"gitHead": "d9db26be111ff5fd837b57ed826d16cc9d432ba8",
|
|
45
46
|
"homepage": "https://github.com/Inist-CNRS/ezs/tree/master/packages/basics#readme",
|
|
46
47
|
"keywords": [
|
|
47
48
|
"ezs"
|