@loaders.gl/polyfills 3.3.0-alpha.9 → 3.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/dist.min.js +64 -15
- package/dist/es5/index.js +17 -11
- package/dist/es5/index.js.map +1 -1
- package/dist/es5/node/fetch/fetch-file.node.js +88 -0
- package/dist/es5/node/fetch/fetch-file.node.js.map +1 -0
- package/dist/es5/node/fetch/fetch.node.js +90 -35
- package/dist/es5/node/fetch/fetch.node.js.map +1 -1
- package/dist/es5/node/fetch/headers.node.js +2 -2
- package/dist/es5/node/fetch/headers.node.js.map +1 -1
- package/dist/es5/node/fetch/response.node.js +5 -4
- package/dist/es5/node/fetch/response.node.js.map +1 -1
- package/dist/es5/node/fetch/utils/stream-utils.node.js +9 -93
- package/dist/es5/node/fetch/utils/stream-utils.node.js.map +1 -1
- package/dist/es5/node/images/parse-image.node.js +30 -18
- package/dist/es5/node/images/parse-image.node.js.map +1 -1
- package/dist/esm/index.js +5 -4
- package/dist/esm/index.js.map +1 -1
- package/dist/esm/node/fetch/fetch-file.node.js +54 -0
- package/dist/esm/node/fetch/fetch-file.node.js.map +1 -0
- package/dist/esm/node/fetch/fetch.node.js +43 -18
- package/dist/esm/node/fetch/fetch.node.js.map +1 -1
- package/dist/esm/node/fetch/headers.node.js +1 -1
- package/dist/esm/node/fetch/headers.node.js.map +1 -1
- package/dist/esm/node/fetch/response.node.js +3 -2
- package/dist/esm/node/fetch/response.node.js.map +1 -1
- package/dist/esm/node/fetch/utils/stream-utils.node.js +3 -46
- package/dist/esm/node/fetch/utils/stream-utils.node.js.map +1 -1
- package/dist/esm/node/images/parse-image.node.js +23 -16
- package/dist/esm/node/images/parse-image.node.js.map +1 -1
- package/dist/index.d.ts +2 -1
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +13 -14
- package/dist/node/fetch/fetch-file.node.d.ts +4 -0
- package/dist/node/fetch/fetch-file.node.d.ts.map +1 -0
- package/dist/node/fetch/fetch-file.node.js +51 -0
- package/dist/node/fetch/fetch.node.d.ts +6 -1
- package/dist/node/fetch/fetch.node.d.ts.map +1 -1
- package/dist/node/fetch/fetch.node.js +57 -31
- package/dist/node/fetch/headers.node.d.ts +1 -1
- package/dist/node/fetch/headers.node.d.ts.map +1 -1
- package/dist/node/fetch/headers.node.js +2 -1
- package/dist/node/fetch/response.node.d.ts +2 -2
- package/dist/node/fetch/response.node.d.ts.map +1 -1
- package/dist/node/fetch/response.node.js +5 -6
- package/dist/node/fetch/utils/stream-utils.node.d.ts +8 -1
- package/dist/node/fetch/utils/stream-utils.node.d.ts.map +1 -1
- package/dist/node/fetch/utils/stream-utils.node.js +10 -54
- package/dist/node/images/parse-image.node.d.ts.map +1 -1
- package/dist/node/images/parse-image.node.js +26 -18
- package/package.json +3 -2
- package/src/index.ts +5 -4
- package/src/node/fetch/fetch-file.node.ts +51 -0
- package/src/node/fetch/fetch.node.ts +64 -30
- package/src/node/fetch/headers.node.ts +1 -1
- package/src/node/fetch/response.node.ts +4 -2
- package/src/node/fetch/utils/stream-utils.node.ts +10 -59
- package/src/node/images/parse-image.node.ts +31 -20
|
@@ -4,7 +4,7 @@ var _interopRequireDefault = require("@babel/runtime/helpers/interopRequireDefau
|
|
|
4
4
|
Object.defineProperty(exports, "__esModule", {
|
|
5
5
|
value: true
|
|
6
6
|
});
|
|
7
|
-
exports.
|
|
7
|
+
exports.Response = void 0;
|
|
8
8
|
var _regenerator = _interopRequireDefault(require("@babel/runtime/regenerator"));
|
|
9
9
|
var _asyncToGenerator2 = _interopRequireDefault(require("@babel/runtime/helpers/asyncToGenerator"));
|
|
10
10
|
var _classCallCheck2 = _interopRequireDefault(require("@babel/runtime/helpers/classCallCheck"));
|
|
@@ -13,8 +13,9 @@ var _defineProperty2 = _interopRequireDefault(require("@babel/runtime/helpers/de
|
|
|
13
13
|
var _typeof2 = _interopRequireDefault(require("@babel/runtime/helpers/typeof"));
|
|
14
14
|
var _assert = require("../../utils/assert");
|
|
15
15
|
var _streamUtils = require("./utils/stream-utils.node");
|
|
16
|
-
var _headers =
|
|
16
|
+
var _headers = require("./headers.node");
|
|
17
17
|
var _stream = require("stream");
|
|
18
|
+
|
|
18
19
|
var isBoolean = function isBoolean(x) {
|
|
19
20
|
return typeof x === 'boolean';
|
|
20
21
|
};
|
|
@@ -49,7 +50,7 @@ var Response = function () {
|
|
|
49
50
|
this.ok = status === 200;
|
|
50
51
|
this.status = status;
|
|
51
52
|
this.statusText = statusText;
|
|
52
|
-
this.headers = new _headers.
|
|
53
|
+
this.headers = new _headers.Headers((options === null || options === void 0 ? void 0 : options.headers) || {});
|
|
53
54
|
|
|
54
55
|
if (isReadableNodeStream(body)) {
|
|
55
56
|
this._body = (0, _streamUtils.decompressReadStream)(body, headers);
|
|
@@ -192,5 +193,5 @@ var Response = function () {
|
|
|
192
193
|
}]);
|
|
193
194
|
return Response;
|
|
194
195
|
}();
|
|
195
|
-
exports.
|
|
196
|
+
exports.Response = Response;
|
|
196
197
|
//# sourceMappingURL=response.node.js.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"response.node.js","names":["isBoolean","x","isFunction","isObject","isReadableNodeStream","read","pipe","readable","Response","body","options","headers","status","statusText","url","ok","Headers","_body","decompressReadStream","Readable","from","TextEncoder","encode","ArrayBuffer","assert","bodyUsed","concatenateReadStream","data","arrayBuffer","textDecoder","TextDecoder","decode","text","JSON","parse","Blob","Error"],"sources":["../../../../src/node/fetch/response.node.ts"],"sourcesContent":["
|
|
1
|
+
{"version":3,"file":"response.node.js","names":["isBoolean","x","isFunction","isObject","isReadableNodeStream","read","pipe","readable","Response","body","options","headers","status","statusText","url","ok","Headers","_body","decompressReadStream","Readable","from","TextEncoder","encode","ArrayBuffer","assert","bodyUsed","concatenateReadStream","data","arrayBuffer","textDecoder","TextDecoder","decode","text","JSON","parse","Blob","Error"],"sources":["../../../../src/node/fetch/response.node.ts"],"sourcesContent":["// loaders.gl, MIT license\n\nimport {assert} from '../../utils/assert';\nimport {decompressReadStream, concatenateReadStream} from './utils/stream-utils.node';\nimport {Headers} from './headers.node';\n\nconst isBoolean = (x) => typeof x === 'boolean';\nconst isFunction = (x) => typeof x === 'function';\nconst isObject = (x) => x !== null && typeof x === 'object';\nconst isReadableNodeStream = (x) =>\n isObject(x) && isFunction(x.read) && isFunction(x.pipe) && isBoolean(x.readable);\n\n/**\n * Polyfill for Browser Response\n *\n * Under Node.js we return a mock \"fetch response object\"\n * so that apps can use the same API as in the browser.\n *\n * Note: This is intended to be a \"lightweight\" implementation and will have limitations.\n *\n * See https://developer.mozilla.org/en-US/docs/Web/API/Response\n */\nimport {Readable} from 'stream';\n\nexport class Response {\n readonly ok: boolean;\n readonly status: number;\n readonly statusText: string;\n readonly headers: Headers;\n readonly url: string;\n bodyUsed: boolean = false;\n private readonly _body;\n\n // TODO - handle ArrayBuffer, ArrayBufferView, Buffer\n constructor(\n body,\n options: {\n headers?;\n status?: number;\n statusText?: string;\n url: string;\n }\n ) {\n const {headers, status = 200, statusText = 'OK', url} = options || {};\n\n this.url = url;\n this.ok = status === 200;\n this.status = status; // TODO - handle errors and set status\n this.statusText = statusText;\n this.headers = new Headers(options?.headers || {});\n\n // Check for content-encoding and create a decompression stream\n if (isReadableNodeStream(body)) {\n this._body = decompressReadStream(body, headers);\n } else if (typeof body === 'string') {\n this._body = Readable.from([new TextEncoder().encode(body)]);\n } else {\n this._body = Readable.from([body || new ArrayBuffer(0)]);\n }\n }\n\n // Subset of Properties\n\n // Returns a readable stream to the \"body\" of the response (or file)\n get body() {\n assert(!this.bodyUsed);\n assert(isReadableNodeStream(this._body)); // Not implemented: conversion of ArrayBuffer etc to stream\n this.bodyUsed = true;\n return this._body;\n }\n\n // Subset of Methods\n\n async arrayBuffer() {\n if (!isReadableNodeStream(this._body)) {\n return this._body || new ArrayBuffer(0);\n }\n const data = await concatenateReadStream(this._body);\n return data;\n }\n\n async text() {\n const arrayBuffer = await this.arrayBuffer();\n const textDecoder = new TextDecoder();\n return textDecoder.decode(arrayBuffer);\n }\n\n async json() {\n const text = await this.text();\n return JSON.parse(text);\n }\n\n async blob() {\n if (typeof Blob === 'undefined') {\n throw new Error('Blob polyfill not installed');\n }\n return new Blob([await this.arrayBuffer()]);\n }\n}\n"],"mappings":";;;;;;;;;;;;;AAEA;AACA;AACA;AAkBA;;AAhBA,IAAMA,SAAS,GAAG,SAAZA,SAAS,CAAIC,CAAC;EAAA,OAAK,OAAOA,CAAC,KAAK,SAAS;AAAA;AAC/C,IAAMC,UAAU,GAAG,SAAbA,UAAU,CAAID,CAAC;EAAA,OAAK,OAAOA,CAAC,KAAK,UAAU;AAAA;AACjD,IAAME,QAAQ,GAAG,SAAXA,QAAQ,CAAIF,CAAC;EAAA,OAAKA,CAAC,KAAK,IAAI,IAAI,sBAAOA,CAAC,MAAK,QAAQ;AAAA;AAC3D,IAAMG,oBAAoB,GAAG,SAAvBA,oBAAoB,CAAIH,CAAC;EAAA,OAC7BE,QAAQ,CAACF,CAAC,CAAC,IAAIC,UAAU,CAACD,CAAC,CAACI,IAAI,CAAC,IAAIH,UAAU,CAACD,CAAC,CAACK,IAAI,CAAC,IAAIN,SAAS,CAACC,CAAC,CAACM,QAAQ,CAAC;AAAA;;AAAC,IActEC,QAAQ;EAUnB,kBACEC,IAAI,EACJC,OAKC,EACD;IAAA;IAAA;IAAA;IAAA;IAAA;IAAA;IAAA,gDAZkB,KAAK;IAAA;IAavB,WAAwDA,OAAO,IAAI,CAAC,CAAC;MAA9DC,OAAO,QAAPA,OAAO;MAAA,mBAAEC,MAAM;MAANA,MAAM,4BAAG,GAAG;MAAA,uBAAEC,UAAU;MAAVA,UAAU,gCAAG,IAAI;MAAEC,GAAG,QAAHA,GAAG;IAEpD,IAAI,CAACA,GAAG,GAAGA,GAAG;IACd,IAAI,CAACC,EAAE,GAAGH,MAAM,KAAK,GAAG;IACxB,IAAI,CAACA,MAAM,GAAGA,MAAM;IACpB,IAAI,CAACC,UAAU,GAAGA,UAAU;IAC5B,IAAI,CAACF,OAAO,GAAG,IAAIK,gBAAO,CAAC,CAAAN,OAAO,aAAPA,OAAO,uBAAPA,OAAO,CAAEC,OAAO,KAAI,CAAC,CAAC,CAAC;;IAGlD,IAAIP,oBAAoB,CAACK,IAAI,CAAC,EAAE;MAC9B,IAAI,CAACQ,KAAK,GAAG,IAAAC,iCAAoB,EAACT,IAAI,EAAEE,OAAO,CAAC;IAClD,CAAC,MAAM,IAAI,OAAOF,IAAI,KAAK,QAAQ,EAAE;MACnC,IAAI,CAACQ,KAAK,GAAGE,gBAAQ,CAACC,IAAI,CAAC,CAAC,IAAIC,WAAW,EAAE,CAACC,MAAM,CAACb,IAAI,CAAC,CAAC,CAAC;IAC9D,CAAC,MAAM;MACL,IAAI,CAACQ,KAAK,GAAGE,gBAAQ,CAACC,IAAI,CAAC,CAACX,IAAI,IAAI,IAAIc,WAAW,CAAC,CAAC,CAAC,CAAC,CAAC;IAC1D;EACF;;EAAC;IAAA;IAAA;;IAKD,eAAW;MACT,IAAAC,cAAM,EAAC,CAAC,IAAI,CAACC,QAAQ,CAAC;MACtB,IAAAD,cAAM,EAACpB,oBAAoB,CAAC,IAAI,CAACa,KAAK,CAAC,CAAC;MACxC,IAAI,CAACQ,QAAQ,GAAG,IAAI;MACpB,OAAO,IAAI,CAACR,KAAK;IACnB;;EAAC;IAAA;IAAA;MAAA,6EAID;QAAA;QAAA;UAAA;YAAA;cAAA;gBAAA,IACOb,oBAAoB,CAAC,IAAI,CAACa,KAAK,CAAC;kBAAA;kBAAA;gBAAA;gBAAA,iCAC5B,IAAI,CAACA,KAAK,IAAI,IAAIM,WAAW,CAAC,CAAC,CAAC;cAAA;gBAAA;gBAAA,OAEtB,IAAAG,kCAAqB,EAAC,IAAI,CAACT,KAAK,CAAC;cAAA;gBAA9CU,IAAI;gBAAA,iCACHA,IAAI;cAAA;cAAA;gBAAA;YAAA;UAAA;QAAA;MAAA,CACZ;MAAA;QAAA;MAAA;MAAA;IAAA;EAAA;IAAA;IAAA;MAAA,sEAED;QAAA;QAAA;UAAA;YAAA;cAAA;gBAAA;gBAAA,OAC4B,IAAI,CAACC,WAAW,EAAE;cAAA;gBAAtCA,WAAW;gBACXC,WAAW,GAAG,IAAIC,WAAW,EAAE;gBAAA,kCAC9BD,WAAW,CAACE,MAAM,CAACH,WAAW,CAAC;cAAA;cAAA;gBAAA;YAAA;UAAA;QAAA;MAAA,CACvC;MAAA;QAAA;MAAA;MAAA;IAAA;EAAA;IAAA;IAAA;MAAA,sEAED;QAAA;QAAA;UAAA;YAAA;cAAA;gBAAA;gBAAA,OACqB,IAAI,CAACI,IAAI,EAAE;cAAA;gBAAxBA,IAAI;gBAAA,kCACHC,IAAI,CAACC,KAAK,CAACF,IAAI,CAAC;cAAA;cAAA;gBAAA;YAAA;UAAA;QAAA;MAAA,CACxB;MAAA;QAAA;MAAA;MAAA;IAAA;EAAA;IAAA;IAAA;MAAA,sEAED;QAAA;UAAA;YAAA;cAAA;gBAAA,MACM,OAAOG,IAAI,KAAK,WAAW;kBAAA;kBAAA;gBAAA;gBAAA,MACvB,IAAIC,KAAK,CAAC,6BAA6B,CAAC;cAAA;gBAAA,eAErCD,IAAI;gBAAA;gBAAA,OAAQ,IAAI,CAACP,WAAW,EAAE;cAAA;gBAAA;gBAAA;gBAAA;cAAA;cAAA;gBAAA;YAAA;UAAA;QAAA;MAAA,CAC1C;MAAA;QAAA;MAAA;MAAA;IAAA;EAAA;EAAA;AAAA;AAAA"}
|
|
@@ -6,78 +6,14 @@ Object.defineProperty(exports, "__esModule", {
|
|
|
6
6
|
});
|
|
7
7
|
exports.concatenateArrayBuffers = concatenateArrayBuffers;
|
|
8
8
|
exports.concatenateReadStream = concatenateReadStream;
|
|
9
|
-
exports.createReadStream = createReadStream;
|
|
10
9
|
exports.decompressReadStream = decompressReadStream;
|
|
11
10
|
var _regenerator = _interopRequireDefault(require("@babel/runtime/regenerator"));
|
|
12
|
-
var _defineProperty2 = _interopRequireDefault(require("@babel/runtime/helpers/defineProperty"));
|
|
13
11
|
var _asyncToGenerator2 = _interopRequireDefault(require("@babel/runtime/helpers/asyncToGenerator"));
|
|
14
|
-
var _fs = _interopRequireDefault(require("fs"));
|
|
15
|
-
var _http = _interopRequireDefault(require("http"));
|
|
16
|
-
var _https = _interopRequireDefault(require("https"));
|
|
17
12
|
var _zlib = _interopRequireDefault(require("zlib"));
|
|
18
13
|
var _decodeDataUri = require("./decode-data-uri.node");
|
|
19
14
|
function _createForOfIteratorHelper(o, allowArrayLike) { var it = typeof Symbol !== "undefined" && o[Symbol.iterator] || o["@@iterator"]; if (!it) { if (Array.isArray(o) || (it = _unsupportedIterableToArray(o)) || allowArrayLike && o && typeof o.length === "number") { if (it) o = it; var i = 0; var F = function F() {}; return { s: F, n: function n() { if (i >= o.length) return { done: true }; return { done: false, value: o[i++] }; }, e: function e(_e) { throw _e; }, f: F }; } throw new TypeError("Invalid attempt to iterate non-iterable instance.\nIn order to be iterable, non-array objects must have a [Symbol.iterator]() method."); } var normalCompletion = true, didErr = false, err; return { s: function s() { it = it.call(o); }, n: function n() { var step = it.next(); normalCompletion = step.done; return step; }, e: function e(_e2) { didErr = true; err = _e2; }, f: function f() { try { if (!normalCompletion && it.return != null) it.return(); } finally { if (didErr) throw err; } } }; }
|
|
20
15
|
function _unsupportedIterableToArray(o, minLen) { if (!o) return; if (typeof o === "string") return _arrayLikeToArray(o, minLen); var n = Object.prototype.toString.call(o).slice(8, -1); if (n === "Object" && o.constructor) n = o.constructor.name; if (n === "Map" || n === "Set") return Array.from(o); if (n === "Arguments" || /^(?:Ui|I)nt(?:8|16|32)(?:Clamped)?Array$/.test(n)) return _arrayLikeToArray(o, minLen); }
|
|
21
16
|
function _arrayLikeToArray(arr, len) { if (len == null || len > arr.length) len = arr.length; for (var i = 0, arr2 = new Array(len); i < len; i++) { arr2[i] = arr[i]; } return arr2; }
|
|
22
|
-
function ownKeys(object, enumerableOnly) { var keys = Object.keys(object); if (Object.getOwnPropertySymbols) { var symbols = Object.getOwnPropertySymbols(object); enumerableOnly && (symbols = symbols.filter(function (sym) { return Object.getOwnPropertyDescriptor(object, sym).enumerable; })), keys.push.apply(keys, symbols); } return keys; }
|
|
23
|
-
function _objectSpread(target) { for (var i = 1; i < arguments.length; i++) { var source = null != arguments[i] ? arguments[i] : {}; i % 2 ? ownKeys(Object(source), !0).forEach(function (key) { (0, _defineProperty2.default)(target, key, source[key]); }) : Object.getOwnPropertyDescriptors ? Object.defineProperties(target, Object.getOwnPropertyDescriptors(source)) : ownKeys(Object(source)).forEach(function (key) { Object.defineProperty(target, key, Object.getOwnPropertyDescriptor(source, key)); }); } return target; }
|
|
24
|
-
function isRequestURL(url) {
|
|
25
|
-
return url.startsWith('http:') || url.startsWith('https:');
|
|
26
|
-
}
|
|
27
|
-
|
|
28
|
-
function createReadStream(_x, _x2) {
|
|
29
|
-
return _createReadStream.apply(this, arguments);
|
|
30
|
-
}
|
|
31
|
-
function _createReadStream() {
|
|
32
|
-
_createReadStream = (0, _asyncToGenerator2.default)(_regenerator.default.mark(function _callee(url, options) {
|
|
33
|
-
var noqueryUrl;
|
|
34
|
-
return _regenerator.default.wrap(function _callee$(_context) {
|
|
35
|
-
while (1) {
|
|
36
|
-
switch (_context.prev = _context.next) {
|
|
37
|
-
case 0:
|
|
38
|
-
if (isRequestURL(url)) {
|
|
39
|
-
_context.next = 5;
|
|
40
|
-
break;
|
|
41
|
-
}
|
|
42
|
-
noqueryUrl = url.split('?')[0];
|
|
43
|
-
_context.next = 4;
|
|
44
|
-
return new Promise(function (resolve, reject) {
|
|
45
|
-
var stream = _fs.default.createReadStream(noqueryUrl, {
|
|
46
|
-
encoding: null
|
|
47
|
-
});
|
|
48
|
-
stream.once('readable', function () {
|
|
49
|
-
return resolve(stream);
|
|
50
|
-
});
|
|
51
|
-
stream.on('error', function (error) {
|
|
52
|
-
return reject(error);
|
|
53
|
-
});
|
|
54
|
-
});
|
|
55
|
-
case 4:
|
|
56
|
-
return _context.abrupt("return", _context.sent);
|
|
57
|
-
case 5:
|
|
58
|
-
_context.next = 7;
|
|
59
|
-
return new Promise(function (resolve, reject) {
|
|
60
|
-
var requestFunction = url.startsWith('https:') ? _https.default.request : _http.default.request;
|
|
61
|
-
var requestOptions = getRequestOptions(url, options);
|
|
62
|
-
var req = requestFunction(requestOptions, function (res) {
|
|
63
|
-
return resolve(res);
|
|
64
|
-
});
|
|
65
|
-
req.on('error', function (error) {
|
|
66
|
-
return reject(error);
|
|
67
|
-
});
|
|
68
|
-
req.end();
|
|
69
|
-
});
|
|
70
|
-
case 7:
|
|
71
|
-
return _context.abrupt("return", _context.sent);
|
|
72
|
-
case 8:
|
|
73
|
-
case "end":
|
|
74
|
-
return _context.stop();
|
|
75
|
-
}
|
|
76
|
-
}
|
|
77
|
-
}, _callee);
|
|
78
|
-
}));
|
|
79
|
-
return _createReadStream.apply(this, arguments);
|
|
80
|
-
}
|
|
81
17
|
function decompressReadStream(readStream, headers) {
|
|
82
18
|
switch (headers.get('content-encoding')) {
|
|
83
19
|
case 'br':
|
|
@@ -90,18 +26,19 @@ function decompressReadStream(readStream, headers) {
|
|
|
90
26
|
return readStream;
|
|
91
27
|
}
|
|
92
28
|
}
|
|
93
|
-
|
|
29
|
+
|
|
30
|
+
function concatenateReadStream(_x) {
|
|
94
31
|
return _concatenateReadStream.apply(this, arguments);
|
|
95
32
|
}
|
|
96
33
|
function _concatenateReadStream() {
|
|
97
|
-
_concatenateReadStream = (0, _asyncToGenerator2.default)(_regenerator.default.mark(function
|
|
34
|
+
_concatenateReadStream = (0, _asyncToGenerator2.default)(_regenerator.default.mark(function _callee(readStream) {
|
|
98
35
|
var arrayBufferChunks;
|
|
99
|
-
return _regenerator.default.wrap(function
|
|
36
|
+
return _regenerator.default.wrap(function _callee$(_context) {
|
|
100
37
|
while (1) {
|
|
101
|
-
switch (
|
|
38
|
+
switch (_context.prev = _context.next) {
|
|
102
39
|
case 0:
|
|
103
40
|
arrayBufferChunks = [];
|
|
104
|
-
|
|
41
|
+
_context.next = 3;
|
|
105
42
|
return new Promise(function (resolve, reject) {
|
|
106
43
|
readStream.on('error', function (error) {
|
|
107
44
|
return reject(error);
|
|
@@ -122,37 +59,16 @@ function _concatenateReadStream() {
|
|
|
122
59
|
});
|
|
123
60
|
});
|
|
124
61
|
case 3:
|
|
125
|
-
return
|
|
62
|
+
return _context.abrupt("return", _context.sent);
|
|
126
63
|
case 4:
|
|
127
64
|
case "end":
|
|
128
|
-
return
|
|
65
|
+
return _context.stop();
|
|
129
66
|
}
|
|
130
67
|
}
|
|
131
|
-
},
|
|
68
|
+
}, _callee);
|
|
132
69
|
}));
|
|
133
70
|
return _concatenateReadStream.apply(this, arguments);
|
|
134
71
|
}
|
|
135
|
-
|
|
136
|
-
function getRequestOptions(url, options) {
|
|
137
|
-
var originalHeaders = (options === null || options === void 0 ? void 0 : options.headers) || {};
|
|
138
|
-
var headers = {};
|
|
139
|
-
for (var _i = 0, _Object$keys = Object.keys(originalHeaders); _i < _Object$keys.length; _i++) {
|
|
140
|
-
var key = _Object$keys[_i];
|
|
141
|
-
headers[key.toLowerCase()] = originalHeaders[key];
|
|
142
|
-
}
|
|
143
|
-
|
|
144
|
-
headers['accept-encoding'] = headers['accept-encoding'] || 'gzip,br,deflate';
|
|
145
|
-
var urlObject = new URL(url);
|
|
146
|
-
return _objectSpread(_objectSpread(_objectSpread({
|
|
147
|
-
hostname: urlObject.hostname,
|
|
148
|
-
path: urlObject.pathname,
|
|
149
|
-
method: 'GET'
|
|
150
|
-
}, options), options === null || options === void 0 ? void 0 : options.fetch), {}, {
|
|
151
|
-
headers: headers,
|
|
152
|
-
port: urlObject.port
|
|
153
|
-
});
|
|
154
|
-
}
|
|
155
|
-
|
|
156
72
|
function concatenateArrayBuffers(sources) {
|
|
157
73
|
var sourceArrays = sources.map(function (source2) {
|
|
158
74
|
return source2 instanceof ArrayBuffer ? new Uint8Array(source2) : source2;
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"stream-utils.node.js","names":["
|
|
1
|
+
{"version":3,"file":"stream-utils.node.js","names":["decompressReadStream","readStream","headers","get","pipe","zlib","createBrotliDecompress","createGunzip","createDeflate","concatenateReadStream","arrayBufferChunks","Promise","resolve","reject","on","error","read","chunk","Error","push","toArrayBuffer","arrayBuffer","concatenateArrayBuffers","sources","sourceArrays","map","source2","ArrayBuffer","Uint8Array","byteLength","reduce","length","typedArray","result","offset","sourceArray","set","buffer"],"sources":["../../../../../src/node/fetch/utils/stream-utils.node.ts"],"sourcesContent":["// loaders.gl, MIT license\n\nimport zlib from 'zlib';\n\nimport {toArrayBuffer} from './decode-data-uri.node';\n\n/**\n *\n */\nexport function decompressReadStream(readStream, headers) {\n switch (headers.get('content-encoding')) {\n case 'br':\n return readStream.pipe(zlib.createBrotliDecompress());\n case 'gzip':\n return readStream.pipe(zlib.createGunzip());\n case 'deflate':\n return readStream.pipe(zlib.createDeflate());\n default:\n // No compression or an unknown one, just return it as is\n return readStream;\n }\n}\n\n/**\n *\n * @param readStream\n * @returns\n */\nexport async function concatenateReadStream(readStream): Promise<ArrayBuffer> {\n const arrayBufferChunks: ArrayBuffer[] = [];\n\n return await new Promise((resolve, reject) => {\n readStream.on('error', (error) => reject(error));\n\n // Once the readable callback has been added, stream switches to \"flowing mode\"\n // In Node 10 (but not 12 and 14) this causes `data` and `end` to never be called unless we read data here\n readStream.on('readable', () => readStream.read());\n\n readStream.on('data', (chunk) => {\n if (typeof chunk === 'string') {\n reject(new Error('Read stream not binary'));\n }\n arrayBufferChunks.push(toArrayBuffer(chunk));\n });\n\n readStream.on('end', () => {\n const arrayBuffer = concatenateArrayBuffers(arrayBufferChunks);\n resolve(arrayBuffer);\n });\n });\n}\n\n/**\n * Concatenate a sequence of ArrayBuffers\n * @return A concatenated ArrayBuffer\n * @note duplicates loader-utils since polyfills should be independent\n */\nexport function concatenateArrayBuffers(sources: (ArrayBuffer | Uint8Array)[]): ArrayBuffer {\n // Make sure all inputs are wrapped in typed arrays\n const sourceArrays = sources.map((source2) =>\n source2 instanceof ArrayBuffer ? new Uint8Array(source2) : source2\n );\n\n // Get length of all inputs\n const byteLength = sourceArrays.reduce((length, typedArray) => length + typedArray.byteLength, 0);\n\n // Allocate array with space for all inputs\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let offset = 0;\n for (const sourceArray of sourceArrays) {\n result.set(sourceArray, offset);\n offset += sourceArray.byteLength;\n }\n\n // We work with ArrayBuffers, discard the typed array wrapper\n return result.buffer;\n}\n"],"mappings":";;;;;;;;;;;AAEA;AAEA;AAAqD;AAAA;AAAA;AAK9C,SAASA,oBAAoB,CAACC,UAAU,EAAEC,OAAO,EAAE;EACxD,QAAQA,OAAO,CAACC,GAAG,CAAC,kBAAkB,CAAC;IACrC,KAAK,IAAI;MACP,OAAOF,UAAU,CAACG,IAAI,CAACC,aAAI,CAACC,sBAAsB,EAAE,CAAC;IACvD,KAAK,MAAM;MACT,OAAOL,UAAU,CAACG,IAAI,CAACC,aAAI,CAACE,YAAY,EAAE,CAAC;IAC7C,KAAK,SAAS;MACZ,OAAON,UAAU,CAACG,IAAI,CAACC,aAAI,CAACG,aAAa,EAAE,CAAC;IAC9C;MAEE,OAAOP,UAAU;EAAC;AAExB;;AAAC,SAOqBQ,qBAAqB;EAAA;AAAA;AAAA;EAAA,mFAApC,iBAAqCR,UAAU;IAAA;IAAA;MAAA;QAAA;UAAA;YAC9CS,iBAAgC,GAAG,EAAE;YAAA;YAAA,OAE9B,IAAIC,OAAO,CAAC,UAACC,OAAO,EAAEC,MAAM,EAAK;cAC5CZ,UAAU,CAACa,EAAE,CAAC,OAAO,EAAE,UAACC,KAAK;gBAAA,OAAKF,MAAM,CAACE,KAAK,CAAC;cAAA,EAAC;;cAIhDd,UAAU,CAACa,EAAE,CAAC,UAAU,EAAE;gBAAA,OAAMb,UAAU,CAACe,IAAI,EAAE;cAAA,EAAC;cAElDf,UAAU,CAACa,EAAE,CAAC,MAAM,EAAE,UAACG,KAAK,EAAK;gBAC/B,IAAI,OAAOA,KAAK,KAAK,QAAQ,EAAE;kBAC7BJ,MAAM,CAAC,IAAIK,KAAK,CAAC,wBAAwB,CAAC,CAAC;gBAC7C;gBACAR,iBAAiB,CAACS,IAAI,CAAC,IAAAC,4BAAa,EAACH,KAAK,CAAC,CAAC;cAC9C,CAAC,CAAC;cAEFhB,UAAU,CAACa,EAAE,CAAC,KAAK,EAAE,YAAM;gBACzB,IAAMO,WAAW,GAAGC,uBAAuB,CAACZ,iBAAiB,CAAC;gBAC9DE,OAAO,CAACS,WAAW,CAAC;cACtB,CAAC,CAAC;YACJ,CAAC,CAAC;UAAA;YAAA;UAAA;UAAA;YAAA;QAAA;MAAA;IAAA;EAAA,CACH;EAAA;AAAA;AAOM,SAASC,uBAAuB,CAACC,OAAqC,EAAe;EAE1F,IAAMC,YAAY,GAAGD,OAAO,CAACE,GAAG,CAAC,UAACC,OAAO;IAAA,OACvCA,OAAO,YAAYC,WAAW,GAAG,IAAIC,UAAU,CAACF,OAAO,CAAC,GAAGA,OAAO;EAAA,EACnE;;EAGD,IAAMG,UAAU,GAAGL,YAAY,CAACM,MAAM,CAAC,UAACC,MAAM,EAAEC,UAAU;IAAA,OAAKD,MAAM,GAAGC,UAAU,CAACH,UAAU;EAAA,GAAE,CAAC,CAAC;;EAGjG,IAAMI,MAAM,GAAG,IAAIL,UAAU,CAACC,UAAU,CAAC;;EAGzC,IAAIK,MAAM,GAAG,CAAC;EAAC,2CACWV,YAAY;IAAA;EAAA;IAAtC,oDAAwC;MAAA,IAA7BW,WAAW;MACpBF,MAAM,CAACG,GAAG,CAACD,WAAW,EAAED,MAAM,CAAC;MAC/BA,MAAM,IAAIC,WAAW,CAACN,UAAU;IAClC;;EAAC;IAAA;EAAA;IAAA;EAAA;EAGD,OAAOI,MAAM,CAACI,MAAM;AACtB"}
|
|
@@ -9,38 +9,29 @@ var _regenerator = _interopRequireDefault(require("@babel/runtime/regenerator"))
|
|
|
9
9
|
var _toConsumableArray2 = _interopRequireDefault(require("@babel/runtime/helpers/toConsumableArray"));
|
|
10
10
|
var _asyncToGenerator2 = _interopRequireDefault(require("@babel/runtime/helpers/asyncToGenerator"));
|
|
11
11
|
var _getPixels = _interopRequireDefault(require("get-pixels"));
|
|
12
|
-
var _assert = require("../../utils/assert");
|
|
13
|
-
var _util = _interopRequireDefault(require("util"));
|
|
14
12
|
function parseImageNode(_x, _x2) {
|
|
15
13
|
return _parseImageNode.apply(this, arguments);
|
|
16
14
|
}
|
|
17
15
|
function _parseImageNode() {
|
|
18
16
|
_parseImageNode = (0, _asyncToGenerator2.default)(_regenerator.default.mark(function _callee(arrayBuffer, mimeType) {
|
|
19
|
-
var
|
|
17
|
+
var buffer, ndarray;
|
|
20
18
|
return _regenerator.default.wrap(function _callee$(_context) {
|
|
21
19
|
while (1) {
|
|
22
20
|
switch (_context.prev = _context.next) {
|
|
23
21
|
case 0:
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
22
|
+
if (mimeType) {
|
|
23
|
+
_context.next = 2;
|
|
24
|
+
break;
|
|
25
|
+
}
|
|
26
|
+
throw new Error('MIMEType is required to parse image under Node.js');
|
|
27
|
+
case 2:
|
|
27
28
|
buffer = arrayBuffer instanceof Buffer ? arrayBuffer : Buffer.from(arrayBuffer);
|
|
28
29
|
_context.next = 5;
|
|
29
30
|
return getPixelsAsync(buffer, mimeType);
|
|
30
31
|
case 5:
|
|
31
32
|
ndarray = _context.sent;
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
data = ndarray.data instanceof Buffer ? new Uint8Array(ndarray.data) : ndarray.data;
|
|
35
|
-
return _context.abrupt("return", {
|
|
36
|
-
shape: shape,
|
|
37
|
-
data: data,
|
|
38
|
-
width: ndarray.shape[0],
|
|
39
|
-
height: ndarray.shape[1],
|
|
40
|
-
components: ndarray.shape[2],
|
|
41
|
-
layers: layers
|
|
42
|
-
});
|
|
43
|
-
case 10:
|
|
33
|
+
return _context.abrupt("return", ndarray);
|
|
34
|
+
case 7:
|
|
44
35
|
case "end":
|
|
45
36
|
return _context.stop();
|
|
46
37
|
}
|
|
@@ -49,4 +40,25 @@ function _parseImageNode() {
|
|
|
49
40
|
}));
|
|
50
41
|
return _parseImageNode.apply(this, arguments);
|
|
51
42
|
}
|
|
43
|
+
function getPixelsAsync(buffer, mimeType) {
|
|
44
|
+
return new Promise(function (resolve) {
|
|
45
|
+
return (0, _getPixels.default)(buffer, mimeType, function (err, ndarray) {
|
|
46
|
+
if (err) {
|
|
47
|
+
throw err;
|
|
48
|
+
}
|
|
49
|
+
var shape = (0, _toConsumableArray2.default)(ndarray.shape);
|
|
50
|
+
var layers = ndarray.shape.length === 4 ? ndarray.shape.shift() : 1;
|
|
51
|
+
var data = ndarray.data instanceof Buffer ? new Uint8Array(ndarray.data) : ndarray.data;
|
|
52
|
+
|
|
53
|
+
resolve({
|
|
54
|
+
shape: shape,
|
|
55
|
+
data: data,
|
|
56
|
+
width: ndarray.shape[0],
|
|
57
|
+
height: ndarray.shape[1],
|
|
58
|
+
components: ndarray.shape[2],
|
|
59
|
+
layers: layers ? [layers] : []
|
|
60
|
+
});
|
|
61
|
+
});
|
|
62
|
+
});
|
|
63
|
+
}
|
|
52
64
|
//# sourceMappingURL=parse-image.node.js.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parse-image.node.js","names":["parseImageNode","arrayBuffer","mimeType","
|
|
1
|
+
{"version":3,"file":"parse-image.node.js","names":["parseImageNode","arrayBuffer","mimeType","Error","buffer","Buffer","from","getPixelsAsync","ndarray","Promise","resolve","getPixels","err","shape","layers","length","shift","data","Uint8Array","width","height","components"],"sources":["../../../../src/node/images/parse-image.node.ts"],"sourcesContent":["import getPixels from 'get-pixels';\n\n// Note: These types are also defined in @loaders.gl/images and need to be kept in sync\ntype NDArray = {\n shape: number[];\n data: Uint8Array;\n width: number;\n height: number;\n components: number;\n layers: number[];\n};\n\nexport async function parseImageNode(arrayBuffer: ArrayBuffer, mimeType: string): Promise<NDArray> {\n if (!mimeType) {\n throw new Error('MIMEType is required to parse image under Node.js');\n }\n\n const buffer = arrayBuffer instanceof Buffer ? arrayBuffer : Buffer.from(arrayBuffer);\n\n const ndarray = await getPixelsAsync(buffer, mimeType);\n\n return ndarray;\n}\n\n// TODO - check if getPixels callback is asynchronous if provided with buffer input\n// if not, parseImage can be a sync function\nfunction getPixelsAsync(buffer: Buffer, mimeType: string): Promise<NDArray> {\n return new Promise<NDArray>((resolve) =>\n getPixels(buffer, mimeType, (err, ndarray) => {\n if (err) {\n throw err;\n }\n\n const shape = [...ndarray.shape];\n const layers = ndarray.shape.length === 4 ? ndarray.shape.shift() : 1;\n const data = ndarray.data instanceof Buffer ? new Uint8Array(ndarray.data) : ndarray.data;\n\n // extract width/height etc\n resolve({\n shape,\n data,\n width: ndarray.shape[0],\n height: ndarray.shape[1],\n components: ndarray.shape[2],\n // TODO - error\n layers: layers ? [layers] : []\n });\n })\n );\n}\n"],"mappings":";;;;;;;;;;AAAA;AAAmC,SAYbA,cAAc;EAAA;AAAA;AAAA;EAAA,4EAA7B,iBAA8BC,WAAwB,EAAEC,QAAgB;IAAA;IAAA;MAAA;QAAA;UAAA;YAAA,IACxEA,QAAQ;cAAA;cAAA;YAAA;YAAA,MACL,IAAIC,KAAK,CAAC,mDAAmD,CAAC;UAAA;YAGhEC,MAAM,GAAGH,WAAW,YAAYI,MAAM,GAAGJ,WAAW,GAAGI,MAAM,CAACC,IAAI,CAACL,WAAW,CAAC;YAAA;YAAA,OAE/DM,cAAc,CAACH,MAAM,EAAEF,QAAQ,CAAC;UAAA;YAAhDM,OAAO;YAAA,iCAENA,OAAO;UAAA;UAAA;YAAA;QAAA;MAAA;IAAA;EAAA,CACf;EAAA;AAAA;AAID,SAASD,cAAc,CAACH,MAAc,EAAEF,QAAgB,EAAoB;EAC1E,OAAO,IAAIO,OAAO,CAAU,UAACC,OAAO;IAAA,OAClC,IAAAC,kBAAS,EAACP,MAAM,EAAEF,QAAQ,EAAE,UAACU,GAAG,EAAEJ,OAAO,EAAK;MAC5C,IAAII,GAAG,EAAE;QACP,MAAMA,GAAG;MACX;MAEA,IAAMC,KAAK,oCAAOL,OAAO,CAACK,KAAK,CAAC;MAChC,IAAMC,MAAM,GAAGN,OAAO,CAACK,KAAK,CAACE,MAAM,KAAK,CAAC,GAAGP,OAAO,CAACK,KAAK,CAACG,KAAK,EAAE,GAAG,CAAC;MACrE,IAAMC,IAAI,GAAGT,OAAO,CAACS,IAAI,YAAYZ,MAAM,GAAG,IAAIa,UAAU,CAACV,OAAO,CAACS,IAAI,CAAC,GAAGT,OAAO,CAACS,IAAI;;MAGzFP,OAAO,CAAC;QACNG,KAAK,EAALA,KAAK;QACLI,IAAI,EAAJA,IAAI;QACJE,KAAK,EAAEX,OAAO,CAACK,KAAK,CAAC,CAAC,CAAC;QACvBO,MAAM,EAAEZ,OAAO,CAACK,KAAK,CAAC,CAAC,CAAC;QACxBQ,UAAU,EAAEb,OAAO,CAACK,KAAK,CAAC,CAAC,CAAC;QAE5BC,MAAM,EAAEA,MAAM,GAAG,CAACA,MAAM,CAAC,GAAG;MAC9B,CAAC,CAAC;IACJ,CAAC,CAAC;EAAA,EACH;AACH"}
|
package/dist/esm/index.js
CHANGED
|
@@ -5,9 +5,9 @@ import { TextDecoder, TextEncoder } from './lib/encoding';
|
|
|
5
5
|
import { allSettled } from './promise/all-settled';
|
|
6
6
|
|
|
7
7
|
import * as base64 from './node/buffer/btoa.node';
|
|
8
|
-
import HeadersNode from './node/fetch/headers.node';
|
|
9
|
-
import ResponseNode from './node/fetch/response.node';
|
|
10
|
-
import fetchNode from './node/fetch/fetch.node';
|
|
8
|
+
import { Headers as HeadersNode } from './node/fetch/headers.node';
|
|
9
|
+
import { Response as ResponseNode } from './node/fetch/response.node';
|
|
10
|
+
import { fetchNode } from './node/fetch/fetch.node';
|
|
11
11
|
import { encodeImageNode } from './node/images/encode-image.node';
|
|
12
12
|
import { parseImageNode } from './node/images/parse-image.node';
|
|
13
13
|
export { ReadableStreamPolyfill } from './node/file/readable-stream';
|
|
@@ -15,7 +15,8 @@ export { BlobPolyfill } from './node/file/blob';
|
|
|
15
15
|
export { FileReaderPolyfill } from './node/file/file-reader';
|
|
16
16
|
export { FilePolyfill } from './node/file/file';
|
|
17
17
|
export { installFilePolyfills } from './node/file/install-file-polyfills';
|
|
18
|
-
export {
|
|
18
|
+
export { fetchNode as _fetchNode } from './node/fetch/fetch.node';
|
|
19
|
+
export { fetchFileNode as _fetchFileNode } from './node/fetch/fetch-file.node';
|
|
19
20
|
|
|
20
21
|
const installTextEncoder = !isBrowser || !('TextEncoder' in global);
|
|
21
22
|
if (installTextEncoder) {
|
package/dist/esm/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","names":["DOMParser","isBrowser","global","TextDecoder","TextEncoder","allSettled","base64","HeadersNode","ResponseNode","fetchNode","encodeImageNode","parseImageNode","ReadableStreamPolyfill","BlobPolyfill","FileReaderPolyfill","FilePolyfill","installFilePolyfills","
|
|
1
|
+
{"version":3,"file":"index.js","names":["DOMParser","isBrowser","global","TextDecoder","TextEncoder","allSettled","base64","Headers","HeadersNode","Response","ResponseNode","fetchNode","encodeImageNode","parseImageNode","ReadableStreamPolyfill","BlobPolyfill","FileReaderPolyfill","FilePolyfill","installFilePolyfills","_fetchNode","fetchFileNode","_fetchFileNode","installTextEncoder","installTextDecoder","atob","btoa","Promise"],"sources":["../../src/index.ts"],"sourcesContent":["/* eslint-disable dot-notation */\nimport {DOMParser} from '@xmldom/xmldom';\nimport {isBrowser, global} from './utils/globals';\n\nimport {TextDecoder, TextEncoder} from './lib/encoding';\nimport {allSettled} from './promise/all-settled';\n\n// Node specific\nimport * as base64 from './node/buffer/btoa.node';\n\nimport {Headers as HeadersNode} from './node/fetch/headers.node';\nimport {Response as ResponseNode} from './node/fetch/response.node';\nimport {fetchNode as fetchNode} from './node/fetch/fetch.node';\n\nimport {encodeImageNode} from './node/images/encode-image.node';\nimport {parseImageNode} from './node/images/parse-image.node';\n\nexport {ReadableStreamPolyfill} from './node/file/readable-stream';\nexport {BlobPolyfill} from './node/file/blob';\nexport {FileReaderPolyfill} from './node/file/file-reader';\nexport {FilePolyfill} from './node/file/file';\nexport {installFilePolyfills} from './node/file/install-file-polyfills';\nexport {fetchNode as _fetchNode} from './node/fetch/fetch.node';\nexport {fetchFileNode as _fetchFileNode} from './node/fetch/fetch-file.node';\n\n// POLYFILLS: TextEncoder, TextDecoder\n// - Recent Node versions have these classes but virtually no encodings unless special build.\n// - Browser: Edge, IE11 do not have these\n\nconst installTextEncoder = !isBrowser || !('TextEncoder' in global);\nif (installTextEncoder) {\n global['TextEncoder'] = TextEncoder;\n}\n\nconst installTextDecoder = !isBrowser || !('TextDecoder' in global);\nif (installTextDecoder) {\n global['TextDecoder'] = TextDecoder;\n}\n\n// POLYFILLS: btoa, atob\n// - Node: Yes\n// - Browser: No\n\nif (!isBrowser && !('atob' in global) && base64.atob) {\n global['atob'] = base64.atob;\n}\nif (!isBrowser && !('btoa' in global) && base64.btoa) {\n global['btoa'] = base64.btoa;\n}\n\n// POLYFILL: fetch\n// - Node: Yes\n// - Browser: No. For This polyfill is node only, IE11 etc, install external polyfill\n\nif (!isBrowser && !('Headers' in global) && HeadersNode) {\n global['Headers'] = HeadersNode;\n}\n\nif (!isBrowser && !('Response' in global) && ResponseNode) {\n global['Response'] = ResponseNode;\n}\n\nif (!isBrowser && !('fetch' in global) && fetchNode) {\n global['fetch'] = fetchNode;\n}\n\n// POLYFILL: DOMParser\n// - Node: Yes\n// - Browser: No\n\nif (!isBrowser && !('DOMParser' in global) && DOMParser) {\n global['DOMParser'] = DOMParser;\n}\n\n// NODE IMAGE FUNCTIONS:\n// These are not official polyfills but used by the @loaders.gl/images module if installed\n// TODO - is there an appropriate Image API we could polyfill using an adapter?\n\nif (!isBrowser && !('_encodeImageNode' in global) && encodeImageNode) {\n global['_encodeImageNode'] = encodeImageNode;\n}\n\nif (!isBrowser && !('_parseImageNode' in global) && parseImageNode) {\n global['_parseImageNode'] = parseImageNode;\n}\n\nif (!('allSettled' in Promise)) {\n // @ts-ignore\n Promise.allSettled = allSettled;\n}\n"],"mappings":";AACA,SAAQA,SAAS,QAAO,gBAAgB;AACxC,SAAQC,SAAS,EAAEC,MAAM,QAAO,iBAAiB;AAEjD,SAAQC,WAAW,EAAEC,WAAW,QAAO,gBAAgB;AACvD,SAAQC,UAAU,QAAO,uBAAuB;;AAGhD,OAAO,KAAKC,MAAM,MAAM,yBAAyB;AAEjD,SAAQC,OAAO,IAAIC,WAAW,QAAO,2BAA2B;AAChE,SAAQC,QAAQ,IAAIC,YAAY,QAAO,4BAA4B;AACnE,SAAQC,SAAsB,QAAO,yBAAyB;AAE9D,SAAQC,eAAe,QAAO,iCAAiC;AAC/D,SAAQC,cAAc,QAAO,gCAAgC;AAE7D,SAAQC,sBAAsB,QAAO,6BAA6B;AAClE,SAAQC,YAAY,QAAO,kBAAkB;AAC7C,SAAQC,kBAAkB,QAAO,yBAAyB;AAC1D,SAAQC,YAAY,QAAO,kBAAkB;AAC7C,SAAQC,oBAAoB,QAAO,oCAAoC;AACvE,SAAQP,SAAS,IAAIQ,UAAU,QAAO,yBAAyB;AAC/D,SAAQC,aAAa,IAAIC,cAAc,QAAO,8BAA8B;;AAM5E,MAAMC,kBAAkB,GAAG,CAACrB,SAAS,IAAI,EAAE,aAAa,IAAIC,MAAM,CAAC;AACnE,IAAIoB,kBAAkB,EAAE;EACtBpB,MAAM,CAAC,aAAa,CAAC,GAAGE,WAAW;AACrC;AAEA,MAAMmB,kBAAkB,GAAG,CAACtB,SAAS,IAAI,EAAE,aAAa,IAAIC,MAAM,CAAC;AACnE,IAAIqB,kBAAkB,EAAE;EACtBrB,MAAM,CAAC,aAAa,CAAC,GAAGC,WAAW;AACrC;;AAMA,IAAI,CAACF,SAAS,IAAI,EAAE,MAAM,IAAIC,MAAM,CAAC,IAAII,MAAM,CAACkB,IAAI,EAAE;EACpDtB,MAAM,CAAC,MAAM,CAAC,GAAGI,MAAM,CAACkB,IAAI;AAC9B;AACA,IAAI,CAACvB,SAAS,IAAI,EAAE,MAAM,IAAIC,MAAM,CAAC,IAAII,MAAM,CAACmB,IAAI,EAAE;EACpDvB,MAAM,CAAC,MAAM,CAAC,GAAGI,MAAM,CAACmB,IAAI;AAC9B;;AAMA,IAAI,CAACxB,SAAS,IAAI,EAAE,SAAS,IAAIC,MAAM,CAAC,IAAIM,WAAW,EAAE;EACvDN,MAAM,CAAC,SAAS,CAAC,GAAGM,WAAW;AACjC;AAEA,IAAI,CAACP,SAAS,IAAI,EAAE,UAAU,IAAIC,MAAM,CAAC,IAAIQ,YAAY,EAAE;EACzDR,MAAM,CAAC,UAAU,CAAC,GAAGQ,YAAY;AACnC;AAEA,IAAI,CAACT,SAAS,IAAI,EAAE,OAAO,IAAIC,MAAM,CAAC,IAAIS,SAAS,EAAE;EACnDT,MAAM,CAAC,OAAO,CAAC,GAAGS,SAAS;AAC7B;;AAMA,IAAI,CAACV,SAAS,IAAI,EAAE,WAAW,IAAIC,MAAM,CAAC,IAAIF,SAAS,EAAE;EACvDE,MAAM,CAAC,WAAW,CAAC,GAAGF,SAAS;AACjC;;AAMA,IAAI,CAACC,SAAS,IAAI,EAAE,kBAAkB,IAAIC,MAAM,CAAC,IAAIU,eAAe,EAAE;EACpEV,MAAM,CAAC,kBAAkB,CAAC,GAAGU,eAAe;AAC9C;AAEA,IAAI,CAACX,SAAS,IAAI,EAAE,iBAAiB,IAAIC,MAAM,CAAC,IAAIW,cAAc,EAAE;EAClEX,MAAM,CAAC,iBAAiB,CAAC,GAAGW,cAAc;AAC5C;AAEA,IAAI,EAAE,YAAY,IAAIa,OAAO,CAAC,EAAE;EAE9BA,OAAO,CAACrB,UAAU,GAAGA,UAAU;AACjC"}
|
|
@@ -0,0 +1,54 @@
|
|
|
1
|
+
|
|
2
|
+
|
|
3
|
+
import fs from 'fs';
|
|
4
|
+
import { Response } from './response.node';
|
|
5
|
+
import { Headers } from './headers.node';
|
|
6
|
+
export function isRequestURL(url) {
|
|
7
|
+
return url.startsWith('http:') || url.startsWith('https:');
|
|
8
|
+
}
|
|
9
|
+
export async function fetchFileNode(url, options) {
|
|
10
|
+
const noqueryUrl = url.split('?')[0];
|
|
11
|
+
try {
|
|
12
|
+
const body = await new Promise((resolve, reject) => {
|
|
13
|
+
const stream = fs.createReadStream(noqueryUrl, {
|
|
14
|
+
encoding: null
|
|
15
|
+
});
|
|
16
|
+
stream.once('readable', () => resolve(stream));
|
|
17
|
+
stream.on('error', error => reject(error));
|
|
18
|
+
});
|
|
19
|
+
const status = 200;
|
|
20
|
+
const statusText = 'OK';
|
|
21
|
+
const headers = getHeadersForFile(noqueryUrl);
|
|
22
|
+
return new Response(body, {
|
|
23
|
+
headers,
|
|
24
|
+
status,
|
|
25
|
+
statusText,
|
|
26
|
+
url
|
|
27
|
+
});
|
|
28
|
+
} catch (error) {
|
|
29
|
+
const status = 400;
|
|
30
|
+
const statusText = error.message;
|
|
31
|
+
const headers = {};
|
|
32
|
+
return new Response(error.message, {
|
|
33
|
+
headers,
|
|
34
|
+
status,
|
|
35
|
+
statusText,
|
|
36
|
+
url
|
|
37
|
+
});
|
|
38
|
+
}
|
|
39
|
+
}
|
|
40
|
+
function getHeadersForFile(noqueryUrl) {
|
|
41
|
+
const headers = {};
|
|
42
|
+
|
|
43
|
+
if (!headers['content-length']) {
|
|
44
|
+
const stats = fs.statSync(noqueryUrl);
|
|
45
|
+
headers['content-length'] = stats.size;
|
|
46
|
+
}
|
|
47
|
+
|
|
48
|
+
if (noqueryUrl.endsWith('.gz')) {
|
|
49
|
+
noqueryUrl = noqueryUrl.slice(0, -3);
|
|
50
|
+
headers['content-encoding'] = 'gzip';
|
|
51
|
+
}
|
|
52
|
+
return new Headers(headers);
|
|
53
|
+
}
|
|
54
|
+
//# sourceMappingURL=fetch-file.node.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"fetch-file.node.js","names":["fs","Response","Headers","isRequestURL","url","startsWith","fetchFileNode","options","noqueryUrl","split","body","Promise","resolve","reject","stream","createReadStream","encoding","once","on","error","status","statusText","headers","getHeadersForFile","message","stats","statSync","size","endsWith","slice"],"sources":["../../../../src/node/fetch/fetch-file.node.ts"],"sourcesContent":["// loaders.gl, MIT license\n\nimport fs from 'fs'; // `fs` will be empty object in browsers (see package.json \"browser\" field).\nimport {Response} from './response.node';\nimport {Headers} from './headers.node';\n\nexport function isRequestURL(url: string): boolean {\n return url.startsWith('http:') || url.startsWith('https:');\n}\n\nexport async function fetchFileNode(url: string, options): Promise<Response> {\n const noqueryUrl = url.split('?')[0];\n\n try {\n // Now open the stream\n const body = await new Promise((resolve, reject) => {\n // @ts-ignore\n const stream = fs.createReadStream(noqueryUrl, {encoding: null});\n stream.once('readable', () => resolve(stream));\n stream.on('error', (error) => reject(error));\n });\n\n const status = 200;\n const statusText = 'OK';\n const headers = getHeadersForFile(noqueryUrl);\n return new Response(body, {headers, status, statusText, url});\n } catch (error) {\n const status = 400;\n const statusText = (error as Error).message;\n const headers = {};\n return new Response((error as Error).message, {headers, status, statusText, url});\n }\n}\n\nfunction getHeadersForFile(noqueryUrl: string): Headers {\n const headers = {};\n\n // Fix up content length if we can for best progress experience\n if (!headers['content-length']) {\n const stats = fs.statSync(noqueryUrl);\n headers['content-length'] = stats.size;\n }\n\n // Automatically decompress gzipped files with .gz extension\n if (noqueryUrl.endsWith('.gz')) {\n noqueryUrl = noqueryUrl.slice(0, -3);\n headers['content-encoding'] = 'gzip';\n }\n\n return new Headers(headers);\n}\n"],"mappings":";;AAEA,OAAOA,EAAE,MAAM,IAAI;AACnB,SAAQC,QAAQ,QAAO,iBAAiB;AACxC,SAAQC,OAAO,QAAO,gBAAgB;AAEtC,OAAO,SAASC,YAAY,CAACC,GAAW,EAAW;EACjD,OAAOA,GAAG,CAACC,UAAU,CAAC,OAAO,CAAC,IAAID,GAAG,CAACC,UAAU,CAAC,QAAQ,CAAC;AAC5D;AAEA,OAAO,eAAeC,aAAa,CAACF,GAAW,EAAEG,OAAO,EAAqB;EAC3E,MAAMC,UAAU,GAAGJ,GAAG,CAACK,KAAK,CAAC,GAAG,CAAC,CAAC,CAAC,CAAC;EAEpC,IAAI;IAEF,MAAMC,IAAI,GAAG,MAAM,IAAIC,OAAO,CAAC,CAACC,OAAO,EAAEC,MAAM,KAAK;MAElD,MAAMC,MAAM,GAAGd,EAAE,CAACe,gBAAgB,CAACP,UAAU,EAAE;QAACQ,QAAQ,EAAE;MAAI,CAAC,CAAC;MAChEF,MAAM,CAACG,IAAI,CAAC,UAAU,EAAE,MAAML,OAAO,CAACE,MAAM,CAAC,CAAC;MAC9CA,MAAM,CAACI,EAAE,CAAC,OAAO,EAAGC,KAAK,IAAKN,MAAM,CAACM,KAAK,CAAC,CAAC;IAC9C,CAAC,CAAC;IAEF,MAAMC,MAAM,GAAG,GAAG;IAClB,MAAMC,UAAU,GAAG,IAAI;IACvB,MAAMC,OAAO,GAAGC,iBAAiB,CAACf,UAAU,CAAC;IAC7C,OAAO,IAAIP,QAAQ,CAACS,IAAI,EAAE;MAACY,OAAO;MAAEF,MAAM;MAAEC,UAAU;MAAEjB;IAAG,CAAC,CAAC;EAC/D,CAAC,CAAC,OAAOe,KAAK,EAAE;IACd,MAAMC,MAAM,GAAG,GAAG;IAClB,MAAMC,UAAU,GAAIF,KAAK,CAAWK,OAAO;IAC3C,MAAMF,OAAO,GAAG,CAAC,CAAC;IAClB,OAAO,IAAIrB,QAAQ,CAAEkB,KAAK,CAAWK,OAAO,EAAE;MAACF,OAAO;MAAEF,MAAM;MAAEC,UAAU;MAAEjB;IAAG,CAAC,CAAC;EACnF;AACF;AAEA,SAASmB,iBAAiB,CAACf,UAAkB,EAAW;EACtD,MAAMc,OAAO,GAAG,CAAC,CAAC;;EAGlB,IAAI,CAACA,OAAO,CAAC,gBAAgB,CAAC,EAAE;IAC9B,MAAMG,KAAK,GAAGzB,EAAE,CAAC0B,QAAQ,CAAClB,UAAU,CAAC;IACrCc,OAAO,CAAC,gBAAgB,CAAC,GAAGG,KAAK,CAACE,IAAI;EACxC;;EAGA,IAAInB,UAAU,CAACoB,QAAQ,CAAC,KAAK,CAAC,EAAE;IAC9BpB,UAAU,GAAGA,UAAU,CAACqB,KAAK,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC;IACpCP,OAAO,CAAC,kBAAkB,CAAC,GAAG,MAAM;EACtC;EAEA,OAAO,IAAIpB,OAAO,CAACoB,OAAO,CAAC;AAC7B"}
|
|
@@ -1,13 +1,20 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
import
|
|
1
|
+
|
|
2
|
+
|
|
3
|
+
import http from 'http';
|
|
4
|
+
import https from 'https';
|
|
5
|
+
import { Response } from './response.node';
|
|
6
|
+
import { Headers } from './headers.node';
|
|
4
7
|
import { decodeDataUri } from './utils/decode-data-uri.node';
|
|
5
|
-
import {
|
|
8
|
+
import { fetchFileNode } from './fetch-file.node';
|
|
6
9
|
const isDataURL = url => url.startsWith('data:');
|
|
7
10
|
const isRequestURL = url => url.startsWith('http:') || url.startsWith('https:');
|
|
8
11
|
|
|
9
|
-
export
|
|
12
|
+
export async function fetchNode(url, options) {
|
|
10
13
|
try {
|
|
14
|
+
if (!isRequestURL(url) && !isDataURL(url)) {
|
|
15
|
+
return await fetchFileNode(url, options);
|
|
16
|
+
}
|
|
17
|
+
|
|
11
18
|
if (isDataURL(url)) {
|
|
12
19
|
const {
|
|
13
20
|
arrayBuffer,
|
|
@@ -29,7 +36,7 @@ export default async function fetchNode(url, options) {
|
|
|
29
36
|
syntheticResponseHeaders['content-encoding'] = 'gzip';
|
|
30
37
|
}
|
|
31
38
|
|
|
32
|
-
const body = await
|
|
39
|
+
const body = await createHTTPRequestReadStream(originalUrl, options);
|
|
33
40
|
const headers = getHeaders(url, body, syntheticResponseHeaders);
|
|
34
41
|
const {
|
|
35
42
|
status,
|
|
@@ -56,6 +63,15 @@ export default async function fetchNode(url, options) {
|
|
|
56
63
|
}
|
|
57
64
|
}
|
|
58
65
|
|
|
66
|
+
export async function createHTTPRequestReadStream(url, options) {
|
|
67
|
+
return await new Promise((resolve, reject) => {
|
|
68
|
+
const requestOptions = getRequestOptions(url, options);
|
|
69
|
+
const req = url.startsWith('https:') ? https.request(requestOptions, res => resolve(res)) : http.request(requestOptions, res => resolve(res));
|
|
70
|
+
req.on('error', error => reject(error));
|
|
71
|
+
req.end();
|
|
72
|
+
});
|
|
73
|
+
}
|
|
74
|
+
|
|
59
75
|
function generateRedirectUrl(originalUrl, location) {
|
|
60
76
|
if (location.startsWith('http')) {
|
|
61
77
|
return location;
|
|
@@ -65,6 +81,25 @@ function generateRedirectUrl(originalUrl, location) {
|
|
|
65
81
|
return url.href;
|
|
66
82
|
}
|
|
67
83
|
|
|
84
|
+
function getRequestOptions(url, options) {
|
|
85
|
+
const originalHeaders = (options === null || options === void 0 ? void 0 : options.headers) || {};
|
|
86
|
+
const headers = {};
|
|
87
|
+
for (const key of Object.keys(originalHeaders)) {
|
|
88
|
+
headers[key.toLowerCase()] = originalHeaders[key];
|
|
89
|
+
}
|
|
90
|
+
|
|
91
|
+
headers['accept-encoding'] = headers['accept-encoding'] || 'gzip,br,deflate';
|
|
92
|
+
const urlObject = new URL(url);
|
|
93
|
+
return {
|
|
94
|
+
hostname: urlObject.hostname,
|
|
95
|
+
path: urlObject.pathname,
|
|
96
|
+
method: 'GET',
|
|
97
|
+
...options,
|
|
98
|
+
...(options === null || options === void 0 ? void 0 : options.fetch),
|
|
99
|
+
headers,
|
|
100
|
+
port: urlObject.port
|
|
101
|
+
};
|
|
102
|
+
}
|
|
68
103
|
function getStatus(httpResponse) {
|
|
69
104
|
if (httpResponse.statusCode) {
|
|
70
105
|
return {
|
|
@@ -97,18 +132,8 @@ function getHeaders(url, httpResponse) {
|
|
|
97
132
|
Object.assign(headers, additionalHeaders);
|
|
98
133
|
return new Headers(headers);
|
|
99
134
|
}
|
|
135
|
+
|
|
100
136
|
function getContentLength(url) {
|
|
101
|
-
|
|
102
|
-
return null;
|
|
103
|
-
} else if (isDataURL(url)) {
|
|
104
|
-
return url.length - 'data:'.length;
|
|
105
|
-
}
|
|
106
|
-
try {
|
|
107
|
-
const noqueryUrl = url.split('?')[0];
|
|
108
|
-
const stats = fs.statSync(noqueryUrl);
|
|
109
|
-
return stats.size;
|
|
110
|
-
} catch (error) {
|
|
111
|
-
}
|
|
112
|
-
return null;
|
|
137
|
+
return isDataURL(url) ? url.length - 'data:'.length : null;
|
|
113
138
|
}
|
|
114
139
|
//# sourceMappingURL=fetch.node.js.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"fetch.node.js","names":["fs","Response","Headers","decodeDataUri","createReadStream","isDataURL","url","startsWith","isRequestURL","fetchNode","options","arrayBuffer","mimeType","response","headers","syntheticResponseHeaders","originalUrl","endsWith","slice","body","getHeaders","status","statusText","getStatus","followRedirect","undefined","has","redirectUrl","generateRedirectUrl","get","error","String","location","URL","pathname","href","httpResponse","statusCode","statusMessage","additionalHeaders","httpHeaders","key","header","toLowerCase","contentLength","getContentLength","Number","isFinite","Object","assign","length","noqueryUrl","split","stats","statSync","size"],"sources":["../../../../src/node/fetch/fetch.node.ts"],"sourcesContent":["import fs from 'fs'; // `fs` will be empty object in browsers (see package.json \"browser\" field).\nimport Response from './response.node';\nimport Headers from './headers.node';\n\nimport {decodeDataUri} from './utils/decode-data-uri.node';\nimport {createReadStream} from './utils/stream-utils.node';\n\nconst isDataURL = (url) => url.startsWith('data:');\nconst isRequestURL = (url) => url.startsWith('http:') || url.startsWith('https:');\n\n/**\n * Emulation of Browser fetch for Node.js\n * @param url\n * @param options\n */\nexport default async function fetchNode(url, options) {\n try {\n // Handle data urls in node, to match `fetch``\n // Note - this loses the MIME type, data URIs are handled directly in fetch\n if (isDataURL(url)) {\n const {arrayBuffer, mimeType} = decodeDataUri(url);\n const response = new Response(arrayBuffer, {\n headers: {'content-type': mimeType},\n url\n });\n return response;\n }\n\n // Automatically decompress gzipped files with .gz extension\n const syntheticResponseHeaders = {};\n const originalUrl = url;\n if (url.endsWith('.gz')) {\n url = url.slice(0, -3);\n syntheticResponseHeaders['content-encoding'] = 'gzip';\n }\n\n // Need to create the stream in advance since Response constructor needs to be sync\n const body = await createReadStream(originalUrl, options);\n const headers = getHeaders(url, body, syntheticResponseHeaders);\n const {status, statusText} = getStatus(body);\n\n const followRedirect =\n !options || options.followRedirect || options.followRedirect === undefined;\n\n if (status >= 300 && status < 400 && headers.has('location') && followRedirect) {\n const redirectUrl = generateRedirectUrl(url, headers.get('location'));\n\n // Redirect\n return await fetchNode(redirectUrl, options);\n }\n return new Response(body, {headers, status, statusText, url});\n } catch (error) {\n // TODO - what error code to use here?\n return new Response(null, {status: 400, statusText: String(error), url});\n }\n}\n\n/**\n * Generate redirect url from location without origin and protocol.\n * @param originalUrl\n * @param redirectUrl\n */\nfunction generateRedirectUrl(originalUrl: string, location: string): string {\n if (location.startsWith('http')) {\n return location;\n }\n // If url doesn't have origin and protocol just extend current url origin with location.\n const url = new URL(originalUrl);\n url.pathname = location;\n\n return url.href;\n}\n\n// HELPER FUNCTIONS\n// PRIVATE\n\nfunction getStatus(httpResponse) {\n if (httpResponse.statusCode) {\n return {status: httpResponse.statusCode, statusText: httpResponse.statusMessage || 'NA'};\n }\n return {status: 200, statusText: 'OK'};\n}\n\nfunction getHeaders(url, httpResponse, additionalHeaders = {}) {\n const headers = {};\n\n if (httpResponse && httpResponse.headers) {\n const httpHeaders = httpResponse.headers;\n for (const key in httpHeaders) {\n const header = httpHeaders[key];\n headers[key.toLowerCase()] = String(header);\n }\n }\n\n // Fix up content length if we can for best progress experience\n if (!headers['content-length']) {\n const contentLength = getContentLength(url);\n if (Number.isFinite(contentLength)) {\n headers['content-length'] = contentLength;\n }\n }\n\n Object.assign(headers, additionalHeaders);\n\n return new Headers(headers);\n}\n\nfunction getContentLength(url) {\n if (isRequestURL(url)) {\n // Needs to be read from actual headers\n return null;\n } else if (isDataURL(url)) {\n // TODO - remove media type etc\n return url.length - 'data:'.length;\n }\n // File URL\n // TODO - how to handle non-existing file, this presumably just throws\n try {\n // strip query params from URL\n const noqueryUrl = url.split('?')[0];\n const stats = fs.statSync(noqueryUrl);\n return stats.size;\n } catch (error) {\n // ignore for now\n }\n\n return null;\n}\n"],"mappings":"AAAA,OAAOA,EAAE,MAAM,IAAI;AACnB,OAAOC,QAAQ,MAAM,iBAAiB;AACtC,OAAOC,OAAO,MAAM,gBAAgB;AAEpC,SAAQC,aAAa,QAAO,8BAA8B;AAC1D,SAAQC,gBAAgB,QAAO,2BAA2B;AAE1D,MAAMC,SAAS,GAAIC,GAAG,IAAKA,GAAG,CAACC,UAAU,CAAC,OAAO,CAAC;AAClD,MAAMC,YAAY,GAAIF,GAAG,IAAKA,GAAG,CAACC,UAAU,CAAC,OAAO,CAAC,IAAID,GAAG,CAACC,UAAU,CAAC,QAAQ,CAAC;;AAOjF,eAAe,eAAeE,SAAS,CAACH,GAAG,EAAEI,OAAO,EAAE;EACpD,IAAI;IAGF,IAAIL,SAAS,CAACC,GAAG,CAAC,EAAE;MAClB,MAAM;QAACK,WAAW;QAAEC;MAAQ,CAAC,GAAGT,aAAa,CAACG,GAAG,CAAC;MAClD,MAAMO,QAAQ,GAAG,IAAIZ,QAAQ,CAACU,WAAW,EAAE;QACzCG,OAAO,EAAE;UAAC,cAAc,EAAEF;QAAQ,CAAC;QACnCN;MACF,CAAC,CAAC;MACF,OAAOO,QAAQ;IACjB;;IAGA,MAAME,wBAAwB,GAAG,CAAC,CAAC;IACnC,MAAMC,WAAW,GAAGV,GAAG;IACvB,IAAIA,GAAG,CAACW,QAAQ,CAAC,KAAK,CAAC,EAAE;MACvBX,GAAG,GAAGA,GAAG,CAACY,KAAK,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC;MACtBH,wBAAwB,CAAC,kBAAkB,CAAC,GAAG,MAAM;IACvD;;IAGA,MAAMI,IAAI,GAAG,MAAMf,gBAAgB,CAACY,WAAW,EAAEN,OAAO,CAAC;IACzD,MAAMI,OAAO,GAAGM,UAAU,CAACd,GAAG,EAAEa,IAAI,EAAEJ,wBAAwB,CAAC;IAC/D,MAAM;MAACM,MAAM;MAAEC;IAAU,CAAC,GAAGC,SAAS,CAACJ,IAAI,CAAC;IAE5C,MAAMK,cAAc,GAClB,CAACd,OAAO,IAAIA,OAAO,CAACc,cAAc,IAAId,OAAO,CAACc,cAAc,KAAKC,SAAS;IAE5E,IAAIJ,MAAM,IAAI,GAAG,IAAIA,MAAM,GAAG,GAAG,IAAIP,OAAO,CAACY,GAAG,CAAC,UAAU,CAAC,IAAIF,cAAc,EAAE;MAC9E,MAAMG,WAAW,GAAGC,mBAAmB,CAACtB,GAAG,EAAEQ,OAAO,CAACe,GAAG,CAAC,UAAU,CAAC,CAAC;;MAGrE,OAAO,MAAMpB,SAAS,CAACkB,WAAW,EAAEjB,OAAO,CAAC;IAC9C;IACA,OAAO,IAAIT,QAAQ,CAACkB,IAAI,EAAE;MAACL,OAAO;MAAEO,MAAM;MAAEC,UAAU;MAAEhB;IAAG,CAAC,CAAC;EAC/D,CAAC,CAAC,OAAOwB,KAAK,EAAE;IAEd,OAAO,IAAI7B,QAAQ,CAAC,IAAI,EAAE;MAACoB,MAAM,EAAE,GAAG;MAAEC,UAAU,EAAES,MAAM,CAACD,KAAK,CAAC;MAAExB;IAAG,CAAC,CAAC;EAC1E;AACF;;AAOA,SAASsB,mBAAmB,CAACZ,WAAmB,EAAEgB,QAAgB,EAAU;EAC1E,IAAIA,QAAQ,CAACzB,UAAU,CAAC,MAAM,CAAC,EAAE;IAC/B,OAAOyB,QAAQ;EACjB;EAEA,MAAM1B,GAAG,GAAG,IAAI2B,GAAG,CAACjB,WAAW,CAAC;EAChCV,GAAG,CAAC4B,QAAQ,GAAGF,QAAQ;EAEvB,OAAO1B,GAAG,CAAC6B,IAAI;AACjB;;AAKA,SAASZ,SAAS,CAACa,YAAY,EAAE;EAC/B,IAAIA,YAAY,CAACC,UAAU,EAAE;IAC3B,OAAO;MAAChB,MAAM,EAAEe,YAAY,CAACC,UAAU;MAAEf,UAAU,EAAEc,YAAY,CAACE,aAAa,IAAI;IAAI,CAAC;EAC1F;EACA,OAAO;IAACjB,MAAM,EAAE,GAAG;IAAEC,UAAU,EAAE;EAAI,CAAC;AACxC;AAEA,SAASF,UAAU,CAACd,GAAG,EAAE8B,YAAY,EAA0B;EAAA,IAAxBG,iBAAiB,uEAAG,CAAC,CAAC;EAC3D,MAAMzB,OAAO,GAAG,CAAC,CAAC;EAElB,IAAIsB,YAAY,IAAIA,YAAY,CAACtB,OAAO,EAAE;IACxC,MAAM0B,WAAW,GAAGJ,YAAY,CAACtB,OAAO;IACxC,KAAK,MAAM2B,GAAG,IAAID,WAAW,EAAE;MAC7B,MAAME,MAAM,GAAGF,WAAW,CAACC,GAAG,CAAC;MAC/B3B,OAAO,CAAC2B,GAAG,CAACE,WAAW,EAAE,CAAC,GAAGZ,MAAM,CAACW,MAAM,CAAC;IAC7C;EACF;;EAGA,IAAI,CAAC5B,OAAO,CAAC,gBAAgB,CAAC,EAAE;IAC9B,MAAM8B,aAAa,GAAGC,gBAAgB,CAACvC,GAAG,CAAC;IAC3C,IAAIwC,MAAM,CAACC,QAAQ,CAACH,aAAa,CAAC,EAAE;MAClC9B,OAAO,CAAC,gBAAgB,CAAC,GAAG8B,aAAa;IAC3C;EACF;EAEAI,MAAM,CAACC,MAAM,CAACnC,OAAO,EAAEyB,iBAAiB,CAAC;EAEzC,OAAO,IAAIrC,OAAO,CAACY,OAAO,CAAC;AAC7B;AAEA,SAAS+B,gBAAgB,CAACvC,GAAG,EAAE;EAC7B,IAAIE,YAAY,CAACF,GAAG,CAAC,EAAE;IAErB,OAAO,IAAI;EACb,CAAC,MAAM,IAAID,SAAS,CAACC,GAAG,CAAC,EAAE;IAEzB,OAAOA,GAAG,CAAC4C,MAAM,GAAG,OAAO,CAACA,MAAM;EACpC;EAGA,IAAI;IAEF,MAAMC,UAAU,GAAG7C,GAAG,CAAC8C,KAAK,CAAC,GAAG,CAAC,CAAC,CAAC,CAAC;IACpC,MAAMC,KAAK,GAAGrD,EAAE,CAACsD,QAAQ,CAACH,UAAU,CAAC;IACrC,OAAOE,KAAK,CAACE,IAAI;EACnB,CAAC,CAAC,OAAOzB,KAAK,EAAE;EAEhB;EAEA,OAAO,IAAI;AACb"}
|
|
1
|
+
{"version":3,"file":"fetch.node.js","names":["http","https","Response","Headers","decodeDataUri","fetchFileNode","isDataURL","url","startsWith","isRequestURL","fetchNode","options","arrayBuffer","mimeType","response","headers","syntheticResponseHeaders","originalUrl","endsWith","slice","body","createHTTPRequestReadStream","getHeaders","status","statusText","getStatus","followRedirect","undefined","has","redirectUrl","generateRedirectUrl","get","error","String","Promise","resolve","reject","requestOptions","getRequestOptions","req","request","res","on","end","location","URL","pathname","href","originalHeaders","key","Object","keys","toLowerCase","urlObject","hostname","path","method","fetch","port","httpResponse","statusCode","statusMessage","additionalHeaders","httpHeaders","header","contentLength","getContentLength","Number","isFinite","assign","length"],"sources":["../../../../src/node/fetch/fetch.node.ts"],"sourcesContent":["// loaders.gl, MIT license\n\nimport http from 'http';\nimport https from 'https';\nimport {Response} from './response.node';\nimport {Headers} from './headers.node';\nimport {decodeDataUri} from './utils/decode-data-uri.node';\n\nimport {fetchFileNode} from './fetch-file.node';\n\nconst isDataURL = (url: string): boolean => url.startsWith('data:');\nconst isRequestURL = (url: string): boolean => url.startsWith('http:') || url.startsWith('https:');\n\n/**\n * Emulation of Browser fetch for Node.js\n * @param url\n * @param options\n */\n// eslint-disable-next-line complexity\nexport async function fetchNode(url: string, options): Promise<Response> {\n try {\n // Handle file streams in node\n if (!isRequestURL(url) && !isDataURL(url)) {\n return await fetchFileNode(url, options);\n }\n\n // Handle data urls in node, to match `fetch``\n // Note - this loses the MIME type, data URIs are handled directly in fetch\n if (isDataURL(url)) {\n const {arrayBuffer, mimeType} = decodeDataUri(url);\n const response = new Response(arrayBuffer, {\n headers: {'content-type': mimeType},\n url\n });\n return response;\n }\n\n // Automatically decompress gzipped files with .gz extension\n const syntheticResponseHeaders = {};\n const originalUrl = url;\n if (url.endsWith('.gz')) {\n url = url.slice(0, -3);\n syntheticResponseHeaders['content-encoding'] = 'gzip';\n }\n\n // Need to create the stream in advance since Response constructor needs to be sync\n const body = await createHTTPRequestReadStream(originalUrl, options);\n const headers = getHeaders(url, body, syntheticResponseHeaders);\n const {status, statusText} = getStatus(body);\n\n const followRedirect =\n !options || options.followRedirect || options.followRedirect === undefined;\n\n if (status >= 300 && status < 400 && headers.has('location') && followRedirect) {\n const redirectUrl = generateRedirectUrl(url, headers.get('location'));\n\n // Redirect\n return await fetchNode(redirectUrl, options);\n }\n return new Response(body, {headers, status, statusText, url});\n } catch (error) {\n // TODO - what error code to use here?\n return new Response(null, {status: 400, statusText: String(error), url});\n }\n}\n\n/** Returns a promise that resolves to a readable stream */\nexport async function createHTTPRequestReadStream(\n url: string,\n options\n): Promise<http.IncomingMessage> {\n // HANDLE HTTP/HTTPS REQUESTS IN NODE\n // TODO: THIS IS BAD SINCE WE RETURN A PROMISE INSTEAD OF A STREAM\n return await new Promise((resolve, reject) => {\n const requestOptions = getRequestOptions(url, options);\n const req = url.startsWith('https:')\n ? https.request(requestOptions, (res) => resolve(res))\n : http.request(requestOptions, (res) => resolve(res));\n req.on('error', (error) => reject(error));\n req.end();\n });\n}\n\n/**\n * Generate redirect url from location without origin and protocol.\n * @param originalUrl\n * @param redirectUrl\n */\nfunction generateRedirectUrl(originalUrl: string, location: string): string {\n if (location.startsWith('http')) {\n return location;\n }\n // If url doesn't have origin and protocol just extend current url origin with location.\n const url = new URL(originalUrl);\n url.pathname = location;\n\n return url.href;\n}\n\n// HELPER FUNCTIONS\n\nfunction getRequestOptions(url: string, options?: {fetch?: typeof fetch; headers?}) {\n // Ensure header keys are lower case so that we can merge without duplicates\n const originalHeaders = options?.headers || {};\n const headers = {};\n for (const key of Object.keys(originalHeaders)) {\n headers[key.toLowerCase()] = originalHeaders[key];\n }\n\n // Add default accept-encoding to headers\n headers['accept-encoding'] = headers['accept-encoding'] || 'gzip,br,deflate';\n\n const urlObject = new URL(url);\n return {\n hostname: urlObject.hostname,\n path: urlObject.pathname,\n method: 'GET',\n // Add options and user provided 'options.fetch' overrides if available\n ...options,\n ...options?.fetch,\n // Override with updated headers with accepted encodings:\n headers,\n port: urlObject.port\n };\n}\n\nfunction getStatus(httpResponse: http.IncomingMessage): {status: number; statusText: string} {\n if (httpResponse.statusCode) {\n return {status: httpResponse.statusCode, statusText: httpResponse.statusMessage || 'NA'};\n }\n return {status: 200, statusText: 'OK'};\n}\n\nfunction getHeaders(url, httpResponse, additionalHeaders = {}) {\n const headers = {};\n\n if (httpResponse && httpResponse.headers) {\n const httpHeaders = httpResponse.headers;\n for (const key in httpHeaders) {\n const header = httpHeaders[key];\n headers[key.toLowerCase()] = String(header);\n }\n }\n\n // Fix up content length if we can for best progress experience\n if (!headers['content-length']) {\n const contentLength = getContentLength(url);\n if (Number.isFinite(contentLength)) {\n headers['content-length'] = contentLength;\n }\n }\n\n Object.assign(headers, additionalHeaders);\n\n return new Headers(headers);\n}\n\n/** Needs to be read from actual headers */\nfunction getContentLength(url: string): number | null {\n // TODO - remove media type etc\n return isDataURL(url) ? url.length - 'data:'.length : null;\n}\n"],"mappings":";;AAEA,OAAOA,IAAI,MAAM,MAAM;AACvB,OAAOC,KAAK,MAAM,OAAO;AACzB,SAAQC,QAAQ,QAAO,iBAAiB;AACxC,SAAQC,OAAO,QAAO,gBAAgB;AACtC,SAAQC,aAAa,QAAO,8BAA8B;AAE1D,SAAQC,aAAa,QAAO,mBAAmB;AAE/C,MAAMC,SAAS,GAAIC,GAAW,IAAcA,GAAG,CAACC,UAAU,CAAC,OAAO,CAAC;AACnE,MAAMC,YAAY,GAAIF,GAAW,IAAcA,GAAG,CAACC,UAAU,CAAC,OAAO,CAAC,IAAID,GAAG,CAACC,UAAU,CAAC,QAAQ,CAAC;;AAQlG,OAAO,eAAeE,SAAS,CAACH,GAAW,EAAEI,OAAO,EAAqB;EACvE,IAAI;IAEF,IAAI,CAACF,YAAY,CAACF,GAAG,CAAC,IAAI,CAACD,SAAS,CAACC,GAAG,CAAC,EAAE;MACzC,OAAO,MAAMF,aAAa,CAACE,GAAG,EAAEI,OAAO,CAAC;IAC1C;;IAIA,IAAIL,SAAS,CAACC,GAAG,CAAC,EAAE;MAClB,MAAM;QAACK,WAAW;QAAEC;MAAQ,CAAC,GAAGT,aAAa,CAACG,GAAG,CAAC;MAClD,MAAMO,QAAQ,GAAG,IAAIZ,QAAQ,CAACU,WAAW,EAAE;QACzCG,OAAO,EAAE;UAAC,cAAc,EAAEF;QAAQ,CAAC;QACnCN;MACF,CAAC,CAAC;MACF,OAAOO,QAAQ;IACjB;;IAGA,MAAME,wBAAwB,GAAG,CAAC,CAAC;IACnC,MAAMC,WAAW,GAAGV,GAAG;IACvB,IAAIA,GAAG,CAACW,QAAQ,CAAC,KAAK,CAAC,EAAE;MACvBX,GAAG,GAAGA,GAAG,CAACY,KAAK,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC;MACtBH,wBAAwB,CAAC,kBAAkB,CAAC,GAAG,MAAM;IACvD;;IAGA,MAAMI,IAAI,GAAG,MAAMC,2BAA2B,CAACJ,WAAW,EAAEN,OAAO,CAAC;IACpE,MAAMI,OAAO,GAAGO,UAAU,CAACf,GAAG,EAAEa,IAAI,EAAEJ,wBAAwB,CAAC;IAC/D,MAAM;MAACO,MAAM;MAAEC;IAAU,CAAC,GAAGC,SAAS,CAACL,IAAI,CAAC;IAE5C,MAAMM,cAAc,GAClB,CAACf,OAAO,IAAIA,OAAO,CAACe,cAAc,IAAIf,OAAO,CAACe,cAAc,KAAKC,SAAS;IAE5E,IAAIJ,MAAM,IAAI,GAAG,IAAIA,MAAM,GAAG,GAAG,IAAIR,OAAO,CAACa,GAAG,CAAC,UAAU,CAAC,IAAIF,cAAc,EAAE;MAC9E,MAAMG,WAAW,GAAGC,mBAAmB,CAACvB,GAAG,EAAEQ,OAAO,CAACgB,GAAG,CAAC,UAAU,CAAC,CAAC;;MAGrE,OAAO,MAAMrB,SAAS,CAACmB,WAAW,EAAElB,OAAO,CAAC;IAC9C;IACA,OAAO,IAAIT,QAAQ,CAACkB,IAAI,EAAE;MAACL,OAAO;MAAEQ,MAAM;MAAEC,UAAU;MAAEjB;IAAG,CAAC,CAAC;EAC/D,CAAC,CAAC,OAAOyB,KAAK,EAAE;IAEd,OAAO,IAAI9B,QAAQ,CAAC,IAAI,EAAE;MAACqB,MAAM,EAAE,GAAG;MAAEC,UAAU,EAAES,MAAM,CAACD,KAAK,CAAC;MAAEzB;IAAG,CAAC,CAAC;EAC1E;AACF;;AAGA,OAAO,eAAec,2BAA2B,CAC/Cd,GAAW,EACXI,OAAO,EACwB;EAG/B,OAAO,MAAM,IAAIuB,OAAO,CAAC,CAACC,OAAO,EAAEC,MAAM,KAAK;IAC5C,MAAMC,cAAc,GAAGC,iBAAiB,CAAC/B,GAAG,EAAEI,OAAO,CAAC;IACtD,MAAM4B,GAAG,GAAGhC,GAAG,CAACC,UAAU,CAAC,QAAQ,CAAC,GAChCP,KAAK,CAACuC,OAAO,CAACH,cAAc,EAAGI,GAAG,IAAKN,OAAO,CAACM,GAAG,CAAC,CAAC,GACpDzC,IAAI,CAACwC,OAAO,CAACH,cAAc,EAAGI,GAAG,IAAKN,OAAO,CAACM,GAAG,CAAC,CAAC;IACvDF,GAAG,CAACG,EAAE,CAAC,OAAO,EAAGV,KAAK,IAAKI,MAAM,CAACJ,KAAK,CAAC,CAAC;IACzCO,GAAG,CAACI,GAAG,EAAE;EACX,CAAC,CAAC;AACJ;;AAOA,SAASb,mBAAmB,CAACb,WAAmB,EAAE2B,QAAgB,EAAU;EAC1E,IAAIA,QAAQ,CAACpC,UAAU,CAAC,MAAM,CAAC,EAAE;IAC/B,OAAOoC,QAAQ;EACjB;EAEA,MAAMrC,GAAG,GAAG,IAAIsC,GAAG,CAAC5B,WAAW,CAAC;EAChCV,GAAG,CAACuC,QAAQ,GAAGF,QAAQ;EAEvB,OAAOrC,GAAG,CAACwC,IAAI;AACjB;;AAIA,SAAST,iBAAiB,CAAC/B,GAAW,EAAEI,OAA0C,EAAE;EAElF,MAAMqC,eAAe,GAAG,CAAArC,OAAO,aAAPA,OAAO,uBAAPA,OAAO,CAAEI,OAAO,KAAI,CAAC,CAAC;EAC9C,MAAMA,OAAO,GAAG,CAAC,CAAC;EAClB,KAAK,MAAMkC,GAAG,IAAIC,MAAM,CAACC,IAAI,CAACH,eAAe,CAAC,EAAE;IAC9CjC,OAAO,CAACkC,GAAG,CAACG,WAAW,EAAE,CAAC,GAAGJ,eAAe,CAACC,GAAG,CAAC;EACnD;;EAGAlC,OAAO,CAAC,iBAAiB,CAAC,GAAGA,OAAO,CAAC,iBAAiB,CAAC,IAAI,iBAAiB;EAE5E,MAAMsC,SAAS,GAAG,IAAIR,GAAG,CAACtC,GAAG,CAAC;EAC9B,OAAO;IACL+C,QAAQ,EAAED,SAAS,CAACC,QAAQ;IAC5BC,IAAI,EAAEF,SAAS,CAACP,QAAQ;IACxBU,MAAM,EAAE,KAAK;IAEb,GAAG7C,OAAO;IACV,IAAGA,OAAO,aAAPA,OAAO,uBAAPA,OAAO,CAAE8C,KAAK;IAEjB1C,OAAO;IACP2C,IAAI,EAAEL,SAAS,CAACK;EAClB,CAAC;AACH;AAEA,SAASjC,SAAS,CAACkC,YAAkC,EAAwC;EAC3F,IAAIA,YAAY,CAACC,UAAU,EAAE;IAC3B,OAAO;MAACrC,MAAM,EAAEoC,YAAY,CAACC,UAAU;MAAEpC,UAAU,EAAEmC,YAAY,CAACE,aAAa,IAAI;IAAI,CAAC;EAC1F;EACA,OAAO;IAACtC,MAAM,EAAE,GAAG;IAAEC,UAAU,EAAE;EAAI,CAAC;AACxC;AAEA,SAASF,UAAU,CAACf,GAAG,EAAEoD,YAAY,EAA0B;EAAA,IAAxBG,iBAAiB,uEAAG,CAAC,CAAC;EAC3D,MAAM/C,OAAO,GAAG,CAAC,CAAC;EAElB,IAAI4C,YAAY,IAAIA,YAAY,CAAC5C,OAAO,EAAE;IACxC,MAAMgD,WAAW,GAAGJ,YAAY,CAAC5C,OAAO;IACxC,KAAK,MAAMkC,GAAG,IAAIc,WAAW,EAAE;MAC7B,MAAMC,MAAM,GAAGD,WAAW,CAACd,GAAG,CAAC;MAC/BlC,OAAO,CAACkC,GAAG,CAACG,WAAW,EAAE,CAAC,GAAGnB,MAAM,CAAC+B,MAAM,CAAC;IAC7C;EACF;;EAGA,IAAI,CAACjD,OAAO,CAAC,gBAAgB,CAAC,EAAE;IAC9B,MAAMkD,aAAa,GAAGC,gBAAgB,CAAC3D,GAAG,CAAC;IAC3C,IAAI4D,MAAM,CAACC,QAAQ,CAACH,aAAa,CAAC,EAAE;MAClClD,OAAO,CAAC,gBAAgB,CAAC,GAAGkD,aAAa;IAC3C;EACF;EAEAf,MAAM,CAACmB,MAAM,CAACtD,OAAO,EAAE+C,iBAAiB,CAAC;EAEzC,OAAO,IAAI3D,OAAO,CAACY,OAAO,CAAC;AAC7B;;AAGA,SAASmD,gBAAgB,CAAC3D,GAAW,EAAiB;EAEpD,OAAOD,SAAS,CAACC,GAAG,CAAC,GAAGA,GAAG,CAAC+D,MAAM,GAAG,OAAO,CAACA,MAAM,GAAG,IAAI;AAC5D"}
|