@loaders.gl/polyfills 3.1.3 → 4.0.0-alpha.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (142) hide show
  1. package/dist/bundle.js +2 -2
  2. package/dist/bundle.js.map +1 -0
  3. package/dist/index.js +49 -84
  4. package/dist/index.js.map +1 -0
  5. package/dist/lib/encoding-indexes.js +29 -34
  6. package/dist/lib/encoding-indexes.js.map +1 -0
  7. package/dist/lib/encoding.js +1281 -2610
  8. package/dist/lib/encoding.js.map +1 -0
  9. package/dist/libs/encoding-indexes-asian.js +8 -9
  10. package/dist/node/buffer/btoa.node.js +5 -12
  11. package/dist/node/buffer/btoa.node.js.map +1 -0
  12. package/dist/node/buffer/to-array-buffer.node.js +8 -11
  13. package/dist/node/buffer/to-array-buffer.node.js.map +1 -0
  14. package/dist/node/fetch/fetch.node.js +112 -103
  15. package/dist/node/fetch/fetch.node.js.map +1 -0
  16. package/dist/node/fetch/headers.node.js +114 -94
  17. package/dist/node/fetch/headers.node.js.map +1 -0
  18. package/dist/node/fetch/response.node.js +84 -73
  19. package/dist/node/fetch/response.node.js.map +1 -0
  20. package/dist/node/fetch/utils/decode-data-uri.node.js +53 -63
  21. package/dist/node/fetch/utils/decode-data-uri.node.js.map +1 -0
  22. package/dist/node/fetch/utils/stream-utils.node.js +87 -105
  23. package/dist/node/fetch/utils/stream-utils.node.js.map +1 -0
  24. package/dist/node/file/blob-stream-controller.js +52 -59
  25. package/dist/node/file/blob-stream-controller.js.map +1 -0
  26. package/dist/node/file/blob-stream.js +25 -36
  27. package/dist/node/file/blob-stream.js.map +1 -0
  28. package/dist/node/file/blob.js +131 -151
  29. package/dist/node/file/blob.js.map +1 -0
  30. package/dist/node/file/file-reader.js +77 -28
  31. package/dist/node/file/file-reader.js.map +1 -0
  32. package/dist/node/file/file.js +25 -36
  33. package/dist/node/file/file.js.map +1 -0
  34. package/dist/node/file/install-file-polyfills.js +21 -26
  35. package/dist/node/file/install-file-polyfills.js.map +1 -0
  36. package/dist/node/file/readable-stream.js +3 -10
  37. package/dist/node/file/readable-stream.js.map +1 -0
  38. package/dist/node/images/encode-image.node.js +17 -38
  39. package/dist/node/images/encode-image.node.js.map +1 -0
  40. package/dist/node/images/parse-image.node.js +19 -28
  41. package/dist/node/images/parse-image.node.js.map +1 -0
  42. package/dist/promise/all-settled.js +17 -22
  43. package/dist/promise/all-settled.js.map +1 -0
  44. package/dist/utils/assert.js +5 -8
  45. package/dist/utils/assert.js.map +1 -0
  46. package/dist/utils/globals.js +7 -34
  47. package/dist/utils/globals.js.map +1 -0
  48. package/package.json +4 -4
  49. package/dist/es5/bundle.js +0 -7
  50. package/dist/es5/bundle.js.map +0 -1
  51. package/dist/es5/index.js +0 -122
  52. package/dist/es5/index.js.map +0 -1
  53. package/dist/es5/lib/encoding-indexes.js +0 -37
  54. package/dist/es5/lib/encoding-indexes.js.map +0 -1
  55. package/dist/es5/lib/encoding.js +0 -1465
  56. package/dist/es5/lib/encoding.js.map +0 -1
  57. package/dist/es5/libs/encoding-indexes-asian.js +0 -13
  58. package/dist/es5/node/buffer/btoa.node.js +0 -16
  59. package/dist/es5/node/buffer/btoa.node.js.map +0 -1
  60. package/dist/es5/node/buffer/to-array-buffer.node.js +0 -16
  61. package/dist/es5/node/buffer/to-array-buffer.node.js.map +0 -1
  62. package/dist/es5/node/fetch/fetch.node.js +0 -183
  63. package/dist/es5/node/fetch/fetch.node.js.map +0 -1
  64. package/dist/es5/node/fetch/headers.node.js +0 -175
  65. package/dist/es5/node/fetch/headers.node.js.map +0 -1
  66. package/dist/es5/node/fetch/response.node.js +0 -231
  67. package/dist/es5/node/fetch/response.node.js.map +0 -1
  68. package/dist/es5/node/fetch/utils/decode-data-uri.node.js +0 -72
  69. package/dist/es5/node/fetch/utils/decode-data-uri.node.js.map +0 -1
  70. package/dist/es5/node/fetch/utils/stream-utils.node.js +0 -213
  71. package/dist/es5/node/fetch/utils/stream-utils.node.js.map +0 -1
  72. package/dist/es5/node/file/blob-stream-controller.js +0 -110
  73. package/dist/es5/node/file/blob-stream-controller.js.map +0 -1
  74. package/dist/es5/node/file/blob-stream.js +0 -93
  75. package/dist/es5/node/file/blob-stream.js.map +0 -1
  76. package/dist/es5/node/file/blob.js +0 -259
  77. package/dist/es5/node/file/blob.js.map +0 -1
  78. package/dist/es5/node/file/file-reader.js +0 -189
  79. package/dist/es5/node/file/file-reader.js.map +0 -1
  80. package/dist/es5/node/file/file.js +0 -63
  81. package/dist/es5/node/file/file.js.map +0 -1
  82. package/dist/es5/node/file/install-file-polyfills.js +0 -33
  83. package/dist/es5/node/file/install-file-polyfills.js.map +0 -1
  84. package/dist/es5/node/file/readable-stream.js +0 -40
  85. package/dist/es5/node/file/readable-stream.js.map +0 -1
  86. package/dist/es5/node/images/encode-image.node.js +0 -36
  87. package/dist/es5/node/images/encode-image.node.js.map +0 -1
  88. package/dist/es5/node/images/parse-image.node.js +0 -61
  89. package/dist/es5/node/images/parse-image.node.js.map +0 -1
  90. package/dist/es5/promise/all-settled.js +0 -29
  91. package/dist/es5/promise/all-settled.js.map +0 -1
  92. package/dist/es5/utils/assert.js +0 -13
  93. package/dist/es5/utils/assert.js.map +0 -1
  94. package/dist/es5/utils/globals.js +0 -21
  95. package/dist/es5/utils/globals.js.map +0 -1
  96. package/dist/esm/bundle.js +0 -5
  97. package/dist/esm/bundle.js.map +0 -1
  98. package/dist/esm/index.js +0 -63
  99. package/dist/esm/index.js.map +0 -1
  100. package/dist/esm/lib/encoding-indexes.js +0 -30
  101. package/dist/esm/lib/encoding-indexes.js.map +0 -1
  102. package/dist/esm/lib/encoding.js +0 -1450
  103. package/dist/esm/lib/encoding.js.map +0 -1
  104. package/dist/esm/libs/encoding-indexes-asian.js +0 -13
  105. package/dist/esm/node/buffer/btoa.node.js +0 -7
  106. package/dist/esm/node/buffer/btoa.node.js.map +0 -1
  107. package/dist/esm/node/buffer/to-array-buffer.node.js +0 -9
  108. package/dist/esm/node/buffer/to-array-buffer.node.js.map +0 -1
  109. package/dist/esm/node/fetch/fetch.node.js +0 -126
  110. package/dist/esm/node/fetch/fetch.node.js.map +0 -1
  111. package/dist/esm/node/fetch/headers.node.js +0 -125
  112. package/dist/esm/node/fetch/headers.node.js.map +0 -1
  113. package/dist/esm/node/fetch/response.node.js +0 -88
  114. package/dist/esm/node/fetch/response.node.js.map +0 -1
  115. package/dist/esm/node/fetch/utils/decode-data-uri.node.js +0 -55
  116. package/dist/esm/node/fetch/utils/decode-data-uri.node.js.map +0 -1
  117. package/dist/esm/node/fetch/utils/stream-utils.node.js +0 -98
  118. package/dist/esm/node/fetch/utils/stream-utils.node.js.map +0 -1
  119. package/dist/esm/node/file/blob-stream-controller.js +0 -56
  120. package/dist/esm/node/file/blob-stream-controller.js.map +0 -1
  121. package/dist/esm/node/file/blob-stream.js +0 -26
  122. package/dist/esm/node/file/blob-stream.js.map +0 -1
  123. package/dist/esm/node/file/blob.js +0 -140
  124. package/dist/esm/node/file/blob.js.map +0 -1
  125. package/dist/esm/node/file/file-reader.js +0 -84
  126. package/dist/esm/node/file/file-reader.js.map +0 -1
  127. package/dist/esm/node/file/file.js +0 -26
  128. package/dist/esm/node/file/file.js.map +0 -1
  129. package/dist/esm/node/file/install-file-polyfills.js +0 -22
  130. package/dist/esm/node/file/install-file-polyfills.js.map +0 -1
  131. package/dist/esm/node/file/readable-stream.js +0 -4
  132. package/dist/esm/node/file/readable-stream.js.map +0 -1
  133. package/dist/esm/node/images/encode-image.node.js +0 -20
  134. package/dist/esm/node/images/encode-image.node.js.map +0 -1
  135. package/dist/esm/node/images/parse-image.node.js +0 -20
  136. package/dist/esm/node/images/parse-image.node.js.map +0 -1
  137. package/dist/esm/promise/all-settled.js +0 -19
  138. package/dist/esm/promise/all-settled.js.map +0 -1
  139. package/dist/esm/utils/assert.js +0 -6
  140. package/dist/esm/utils/assert.js.map +0 -1
  141. package/dist/esm/utils/globals.js +0 -9
  142. package/dist/esm/utils/globals.js.map +0 -1
@@ -1,77 +1,88 @@
1
- "use strict";
2
- var __importDefault = (this && this.__importDefault) || function (mod) {
3
- return (mod && mod.__esModule) ? mod : { "default": mod };
4
- };
5
- Object.defineProperty(exports, "__esModule", { value: true });
6
- const assert_1 = require("../../utils/assert");
7
- const stream_utils_node_1 = require("./utils/stream-utils.node");
8
- const headers_node_1 = __importDefault(require("./headers.node"));
9
- const isBoolean = (x) => typeof x === 'boolean';
10
- const isFunction = (x) => typeof x === 'function';
11
- const isObject = (x) => x !== null && typeof x === 'object';
12
- const isReadableNodeStream = (x) => isObject(x) && isFunction(x.read) && isFunction(x.pipe) && isBoolean(x.readable);
13
- /**
14
- * Polyfill for Browser Response
15
- *
16
- * Under Node.js we return a mock "fetch response object"
17
- * so that apps can use the same API as in the browser.
18
- *
19
- * Note: This is intended to be a "lightweight" implementation and will have limitations.
20
- *
21
- * See https://developer.mozilla.org/en-US/docs/Web/API/Response
22
- */
23
- const stream_1 = require("stream");
24
- class Response {
25
- // TODO - handle ArrayBuffer, ArrayBufferView, Buffer
26
- constructor(body, options) {
27
- this.bodyUsed = false;
28
- const { headers, status = 200, statusText = 'OK', url } = options || {};
29
- this.url = url;
30
- this.ok = status === 200;
31
- this.status = status; // TODO - handle errors and set status
32
- this.statusText = statusText;
33
- this.headers = new headers_node_1.default(options?.headers || {});
34
- // Check for content-encoding and create a decompression stream
35
- if (isReadableNodeStream(body)) {
36
- this._body = (0, stream_utils_node_1.decompressReadStream)(body, headers);
37
- }
38
- else if (typeof body === 'string') {
39
- this._body = stream_1.Readable.from([new TextEncoder().encode(body)]);
40
- }
41
- else {
42
- this._body = stream_1.Readable.from([body || new ArrayBuffer(0)]);
43
- }
1
+ import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
2
+ import { assert } from '../../utils/assert';
3
+ import { decompressReadStream, concatenateReadStream } from './utils/stream-utils.node';
4
+ import Headers from './headers.node';
5
+
6
+ const isBoolean = x => typeof x === 'boolean';
7
+
8
+ const isFunction = x => typeof x === 'function';
9
+
10
+ const isObject = x => x !== null && typeof x === 'object';
11
+
12
+ const isReadableNodeStream = x => isObject(x) && isFunction(x.read) && isFunction(x.pipe) && isBoolean(x.readable);
13
+
14
+ import { Readable } from 'stream';
15
+ export default class Response {
16
+ constructor(body, options) {
17
+ _defineProperty(this, "ok", void 0);
18
+
19
+ _defineProperty(this, "status", void 0);
20
+
21
+ _defineProperty(this, "statusText", void 0);
22
+
23
+ _defineProperty(this, "headers", void 0);
24
+
25
+ _defineProperty(this, "url", void 0);
26
+
27
+ _defineProperty(this, "bodyUsed", false);
28
+
29
+ _defineProperty(this, "_body", void 0);
30
+
31
+ const {
32
+ headers,
33
+ status = 200,
34
+ statusText = 'OK',
35
+ url
36
+ } = options || {};
37
+ this.url = url;
38
+ this.ok = status === 200;
39
+ this.status = status;
40
+ this.statusText = statusText;
41
+ this.headers = new Headers((options === null || options === void 0 ? void 0 : options.headers) || {});
42
+
43
+ if (isReadableNodeStream(body)) {
44
+ this._body = decompressReadStream(body, headers);
45
+ } else if (typeof body === 'string') {
46
+ this._body = Readable.from([new TextEncoder().encode(body)]);
47
+ } else {
48
+ this._body = Readable.from([body || new ArrayBuffer(0)]);
44
49
  }
45
- // Subset of Properties
46
- // Returns a readable stream to the "body" of the response (or file)
47
- get body() {
48
- (0, assert_1.assert)(!this.bodyUsed);
49
- (0, assert_1.assert)(isReadableNodeStream(this._body)); // Not implemented: conversion of ArrayBuffer etc to stream
50
- this.bodyUsed = true;
51
- return this._body;
50
+ }
51
+
52
+ get body() {
53
+ assert(!this.bodyUsed);
54
+ assert(isReadableNodeStream(this._body));
55
+ this.bodyUsed = true;
56
+ return this._body;
57
+ }
58
+
59
+ async arrayBuffer() {
60
+ if (!isReadableNodeStream(this._body)) {
61
+ return this._body || new ArrayBuffer(0);
52
62
  }
53
- // Subset of Methods
54
- async arrayBuffer() {
55
- if (!isReadableNodeStream(this._body)) {
56
- return this._body || new ArrayBuffer(0);
57
- }
58
- const data = await (0, stream_utils_node_1.concatenateReadStream)(this._body);
59
- return data;
60
- }
61
- async text() {
62
- const arrayBuffer = await this.arrayBuffer();
63
- const textDecoder = new TextDecoder();
64
- return textDecoder.decode(arrayBuffer);
65
- }
66
- async json() {
67
- const text = await this.text();
68
- return JSON.parse(text);
69
- }
70
- async blob() {
71
- if (typeof Blob === 'undefined') {
72
- throw new Error('Blob polyfill not installed');
73
- }
74
- return new Blob([await this.arrayBuffer()]);
63
+
64
+ const data = await concatenateReadStream(this._body);
65
+ return data;
66
+ }
67
+
68
+ async text() {
69
+ const arrayBuffer = await this.arrayBuffer();
70
+ const textDecoder = new TextDecoder();
71
+ return textDecoder.decode(arrayBuffer);
72
+ }
73
+
74
+ async json() {
75
+ const text = await this.text();
76
+ return JSON.parse(text);
77
+ }
78
+
79
+ async blob() {
80
+ if (typeof Blob === 'undefined') {
81
+ throw new Error('Blob polyfill not installed');
75
82
  }
83
+
84
+ return new Blob([await this.arrayBuffer()]);
85
+ }
86
+
76
87
  }
77
- exports.default = Response;
88
+ //# sourceMappingURL=response.node.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../../src/node/fetch/response.node.ts"],"names":["assert","decompressReadStream","concatenateReadStream","Headers","isBoolean","x","isFunction","isObject","isReadableNodeStream","read","pipe","readable","Readable","Response","constructor","body","options","headers","status","statusText","url","ok","_body","from","TextEncoder","encode","ArrayBuffer","bodyUsed","arrayBuffer","data","text","textDecoder","TextDecoder","decode","json","JSON","parse","blob","Blob","Error"],"mappings":";AAAA,SAAQA,MAAR,QAAqB,oBAArB;AACA,SAAQC,oBAAR,EAA8BC,qBAA9B,QAA0D,2BAA1D;AACA,OAAOC,OAAP,MAAoB,gBAApB;;AAEA,MAAMC,SAAS,GAAIC,CAAD,IAAO,OAAOA,CAAP,KAAa,SAAtC;;AACA,MAAMC,UAAU,GAAID,CAAD,IAAO,OAAOA,CAAP,KAAa,UAAvC;;AACA,MAAME,QAAQ,GAAIF,CAAD,IAAOA,CAAC,KAAK,IAAN,IAAc,OAAOA,CAAP,KAAa,QAAnD;;AACA,MAAMG,oBAAoB,GAAIH,CAAD,IAC3BE,QAAQ,CAACF,CAAD,CAAR,IAAeC,UAAU,CAACD,CAAC,CAACI,IAAH,CAAzB,IAAqCH,UAAU,CAACD,CAAC,CAACK,IAAH,CAA/C,IAA2DN,SAAS,CAACC,CAAC,CAACM,QAAH,CADtE;;AAaA,SAAQC,QAAR,QAAuB,QAAvB;AAEA,eAAe,MAAMC,QAAN,CAAe;AAU5BC,EAAAA,WAAW,CACTC,IADS,EAETC,OAFS,EAQT;AAAA;;AAAA;;AAAA;;AAAA;;AAAA;;AAAA,sCAZkB,KAYlB;;AAAA;;AACA,UAAM;AAACC,MAAAA,OAAD;AAAUC,MAAAA,MAAM,GAAG,GAAnB;AAAwBC,MAAAA,UAAU,GAAG,IAArC;AAA2CC,MAAAA;AAA3C,QAAkDJ,OAAO,IAAI,EAAnE;AAEA,SAAKI,GAAL,GAAWA,GAAX;AACA,SAAKC,EAAL,GAAUH,MAAM,KAAK,GAArB;AACA,SAAKA,MAAL,GAAcA,MAAd;AACA,SAAKC,UAAL,GAAkBA,UAAlB;AACA,SAAKF,OAAL,GAAe,IAAId,OAAJ,CAAY,CAAAa,OAAO,SAAP,IAAAA,OAAO,WAAP,YAAAA,OAAO,CAAEC,OAAT,KAAoB,EAAhC,CAAf;;AAGA,QAAIT,oBAAoB,CAACO,IAAD,CAAxB,EAAgC;AAC9B,WAAKO,KAAL,GAAarB,oBAAoB,CAACc,IAAD,EAAOE,OAAP,CAAjC;AACD,KAFD,MAEO,IAAI,OAAOF,IAAP,KAAgB,QAApB,EAA8B;AACnC,WAAKO,KAAL,GAAaV,QAAQ,CAACW,IAAT,CAAc,CAAC,IAAIC,WAAJ,GAAkBC,MAAlB,CAAyBV,IAAzB,CAAD,CAAd,CAAb;AACD,KAFM,MAEA;AACL,WAAKO,KAAL,GAAaV,QAAQ,CAACW,IAAT,CAAc,CAACR,IAAI,IAAI,IAAIW,WAAJ,CAAgB,CAAhB,CAAT,CAAd,CAAb;AACD;AACF;;AAKO,MAAJX,IAAI,GAAG;AACTf,IAAAA,MAAM,CAAC,CAAC,KAAK2B,QAAP,CAAN;AACA3B,IAAAA,MAAM,CAACQ,oBAAoB,CAAC,KAAKc,KAAN,CAArB,CAAN;AACA,SAAKK,QAAL,GAAgB,IAAhB;AACA,WAAO,KAAKL,KAAZ;AACD;;AAIgB,QAAXM,WAAW,GAAG;AAClB,QAAI,CAACpB,oBAAoB,CAAC,KAAKc,KAAN,CAAzB,EAAuC;AACrC,aAAO,KAAKA,KAAL,IAAc,IAAII,WAAJ,CAAgB,CAAhB,CAArB;AACD;;AACD,UAAMG,IAAI,GAAG,MAAM3B,qBAAqB,CAAC,KAAKoB,KAAN,CAAxC;AACA,WAAOO,IAAP;AACD;;AAES,QAAJC,IAAI,GAAG;AACX,UAAMF,WAAW,GAAG,MAAM,KAAKA,WAAL,EAA1B;AACA,UAAMG,WAAW,GAAG,IAAIC,WAAJ,EAApB;AACA,WAAOD,WAAW,CAACE,MAAZ,CAAmBL,WAAnB,CAAP;AACD;;AAES,QAAJM,IAAI,GAAG;AACX,UAAMJ,IAAI,GAAG,MAAM,KAAKA,IAAL,EAAnB;AACA,WAAOK,IAAI,CAACC,KAAL,CAAWN,IAAX,CAAP;AACD;;AAES,QAAJO,IAAI,GAAG;AACX,QAAI,OAAOC,IAAP,KAAgB,WAApB,EAAiC;AAC/B,YAAM,IAAIC,KAAJ,CAAU,6BAAV,CAAN;AACD;;AACD,WAAO,IAAID,IAAJ,CAAS,CAAC,MAAM,KAAKV,WAAL,EAAP,CAAT,CAAP;AACD;;AAzE2B","sourcesContent":["import {assert} from '../../utils/assert';\nimport {decompressReadStream, concatenateReadStream} from './utils/stream-utils.node';\nimport Headers from './headers.node';\n\nconst isBoolean = (x) => typeof x === 'boolean';\nconst isFunction = (x) => typeof x === 'function';\nconst isObject = (x) => x !== null && typeof x === 'object';\nconst isReadableNodeStream = (x) =>\n isObject(x) && isFunction(x.read) && isFunction(x.pipe) && isBoolean(x.readable);\n\n/**\n * Polyfill for Browser Response\n *\n * Under Node.js we return a mock \"fetch response object\"\n * so that apps can use the same API as in the browser.\n *\n * Note: This is intended to be a \"lightweight\" implementation and will have limitations.\n *\n * See https://developer.mozilla.org/en-US/docs/Web/API/Response\n */\nimport {Readable} from 'stream';\n\nexport default class Response {\n readonly ok: boolean;\n readonly status: number;\n readonly statusText: string;\n readonly headers: Headers;\n readonly url: string;\n bodyUsed: boolean = false;\n private readonly _body;\n\n // TODO - handle ArrayBuffer, ArrayBufferView, Buffer\n constructor(\n body,\n options: {\n headers?;\n status?: number;\n statusText?: string;\n url: string;\n }\n ) {\n const {headers, status = 200, statusText = 'OK', url} = options || {};\n\n this.url = url;\n this.ok = status === 200;\n this.status = status; // TODO - handle errors and set status\n this.statusText = statusText;\n this.headers = new Headers(options?.headers || {});\n\n // Check for content-encoding and create a decompression stream\n if (isReadableNodeStream(body)) {\n this._body = decompressReadStream(body, headers);\n } else if (typeof body === 'string') {\n this._body = Readable.from([new TextEncoder().encode(body)]);\n } else {\n this._body = Readable.from([body || new ArrayBuffer(0)]);\n }\n }\n\n // Subset of Properties\n\n // Returns a readable stream to the \"body\" of the response (or file)\n get body() {\n assert(!this.bodyUsed);\n assert(isReadableNodeStream(this._body)); // Not implemented: conversion of ArrayBuffer etc to stream\n this.bodyUsed = true;\n return this._body;\n }\n\n // Subset of Methods\n\n async arrayBuffer() {\n if (!isReadableNodeStream(this._body)) {\n return this._body || new ArrayBuffer(0);\n }\n const data = await concatenateReadStream(this._body);\n return data;\n }\n\n async text() {\n const arrayBuffer = await this.arrayBuffer();\n const textDecoder = new TextDecoder();\n return textDecoder.decode(arrayBuffer);\n }\n\n async json() {\n const text = await this.text();\n return JSON.parse(text);\n }\n\n async blob() {\n if (typeof Blob === 'undefined') {\n throw new Error('Blob polyfill not installed');\n }\n return new Blob([await this.arrayBuffer()]);\n }\n}\n"],"file":"response.node.js"}
@@ -1,65 +1,55 @@
1
- "use strict";
2
- // Based on binary-gltf-utils under MIT license: Copyright (c) 2016-17 Karl Cheng
3
- Object.defineProperty(exports, "__esModule", { value: true });
4
- exports.toArrayBuffer = exports.decodeDataUri = void 0;
5
- const isArrayBuffer = (x) => x && x instanceof ArrayBuffer;
6
- const isBuffer = (x) => x && x instanceof Buffer;
7
- /**
8
- * Parses a data URI into a buffer, as well as retrieving its declared MIME type.
9
- *
10
- * @param {string} uri - a data URI (assumed to be valid)
11
- * @returns {Object} { buffer, mimeType }
12
- */
13
- function decodeDataUri(uri) {
14
- const dataIndex = uri.indexOf(',');
15
- let buffer;
16
- let mimeType;
17
- if (uri.slice(dataIndex - 7, dataIndex) === ';base64') {
18
- buffer = Buffer.from(uri.slice(dataIndex + 1), 'base64');
19
- mimeType = uri.slice(5, dataIndex - 7).trim();
20
- }
21
- else {
22
- buffer = Buffer.from(decodeURIComponent(uri.slice(dataIndex + 1)));
23
- mimeType = uri.slice(5, dataIndex).trim();
24
- }
25
- if (!mimeType) {
26
- mimeType = 'text/plain;charset=US-ASCII';
27
- }
28
- else if (mimeType.startsWith(';')) {
29
- mimeType = `text/plain${mimeType}`;
30
- }
31
- return { arrayBuffer: toArrayBuffer(buffer), mimeType };
1
+ const isArrayBuffer = x => x && x instanceof ArrayBuffer;
2
+
3
+ const isBuffer = x => x && x instanceof Buffer;
4
+
5
+ export function decodeDataUri(uri) {
6
+ const dataIndex = uri.indexOf(',');
7
+ let buffer;
8
+ let mimeType;
9
+
10
+ if (uri.slice(dataIndex - 7, dataIndex) === ';base64') {
11
+ buffer = Buffer.from(uri.slice(dataIndex + 1), 'base64');
12
+ mimeType = uri.slice(5, dataIndex - 7).trim();
13
+ } else {
14
+ buffer = Buffer.from(decodeURIComponent(uri.slice(dataIndex + 1)));
15
+ mimeType = uri.slice(5, dataIndex).trim();
16
+ }
17
+
18
+ if (!mimeType) {
19
+ mimeType = 'text/plain;charset=US-ASCII';
20
+ } else if (mimeType.startsWith(';')) {
21
+ mimeType = "text/plain".concat(mimeType);
22
+ }
23
+
24
+ return {
25
+ arrayBuffer: toArrayBuffer(buffer),
26
+ mimeType
27
+ };
32
28
  }
33
- exports.decodeDataUri = decodeDataUri;
34
- /**
35
- * @param data
36
- * @todo Duplicate of core
37
- */
38
- function toArrayBuffer(data) {
39
- if (isArrayBuffer(data)) {
40
- return data;
41
- }
42
- // TODO - per docs we should just be able to call buffer.buffer, but there are issues
43
- if (isBuffer(data)) {
44
- // @ts-expect-error
45
- const typedArray = new Uint8Array(data);
46
- return typedArray.buffer;
47
- }
48
- // Careful - Node Buffers will look like ArrayBuffers (keep after isBuffer)
49
- if (ArrayBuffer.isView(data)) {
50
- return data.buffer;
51
- }
52
- if (typeof data === 'string') {
53
- const text = data;
54
- const uint8Array = new TextEncoder().encode(text);
55
- return uint8Array.buffer;
56
- }
57
- // HACK to support Blob polyfill
58
- // @ts-expect-error
59
- if (data && typeof data === 'object' && data._toArrayBuffer) {
60
- // @ts-expect-error
61
- return data._toArrayBuffer();
62
- }
63
- throw new Error(`toArrayBuffer(${JSON.stringify(data, null, 2).slice(10)})`);
29
+ export function toArrayBuffer(data) {
30
+ if (isArrayBuffer(data)) {
31
+ return data;
32
+ }
33
+
34
+ if (isBuffer(data)) {
35
+ const typedArray = new Uint8Array(data);
36
+ return typedArray.buffer;
37
+ }
38
+
39
+ if (ArrayBuffer.isView(data)) {
40
+ return data.buffer;
41
+ }
42
+
43
+ if (typeof data === 'string') {
44
+ const text = data;
45
+ const uint8Array = new TextEncoder().encode(text);
46
+ return uint8Array.buffer;
47
+ }
48
+
49
+ if (data && typeof data === 'object' && data._toArrayBuffer) {
50
+ return data._toArrayBuffer();
51
+ }
52
+
53
+ throw new Error("toArrayBuffer(".concat(JSON.stringify(data, null, 2).slice(10), ")"));
64
54
  }
65
- exports.toArrayBuffer = toArrayBuffer;
55
+ //# sourceMappingURL=decode-data-uri.node.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../../../src/node/fetch/utils/decode-data-uri.node.ts"],"names":["isArrayBuffer","x","ArrayBuffer","isBuffer","Buffer","decodeDataUri","uri","dataIndex","indexOf","buffer","mimeType","slice","from","trim","decodeURIComponent","startsWith","arrayBuffer","toArrayBuffer","data","typedArray","Uint8Array","isView","text","uint8Array","TextEncoder","encode","_toArrayBuffer","Error","JSON","stringify"],"mappings":"AAEA,MAAMA,aAAa,GAAIC,CAAD,IAAOA,CAAC,IAAIA,CAAC,YAAYC,WAA/C;;AACA,MAAMC,QAAQ,GAAIF,CAAD,IAAOA,CAAC,IAAIA,CAAC,YAAYG,MAA1C;;AAQA,OAAO,SAASC,aAAT,CAAuBC,GAAvB,EAAkF;AACvF,QAAMC,SAAS,GAAGD,GAAG,CAACE,OAAJ,CAAY,GAAZ,CAAlB;AAEA,MAAIC,MAAJ;AACA,MAAIC,QAAJ;;AACA,MAAIJ,GAAG,CAACK,KAAJ,CAAUJ,SAAS,GAAG,CAAtB,EAAyBA,SAAzB,MAAwC,SAA5C,EAAuD;AACrDE,IAAAA,MAAM,GAAGL,MAAM,CAACQ,IAAP,CAAYN,GAAG,CAACK,KAAJ,CAAUJ,SAAS,GAAG,CAAtB,CAAZ,EAAsC,QAAtC,CAAT;AACAG,IAAAA,QAAQ,GAAGJ,GAAG,CAACK,KAAJ,CAAU,CAAV,EAAaJ,SAAS,GAAG,CAAzB,EAA4BM,IAA5B,EAAX;AACD,GAHD,MAGO;AACLJ,IAAAA,MAAM,GAAGL,MAAM,CAACQ,IAAP,CAAYE,kBAAkB,CAACR,GAAG,CAACK,KAAJ,CAAUJ,SAAS,GAAG,CAAtB,CAAD,CAA9B,CAAT;AACAG,IAAAA,QAAQ,GAAGJ,GAAG,CAACK,KAAJ,CAAU,CAAV,EAAaJ,SAAb,EAAwBM,IAAxB,EAAX;AACD;;AAED,MAAI,CAACH,QAAL,EAAe;AACbA,IAAAA,QAAQ,GAAG,6BAAX;AACD,GAFD,MAEO,IAAIA,QAAQ,CAACK,UAAT,CAAoB,GAApB,CAAJ,EAA8B;AACnCL,IAAAA,QAAQ,uBAAgBA,QAAhB,CAAR;AACD;;AAED,SAAO;AAACM,IAAAA,WAAW,EAAEC,aAAa,CAACR,MAAD,CAA3B;AAAqCC,IAAAA;AAArC,GAAP;AACD;AAMD,OAAO,SAASO,aAAT,CAAuBC,IAAvB,EAAmD;AACxD,MAAIlB,aAAa,CAACkB,IAAD,CAAjB,EAAyB;AACvB,WAAOA,IAAP;AACD;;AAGD,MAAIf,QAAQ,CAACe,IAAD,CAAZ,EAAoB;AAElB,UAAMC,UAAU,GAAG,IAAIC,UAAJ,CAAeF,IAAf,CAAnB;AACA,WAAOC,UAAU,CAACV,MAAlB;AACD;;AAGD,MAAIP,WAAW,CAACmB,MAAZ,CAAmBH,IAAnB,CAAJ,EAA8B;AAC5B,WAAOA,IAAI,CAACT,MAAZ;AACD;;AAED,MAAI,OAAOS,IAAP,KAAgB,QAApB,EAA8B;AAC5B,UAAMI,IAAI,GAAGJ,IAAb;AACA,UAAMK,UAAU,GAAG,IAAIC,WAAJ,GAAkBC,MAAlB,CAAyBH,IAAzB,CAAnB;AACA,WAAOC,UAAU,CAACd,MAAlB;AACD;;AAID,MAAIS,IAAI,IAAI,OAAOA,IAAP,KAAgB,QAAxB,IAAoCA,IAAI,CAACQ,cAA7C,EAA6D;AAE3D,WAAOR,IAAI,CAACQ,cAAL,EAAP;AACD;;AAED,QAAM,IAAIC,KAAJ,yBAA2BC,IAAI,CAACC,SAAL,CAAeX,IAAf,EAAqB,IAArB,EAA2B,CAA3B,EAA8BP,KAA9B,CAAoC,EAApC,CAA3B,OAAN;AACD","sourcesContent":["// Based on binary-gltf-utils under MIT license: Copyright (c) 2016-17 Karl Cheng\n\nconst isArrayBuffer = (x) => x && x instanceof ArrayBuffer;\nconst isBuffer = (x) => x && x instanceof Buffer;\n\n/**\n * Parses a data URI into a buffer, as well as retrieving its declared MIME type.\n *\n * @param {string} uri - a data URI (assumed to be valid)\n * @returns {Object} { buffer, mimeType }\n */\nexport function decodeDataUri(uri: string): {arrayBuffer: ArrayBuffer; mimeType: string} {\n const dataIndex = uri.indexOf(',');\n\n let buffer;\n let mimeType;\n if (uri.slice(dataIndex - 7, dataIndex) === ';base64') {\n buffer = Buffer.from(uri.slice(dataIndex + 1), 'base64');\n mimeType = uri.slice(5, dataIndex - 7).trim();\n } else {\n buffer = Buffer.from(decodeURIComponent(uri.slice(dataIndex + 1)));\n mimeType = uri.slice(5, dataIndex).trim();\n }\n\n if (!mimeType) {\n mimeType = 'text/plain;charset=US-ASCII';\n } else if (mimeType.startsWith(';')) {\n mimeType = `text/plain${mimeType}`;\n }\n\n return {arrayBuffer: toArrayBuffer(buffer), mimeType};\n}\n\n/**\n * @param data\n * @todo Duplicate of core\n */\nexport function toArrayBuffer(data: unknown): ArrayBuffer {\n if (isArrayBuffer(data)) {\n return data as ArrayBuffer;\n }\n\n // TODO - per docs we should just be able to call buffer.buffer, but there are issues\n if (isBuffer(data)) {\n // @ts-expect-error\n const typedArray = new Uint8Array(data);\n return typedArray.buffer;\n }\n\n // Careful - Node Buffers will look like ArrayBuffers (keep after isBuffer)\n if (ArrayBuffer.isView(data)) {\n return data.buffer;\n }\n\n if (typeof data === 'string') {\n const text = data;\n const uint8Array = new TextEncoder().encode(text);\n return uint8Array.buffer;\n }\n\n // HACK to support Blob polyfill\n // @ts-expect-error\n if (data && typeof data === 'object' && data._toArrayBuffer) {\n // @ts-expect-error\n return data._toArrayBuffer();\n }\n\n throw new Error(`toArrayBuffer(${JSON.stringify(data, null, 2).slice(10)})`);\n}\n"],"file":"decode-data-uri.node.js"}
@@ -1,116 +1,98 @@
1
- "use strict";
2
- var __importDefault = (this && this.__importDefault) || function (mod) {
3
- return (mod && mod.__esModule) ? mod : { "default": mod };
4
- };
5
- Object.defineProperty(exports, "__esModule", { value: true });
6
- exports.concatenateArrayBuffers = exports.concatenateReadStream = exports.decompressReadStream = exports.createReadStream = void 0;
7
- const fs_1 = __importDefault(require("fs")); // `fs` will be empty object in browsers (see package.json "browser" field).
8
- const http_1 = __importDefault(require("http"));
9
- const https_1 = __importDefault(require("https"));
10
- const zlib_1 = __importDefault(require("zlib"));
11
- const decode_data_uri_node_1 = require("./decode-data-uri.node");
1
+ import fs from 'fs';
2
+ import http from 'http';
3
+ import https from 'https';
4
+ import zlib from 'zlib';
5
+ import { toArrayBuffer } from './decode-data-uri.node';
6
+
12
7
  function isRequestURL(url) {
13
- return url.startsWith('http:') || url.startsWith('https:');
8
+ return url.startsWith('http:') || url.startsWith('https:');
14
9
  }
15
- // Returns a promise that resolves to a readable stream
16
- async function createReadStream(url, options) {
17
- // Handle file streams in node
18
- if (!isRequestURL(url)) {
19
- const noqueryUrl = url.split('?')[0];
20
- // Now open the stream
21
- return await new Promise((resolve, reject) => {
22
- // @ts-ignore
23
- const stream = fs_1.default.createReadStream(noqueryUrl, { encoding: null });
24
- stream.once('readable', () => resolve(stream));
25
- stream.on('error', (error) => reject(error));
26
- });
27
- }
28
- // HANDLE HTTP/HTTPS REQUESTS IN NODE
29
- // TODO: THIS IS BAD SINCE WE RETURN A PROMISE INSTEAD OF A STREAM
10
+
11
+ export async function createReadStream(url, options) {
12
+ if (!isRequestURL(url)) {
13
+ const noqueryUrl = url.split('?')[0];
30
14
  return await new Promise((resolve, reject) => {
31
- const requestFunction = url.startsWith('https:') ? https_1.default.request : http_1.default.request;
32
- const requestOptions = getRequestOptions(url, options);
33
- const req = requestFunction(requestOptions, (res) => resolve(res));
34
- req.on('error', (error) => reject(error));
35
- req.end();
15
+ const stream = fs.createReadStream(noqueryUrl, {
16
+ encoding: null
17
+ });
18
+ stream.once('readable', () => resolve(stream));
19
+ stream.on('error', error => reject(error));
36
20
  });
21
+ }
22
+
23
+ return await new Promise((resolve, reject) => {
24
+ const requestFunction = url.startsWith('https:') ? https.request : http.request;
25
+ const requestOptions = getRequestOptions(url, options);
26
+ const req = requestFunction(requestOptions, res => resolve(res));
27
+ req.on('error', error => reject(error));
28
+ req.end();
29
+ });
37
30
  }
38
- exports.createReadStream = createReadStream;
39
- function decompressReadStream(readStream, headers) {
40
- switch (headers.get('content-encoding')) {
41
- case 'br':
42
- return readStream.pipe(zlib_1.default.createBrotliDecompress());
43
- case 'gzip':
44
- return readStream.pipe(zlib_1.default.createGunzip());
45
- case 'deflate':
46
- return readStream.pipe(zlib_1.default.createDeflate());
47
- default:
48
- // No compression or an unknown one, just return it as is
49
- return readStream;
50
- }
31
+ export function decompressReadStream(readStream, headers) {
32
+ switch (headers.get('content-encoding')) {
33
+ case 'br':
34
+ return readStream.pipe(zlib.createBrotliDecompress());
35
+
36
+ case 'gzip':
37
+ return readStream.pipe(zlib.createGunzip());
38
+
39
+ case 'deflate':
40
+ return readStream.pipe(zlib.createDeflate());
41
+
42
+ default:
43
+ return readStream;
44
+ }
51
45
  }
52
- exports.decompressReadStream = decompressReadStream;
53
- async function concatenateReadStream(readStream) {
54
- const arrayBufferChunks = [];
55
- return await new Promise((resolve, reject) => {
56
- readStream.on('error', (error) => reject(error));
57
- // Once the readable callback has been added, stream switches to "flowing mode"
58
- // In Node 10 (but not 12 and 14) this causes `data` and `end` to never be called unless we read data here
59
- readStream.on('readable', () => readStream.read());
60
- readStream.on('data', (chunk) => {
61
- if (typeof chunk === 'string') {
62
- reject(new Error('Read stream not binary'));
63
- }
64
- arrayBufferChunks.push((0, decode_data_uri_node_1.toArrayBuffer)(chunk));
65
- });
66
- readStream.on('end', () => {
67
- const arrayBuffer = concatenateArrayBuffers(arrayBufferChunks);
68
- resolve(arrayBuffer);
69
- });
46
+ export async function concatenateReadStream(readStream) {
47
+ const arrayBufferChunks = [];
48
+ return await new Promise((resolve, reject) => {
49
+ readStream.on('error', error => reject(error));
50
+ readStream.on('readable', () => readStream.read());
51
+ readStream.on('data', chunk => {
52
+ if (typeof chunk === 'string') {
53
+ reject(new Error('Read stream not binary'));
54
+ }
55
+
56
+ arrayBufferChunks.push(toArrayBuffer(chunk));
57
+ });
58
+ readStream.on('end', () => {
59
+ const arrayBuffer = concatenateArrayBuffers(arrayBufferChunks);
60
+ resolve(arrayBuffer);
70
61
  });
62
+ });
71
63
  }
72
- exports.concatenateReadStream = concatenateReadStream;
73
- // HELPERS
64
+
74
65
  function getRequestOptions(url, options) {
75
- // Ensure header keys are lower case so that we can merge without duplicates
76
- const originalHeaders = options?.headers || {};
77
- const headers = {};
78
- for (const key of Object.keys(originalHeaders)) {
79
- headers[key.toLowerCase()] = originalHeaders[key];
80
- }
81
- // Add default accept-encoding to headers
82
- headers['accept-encoding'] = headers['accept-encoding'] || 'gzip,br,deflate';
83
- const urlObject = new URL(url);
84
- return {
85
- hostname: urlObject.hostname,
86
- path: urlObject.pathname,
87
- method: 'GET',
88
- // Add options and user provided 'options.fetch' overrides if available
89
- ...options,
90
- ...options?.fetch,
91
- // Override with updated headers with accepted encodings:
92
- headers
93
- };
66
+ const originalHeaders = (options === null || options === void 0 ? void 0 : options.headers) || {};
67
+ const headers = {};
68
+
69
+ for (const key of Object.keys(originalHeaders)) {
70
+ headers[key.toLowerCase()] = originalHeaders[key];
71
+ }
72
+
73
+ headers['accept-encoding'] = headers['accept-encoding'] || 'gzip,br,deflate';
74
+ const urlObject = new URL(url);
75
+ return {
76
+ hostname: urlObject.hostname,
77
+ path: urlObject.pathname,
78
+ method: 'GET',
79
+ ...options,
80
+ ...(options === null || options === void 0 ? void 0 : options.fetch),
81
+ headers
82
+ };
94
83
  }
95
- /**
96
- * Concatenate a sequence of ArrayBuffers
97
- * @return A concatenated ArrayBuffer
98
- * @note duplicates loader-utils since polyfills should be independent
99
- */
100
- function concatenateArrayBuffers(sources) {
101
- // Make sure all inputs are wrapped in typed arrays
102
- const sourceArrays = sources.map((source2) => source2 instanceof ArrayBuffer ? new Uint8Array(source2) : source2);
103
- // Get length of all inputs
104
- const byteLength = sourceArrays.reduce((length, typedArray) => length + typedArray.byteLength, 0);
105
- // Allocate array with space for all inputs
106
- const result = new Uint8Array(byteLength);
107
- // Copy the subarrays
108
- let offset = 0;
109
- for (const sourceArray of sourceArrays) {
110
- result.set(sourceArray, offset);
111
- offset += sourceArray.byteLength;
112
- }
113
- // We work with ArrayBuffers, discard the typed array wrapper
114
- return result.buffer;
84
+
85
+ export function concatenateArrayBuffers(sources) {
86
+ const sourceArrays = sources.map(source2 => source2 instanceof ArrayBuffer ? new Uint8Array(source2) : source2);
87
+ const byteLength = sourceArrays.reduce((length, typedArray) => length + typedArray.byteLength, 0);
88
+ const result = new Uint8Array(byteLength);
89
+ let offset = 0;
90
+
91
+ for (const sourceArray of sourceArrays) {
92
+ result.set(sourceArray, offset);
93
+ offset += sourceArray.byteLength;
94
+ }
95
+
96
+ return result.buffer;
115
97
  }
116
- exports.concatenateArrayBuffers = concatenateArrayBuffers;
98
+ //# sourceMappingURL=stream-utils.node.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../../../../src/node/fetch/utils/stream-utils.node.ts"],"names":["fs","http","https","zlib","toArrayBuffer","isRequestURL","url","startsWith","createReadStream","options","noqueryUrl","split","Promise","resolve","reject","stream","encoding","once","on","error","requestFunction","request","requestOptions","getRequestOptions","req","res","end","decompressReadStream","readStream","headers","get","pipe","createBrotliDecompress","createGunzip","createDeflate","concatenateReadStream","arrayBufferChunks","read","chunk","Error","push","arrayBuffer","concatenateArrayBuffers","originalHeaders","key","Object","keys","toLowerCase","urlObject","URL","hostname","path","pathname","method","fetch","sources","sourceArrays","map","source2","ArrayBuffer","Uint8Array","byteLength","reduce","length","typedArray","result","offset","sourceArray","set","buffer"],"mappings":"AAAA,OAAOA,EAAP,MAAe,IAAf;AACA,OAAOC,IAAP,MAAiB,MAAjB;AACA,OAAOC,KAAP,MAAkB,OAAlB;AACA,OAAOC,IAAP,MAAiB,MAAjB;AAEA,SAAQC,aAAR,QAA4B,wBAA5B;;AAEA,SAASC,YAAT,CAAsBC,GAAtB,EAA4C;AAC1C,SAAOA,GAAG,CAACC,UAAJ,CAAe,OAAf,KAA2BD,GAAG,CAACC,UAAJ,CAAe,QAAf,CAAlC;AACD;;AAGD,OAAO,eAAeC,gBAAf,CAAgCF,GAAhC,EAAqCG,OAArC,EAA4D;AAEjE,MAAI,CAACJ,YAAY,CAACC,GAAD,CAAjB,EAAwB;AACtB,UAAMI,UAAU,GAAGJ,GAAG,CAACK,KAAJ,CAAU,GAAV,EAAe,CAAf,CAAnB;AAEA,WAAO,MAAM,IAAIC,OAAJ,CAAY,CAACC,OAAD,EAAUC,MAAV,KAAqB;AAE5C,YAAMC,MAAM,GAAGf,EAAE,CAACQ,gBAAH,CAAoBE,UAApB,EAAgC;AAACM,QAAAA,QAAQ,EAAE;AAAX,OAAhC,CAAf;AACAD,MAAAA,MAAM,CAACE,IAAP,CAAY,UAAZ,EAAwB,MAAMJ,OAAO,CAACE,MAAD,CAArC;AACAA,MAAAA,MAAM,CAACG,EAAP,CAAU,OAAV,EAAoBC,KAAD,IAAWL,MAAM,CAACK,KAAD,CAApC;AACD,KALY,CAAb;AAMD;;AAID,SAAO,MAAM,IAAIP,OAAJ,CAAY,CAACC,OAAD,EAAUC,MAAV,KAAqB;AAC5C,UAAMM,eAAe,GAAGd,GAAG,CAACC,UAAJ,CAAe,QAAf,IAA2BL,KAAK,CAACmB,OAAjC,GAA2CpB,IAAI,CAACoB,OAAxE;AACA,UAAMC,cAAc,GAAGC,iBAAiB,CAACjB,GAAD,EAAMG,OAAN,CAAxC;AACA,UAAMe,GAAG,GAAGJ,eAAe,CAACE,cAAD,EAAkBG,GAAD,IAASZ,OAAO,CAACY,GAAD,CAAjC,CAA3B;AACAD,IAAAA,GAAG,CAACN,EAAJ,CAAO,OAAP,EAAiBC,KAAD,IAAWL,MAAM,CAACK,KAAD,CAAjC;AACAK,IAAAA,GAAG,CAACE,GAAJ;AACD,GANY,CAAb;AAOD;AAED,OAAO,SAASC,oBAAT,CAA8BC,UAA9B,EAA0CC,OAA1C,EAAmD;AACxD,UAAQA,OAAO,CAACC,GAAR,CAAY,kBAAZ,CAAR;AACE,SAAK,IAAL;AACE,aAAOF,UAAU,CAACG,IAAX,CAAgB5B,IAAI,CAAC6B,sBAAL,EAAhB,CAAP;;AACF,SAAK,MAAL;AACE,aAAOJ,UAAU,CAACG,IAAX,CAAgB5B,IAAI,CAAC8B,YAAL,EAAhB,CAAP;;AACF,SAAK,SAAL;AACE,aAAOL,UAAU,CAACG,IAAX,CAAgB5B,IAAI,CAAC+B,aAAL,EAAhB,CAAP;;AACF;AAEE,aAAON,UAAP;AATJ;AAWD;AAED,OAAO,eAAeO,qBAAf,CAAqCP,UAArC,EAAuE;AAC5E,QAAMQ,iBAAgC,GAAG,EAAzC;AAEA,SAAO,MAAM,IAAIxB,OAAJ,CAAY,CAACC,OAAD,EAAUC,MAAV,KAAqB;AAC5Cc,IAAAA,UAAU,CAACV,EAAX,CAAc,OAAd,EAAwBC,KAAD,IAAWL,MAAM,CAACK,KAAD,CAAxC;AAIAS,IAAAA,UAAU,CAACV,EAAX,CAAc,UAAd,EAA0B,MAAMU,UAAU,CAACS,IAAX,EAAhC;AAEAT,IAAAA,UAAU,CAACV,EAAX,CAAc,MAAd,EAAuBoB,KAAD,IAAW;AAC/B,UAAI,OAAOA,KAAP,KAAiB,QAArB,EAA+B;AAC7BxB,QAAAA,MAAM,CAAC,IAAIyB,KAAJ,CAAU,wBAAV,CAAD,CAAN;AACD;;AACDH,MAAAA,iBAAiB,CAACI,IAAlB,CAAuBpC,aAAa,CAACkC,KAAD,CAApC;AACD,KALD;AAOAV,IAAAA,UAAU,CAACV,EAAX,CAAc,KAAd,EAAqB,MAAM;AACzB,YAAMuB,WAAW,GAAGC,uBAAuB,CAACN,iBAAD,CAA3C;AACAvB,MAAAA,OAAO,CAAC4B,WAAD,CAAP;AACD,KAHD;AAID,GAlBY,CAAb;AAmBD;;AAID,SAASlB,iBAAT,CAA2BjB,GAA3B,EAAwCG,OAAxC,EAAoF;AAElF,QAAMkC,eAAe,GAAG,CAAAlC,OAAO,SAAP,IAAAA,OAAO,WAAP,YAAAA,OAAO,CAAEoB,OAAT,KAAoB,EAA5C;AACA,QAAMA,OAAO,GAAG,EAAhB;;AACA,OAAK,MAAMe,GAAX,IAAkBC,MAAM,CAACC,IAAP,CAAYH,eAAZ,CAAlB,EAAgD;AAC9Cd,IAAAA,OAAO,CAACe,GAAG,CAACG,WAAJ,EAAD,CAAP,GAA6BJ,eAAe,CAACC,GAAD,CAA5C;AACD;;AAGDf,EAAAA,OAAO,CAAC,iBAAD,CAAP,GAA6BA,OAAO,CAAC,iBAAD,CAAP,IAA8B,iBAA3D;AAEA,QAAMmB,SAAS,GAAG,IAAIC,GAAJ,CAAQ3C,GAAR,CAAlB;AACA,SAAO;AACL4C,IAAAA,QAAQ,EAAEF,SAAS,CAACE,QADf;AAELC,IAAAA,IAAI,EAAEH,SAAS,CAACI,QAFX;AAGLC,IAAAA,MAAM,EAAE,KAHH;AAKL,OAAG5C,OALE;AAML,QAAGA,OAAH,aAAGA,OAAH,uBAAGA,OAAO,CAAE6C,KAAZ,CANK;AAQLzB,IAAAA;AARK,GAAP;AAUD;;AAOD,OAAO,SAASa,uBAAT,CAAiCa,OAAjC,EAAqF;AAE1F,QAAMC,YAAY,GAAGD,OAAO,CAACE,GAAR,CAAaC,OAAD,IAC/BA,OAAO,YAAYC,WAAnB,GAAiC,IAAIC,UAAJ,CAAeF,OAAf,CAAjC,GAA2DA,OADxC,CAArB;AAKA,QAAMG,UAAU,GAAGL,YAAY,CAACM,MAAb,CAAoB,CAACC,MAAD,EAASC,UAAT,KAAwBD,MAAM,GAAGC,UAAU,CAACH,UAAhE,EAA4E,CAA5E,CAAnB;AAGA,QAAMI,MAAM,GAAG,IAAIL,UAAJ,CAAeC,UAAf,CAAf;AAGA,MAAIK,MAAM,GAAG,CAAb;;AACA,OAAK,MAAMC,WAAX,IAA0BX,YAA1B,EAAwC;AACtCS,IAAAA,MAAM,CAACG,GAAP,CAAWD,WAAX,EAAwBD,MAAxB;AACAA,IAAAA,MAAM,IAAIC,WAAW,CAACN,UAAtB;AACD;;AAGD,SAAOI,MAAM,CAACI,MAAd;AACD","sourcesContent":["import fs from 'fs'; // `fs` will be empty object in browsers (see package.json \"browser\" field).\nimport http from 'http';\nimport https from 'https';\nimport zlib from 'zlib';\n\nimport {toArrayBuffer} from './decode-data-uri.node';\n\nfunction isRequestURL(url: string): boolean {\n return url.startsWith('http:') || url.startsWith('https:');\n}\n\n// Returns a promise that resolves to a readable stream\nexport async function createReadStream(url, options): Promise<any> {\n // Handle file streams in node\n if (!isRequestURL(url)) {\n const noqueryUrl = url.split('?')[0];\n // Now open the stream\n return await new Promise((resolve, reject) => {\n // @ts-ignore\n const stream = fs.createReadStream(noqueryUrl, {encoding: null});\n stream.once('readable', () => resolve(stream));\n stream.on('error', (error) => reject(error));\n });\n }\n\n // HANDLE HTTP/HTTPS REQUESTS IN NODE\n // TODO: THIS IS BAD SINCE WE RETURN A PROMISE INSTEAD OF A STREAM\n return await new Promise((resolve, reject) => {\n const requestFunction = url.startsWith('https:') ? https.request : http.request;\n const requestOptions = getRequestOptions(url, options);\n const req = requestFunction(requestOptions, (res) => resolve(res));\n req.on('error', (error) => reject(error));\n req.end();\n });\n}\n\nexport function decompressReadStream(readStream, headers) {\n switch (headers.get('content-encoding')) {\n case 'br':\n return readStream.pipe(zlib.createBrotliDecompress());\n case 'gzip':\n return readStream.pipe(zlib.createGunzip());\n case 'deflate':\n return readStream.pipe(zlib.createDeflate());\n default:\n // No compression or an unknown one, just return it as is\n return readStream;\n }\n}\n\nexport async function concatenateReadStream(readStream): Promise<ArrayBuffer> {\n const arrayBufferChunks: ArrayBuffer[] = [];\n\n return await new Promise((resolve, reject) => {\n readStream.on('error', (error) => reject(error));\n\n // Once the readable callback has been added, stream switches to \"flowing mode\"\n // In Node 10 (but not 12 and 14) this causes `data` and `end` to never be called unless we read data here\n readStream.on('readable', () => readStream.read());\n\n readStream.on('data', (chunk) => {\n if (typeof chunk === 'string') {\n reject(new Error('Read stream not binary'));\n }\n arrayBufferChunks.push(toArrayBuffer(chunk));\n });\n\n readStream.on('end', () => {\n const arrayBuffer = concatenateArrayBuffers(arrayBufferChunks);\n resolve(arrayBuffer);\n });\n });\n}\n\n// HELPERS\n\nfunction getRequestOptions(url: string, options?: {fetch?: typeof fetch; headers?}) {\n // Ensure header keys are lower case so that we can merge without duplicates\n const originalHeaders = options?.headers || {};\n const headers = {};\n for (const key of Object.keys(originalHeaders)) {\n headers[key.toLowerCase()] = originalHeaders[key];\n }\n\n // Add default accept-encoding to headers\n headers['accept-encoding'] = headers['accept-encoding'] || 'gzip,br,deflate';\n\n const urlObject = new URL(url);\n return {\n hostname: urlObject.hostname,\n path: urlObject.pathname,\n method: 'GET',\n // Add options and user provided 'options.fetch' overrides if available\n ...options,\n ...options?.fetch,\n // Override with updated headers with accepted encodings:\n headers\n };\n}\n\n/**\n * Concatenate a sequence of ArrayBuffers\n * @return A concatenated ArrayBuffer\n * @note duplicates loader-utils since polyfills should be independent\n */\nexport function concatenateArrayBuffers(sources: (ArrayBuffer | Uint8Array)[]): ArrayBuffer {\n // Make sure all inputs are wrapped in typed arrays\n const sourceArrays = sources.map((source2) =>\n source2 instanceof ArrayBuffer ? new Uint8Array(source2) : source2\n );\n\n // Get length of all inputs\n const byteLength = sourceArrays.reduce((length, typedArray) => length + typedArray.byteLength, 0);\n\n // Allocate array with space for all inputs\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let offset = 0;\n for (const sourceArray of sourceArrays) {\n result.set(sourceArray, offset);\n offset += sourceArray.byteLength;\n }\n\n // We work with ArrayBuffers, discard the typed array wrapper\n return result.buffer;\n}\n"],"file":"stream-utils.node.js"}