@loaders.gl/polyfills 3.3.0-alpha.5 → 3.3.0-alpha.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (94) hide show
  1. package/dist/dist.min.js +22 -10
  2. package/dist/es5/bundle.js +0 -1
  3. package/dist/es5/bundle.js.map +1 -1
  4. package/dist/es5/index.js +10 -37
  5. package/dist/es5/index.js.map +1 -1
  6. package/dist/es5/lib/encoding-indexes.js.map +1 -1
  7. package/dist/es5/lib/encoding.js +255 -135
  8. package/dist/es5/lib/encoding.js.map +1 -1
  9. package/dist/es5/node/buffer/btoa.node.js +0 -1
  10. package/dist/es5/node/buffer/btoa.node.js.map +1 -1
  11. package/dist/es5/node/buffer/to-array-buffer.node.js +0 -2
  12. package/dist/es5/node/buffer/to-array-buffer.node.js.map +1 -1
  13. package/dist/es5/node/fetch/fetch.node.js +2 -36
  14. package/dist/es5/node/fetch/fetch.node.js.map +1 -1
  15. package/dist/es5/node/fetch/headers.node.js +0 -21
  16. package/dist/es5/node/fetch/headers.node.js.map +1 -1
  17. package/dist/es5/node/fetch/response.node.js +10 -45
  18. package/dist/es5/node/fetch/response.node.js.map +1 -1
  19. package/dist/es5/node/fetch/utils/decode-data-uri.node.js +0 -8
  20. package/dist/es5/node/fetch/utils/decode-data-uri.node.js.map +1 -1
  21. package/dist/es5/node/fetch/utils/stream-utils.node.js +10 -40
  22. package/dist/es5/node/fetch/utils/stream-utils.node.js.map +1 -1
  23. package/dist/es5/node/file/blob-stream-controller.js +5 -19
  24. package/dist/es5/node/file/blob-stream-controller.js.map +1 -1
  25. package/dist/es5/node/file/blob-stream.js +5 -30
  26. package/dist/es5/node/file/blob-stream.js.map +1 -1
  27. package/dist/es5/node/file/blob.js +17 -50
  28. package/dist/es5/node/file/blob.js.map +1 -1
  29. package/dist/es5/node/file/file-reader.js +1 -28
  30. package/dist/es5/node/file/file-reader.js.map +1 -1
  31. package/dist/es5/node/file/file.js +0 -18
  32. package/dist/es5/node/file/file.js.map +1 -1
  33. package/dist/es5/node/file/install-file-polyfills.js +0 -7
  34. package/dist/es5/node/file/install-file-polyfills.js.map +1 -1
  35. package/dist/es5/node/file/readable-stream.js +2 -14
  36. package/dist/es5/node/file/readable-stream.js.map +1 -1
  37. package/dist/es5/node/images/encode-image.node.js +2 -6
  38. package/dist/es5/node/images/encode-image.node.js.map +1 -1
  39. package/dist/es5/node/images/parse-image.node.js +1 -11
  40. package/dist/es5/node/images/parse-image.node.js.map +1 -1
  41. package/dist/es5/promise/all-settled.js +2 -2
  42. package/dist/es5/promise/all-settled.js.map +1 -1
  43. package/dist/es5/utils/assert.js +0 -1
  44. package/dist/es5/utils/assert.js.map +1 -1
  45. package/dist/es5/utils/globals.js +3 -4
  46. package/dist/es5/utils/globals.js.map +1 -1
  47. package/dist/esm/bundle.js +1 -1
  48. package/dist/esm/bundle.js.map +1 -1
  49. package/dist/esm/index.js +3 -8
  50. package/dist/esm/index.js.map +1 -1
  51. package/dist/esm/lib/encoding-indexes.js +2 -0
  52. package/dist/esm/lib/encoding-indexes.js.map +1 -1
  53. package/dist/esm/lib/encoding.js +256 -131
  54. package/dist/esm/lib/encoding.js.map +1 -1
  55. package/dist/esm/node/buffer/btoa.node.js +2 -0
  56. package/dist/esm/node/buffer/btoa.node.js.map +1 -1
  57. package/dist/esm/node/buffer/to-array-buffer.node.js +0 -1
  58. package/dist/esm/node/buffer/to-array-buffer.node.js.map +1 -1
  59. package/dist/esm/node/fetch/fetch.node.js +5 -17
  60. package/dist/esm/node/fetch/fetch.node.js.map +1 -1
  61. package/dist/esm/node/fetch/headers.node.js +2 -24
  62. package/dist/esm/node/fetch/headers.node.js.map +1 -1
  63. package/dist/esm/node/fetch/response.node.js +0 -17
  64. package/dist/esm/node/fetch/response.node.js.map +1 -1
  65. package/dist/esm/node/fetch/utils/decode-data-uri.node.js +3 -6
  66. package/dist/esm/node/fetch/utils/decode-data-uri.node.js.map +1 -1
  67. package/dist/esm/node/fetch/utils/stream-utils.node.js +4 -7
  68. package/dist/esm/node/fetch/utils/stream-utils.node.js.map +1 -1
  69. package/dist/esm/node/file/blob-stream-controller.js +0 -8
  70. package/dist/esm/node/file/blob-stream-controller.js.map +1 -1
  71. package/dist/esm/node/file/blob-stream.js +1 -5
  72. package/dist/esm/node/file/blob-stream.js.map +1 -1
  73. package/dist/esm/node/file/blob.js +12 -21
  74. package/dist/esm/node/file/blob.js.map +1 -1
  75. package/dist/esm/node/file/file-reader.js +1 -24
  76. package/dist/esm/node/file/file-reader.js.map +1 -1
  77. package/dist/esm/node/file/file.js +4 -9
  78. package/dist/esm/node/file/file.js.map +1 -1
  79. package/dist/esm/node/file/install-file-polyfills.js +0 -2
  80. package/dist/esm/node/file/install-file-polyfills.js.map +1 -1
  81. package/dist/esm/node/file/readable-stream.js +2 -0
  82. package/dist/esm/node/file/readable-stream.js.map +1 -1
  83. package/dist/esm/node/images/encode-image.node.js +4 -0
  84. package/dist/esm/node/images/encode-image.node.js.map +1 -1
  85. package/dist/esm/node/images/parse-image.node.js +3 -0
  86. package/dist/esm/node/images/parse-image.node.js.map +1 -1
  87. package/dist/esm/promise/all-settled.js +1 -0
  88. package/dist/esm/promise/all-settled.js.map +1 -1
  89. package/dist/esm/utils/assert.js.map +1 -1
  90. package/dist/esm/utils/globals.js +4 -1
  91. package/dist/esm/utils/globals.js.map +1 -1
  92. package/dist/index.js +5 -1
  93. package/dist/libs/encoding-indexes-asian.d.ts +1 -1
  94. package/package.json +2 -2
@@ -1,3 +1,5 @@
1
+
2
+
1
3
  export function atob(string) {
2
4
  return Buffer.from(string).toString('base64');
3
5
  }
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/node/buffer/btoa.node.ts"],"names":["atob","string","Buffer","from","toString","btoa","base64"],"mappings":"AAIA,OAAO,SAASA,IAAT,CAAcC,MAAd,EAAsB;AAC3B,SAAOC,MAAM,CAACC,IAAP,CAAYF,MAAZ,EAAoBG,QAApB,CAA6B,QAA7B,CAAP;AACD;AAED,OAAO,SAASC,IAAT,CAAcC,MAAd,EAAsB;AAC3B,SAAOJ,MAAM,CAACC,IAAP,CAAYG,MAAZ,EAAoB,QAApB,EAA8BF,QAA9B,CAAuC,OAAvC,CAAP;AACD","sourcesContent":["// btoa, atob polyfills for Node.js\n// Note: The atob and btoa functions (not just the polyfills!) are not unicode safe\n// But still useful for unit testing\n\nexport function atob(string) {\n return Buffer.from(string).toString('base64');\n}\n\nexport function btoa(base64) {\n return Buffer.from(base64, 'base64').toString('ascii');\n}\n"],"file":"btoa.node.js"}
1
+ {"version":3,"file":"btoa.node.js","names":["atob","string","Buffer","from","toString","btoa","base64"],"sources":["../../../../src/node/buffer/btoa.node.ts"],"sourcesContent":["// btoa, atob polyfills for Node.js\n// Note: The atob and btoa functions (not just the polyfills!) are not unicode safe\n// But still useful for unit testing\n\nexport function atob(string) {\n return Buffer.from(string).toString('base64');\n}\n\nexport function btoa(base64) {\n return Buffer.from(base64, 'base64').toString('ascii');\n}\n"],"mappings":";;AAIA,OAAO,SAASA,IAAI,CAACC,MAAM,EAAE;EAC3B,OAAOC,MAAM,CAACC,IAAI,CAACF,MAAM,CAAC,CAACG,QAAQ,CAAC,QAAQ,CAAC;AAC/C;AAEA,OAAO,SAASC,IAAI,CAACC,MAAM,EAAE;EAC3B,OAAOJ,MAAM,CAACC,IAAI,CAACG,MAAM,EAAE,QAAQ,CAAC,CAACF,QAAQ,CAAC,OAAO,CAAC;AACxD"}
@@ -3,7 +3,6 @@ export function bufferToArrayBuffer(buffer) {
3
3
  const typedArray = new Uint8Array(buffer);
4
4
  return typedArray.buffer;
5
5
  }
6
-
7
6
  return buffer;
8
7
  }
9
8
  //# sourceMappingURL=to-array-buffer.node.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/node/buffer/to-array-buffer.node.ts"],"names":["bufferToArrayBuffer","buffer","Buffer","isBuffer","typedArray","Uint8Array"],"mappings":"AAAA,OAAO,SAASA,mBAAT,CAA6BC,MAA7B,EAAqC;AAE1C,MAAIC,MAAM,CAACC,QAAP,CAAgBF,MAAhB,CAAJ,EAA6B;AAC3B,UAAMG,UAAU,GAAG,IAAIC,UAAJ,CAAeJ,MAAf,CAAnB;AACA,WAAOG,UAAU,CAACH,MAAlB;AACD;;AACD,SAAOA,MAAP;AACD","sourcesContent":["export function bufferToArrayBuffer(buffer) {\n // TODO - per docs we should just be able to call buffer.buffer, but there are issues\n if (Buffer.isBuffer(buffer)) {\n const typedArray = new Uint8Array(buffer);\n return typedArray.buffer;\n }\n return buffer;\n}\n"],"file":"to-array-buffer.node.js"}
1
+ {"version":3,"file":"to-array-buffer.node.js","names":["bufferToArrayBuffer","buffer","Buffer","isBuffer","typedArray","Uint8Array"],"sources":["../../../../src/node/buffer/to-array-buffer.node.ts"],"sourcesContent":["export function bufferToArrayBuffer(buffer) {\n // TODO - per docs we should just be able to call buffer.buffer, but there are issues\n if (Buffer.isBuffer(buffer)) {\n const typedArray = new Uint8Array(buffer);\n return typedArray.buffer;\n }\n return buffer;\n}\n"],"mappings":"AAAA,OAAO,SAASA,mBAAmB,CAACC,MAAM,EAAE;EAE1C,IAAIC,MAAM,CAACC,QAAQ,CAACF,MAAM,CAAC,EAAE;IAC3B,MAAMG,UAAU,GAAG,IAAIC,UAAU,CAACJ,MAAM,CAAC;IACzC,OAAOG,UAAU,CAACH,MAAM;EAC1B;EACA,OAAOA,MAAM;AACf"}
@@ -3,9 +3,7 @@ import Response from './response.node';
3
3
  import Headers from './headers.node';
4
4
  import { decodeDataUri } from './utils/decode-data-uri.node';
5
5
  import { createReadStream } from './utils/stream-utils.node';
6
-
7
6
  const isDataURL = url => url.startsWith('data:');
8
-
9
7
  const isRequestURL = url => url.startsWith('http:') || url.startsWith('https:');
10
8
 
11
9
  export default async function fetchNode(url, options) {
@@ -26,7 +24,6 @@ export default async function fetchNode(url, options) {
26
24
 
27
25
  const syntheticResponseHeaders = {};
28
26
  const originalUrl = url;
29
-
30
27
  if (url.endsWith('.gz')) {
31
28
  url = url.slice(0, -3);
32
29
  syntheticResponseHeaders['content-encoding'] = 'gzip';
@@ -39,12 +36,11 @@ export default async function fetchNode(url, options) {
39
36
  statusText
40
37
  } = getStatus(body);
41
38
  const followRedirect = !options || options.followRedirect || options.followRedirect === undefined;
42
-
43
39
  if (status >= 300 && status < 400 && headers.has('location') && followRedirect) {
44
40
  const redirectUrl = generateRedirectUrl(url, headers.get('location'));
41
+
45
42
  return await fetchNode(redirectUrl, options);
46
43
  }
47
-
48
44
  return new Response(body, {
49
45
  headers,
50
46
  status,
@@ -64,7 +60,6 @@ function generateRedirectUrl(originalUrl, location) {
64
60
  if (location.startsWith('http')) {
65
61
  return location;
66
62
  }
67
-
68
63
  const url = new URL(originalUrl);
69
64
  url.pathname = location;
70
65
  return url.href;
@@ -77,19 +72,16 @@ function getStatus(httpResponse) {
77
72
  statusText: httpResponse.statusMessage || 'NA'
78
73
  };
79
74
  }
80
-
81
75
  return {
82
76
  status: 200,
83
77
  statusText: 'OK'
84
78
  };
85
79
  }
86
-
87
- function getHeaders(url, httpResponse, additionalHeaders = {}) {
80
+ function getHeaders(url, httpResponse) {
81
+ let additionalHeaders = arguments.length > 2 && arguments[2] !== undefined ? arguments[2] : {};
88
82
  const headers = {};
89
-
90
83
  if (httpResponse && httpResponse.headers) {
91
84
  const httpHeaders = httpResponse.headers;
92
-
93
85
  for (const key in httpHeaders) {
94
86
  const header = httpHeaders[key];
95
87
  headers[key.toLowerCase()] = String(header);
@@ -98,29 +90,25 @@ function getHeaders(url, httpResponse, additionalHeaders = {}) {
98
90
 
99
91
  if (!headers['content-length']) {
100
92
  const contentLength = getContentLength(url);
101
-
102
93
  if (Number.isFinite(contentLength)) {
103
94
  headers['content-length'] = contentLength;
104
95
  }
105
96
  }
106
-
107
97
  Object.assign(headers, additionalHeaders);
108
98
  return new Headers(headers);
109
99
  }
110
-
111
100
  function getContentLength(url) {
112
101
  if (isRequestURL(url)) {
113
102
  return null;
114
103
  } else if (isDataURL(url)) {
115
104
  return url.length - 'data:'.length;
116
105
  }
117
-
118
106
  try {
119
107
  const noqueryUrl = url.split('?')[0];
120
108
  const stats = fs.statSync(noqueryUrl);
121
109
  return stats.size;
122
- } catch (error) {}
123
-
110
+ } catch (error) {
111
+ }
124
112
  return null;
125
113
  }
126
114
  //# sourceMappingURL=fetch.node.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/node/fetch/fetch.node.ts"],"names":["fs","Response","Headers","decodeDataUri","createReadStream","isDataURL","url","startsWith","isRequestURL","fetchNode","options","arrayBuffer","mimeType","response","headers","syntheticResponseHeaders","originalUrl","endsWith","slice","body","getHeaders","status","statusText","getStatus","followRedirect","undefined","has","redirectUrl","generateRedirectUrl","get","error","String","location","URL","pathname","href","httpResponse","statusCode","statusMessage","additionalHeaders","httpHeaders","key","header","toLowerCase","contentLength","getContentLength","Number","isFinite","Object","assign","length","noqueryUrl","split","stats","statSync","size"],"mappings":"AAAA,OAAOA,EAAP,MAAe,IAAf;AACA,OAAOC,QAAP,MAAqB,iBAArB;AACA,OAAOC,OAAP,MAAoB,gBAApB;AAEA,SAAQC,aAAR,QAA4B,8BAA5B;AACA,SAAQC,gBAAR,QAA+B,2BAA/B;;AAEA,MAAMC,SAAS,GAAIC,GAAD,IAASA,GAAG,CAACC,UAAJ,CAAe,OAAf,CAA3B;;AACA,MAAMC,YAAY,GAAIF,GAAD,IAASA,GAAG,CAACC,UAAJ,CAAe,OAAf,KAA2BD,GAAG,CAACC,UAAJ,CAAe,QAAf,CAAzD;;AAOA,eAAe,eAAeE,SAAf,CAAyBH,GAAzB,EAA8BI,OAA9B,EAAuC;AACpD,MAAI;AAGF,QAAIL,SAAS,CAACC,GAAD,CAAb,EAAoB;AAClB,YAAM;AAACK,QAAAA,WAAD;AAAcC,QAAAA;AAAd,UAA0BT,aAAa,CAACG,GAAD,CAA7C;AACA,YAAMO,QAAQ,GAAG,IAAIZ,QAAJ,CAAaU,WAAb,EAA0B;AACzCG,QAAAA,OAAO,EAAE;AAAC,0BAAgBF;AAAjB,SADgC;AAEzCN,QAAAA;AAFyC,OAA1B,CAAjB;AAIA,aAAOO,QAAP;AACD;;AAGD,UAAME,wBAAwB,GAAG,EAAjC;AACA,UAAMC,WAAW,GAAGV,GAApB;;AACA,QAAIA,GAAG,CAACW,QAAJ,CAAa,KAAb,CAAJ,EAAyB;AACvBX,MAAAA,GAAG,GAAGA,GAAG,CAACY,KAAJ,CAAU,CAAV,EAAa,CAAC,CAAd,CAAN;AACAH,MAAAA,wBAAwB,CAAC,kBAAD,CAAxB,GAA+C,MAA/C;AACD;;AAGD,UAAMI,IAAI,GAAG,MAAMf,gBAAgB,CAACY,WAAD,EAAcN,OAAd,CAAnC;AACA,UAAMI,OAAO,GAAGM,UAAU,CAACd,GAAD,EAAMa,IAAN,EAAYJ,wBAAZ,CAA1B;AACA,UAAM;AAACM,MAAAA,MAAD;AAASC,MAAAA;AAAT,QAAuBC,SAAS,CAACJ,IAAD,CAAtC;AAEA,UAAMK,cAAc,GAClB,CAACd,OAAD,IAAYA,OAAO,CAACc,cAApB,IAAsCd,OAAO,CAACc,cAAR,KAA2BC,SADnE;;AAGA,QAAIJ,MAAM,IAAI,GAAV,IAAiBA,MAAM,GAAG,GAA1B,IAAiCP,OAAO,CAACY,GAAR,CAAY,UAAZ,CAAjC,IAA4DF,cAAhE,EAAgF;AAC9E,YAAMG,WAAW,GAAGC,mBAAmB,CAACtB,GAAD,EAAMQ,OAAO,CAACe,GAAR,CAAY,UAAZ,CAAN,CAAvC;AAGA,aAAO,MAAMpB,SAAS,CAACkB,WAAD,EAAcjB,OAAd,CAAtB;AACD;;AACD,WAAO,IAAIT,QAAJ,CAAakB,IAAb,EAAmB;AAACL,MAAAA,OAAD;AAAUO,MAAAA,MAAV;AAAkBC,MAAAA,UAAlB;AAA8BhB,MAAAA;AAA9B,KAAnB,CAAP;AACD,GAnCD,CAmCE,OAAOwB,KAAP,EAAc;AAEd,WAAO,IAAI7B,QAAJ,CAAa,IAAb,EAAmB;AAACoB,MAAAA,MAAM,EAAE,GAAT;AAAcC,MAAAA,UAAU,EAAES,MAAM,CAACD,KAAD,CAAhC;AAAyCxB,MAAAA;AAAzC,KAAnB,CAAP;AACD;AACF;;AAOD,SAASsB,mBAAT,CAA6BZ,WAA7B,EAAkDgB,QAAlD,EAA4E;AAC1E,MAAIA,QAAQ,CAACzB,UAAT,CAAoB,MAApB,CAAJ,EAAiC;AAC/B,WAAOyB,QAAP;AACD;;AAED,QAAM1B,GAAG,GAAG,IAAI2B,GAAJ,CAAQjB,WAAR,CAAZ;AACAV,EAAAA,GAAG,CAAC4B,QAAJ,GAAeF,QAAf;AAEA,SAAO1B,GAAG,CAAC6B,IAAX;AACD;;AAKD,SAASZ,SAAT,CAAmBa,YAAnB,EAAiC;AAC/B,MAAIA,YAAY,CAACC,UAAjB,EAA6B;AAC3B,WAAO;AAAChB,MAAAA,MAAM,EAAEe,YAAY,CAACC,UAAtB;AAAkCf,MAAAA,UAAU,EAAEc,YAAY,CAACE,aAAb,IAA8B;AAA5E,KAAP;AACD;;AACD,SAAO;AAACjB,IAAAA,MAAM,EAAE,GAAT;AAAcC,IAAAA,UAAU,EAAE;AAA1B,GAAP;AACD;;AAED,SAASF,UAAT,CAAoBd,GAApB,EAAyB8B,YAAzB,EAAuCG,iBAAiB,GAAG,EAA3D,EAA+D;AAC7D,QAAMzB,OAAO,GAAG,EAAhB;;AAEA,MAAIsB,YAAY,IAAIA,YAAY,CAACtB,OAAjC,EAA0C;AACxC,UAAM0B,WAAW,GAAGJ,YAAY,CAACtB,OAAjC;;AACA,SAAK,MAAM2B,GAAX,IAAkBD,WAAlB,EAA+B;AAC7B,YAAME,MAAM,GAAGF,WAAW,CAACC,GAAD,CAA1B;AACA3B,MAAAA,OAAO,CAAC2B,GAAG,CAACE,WAAJ,EAAD,CAAP,GAA6BZ,MAAM,CAACW,MAAD,CAAnC;AACD;AACF;;AAGD,MAAI,CAAC5B,OAAO,CAAC,gBAAD,CAAZ,EAAgC;AAC9B,UAAM8B,aAAa,GAAGC,gBAAgB,CAACvC,GAAD,CAAtC;;AACA,QAAIwC,MAAM,CAACC,QAAP,CAAgBH,aAAhB,CAAJ,EAAoC;AAClC9B,MAAAA,OAAO,CAAC,gBAAD,CAAP,GAA4B8B,aAA5B;AACD;AACF;;AAEDI,EAAAA,MAAM,CAACC,MAAP,CAAcnC,OAAd,EAAuByB,iBAAvB;AAEA,SAAO,IAAIrC,OAAJ,CAAYY,OAAZ,CAAP;AACD;;AAED,SAAS+B,gBAAT,CAA0BvC,GAA1B,EAA+B;AAC7B,MAAIE,YAAY,CAACF,GAAD,CAAhB,EAAuB;AAErB,WAAO,IAAP;AACD,GAHD,MAGO,IAAID,SAAS,CAACC,GAAD,CAAb,EAAoB;AAEzB,WAAOA,GAAG,CAAC4C,MAAJ,GAAa,QAAQA,MAA5B;AACD;;AAGD,MAAI;AAEF,UAAMC,UAAU,GAAG7C,GAAG,CAAC8C,KAAJ,CAAU,GAAV,EAAe,CAAf,CAAnB;AACA,UAAMC,KAAK,GAAGrD,EAAE,CAACsD,QAAH,CAAYH,UAAZ,CAAd;AACA,WAAOE,KAAK,CAACE,IAAb;AACD,GALD,CAKE,OAAOzB,KAAP,EAAc,CAEf;;AAED,SAAO,IAAP;AACD","sourcesContent":["import fs from 'fs'; // `fs` will be empty object in browsers (see package.json \"browser\" field).\nimport Response from './response.node';\nimport Headers from './headers.node';\n\nimport {decodeDataUri} from './utils/decode-data-uri.node';\nimport {createReadStream} from './utils/stream-utils.node';\n\nconst isDataURL = (url) => url.startsWith('data:');\nconst isRequestURL = (url) => url.startsWith('http:') || url.startsWith('https:');\n\n/**\n * Emulation of Browser fetch for Node.js\n * @param url\n * @param options\n */\nexport default async function fetchNode(url, options) {\n try {\n // Handle data urls in node, to match `fetch``\n // Note - this loses the MIME type, data URIs are handled directly in fetch\n if (isDataURL(url)) {\n const {arrayBuffer, mimeType} = decodeDataUri(url);\n const response = new Response(arrayBuffer, {\n headers: {'content-type': mimeType},\n url\n });\n return response;\n }\n\n // Automatically decompress gzipped files with .gz extension\n const syntheticResponseHeaders = {};\n const originalUrl = url;\n if (url.endsWith('.gz')) {\n url = url.slice(0, -3);\n syntheticResponseHeaders['content-encoding'] = 'gzip';\n }\n\n // Need to create the stream in advance since Response constructor needs to be sync\n const body = await createReadStream(originalUrl, options);\n const headers = getHeaders(url, body, syntheticResponseHeaders);\n const {status, statusText} = getStatus(body);\n\n const followRedirect =\n !options || options.followRedirect || options.followRedirect === undefined;\n\n if (status >= 300 && status < 400 && headers.has('location') && followRedirect) {\n const redirectUrl = generateRedirectUrl(url, headers.get('location'));\n\n // Redirect\n return await fetchNode(redirectUrl, options);\n }\n return new Response(body, {headers, status, statusText, url});\n } catch (error) {\n // TODO - what error code to use here?\n return new Response(null, {status: 400, statusText: String(error), url});\n }\n}\n\n/**\n * Generate redirect url from location without origin and protocol.\n * @param originalUrl\n * @param redirectUrl\n */\nfunction generateRedirectUrl(originalUrl: string, location: string): string {\n if (location.startsWith('http')) {\n return location;\n }\n // If url doesn't have origin and protocol just extend current url origin with location.\n const url = new URL(originalUrl);\n url.pathname = location;\n\n return url.href;\n}\n\n// HELPER FUNCTIONS\n// PRIVATE\n\nfunction getStatus(httpResponse) {\n if (httpResponse.statusCode) {\n return {status: httpResponse.statusCode, statusText: httpResponse.statusMessage || 'NA'};\n }\n return {status: 200, statusText: 'OK'};\n}\n\nfunction getHeaders(url, httpResponse, additionalHeaders = {}) {\n const headers = {};\n\n if (httpResponse && httpResponse.headers) {\n const httpHeaders = httpResponse.headers;\n for (const key in httpHeaders) {\n const header = httpHeaders[key];\n headers[key.toLowerCase()] = String(header);\n }\n }\n\n // Fix up content length if we can for best progress experience\n if (!headers['content-length']) {\n const contentLength = getContentLength(url);\n if (Number.isFinite(contentLength)) {\n headers['content-length'] = contentLength;\n }\n }\n\n Object.assign(headers, additionalHeaders);\n\n return new Headers(headers);\n}\n\nfunction getContentLength(url) {\n if (isRequestURL(url)) {\n // Needs to be read from actual headers\n return null;\n } else if (isDataURL(url)) {\n // TODO - remove media type etc\n return url.length - 'data:'.length;\n }\n // File URL\n // TODO - how to handle non-existing file, this presumably just throws\n try {\n // strip query params from URL\n const noqueryUrl = url.split('?')[0];\n const stats = fs.statSync(noqueryUrl);\n return stats.size;\n } catch (error) {\n // ignore for now\n }\n\n return null;\n}\n"],"file":"fetch.node.js"}
1
+ {"version":3,"file":"fetch.node.js","names":["fs","Response","Headers","decodeDataUri","createReadStream","isDataURL","url","startsWith","isRequestURL","fetchNode","options","arrayBuffer","mimeType","response","headers","syntheticResponseHeaders","originalUrl","endsWith","slice","body","getHeaders","status","statusText","getStatus","followRedirect","undefined","has","redirectUrl","generateRedirectUrl","get","error","String","location","URL","pathname","href","httpResponse","statusCode","statusMessage","additionalHeaders","httpHeaders","key","header","toLowerCase","contentLength","getContentLength","Number","isFinite","Object","assign","length","noqueryUrl","split","stats","statSync","size"],"sources":["../../../../src/node/fetch/fetch.node.ts"],"sourcesContent":["import fs from 'fs'; // `fs` will be empty object in browsers (see package.json \"browser\" field).\nimport Response from './response.node';\nimport Headers from './headers.node';\n\nimport {decodeDataUri} from './utils/decode-data-uri.node';\nimport {createReadStream} from './utils/stream-utils.node';\n\nconst isDataURL = (url) => url.startsWith('data:');\nconst isRequestURL = (url) => url.startsWith('http:') || url.startsWith('https:');\n\n/**\n * Emulation of Browser fetch for Node.js\n * @param url\n * @param options\n */\nexport default async function fetchNode(url, options) {\n try {\n // Handle data urls in node, to match `fetch``\n // Note - this loses the MIME type, data URIs are handled directly in fetch\n if (isDataURL(url)) {\n const {arrayBuffer, mimeType} = decodeDataUri(url);\n const response = new Response(arrayBuffer, {\n headers: {'content-type': mimeType},\n url\n });\n return response;\n }\n\n // Automatically decompress gzipped files with .gz extension\n const syntheticResponseHeaders = {};\n const originalUrl = url;\n if (url.endsWith('.gz')) {\n url = url.slice(0, -3);\n syntheticResponseHeaders['content-encoding'] = 'gzip';\n }\n\n // Need to create the stream in advance since Response constructor needs to be sync\n const body = await createReadStream(originalUrl, options);\n const headers = getHeaders(url, body, syntheticResponseHeaders);\n const {status, statusText} = getStatus(body);\n\n const followRedirect =\n !options || options.followRedirect || options.followRedirect === undefined;\n\n if (status >= 300 && status < 400 && headers.has('location') && followRedirect) {\n const redirectUrl = generateRedirectUrl(url, headers.get('location'));\n\n // Redirect\n return await fetchNode(redirectUrl, options);\n }\n return new Response(body, {headers, status, statusText, url});\n } catch (error) {\n // TODO - what error code to use here?\n return new Response(null, {status: 400, statusText: String(error), url});\n }\n}\n\n/**\n * Generate redirect url from location without origin and protocol.\n * @param originalUrl\n * @param redirectUrl\n */\nfunction generateRedirectUrl(originalUrl: string, location: string): string {\n if (location.startsWith('http')) {\n return location;\n }\n // If url doesn't have origin and protocol just extend current url origin with location.\n const url = new URL(originalUrl);\n url.pathname = location;\n\n return url.href;\n}\n\n// HELPER FUNCTIONS\n// PRIVATE\n\nfunction getStatus(httpResponse) {\n if (httpResponse.statusCode) {\n return {status: httpResponse.statusCode, statusText: httpResponse.statusMessage || 'NA'};\n }\n return {status: 200, statusText: 'OK'};\n}\n\nfunction getHeaders(url, httpResponse, additionalHeaders = {}) {\n const headers = {};\n\n if (httpResponse && httpResponse.headers) {\n const httpHeaders = httpResponse.headers;\n for (const key in httpHeaders) {\n const header = httpHeaders[key];\n headers[key.toLowerCase()] = String(header);\n }\n }\n\n // Fix up content length if we can for best progress experience\n if (!headers['content-length']) {\n const contentLength = getContentLength(url);\n if (Number.isFinite(contentLength)) {\n headers['content-length'] = contentLength;\n }\n }\n\n Object.assign(headers, additionalHeaders);\n\n return new Headers(headers);\n}\n\nfunction getContentLength(url) {\n if (isRequestURL(url)) {\n // Needs to be read from actual headers\n return null;\n } else if (isDataURL(url)) {\n // TODO - remove media type etc\n return url.length - 'data:'.length;\n }\n // File URL\n // TODO - how to handle non-existing file, this presumably just throws\n try {\n // strip query params from URL\n const noqueryUrl = url.split('?')[0];\n const stats = fs.statSync(noqueryUrl);\n return stats.size;\n } catch (error) {\n // ignore for now\n }\n\n return null;\n}\n"],"mappings":"AAAA,OAAOA,EAAE,MAAM,IAAI;AACnB,OAAOC,QAAQ,MAAM,iBAAiB;AACtC,OAAOC,OAAO,MAAM,gBAAgB;AAEpC,SAAQC,aAAa,QAAO,8BAA8B;AAC1D,SAAQC,gBAAgB,QAAO,2BAA2B;AAE1D,MAAMC,SAAS,GAAIC,GAAG,IAAKA,GAAG,CAACC,UAAU,CAAC,OAAO,CAAC;AAClD,MAAMC,YAAY,GAAIF,GAAG,IAAKA,GAAG,CAACC,UAAU,CAAC,OAAO,CAAC,IAAID,GAAG,CAACC,UAAU,CAAC,QAAQ,CAAC;;AAOjF,eAAe,eAAeE,SAAS,CAACH,GAAG,EAAEI,OAAO,EAAE;EACpD,IAAI;IAGF,IAAIL,SAAS,CAACC,GAAG,CAAC,EAAE;MAClB,MAAM;QAACK,WAAW;QAAEC;MAAQ,CAAC,GAAGT,aAAa,CAACG,GAAG,CAAC;MAClD,MAAMO,QAAQ,GAAG,IAAIZ,QAAQ,CAACU,WAAW,EAAE;QACzCG,OAAO,EAAE;UAAC,cAAc,EAAEF;QAAQ,CAAC;QACnCN;MACF,CAAC,CAAC;MACF,OAAOO,QAAQ;IACjB;;IAGA,MAAME,wBAAwB,GAAG,CAAC,CAAC;IACnC,MAAMC,WAAW,GAAGV,GAAG;IACvB,IAAIA,GAAG,CAACW,QAAQ,CAAC,KAAK,CAAC,EAAE;MACvBX,GAAG,GAAGA,GAAG,CAACY,KAAK,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC;MACtBH,wBAAwB,CAAC,kBAAkB,CAAC,GAAG,MAAM;IACvD;;IAGA,MAAMI,IAAI,GAAG,MAAMf,gBAAgB,CAACY,WAAW,EAAEN,OAAO,CAAC;IACzD,MAAMI,OAAO,GAAGM,UAAU,CAACd,GAAG,EAAEa,IAAI,EAAEJ,wBAAwB,CAAC;IAC/D,MAAM;MAACM,MAAM;MAAEC;IAAU,CAAC,GAAGC,SAAS,CAACJ,IAAI,CAAC;IAE5C,MAAMK,cAAc,GAClB,CAACd,OAAO,IAAIA,OAAO,CAACc,cAAc,IAAId,OAAO,CAACc,cAAc,KAAKC,SAAS;IAE5E,IAAIJ,MAAM,IAAI,GAAG,IAAIA,MAAM,GAAG,GAAG,IAAIP,OAAO,CAACY,GAAG,CAAC,UAAU,CAAC,IAAIF,cAAc,EAAE;MAC9E,MAAMG,WAAW,GAAGC,mBAAmB,CAACtB,GAAG,EAAEQ,OAAO,CAACe,GAAG,CAAC,UAAU,CAAC,CAAC;;MAGrE,OAAO,MAAMpB,SAAS,CAACkB,WAAW,EAAEjB,OAAO,CAAC;IAC9C;IACA,OAAO,IAAIT,QAAQ,CAACkB,IAAI,EAAE;MAACL,OAAO;MAAEO,MAAM;MAAEC,UAAU;MAAEhB;IAAG,CAAC,CAAC;EAC/D,CAAC,CAAC,OAAOwB,KAAK,EAAE;IAEd,OAAO,IAAI7B,QAAQ,CAAC,IAAI,EAAE;MAACoB,MAAM,EAAE,GAAG;MAAEC,UAAU,EAAES,MAAM,CAACD,KAAK,CAAC;MAAExB;IAAG,CAAC,CAAC;EAC1E;AACF;;AAOA,SAASsB,mBAAmB,CAACZ,WAAmB,EAAEgB,QAAgB,EAAU;EAC1E,IAAIA,QAAQ,CAACzB,UAAU,CAAC,MAAM,CAAC,EAAE;IAC/B,OAAOyB,QAAQ;EACjB;EAEA,MAAM1B,GAAG,GAAG,IAAI2B,GAAG,CAACjB,WAAW,CAAC;EAChCV,GAAG,CAAC4B,QAAQ,GAAGF,QAAQ;EAEvB,OAAO1B,GAAG,CAAC6B,IAAI;AACjB;;AAKA,SAASZ,SAAS,CAACa,YAAY,EAAE;EAC/B,IAAIA,YAAY,CAACC,UAAU,EAAE;IAC3B,OAAO;MAAChB,MAAM,EAAEe,YAAY,CAACC,UAAU;MAAEf,UAAU,EAAEc,YAAY,CAACE,aAAa,IAAI;IAAI,CAAC;EAC1F;EACA,OAAO;IAACjB,MAAM,EAAE,GAAG;IAAEC,UAAU,EAAE;EAAI,CAAC;AACxC;AAEA,SAASF,UAAU,CAACd,GAAG,EAAE8B,YAAY,EAA0B;EAAA,IAAxBG,iBAAiB,uEAAG,CAAC,CAAC;EAC3D,MAAMzB,OAAO,GAAG,CAAC,CAAC;EAElB,IAAIsB,YAAY,IAAIA,YAAY,CAACtB,OAAO,EAAE;IACxC,MAAM0B,WAAW,GAAGJ,YAAY,CAACtB,OAAO;IACxC,KAAK,MAAM2B,GAAG,IAAID,WAAW,EAAE;MAC7B,MAAME,MAAM,GAAGF,WAAW,CAACC,GAAG,CAAC;MAC/B3B,OAAO,CAAC2B,GAAG,CAACE,WAAW,EAAE,CAAC,GAAGZ,MAAM,CAACW,MAAM,CAAC;IAC7C;EACF;;EAGA,IAAI,CAAC5B,OAAO,CAAC,gBAAgB,CAAC,EAAE;IAC9B,MAAM8B,aAAa,GAAGC,gBAAgB,CAACvC,GAAG,CAAC;IAC3C,IAAIwC,MAAM,CAACC,QAAQ,CAACH,aAAa,CAAC,EAAE;MAClC9B,OAAO,CAAC,gBAAgB,CAAC,GAAG8B,aAAa;IAC3C;EACF;EAEAI,MAAM,CAACC,MAAM,CAACnC,OAAO,EAAEyB,iBAAiB,CAAC;EAEzC,OAAO,IAAIrC,OAAO,CAACY,OAAO,CAAC;AAC7B;AAEA,SAAS+B,gBAAgB,CAACvC,GAAG,EAAE;EAC7B,IAAIE,YAAY,CAACF,GAAG,CAAC,EAAE;IAErB,OAAO,IAAI;EACb,CAAC,MAAM,IAAID,SAAS,CAACC,GAAG,CAAC,EAAE;IAEzB,OAAOA,GAAG,CAAC4C,MAAM,GAAG,OAAO,CAACA,MAAM;EACpC;EAGA,IAAI;IAEF,MAAMC,UAAU,GAAG7C,GAAG,CAAC8C,KAAK,CAAC,GAAG,CAAC,CAAC,CAAC,CAAC;IACpC,MAAMC,KAAK,GAAGrD,EAAE,CAACsD,QAAQ,CAACH,UAAU,CAAC;IACrC,OAAOE,KAAK,CAACE,IAAI;EACnB,CAAC,CAAC,OAAOzB,KAAK,EAAE;EAEhB;EAEA,OAAO,IAAI;AACb"}
@@ -1,14 +1,10 @@
1
1
  import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
2
-
3
2
  let _Symbol$iterator;
4
-
5
3
  _Symbol$iterator = Symbol.iterator;
6
4
  export default class Headers {
7
5
  constructor(headers) {
8
6
  _defineProperty(this, "map", void 0);
9
-
10
7
  this.map = {};
11
-
12
8
  if (headers instanceof Headers) {
13
9
  headers.forEach((value, name) => this.append(name, value));
14
10
  } else if (Array.isArray(headers)) {
@@ -17,32 +13,27 @@ export default class Headers {
17
13
  Object.getOwnPropertyNames(headers).forEach(name => this.append(name, headers[name]));
18
14
  }
19
15
  }
20
-
21
16
  append(name, value) {
22
17
  name = normalizeName(name);
23
18
  value = normalizeValue(value);
24
19
  const oldValue = this.map[name];
25
20
  this.map[name] = oldValue ? "".concat(oldValue, ", ").concat(value) : value;
26
21
  }
27
-
28
22
  delete(name) {
29
23
  delete this.map[normalizeName(name)];
30
24
  }
31
-
32
25
  get(name) {
33
26
  name = normalizeName(name);
34
27
  return this.has(name) ? this.map[name] : null;
35
28
  }
36
-
37
29
  has(name) {
38
30
  return this.map.hasOwnProperty(normalizeName(name));
39
31
  }
40
-
41
32
  set(name, value) {
42
33
  this.map[normalizeName(name)] = normalizeValue(value);
43
34
  }
44
-
45
- forEach(visitor, thisArg = null) {
35
+ forEach(visitor) {
36
+ let thisArg = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : null;
46
37
  for (const name in this.map) {
47
38
  if (this.map.hasOwnProperty(name)) {
48
39
  if (thisArg) {
@@ -53,7 +44,6 @@ export default class Headers {
53
44
  }
54
45
  }
55
46
  }
56
-
57
47
  keys() {
58
48
  const items = [];
59
49
  this.forEach(function (value, name) {
@@ -61,7 +51,6 @@ export default class Headers {
61
51
  });
62
52
  return iteratorFor(items);
63
53
  }
64
-
65
54
  values() {
66
55
  const items = [];
67
56
  this.forEach(function (value) {
@@ -69,7 +58,6 @@ export default class Headers {
69
58
  });
70
59
  return iteratorFor(items);
71
60
  }
72
-
73
61
  entries() {
74
62
  const items = [];
75
63
  this.forEach(function (value, name) {
@@ -77,30 +65,23 @@ export default class Headers {
77
65
  });
78
66
  return iteratorFor(items);
79
67
  }
80
-
81
68
  *[_Symbol$iterator]() {
82
69
  yield* this.entries();
83
70
  }
84
-
85
71
  }
86
-
87
72
  function normalizeName(name) {
88
73
  if (typeof name !== 'string') {
89
74
  name = String(name);
90
75
  }
91
-
92
76
  if (/[^a-z0-9\-#$%&'*+.^_`|~]/i.test(name) || name === '') {
93
77
  throw new TypeError('Invalid character in header field name');
94
78
  }
95
-
96
79
  return name.toLowerCase();
97
80
  }
98
-
99
81
  function normalizeValue(value) {
100
82
  if (typeof value !== 'string') {
101
83
  value = String(value);
102
84
  }
103
-
104
85
  return value;
105
86
  }
106
87
 
@@ -113,13 +94,10 @@ function iteratorFor(items) {
113
94
  value
114
95
  };
115
96
  }
116
-
117
97
  };
118
-
119
98
  iterator[Symbol.iterator] = function () {
120
99
  return iterator;
121
100
  };
122
-
123
101
  return iterator;
124
102
  }
125
103
  //# sourceMappingURL=headers.node.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/node/fetch/headers.node.ts"],"names":["Symbol","iterator","Headers","constructor","headers","map","forEach","value","name","append","Array","isArray","header","Object","getOwnPropertyNames","normalizeName","normalizeValue","oldValue","delete","get","has","hasOwnProperty","set","visitor","thisArg","call","keys","items","push","iteratorFor","values","entries","String","test","TypeError","toLowerCase","next","shift","done","undefined"],"mappings":";;;;mBA+EIA,MAAM,CAACC,Q;AA3EX,eAAe,MAAMC,OAAN,CAAc;AAG3BC,EAAAA,WAAW,CAACC,OAAD,EAAU;AAAA;;AACnB,SAAKC,GAAL,GAAW,EAAX;;AAEA,QAAID,OAAO,YAAYF,OAAvB,EAAgC;AAC9BE,MAAAA,OAAO,CAACE,OAAR,CAAgB,CAACC,KAAD,EAAQC,IAAR,KAAiB,KAAKC,MAAL,CAAYD,IAAZ,EAAkBD,KAAlB,CAAjC;AACD,KAFD,MAEO,IAAIG,KAAK,CAACC,OAAN,CAAcP,OAAd,CAAJ,EAA4B;AACjCA,MAAAA,OAAO,CAACE,OAAR,CAAiBM,MAAD,IAAY,KAAKH,MAAL,CAAYG,MAAM,CAAC,CAAD,CAAlB,EAAuBA,MAAM,CAAC,CAAD,CAA7B,CAA5B;AACD,KAFM,MAEA,IAAIR,OAAJ,EAAa;AAClBS,MAAAA,MAAM,CAACC,mBAAP,CAA2BV,OAA3B,EAAoCE,OAApC,CAA6CE,IAAD,IAAU,KAAKC,MAAL,CAAYD,IAAZ,EAAkBJ,OAAO,CAACI,IAAD,CAAzB,CAAtD;AACD;AACF;;AAEDC,EAAAA,MAAM,CAACD,IAAD,EAAOD,KAAP,EAAc;AAClBC,IAAAA,IAAI,GAAGO,aAAa,CAACP,IAAD,CAApB;AACAD,IAAAA,KAAK,GAAGS,cAAc,CAACT,KAAD,CAAtB;AACA,UAAMU,QAAQ,GAAG,KAAKZ,GAAL,CAASG,IAAT,CAAjB;AACA,SAAKH,GAAL,CAASG,IAAT,IAAiBS,QAAQ,aAAMA,QAAN,eAAmBV,KAAnB,IAA6BA,KAAtD;AACD;;AAEDW,EAAAA,MAAM,CAACV,IAAD,EAAO;AACX,WAAO,KAAKH,GAAL,CAASU,aAAa,CAACP,IAAD,CAAtB,CAAP;AACD;;AAEDW,EAAAA,GAAG,CAACX,IAAD,EAAO;AACRA,IAAAA,IAAI,GAAGO,aAAa,CAACP,IAAD,CAApB;AACA,WAAO,KAAKY,GAAL,CAASZ,IAAT,IAAiB,KAAKH,GAAL,CAASG,IAAT,CAAjB,GAAkC,IAAzC;AACD;;AAEDY,EAAAA,GAAG,CAACZ,IAAD,EAAO;AACR,WAAO,KAAKH,GAAL,CAASgB,cAAT,CAAwBN,aAAa,CAACP,IAAD,CAArC,CAAP;AACD;;AAEDc,EAAAA,GAAG,CAACd,IAAD,EAAOD,KAAP,EAAc;AACf,SAAKF,GAAL,CAASU,aAAa,CAACP,IAAD,CAAtB,IAAgCQ,cAAc,CAACT,KAAD,CAA9C;AACD;;AAEDD,EAAAA,OAAO,CAACiB,OAAD,EAAUC,OAAO,GAAG,IAApB,EAA0B;AAC/B,SAAK,MAAMhB,IAAX,IAAmB,KAAKH,GAAxB,EAA6B;AAC3B,UAAI,KAAKA,GAAL,CAASgB,cAAT,CAAwBb,IAAxB,CAAJ,EAAmC;AACjC,YAAIgB,OAAJ,EAAa;AACXD,UAAAA,OAAO,CAACE,IAAR,CAAaD,OAAb,EAAsB,KAAKnB,GAAL,CAASG,IAAT,CAAtB,EAAsCA,IAAtC,EAA4C,IAA5C;AACD,SAFD,MAEO;AACLe,UAAAA,OAAO,CAAC,KAAKlB,GAAL,CAASG,IAAT,CAAD,EAAiBA,IAAjB,EAAuB,IAAvB,CAAP;AACD;AACF;AACF;AACF;;AAEDkB,EAAAA,IAAI,GAAG;AACL,UAAMC,KAAY,GAAG,EAArB;AACA,SAAKrB,OAAL,CAAa,UAAUC,KAAV,EAAiBC,IAAjB,EAAuB;AAClCmB,MAAAA,KAAK,CAACC,IAAN,CAAWpB,IAAX;AACD,KAFD;AAGA,WAAOqB,WAAW,CAACF,KAAD,CAAlB;AACD;;AAEDG,EAAAA,MAAM,GAAG;AACP,UAAMH,KAAY,GAAG,EAArB;AACA,SAAKrB,OAAL,CAAa,UAAUC,KAAV,EAAiB;AAC5BoB,MAAAA,KAAK,CAACC,IAAN,CAAWrB,KAAX;AACD,KAFD;AAGA,WAAOsB,WAAW,CAACF,KAAD,CAAlB;AACD;;AAEDI,EAAAA,OAAO,GAAG;AACR,UAAMJ,KAAY,GAAG,EAArB;AACA,SAAKrB,OAAL,CAAa,UAAUC,KAAV,EAAiBC,IAAjB,EAAuB;AAClCmB,MAAAA,KAAK,CAACC,IAAN,CAAW,CAACpB,IAAD,EAAOD,KAAP,CAAX;AACD,KAFD;AAGA,WAAOsB,WAAW,CAACF,KAAD,CAAlB;AACD;;AAED,wBAAqB;AAEnB,WAAO,KAAKI,OAAL,EAAP;AACD;;AA9E0B;;AAiF7B,SAAShB,aAAT,CAAuBP,IAAvB,EAA6B;AAC3B,MAAI,OAAOA,IAAP,KAAgB,QAApB,EAA8B;AAC5BA,IAAAA,IAAI,GAAGwB,MAAM,CAACxB,IAAD,CAAb;AACD;;AACD,MAAI,4BAA4ByB,IAA5B,CAAiCzB,IAAjC,KAA0CA,IAAI,KAAK,EAAvD,EAA2D;AACzD,UAAM,IAAI0B,SAAJ,CAAc,wCAAd,CAAN;AACD;;AACD,SAAO1B,IAAI,CAAC2B,WAAL,EAAP;AACD;;AAED,SAASnB,cAAT,CAAwBT,KAAxB,EAA+B;AAC7B,MAAI,OAAOA,KAAP,KAAiB,QAArB,EAA+B;AAC7BA,IAAAA,KAAK,GAAGyB,MAAM,CAACzB,KAAD,CAAd;AACD;;AACD,SAAOA,KAAP;AACD;;AAGD,SAASsB,WAAT,CAAqBF,KAArB,EAA4B;AAC1B,QAAM1B,QAAQ,GAAG;AACfmC,IAAAA,IAAI,GAAG;AACL,YAAM7B,KAAK,GAAGoB,KAAK,CAACU,KAAN,EAAd;AACA,aAAO;AAACC,QAAAA,IAAI,EAAE/B,KAAK,KAAKgC,SAAjB;AAA4BhC,QAAAA;AAA5B,OAAP;AACD;;AAJc,GAAjB;;AAOAN,EAAAA,QAAQ,CAACD,MAAM,CAACC,QAAR,CAAR,GAA4B,YAAY;AACtC,WAAOA,QAAP;AACD,GAFD;;AAIA,SAAOA,QAAP;AACD","sourcesContent":["/**\n * Polyfill for Browser Headers\n * Based on https://github.com/github/fetch under MIT license\n */\nexport default class Headers {\n map: {};\n\n constructor(headers) {\n this.map = {};\n\n if (headers instanceof Headers) {\n headers.forEach((value, name) => this.append(name, value));\n } else if (Array.isArray(headers)) {\n headers.forEach((header) => this.append(header[0], header[1]));\n } else if (headers) {\n Object.getOwnPropertyNames(headers).forEach((name) => this.append(name, headers[name]));\n }\n }\n\n append(name, value) {\n name = normalizeName(name);\n value = normalizeValue(value);\n const oldValue = this.map[name];\n this.map[name] = oldValue ? `${oldValue}, ${value}` : value;\n }\n\n delete(name) {\n delete this.map[normalizeName(name)];\n }\n\n get(name) {\n name = normalizeName(name);\n return this.has(name) ? this.map[name] : null;\n }\n\n has(name) {\n return this.map.hasOwnProperty(normalizeName(name));\n }\n\n set(name, value) {\n this.map[normalizeName(name)] = normalizeValue(value);\n }\n\n forEach(visitor, thisArg = null) {\n for (const name in this.map) {\n if (this.map.hasOwnProperty(name)) {\n if (thisArg) {\n visitor.call(thisArg, this.map[name], name, this);\n } else {\n visitor(this.map[name], name, this);\n }\n }\n }\n }\n\n keys() {\n const items: any[] = [];\n this.forEach(function (value, name) {\n items.push(name);\n });\n return iteratorFor(items);\n }\n\n values() {\n const items: any[] = [];\n this.forEach(function (value) {\n items.push(value);\n });\n return iteratorFor(items);\n }\n\n entries() {\n const items: any[] = [];\n this.forEach(function (value, name) {\n items.push([name, value]);\n });\n return iteratorFor(items);\n }\n\n *[Symbol.iterator]() {\n // @ts-ignore must have a '[Symbol.iterator]()' method that returns an iterator.\n yield* this.entries();\n }\n}\n\nfunction normalizeName(name) {\n if (typeof name !== 'string') {\n name = String(name);\n }\n if (/[^a-z0-9\\-#$%&'*+.^_`|~]/i.test(name) || name === '') {\n throw new TypeError('Invalid character in header field name');\n }\n return name.toLowerCase();\n}\n\nfunction normalizeValue(value) {\n if (typeof value !== 'string') {\n value = String(value);\n }\n return value;\n}\n\n// Build a destructive iterator for the value list\nfunction iteratorFor(items) {\n const iterator = {\n next() {\n const value = items.shift();\n return {done: value === undefined, value};\n }\n };\n\n iterator[Symbol.iterator] = function () {\n return iterator;\n };\n\n return iterator;\n}\n"],"file":"headers.node.js"}
1
+ {"version":3,"file":"headers.node.js","names":["Symbol","iterator","Headers","constructor","headers","map","forEach","value","name","append","Array","isArray","header","Object","getOwnPropertyNames","normalizeName","normalizeValue","oldValue","delete","get","has","hasOwnProperty","set","visitor","thisArg","call","keys","items","push","iteratorFor","values","entries","String","test","TypeError","toLowerCase","next","shift","done","undefined"],"sources":["../../../../src/node/fetch/headers.node.ts"],"sourcesContent":["/**\n * Polyfill for Browser Headers\n * Based on https://github.com/github/fetch under MIT license\n */\nexport default class Headers {\n map: {};\n\n constructor(headers) {\n this.map = {};\n\n if (headers instanceof Headers) {\n headers.forEach((value, name) => this.append(name, value));\n } else if (Array.isArray(headers)) {\n headers.forEach((header) => this.append(header[0], header[1]));\n } else if (headers) {\n Object.getOwnPropertyNames(headers).forEach((name) => this.append(name, headers[name]));\n }\n }\n\n append(name, value) {\n name = normalizeName(name);\n value = normalizeValue(value);\n const oldValue = this.map[name];\n this.map[name] = oldValue ? `${oldValue}, ${value}` : value;\n }\n\n delete(name) {\n delete this.map[normalizeName(name)];\n }\n\n get(name) {\n name = normalizeName(name);\n return this.has(name) ? this.map[name] : null;\n }\n\n has(name) {\n return this.map.hasOwnProperty(normalizeName(name));\n }\n\n set(name, value) {\n this.map[normalizeName(name)] = normalizeValue(value);\n }\n\n forEach(visitor, thisArg = null) {\n for (const name in this.map) {\n if (this.map.hasOwnProperty(name)) {\n if (thisArg) {\n visitor.call(thisArg, this.map[name], name, this);\n } else {\n visitor(this.map[name], name, this);\n }\n }\n }\n }\n\n keys() {\n const items: any[] = [];\n this.forEach(function (value, name) {\n items.push(name);\n });\n return iteratorFor(items);\n }\n\n values() {\n const items: any[] = [];\n this.forEach(function (value) {\n items.push(value);\n });\n return iteratorFor(items);\n }\n\n entries() {\n const items: any[] = [];\n this.forEach(function (value, name) {\n items.push([name, value]);\n });\n return iteratorFor(items);\n }\n\n *[Symbol.iterator]() {\n // @ts-ignore must have a '[Symbol.iterator]()' method that returns an iterator.\n yield* this.entries();\n }\n}\n\nfunction normalizeName(name) {\n if (typeof name !== 'string') {\n name = String(name);\n }\n if (/[^a-z0-9\\-#$%&'*+.^_`|~]/i.test(name) || name === '') {\n throw new TypeError('Invalid character in header field name');\n }\n return name.toLowerCase();\n}\n\nfunction normalizeValue(value) {\n if (typeof value !== 'string') {\n value = String(value);\n }\n return value;\n}\n\n// Build a destructive iterator for the value list\nfunction iteratorFor(items) {\n const iterator = {\n next() {\n const value = items.shift();\n return {done: value === undefined, value};\n }\n };\n\n iterator[Symbol.iterator] = function () {\n return iterator;\n };\n\n return iterator;\n}\n"],"mappings":";;mBA+EIA,MAAM,CAACC,QAAQ;AA3EnB,eAAe,MAAMC,OAAO,CAAC;EAG3BC,WAAW,CAACC,OAAO,EAAE;IAAA;IACnB,IAAI,CAACC,GAAG,GAAG,CAAC,CAAC;IAEb,IAAID,OAAO,YAAYF,OAAO,EAAE;MAC9BE,OAAO,CAACE,OAAO,CAAC,CAACC,KAAK,EAAEC,IAAI,KAAK,IAAI,CAACC,MAAM,CAACD,IAAI,EAAED,KAAK,CAAC,CAAC;IAC5D,CAAC,MAAM,IAAIG,KAAK,CAACC,OAAO,CAACP,OAAO,CAAC,EAAE;MACjCA,OAAO,CAACE,OAAO,CAAEM,MAAM,IAAK,IAAI,CAACH,MAAM,CAACG,MAAM,CAAC,CAAC,CAAC,EAAEA,MAAM,CAAC,CAAC,CAAC,CAAC,CAAC;IAChE,CAAC,MAAM,IAAIR,OAAO,EAAE;MAClBS,MAAM,CAACC,mBAAmB,CAACV,OAAO,CAAC,CAACE,OAAO,CAAEE,IAAI,IAAK,IAAI,CAACC,MAAM,CAACD,IAAI,EAAEJ,OAAO,CAACI,IAAI,CAAC,CAAC,CAAC;IACzF;EACF;EAEAC,MAAM,CAACD,IAAI,EAAED,KAAK,EAAE;IAClBC,IAAI,GAAGO,aAAa,CAACP,IAAI,CAAC;IAC1BD,KAAK,GAAGS,cAAc,CAACT,KAAK,CAAC;IAC7B,MAAMU,QAAQ,GAAG,IAAI,CAACZ,GAAG,CAACG,IAAI,CAAC;IAC/B,IAAI,CAACH,GAAG,CAACG,IAAI,CAAC,GAAGS,QAAQ,aAAMA,QAAQ,eAAKV,KAAK,IAAKA,KAAK;EAC7D;EAEAW,MAAM,CAACV,IAAI,EAAE;IACX,OAAO,IAAI,CAACH,GAAG,CAACU,aAAa,CAACP,IAAI,CAAC,CAAC;EACtC;EAEAW,GAAG,CAACX,IAAI,EAAE;IACRA,IAAI,GAAGO,aAAa,CAACP,IAAI,CAAC;IAC1B,OAAO,IAAI,CAACY,GAAG,CAACZ,IAAI,CAAC,GAAG,IAAI,CAACH,GAAG,CAACG,IAAI,CAAC,GAAG,IAAI;EAC/C;EAEAY,GAAG,CAACZ,IAAI,EAAE;IACR,OAAO,IAAI,CAACH,GAAG,CAACgB,cAAc,CAACN,aAAa,CAACP,IAAI,CAAC,CAAC;EACrD;EAEAc,GAAG,CAACd,IAAI,EAAED,KAAK,EAAE;IACf,IAAI,CAACF,GAAG,CAACU,aAAa,CAACP,IAAI,CAAC,CAAC,GAAGQ,cAAc,CAACT,KAAK,CAAC;EACvD;EAEAD,OAAO,CAACiB,OAAO,EAAkB;IAAA,IAAhBC,OAAO,uEAAG,IAAI;IAC7B,KAAK,MAAMhB,IAAI,IAAI,IAAI,CAACH,GAAG,EAAE;MAC3B,IAAI,IAAI,CAACA,GAAG,CAACgB,cAAc,CAACb,IAAI,CAAC,EAAE;QACjC,IAAIgB,OAAO,EAAE;UACXD,OAAO,CAACE,IAAI,CAACD,OAAO,EAAE,IAAI,CAACnB,GAAG,CAACG,IAAI,CAAC,EAAEA,IAAI,EAAE,IAAI,CAAC;QACnD,CAAC,MAAM;UACLe,OAAO,CAAC,IAAI,CAAClB,GAAG,CAACG,IAAI,CAAC,EAAEA,IAAI,EAAE,IAAI,CAAC;QACrC;MACF;IACF;EACF;EAEAkB,IAAI,GAAG;IACL,MAAMC,KAAY,GAAG,EAAE;IACvB,IAAI,CAACrB,OAAO,CAAC,UAAUC,KAAK,EAAEC,IAAI,EAAE;MAClCmB,KAAK,CAACC,IAAI,CAACpB,IAAI,CAAC;IAClB,CAAC,CAAC;IACF,OAAOqB,WAAW,CAACF,KAAK,CAAC;EAC3B;EAEAG,MAAM,GAAG;IACP,MAAMH,KAAY,GAAG,EAAE;IACvB,IAAI,CAACrB,OAAO,CAAC,UAAUC,KAAK,EAAE;MAC5BoB,KAAK,CAACC,IAAI,CAACrB,KAAK,CAAC;IACnB,CAAC,CAAC;IACF,OAAOsB,WAAW,CAACF,KAAK,CAAC;EAC3B;EAEAI,OAAO,GAAG;IACR,MAAMJ,KAAY,GAAG,EAAE;IACvB,IAAI,CAACrB,OAAO,CAAC,UAAUC,KAAK,EAAEC,IAAI,EAAE;MAClCmB,KAAK,CAACC,IAAI,CAAC,CAACpB,IAAI,EAAED,KAAK,CAAC,CAAC;IAC3B,CAAC,CAAC;IACF,OAAOsB,WAAW,CAACF,KAAK,CAAC;EAC3B;EAEA,sBAAqB;IAEnB,OAAO,IAAI,CAACI,OAAO,EAAE;EACvB;AACF;AAEA,SAAShB,aAAa,CAACP,IAAI,EAAE;EAC3B,IAAI,OAAOA,IAAI,KAAK,QAAQ,EAAE;IAC5BA,IAAI,GAAGwB,MAAM,CAACxB,IAAI,CAAC;EACrB;EACA,IAAI,2BAA2B,CAACyB,IAAI,CAACzB,IAAI,CAAC,IAAIA,IAAI,KAAK,EAAE,EAAE;IACzD,MAAM,IAAI0B,SAAS,CAAC,wCAAwC,CAAC;EAC/D;EACA,OAAO1B,IAAI,CAAC2B,WAAW,EAAE;AAC3B;AAEA,SAASnB,cAAc,CAACT,KAAK,EAAE;EAC7B,IAAI,OAAOA,KAAK,KAAK,QAAQ,EAAE;IAC7BA,KAAK,GAAGyB,MAAM,CAACzB,KAAK,CAAC;EACvB;EACA,OAAOA,KAAK;AACd;;AAGA,SAASsB,WAAW,CAACF,KAAK,EAAE;EAC1B,MAAM1B,QAAQ,GAAG;IACfmC,IAAI,GAAG;MACL,MAAM7B,KAAK,GAAGoB,KAAK,CAACU,KAAK,EAAE;MAC3B,OAAO;QAACC,IAAI,EAAE/B,KAAK,KAAKgC,SAAS;QAAEhC;MAAK,CAAC;IAC3C;EACF,CAAC;EAEDN,QAAQ,CAACD,MAAM,CAACC,QAAQ,CAAC,GAAG,YAAY;IACtC,OAAOA,QAAQ;EACjB,CAAC;EAED,OAAOA,QAAQ;AACjB"}
@@ -2,32 +2,21 @@ import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
2
2
  import { assert } from '../../utils/assert';
3
3
  import { decompressReadStream, concatenateReadStream } from './utils/stream-utils.node';
4
4
  import Headers from './headers.node';
5
-
6
5
  const isBoolean = x => typeof x === 'boolean';
7
-
8
6
  const isFunction = x => typeof x === 'function';
9
-
10
7
  const isObject = x => x !== null && typeof x === 'object';
11
-
12
8
  const isReadableNodeStream = x => isObject(x) && isFunction(x.read) && isFunction(x.pipe) && isBoolean(x.readable);
13
9
 
14
10
  import { Readable } from 'stream';
15
11
  export default class Response {
16
12
  constructor(body, options) {
17
13
  _defineProperty(this, "ok", void 0);
18
-
19
14
  _defineProperty(this, "status", void 0);
20
-
21
15
  _defineProperty(this, "statusText", void 0);
22
-
23
16
  _defineProperty(this, "headers", void 0);
24
-
25
17
  _defineProperty(this, "url", void 0);
26
-
27
18
  _defineProperty(this, "bodyUsed", false);
28
-
29
19
  _defineProperty(this, "_body", void 0);
30
-
31
20
  const {
32
21
  headers,
33
22
  status = 200,
@@ -60,29 +49,23 @@ export default class Response {
60
49
  if (!isReadableNodeStream(this._body)) {
61
50
  return this._body || new ArrayBuffer(0);
62
51
  }
63
-
64
52
  const data = await concatenateReadStream(this._body);
65
53
  return data;
66
54
  }
67
-
68
55
  async text() {
69
56
  const arrayBuffer = await this.arrayBuffer();
70
57
  const textDecoder = new TextDecoder();
71
58
  return textDecoder.decode(arrayBuffer);
72
59
  }
73
-
74
60
  async json() {
75
61
  const text = await this.text();
76
62
  return JSON.parse(text);
77
63
  }
78
-
79
64
  async blob() {
80
65
  if (typeof Blob === 'undefined') {
81
66
  throw new Error('Blob polyfill not installed');
82
67
  }
83
-
84
68
  return new Blob([await this.arrayBuffer()]);
85
69
  }
86
-
87
70
  }
88
71
  //# sourceMappingURL=response.node.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../src/node/fetch/response.node.ts"],"names":["assert","decompressReadStream","concatenateReadStream","Headers","isBoolean","x","isFunction","isObject","isReadableNodeStream","read","pipe","readable","Readable","Response","constructor","body","options","headers","status","statusText","url","ok","_body","from","TextEncoder","encode","ArrayBuffer","bodyUsed","arrayBuffer","data","text","textDecoder","TextDecoder","decode","json","JSON","parse","blob","Blob","Error"],"mappings":";AAAA,SAAQA,MAAR,QAAqB,oBAArB;AACA,SAAQC,oBAAR,EAA8BC,qBAA9B,QAA0D,2BAA1D;AACA,OAAOC,OAAP,MAAoB,gBAApB;;AAEA,MAAMC,SAAS,GAAIC,CAAD,IAAO,OAAOA,CAAP,KAAa,SAAtC;;AACA,MAAMC,UAAU,GAAID,CAAD,IAAO,OAAOA,CAAP,KAAa,UAAvC;;AACA,MAAME,QAAQ,GAAIF,CAAD,IAAOA,CAAC,KAAK,IAAN,IAAc,OAAOA,CAAP,KAAa,QAAnD;;AACA,MAAMG,oBAAoB,GAAIH,CAAD,IAC3BE,QAAQ,CAACF,CAAD,CAAR,IAAeC,UAAU,CAACD,CAAC,CAACI,IAAH,CAAzB,IAAqCH,UAAU,CAACD,CAAC,CAACK,IAAH,CAA/C,IAA2DN,SAAS,CAACC,CAAC,CAACM,QAAH,CADtE;;AAaA,SAAQC,QAAR,QAAuB,QAAvB;AAEA,eAAe,MAAMC,QAAN,CAAe;AAU5BC,EAAAA,WAAW,CACTC,IADS,EAETC,OAFS,EAQT;AAAA;;AAAA;;AAAA;;AAAA;;AAAA;;AAAA,sCAZkB,KAYlB;;AAAA;;AACA,UAAM;AAACC,MAAAA,OAAD;AAAUC,MAAAA,MAAM,GAAG,GAAnB;AAAwBC,MAAAA,UAAU,GAAG,IAArC;AAA2CC,MAAAA;AAA3C,QAAkDJ,OAAO,IAAI,EAAnE;AAEA,SAAKI,GAAL,GAAWA,GAAX;AACA,SAAKC,EAAL,GAAUH,MAAM,KAAK,GAArB;AACA,SAAKA,MAAL,GAAcA,MAAd;AACA,SAAKC,UAAL,GAAkBA,UAAlB;AACA,SAAKF,OAAL,GAAe,IAAId,OAAJ,CAAY,CAAAa,OAAO,SAAP,IAAAA,OAAO,WAAP,YAAAA,OAAO,CAAEC,OAAT,KAAoB,EAAhC,CAAf;;AAGA,QAAIT,oBAAoB,CAACO,IAAD,CAAxB,EAAgC;AAC9B,WAAKO,KAAL,GAAarB,oBAAoB,CAACc,IAAD,EAAOE,OAAP,CAAjC;AACD,KAFD,MAEO,IAAI,OAAOF,IAAP,KAAgB,QAApB,EAA8B;AACnC,WAAKO,KAAL,GAAaV,QAAQ,CAACW,IAAT,CAAc,CAAC,IAAIC,WAAJ,GAAkBC,MAAlB,CAAyBV,IAAzB,CAAD,CAAd,CAAb;AACD,KAFM,MAEA;AACL,WAAKO,KAAL,GAAaV,QAAQ,CAACW,IAAT,CAAc,CAACR,IAAI,IAAI,IAAIW,WAAJ,CAAgB,CAAhB,CAAT,CAAd,CAAb;AACD;AACF;;AAKO,MAAJX,IAAI,GAAG;AACTf,IAAAA,MAAM,CAAC,CAAC,KAAK2B,QAAP,CAAN;AACA3B,IAAAA,MAAM,CAACQ,oBAAoB,CAAC,KAAKc,KAAN,CAArB,CAAN;AACA,SAAKK,QAAL,GAAgB,IAAhB;AACA,WAAO,KAAKL,KAAZ;AACD;;AAIgB,QAAXM,WAAW,GAAG;AAClB,QAAI,CAACpB,oBAAoB,CAAC,KAAKc,KAAN,CAAzB,EAAuC;AACrC,aAAO,KAAKA,KAAL,IAAc,IAAII,WAAJ,CAAgB,CAAhB,CAArB;AACD;;AACD,UAAMG,IAAI,GAAG,MAAM3B,qBAAqB,CAAC,KAAKoB,KAAN,CAAxC;AACA,WAAOO,IAAP;AACD;;AAES,QAAJC,IAAI,GAAG;AACX,UAAMF,WAAW,GAAG,MAAM,KAAKA,WAAL,EAA1B;AACA,UAAMG,WAAW,GAAG,IAAIC,WAAJ,EAApB;AACA,WAAOD,WAAW,CAACE,MAAZ,CAAmBL,WAAnB,CAAP;AACD;;AAES,QAAJM,IAAI,GAAG;AACX,UAAMJ,IAAI,GAAG,MAAM,KAAKA,IAAL,EAAnB;AACA,WAAOK,IAAI,CAACC,KAAL,CAAWN,IAAX,CAAP;AACD;;AAES,QAAJO,IAAI,GAAG;AACX,QAAI,OAAOC,IAAP,KAAgB,WAApB,EAAiC;AAC/B,YAAM,IAAIC,KAAJ,CAAU,6BAAV,CAAN;AACD;;AACD,WAAO,IAAID,IAAJ,CAAS,CAAC,MAAM,KAAKV,WAAL,EAAP,CAAT,CAAP;AACD;;AAzE2B","sourcesContent":["import {assert} from '../../utils/assert';\nimport {decompressReadStream, concatenateReadStream} from './utils/stream-utils.node';\nimport Headers from './headers.node';\n\nconst isBoolean = (x) => typeof x === 'boolean';\nconst isFunction = (x) => typeof x === 'function';\nconst isObject = (x) => x !== null && typeof x === 'object';\nconst isReadableNodeStream = (x) =>\n isObject(x) && isFunction(x.read) && isFunction(x.pipe) && isBoolean(x.readable);\n\n/**\n * Polyfill for Browser Response\n *\n * Under Node.js we return a mock \"fetch response object\"\n * so that apps can use the same API as in the browser.\n *\n * Note: This is intended to be a \"lightweight\" implementation and will have limitations.\n *\n * See https://developer.mozilla.org/en-US/docs/Web/API/Response\n */\nimport {Readable} from 'stream';\n\nexport default class Response {\n readonly ok: boolean;\n readonly status: number;\n readonly statusText: string;\n readonly headers: Headers;\n readonly url: string;\n bodyUsed: boolean = false;\n private readonly _body;\n\n // TODO - handle ArrayBuffer, ArrayBufferView, Buffer\n constructor(\n body,\n options: {\n headers?;\n status?: number;\n statusText?: string;\n url: string;\n }\n ) {\n const {headers, status = 200, statusText = 'OK', url} = options || {};\n\n this.url = url;\n this.ok = status === 200;\n this.status = status; // TODO - handle errors and set status\n this.statusText = statusText;\n this.headers = new Headers(options?.headers || {});\n\n // Check for content-encoding and create a decompression stream\n if (isReadableNodeStream(body)) {\n this._body = decompressReadStream(body, headers);\n } else if (typeof body === 'string') {\n this._body = Readable.from([new TextEncoder().encode(body)]);\n } else {\n this._body = Readable.from([body || new ArrayBuffer(0)]);\n }\n }\n\n // Subset of Properties\n\n // Returns a readable stream to the \"body\" of the response (or file)\n get body() {\n assert(!this.bodyUsed);\n assert(isReadableNodeStream(this._body)); // Not implemented: conversion of ArrayBuffer etc to stream\n this.bodyUsed = true;\n return this._body;\n }\n\n // Subset of Methods\n\n async arrayBuffer() {\n if (!isReadableNodeStream(this._body)) {\n return this._body || new ArrayBuffer(0);\n }\n const data = await concatenateReadStream(this._body);\n return data;\n }\n\n async text() {\n const arrayBuffer = await this.arrayBuffer();\n const textDecoder = new TextDecoder();\n return textDecoder.decode(arrayBuffer);\n }\n\n async json() {\n const text = await this.text();\n return JSON.parse(text);\n }\n\n async blob() {\n if (typeof Blob === 'undefined') {\n throw new Error('Blob polyfill not installed');\n }\n return new Blob([await this.arrayBuffer()]);\n }\n}\n"],"file":"response.node.js"}
1
+ {"version":3,"file":"response.node.js","names":["assert","decompressReadStream","concatenateReadStream","Headers","isBoolean","x","isFunction","isObject","isReadableNodeStream","read","pipe","readable","Readable","Response","constructor","body","options","headers","status","statusText","url","ok","_body","from","TextEncoder","encode","ArrayBuffer","bodyUsed","arrayBuffer","data","text","textDecoder","TextDecoder","decode","json","JSON","parse","blob","Blob","Error"],"sources":["../../../../src/node/fetch/response.node.ts"],"sourcesContent":["import {assert} from '../../utils/assert';\nimport {decompressReadStream, concatenateReadStream} from './utils/stream-utils.node';\nimport Headers from './headers.node';\n\nconst isBoolean = (x) => typeof x === 'boolean';\nconst isFunction = (x) => typeof x === 'function';\nconst isObject = (x) => x !== null && typeof x === 'object';\nconst isReadableNodeStream = (x) =>\n isObject(x) && isFunction(x.read) && isFunction(x.pipe) && isBoolean(x.readable);\n\n/**\n * Polyfill for Browser Response\n *\n * Under Node.js we return a mock \"fetch response object\"\n * so that apps can use the same API as in the browser.\n *\n * Note: This is intended to be a \"lightweight\" implementation and will have limitations.\n *\n * See https://developer.mozilla.org/en-US/docs/Web/API/Response\n */\nimport {Readable} from 'stream';\n\nexport default class Response {\n readonly ok: boolean;\n readonly status: number;\n readonly statusText: string;\n readonly headers: Headers;\n readonly url: string;\n bodyUsed: boolean = false;\n private readonly _body;\n\n // TODO - handle ArrayBuffer, ArrayBufferView, Buffer\n constructor(\n body,\n options: {\n headers?;\n status?: number;\n statusText?: string;\n url: string;\n }\n ) {\n const {headers, status = 200, statusText = 'OK', url} = options || {};\n\n this.url = url;\n this.ok = status === 200;\n this.status = status; // TODO - handle errors and set status\n this.statusText = statusText;\n this.headers = new Headers(options?.headers || {});\n\n // Check for content-encoding and create a decompression stream\n if (isReadableNodeStream(body)) {\n this._body = decompressReadStream(body, headers);\n } else if (typeof body === 'string') {\n this._body = Readable.from([new TextEncoder().encode(body)]);\n } else {\n this._body = Readable.from([body || new ArrayBuffer(0)]);\n }\n }\n\n // Subset of Properties\n\n // Returns a readable stream to the \"body\" of the response (or file)\n get body() {\n assert(!this.bodyUsed);\n assert(isReadableNodeStream(this._body)); // Not implemented: conversion of ArrayBuffer etc to stream\n this.bodyUsed = true;\n return this._body;\n }\n\n // Subset of Methods\n\n async arrayBuffer() {\n if (!isReadableNodeStream(this._body)) {\n return this._body || new ArrayBuffer(0);\n }\n const data = await concatenateReadStream(this._body);\n return data;\n }\n\n async text() {\n const arrayBuffer = await this.arrayBuffer();\n const textDecoder = new TextDecoder();\n return textDecoder.decode(arrayBuffer);\n }\n\n async json() {\n const text = await this.text();\n return JSON.parse(text);\n }\n\n async blob() {\n if (typeof Blob === 'undefined') {\n throw new Error('Blob polyfill not installed');\n }\n return new Blob([await this.arrayBuffer()]);\n }\n}\n"],"mappings":";AAAA,SAAQA,MAAM,QAAO,oBAAoB;AACzC,SAAQC,oBAAoB,EAAEC,qBAAqB,QAAO,2BAA2B;AACrF,OAAOC,OAAO,MAAM,gBAAgB;AAEpC,MAAMC,SAAS,GAAIC,CAAC,IAAK,OAAOA,CAAC,KAAK,SAAS;AAC/C,MAAMC,UAAU,GAAID,CAAC,IAAK,OAAOA,CAAC,KAAK,UAAU;AACjD,MAAME,QAAQ,GAAIF,CAAC,IAAKA,CAAC,KAAK,IAAI,IAAI,OAAOA,CAAC,KAAK,QAAQ;AAC3D,MAAMG,oBAAoB,GAAIH,CAAC,IAC7BE,QAAQ,CAACF,CAAC,CAAC,IAAIC,UAAU,CAACD,CAAC,CAACI,IAAI,CAAC,IAAIH,UAAU,CAACD,CAAC,CAACK,IAAI,CAAC,IAAIN,SAAS,CAACC,CAAC,CAACM,QAAQ,CAAC;;AAYlF,SAAQC,QAAQ,QAAO,QAAQ;AAE/B,eAAe,MAAMC,QAAQ,CAAC;EAU5BC,WAAW,CACTC,IAAI,EACJC,OAKC,EACD;IAAA;IAAA;IAAA;IAAA;IAAA;IAAA,kCAZkB,KAAK;IAAA;IAavB,MAAM;MAACC,OAAO;MAAEC,MAAM,GAAG,GAAG;MAAEC,UAAU,GAAG,IAAI;MAAEC;IAAG,CAAC,GAAGJ,OAAO,IAAI,CAAC,CAAC;IAErE,IAAI,CAACI,GAAG,GAAGA,GAAG;IACd,IAAI,CAACC,EAAE,GAAGH,MAAM,KAAK,GAAG;IACxB,IAAI,CAACA,MAAM,GAAGA,MAAM;IACpB,IAAI,CAACC,UAAU,GAAGA,UAAU;IAC5B,IAAI,CAACF,OAAO,GAAG,IAAId,OAAO,CAAC,CAAAa,OAAO,aAAPA,OAAO,uBAAPA,OAAO,CAAEC,OAAO,KAAI,CAAC,CAAC,CAAC;;IAGlD,IAAIT,oBAAoB,CAACO,IAAI,CAAC,EAAE;MAC9B,IAAI,CAACO,KAAK,GAAGrB,oBAAoB,CAACc,IAAI,EAAEE,OAAO,CAAC;IAClD,CAAC,MAAM,IAAI,OAAOF,IAAI,KAAK,QAAQ,EAAE;MACnC,IAAI,CAACO,KAAK,GAAGV,QAAQ,CAACW,IAAI,CAAC,CAAC,IAAIC,WAAW,EAAE,CAACC,MAAM,CAACV,IAAI,CAAC,CAAC,CAAC;IAC9D,CAAC,MAAM;MACL,IAAI,CAACO,KAAK,GAAGV,QAAQ,CAACW,IAAI,CAAC,CAACR,IAAI,IAAI,IAAIW,WAAW,CAAC,CAAC,CAAC,CAAC,CAAC;IAC1D;EACF;;EAKA,IAAIX,IAAI,GAAG;IACTf,MAAM,CAAC,CAAC,IAAI,CAAC2B,QAAQ,CAAC;IACtB3B,MAAM,CAACQ,oBAAoB,CAAC,IAAI,CAACc,KAAK,CAAC,CAAC;IACxC,IAAI,CAACK,QAAQ,GAAG,IAAI;IACpB,OAAO,IAAI,CAACL,KAAK;EACnB;;EAIA,MAAMM,WAAW,GAAG;IAClB,IAAI,CAACpB,oBAAoB,CAAC,IAAI,CAACc,KAAK,CAAC,EAAE;MACrC,OAAO,IAAI,CAACA,KAAK,IAAI,IAAII,WAAW,CAAC,CAAC,CAAC;IACzC;IACA,MAAMG,IAAI,GAAG,MAAM3B,qBAAqB,CAAC,IAAI,CAACoB,KAAK,CAAC;IACpD,OAAOO,IAAI;EACb;EAEA,MAAMC,IAAI,GAAG;IACX,MAAMF,WAAW,GAAG,MAAM,IAAI,CAACA,WAAW,EAAE;IAC5C,MAAMG,WAAW,GAAG,IAAIC,WAAW,EAAE;IACrC,OAAOD,WAAW,CAACE,MAAM,CAACL,WAAW,CAAC;EACxC;EAEA,MAAMM,IAAI,GAAG;IACX,MAAMJ,IAAI,GAAG,MAAM,IAAI,CAACA,IAAI,EAAE;IAC9B,OAAOK,IAAI,CAACC,KAAK,CAACN,IAAI,CAAC;EACzB;EAEA,MAAMO,IAAI,GAAG;IACX,IAAI,OAAOC,IAAI,KAAK,WAAW,EAAE;MAC/B,MAAM,IAAIC,KAAK,CAAC,6BAA6B,CAAC;IAChD;IACA,OAAO,IAAID,IAAI,CAAC,CAAC,MAAM,IAAI,CAACV,WAAW,EAAE,CAAC,CAAC;EAC7C;AACF"}
@@ -1,12 +1,12 @@
1
- const isArrayBuffer = x => x && x instanceof ArrayBuffer;
2
1
 
2
+
3
+ const isArrayBuffer = x => x && x instanceof ArrayBuffer;
3
4
  const isBuffer = x => x && x instanceof Buffer;
4
5
 
5
6
  export function decodeDataUri(uri) {
6
7
  const dataIndex = uri.indexOf(',');
7
8
  let buffer;
8
9
  let mimeType;
9
-
10
10
  if (uri.slice(dataIndex - 7, dataIndex) === ';base64') {
11
11
  buffer = Buffer.from(uri.slice(dataIndex + 1), 'base64');
12
12
  mimeType = uri.slice(5, dataIndex - 7).trim();
@@ -14,18 +14,17 @@ export function decodeDataUri(uri) {
14
14
  buffer = Buffer.from(decodeURIComponent(uri.slice(dataIndex + 1)));
15
15
  mimeType = uri.slice(5, dataIndex).trim();
16
16
  }
17
-
18
17
  if (!mimeType) {
19
18
  mimeType = 'text/plain;charset=US-ASCII';
20
19
  } else if (mimeType.startsWith(';')) {
21
20
  mimeType = "text/plain".concat(mimeType);
22
21
  }
23
-
24
22
  return {
25
23
  arrayBuffer: toArrayBuffer(buffer),
26
24
  mimeType
27
25
  };
28
26
  }
27
+
29
28
  export function toArrayBuffer(data) {
30
29
  if (isArrayBuffer(data)) {
31
30
  return data;
@@ -39,7 +38,6 @@ export function toArrayBuffer(data) {
39
38
  if (ArrayBuffer.isView(data)) {
40
39
  return data.buffer;
41
40
  }
42
-
43
41
  if (typeof data === 'string') {
44
42
  const text = data;
45
43
  const uint8Array = new TextEncoder().encode(text);
@@ -49,7 +47,6 @@ export function toArrayBuffer(data) {
49
47
  if (data && typeof data === 'object' && data._toArrayBuffer) {
50
48
  return data._toArrayBuffer();
51
49
  }
52
-
53
50
  throw new Error("toArrayBuffer(".concat(JSON.stringify(data, null, 2).slice(10), ")"));
54
51
  }
55
52
  //# sourceMappingURL=decode-data-uri.node.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../../src/node/fetch/utils/decode-data-uri.node.ts"],"names":["isArrayBuffer","x","ArrayBuffer","isBuffer","Buffer","decodeDataUri","uri","dataIndex","indexOf","buffer","mimeType","slice","from","trim","decodeURIComponent","startsWith","arrayBuffer","toArrayBuffer","data","typedArray","Uint8Array","isView","text","uint8Array","TextEncoder","encode","_toArrayBuffer","Error","JSON","stringify"],"mappings":"AAEA,MAAMA,aAAa,GAAIC,CAAD,IAAOA,CAAC,IAAIA,CAAC,YAAYC,WAA/C;;AACA,MAAMC,QAAQ,GAAIF,CAAD,IAAOA,CAAC,IAAIA,CAAC,YAAYG,MAA1C;;AAQA,OAAO,SAASC,aAAT,CAAuBC,GAAvB,EAAkF;AACvF,QAAMC,SAAS,GAAGD,GAAG,CAACE,OAAJ,CAAY,GAAZ,CAAlB;AAEA,MAAIC,MAAJ;AACA,MAAIC,QAAJ;;AACA,MAAIJ,GAAG,CAACK,KAAJ,CAAUJ,SAAS,GAAG,CAAtB,EAAyBA,SAAzB,MAAwC,SAA5C,EAAuD;AACrDE,IAAAA,MAAM,GAAGL,MAAM,CAACQ,IAAP,CAAYN,GAAG,CAACK,KAAJ,CAAUJ,SAAS,GAAG,CAAtB,CAAZ,EAAsC,QAAtC,CAAT;AACAG,IAAAA,QAAQ,GAAGJ,GAAG,CAACK,KAAJ,CAAU,CAAV,EAAaJ,SAAS,GAAG,CAAzB,EAA4BM,IAA5B,EAAX;AACD,GAHD,MAGO;AACLJ,IAAAA,MAAM,GAAGL,MAAM,CAACQ,IAAP,CAAYE,kBAAkB,CAACR,GAAG,CAACK,KAAJ,CAAUJ,SAAS,GAAG,CAAtB,CAAD,CAA9B,CAAT;AACAG,IAAAA,QAAQ,GAAGJ,GAAG,CAACK,KAAJ,CAAU,CAAV,EAAaJ,SAAb,EAAwBM,IAAxB,EAAX;AACD;;AAED,MAAI,CAACH,QAAL,EAAe;AACbA,IAAAA,QAAQ,GAAG,6BAAX;AACD,GAFD,MAEO,IAAIA,QAAQ,CAACK,UAAT,CAAoB,GAApB,CAAJ,EAA8B;AACnCL,IAAAA,QAAQ,uBAAgBA,QAAhB,CAAR;AACD;;AAED,SAAO;AAACM,IAAAA,WAAW,EAAEC,aAAa,CAACR,MAAD,CAA3B;AAAqCC,IAAAA;AAArC,GAAP;AACD;AAMD,OAAO,SAASO,aAAT,CAAuBC,IAAvB,EAAmD;AACxD,MAAIlB,aAAa,CAACkB,IAAD,CAAjB,EAAyB;AACvB,WAAOA,IAAP;AACD;;AAGD,MAAIf,QAAQ,CAACe,IAAD,CAAZ,EAAoB;AAElB,UAAMC,UAAU,GAAG,IAAIC,UAAJ,CAAeF,IAAf,CAAnB;AACA,WAAOC,UAAU,CAACV,MAAlB;AACD;;AAGD,MAAIP,WAAW,CAACmB,MAAZ,CAAmBH,IAAnB,CAAJ,EAA8B;AAC5B,WAAOA,IAAI,CAACT,MAAZ;AACD;;AAED,MAAI,OAAOS,IAAP,KAAgB,QAApB,EAA8B;AAC5B,UAAMI,IAAI,GAAGJ,IAAb;AACA,UAAMK,UAAU,GAAG,IAAIC,WAAJ,GAAkBC,MAAlB,CAAyBH,IAAzB,CAAnB;AACA,WAAOC,UAAU,CAACd,MAAlB;AACD;;AAID,MAAIS,IAAI,IAAI,OAAOA,IAAP,KAAgB,QAAxB,IAAoCA,IAAI,CAACQ,cAA7C,EAA6D;AAE3D,WAAOR,IAAI,CAACQ,cAAL,EAAP;AACD;;AAED,QAAM,IAAIC,KAAJ,yBAA2BC,IAAI,CAACC,SAAL,CAAeX,IAAf,EAAqB,IAArB,EAA2B,CAA3B,EAA8BP,KAA9B,CAAoC,EAApC,CAA3B,OAAN;AACD","sourcesContent":["// Based on binary-gltf-utils under MIT license: Copyright (c) 2016-17 Karl Cheng\n\nconst isArrayBuffer = (x) => x && x instanceof ArrayBuffer;\nconst isBuffer = (x) => x && x instanceof Buffer;\n\n/**\n * Parses a data URI into a buffer, as well as retrieving its declared MIME type.\n *\n * @param {string} uri - a data URI (assumed to be valid)\n * @returns {Object} { buffer, mimeType }\n */\nexport function decodeDataUri(uri: string): {arrayBuffer: ArrayBuffer; mimeType: string} {\n const dataIndex = uri.indexOf(',');\n\n let buffer;\n let mimeType;\n if (uri.slice(dataIndex - 7, dataIndex) === ';base64') {\n buffer = Buffer.from(uri.slice(dataIndex + 1), 'base64');\n mimeType = uri.slice(5, dataIndex - 7).trim();\n } else {\n buffer = Buffer.from(decodeURIComponent(uri.slice(dataIndex + 1)));\n mimeType = uri.slice(5, dataIndex).trim();\n }\n\n if (!mimeType) {\n mimeType = 'text/plain;charset=US-ASCII';\n } else if (mimeType.startsWith(';')) {\n mimeType = `text/plain${mimeType}`;\n }\n\n return {arrayBuffer: toArrayBuffer(buffer), mimeType};\n}\n\n/**\n * @param data\n * @todo Duplicate of core\n */\nexport function toArrayBuffer(data: unknown): ArrayBuffer {\n if (isArrayBuffer(data)) {\n return data as ArrayBuffer;\n }\n\n // TODO - per docs we should just be able to call buffer.buffer, but there are issues\n if (isBuffer(data)) {\n // @ts-expect-error\n const typedArray = new Uint8Array(data);\n return typedArray.buffer;\n }\n\n // Careful - Node Buffers will look like ArrayBuffers (keep after isBuffer)\n if (ArrayBuffer.isView(data)) {\n return data.buffer;\n }\n\n if (typeof data === 'string') {\n const text = data;\n const uint8Array = new TextEncoder().encode(text);\n return uint8Array.buffer;\n }\n\n // HACK to support Blob polyfill\n // @ts-expect-error\n if (data && typeof data === 'object' && data._toArrayBuffer) {\n // @ts-expect-error\n return data._toArrayBuffer();\n }\n\n throw new Error(`toArrayBuffer(${JSON.stringify(data, null, 2).slice(10)})`);\n}\n"],"file":"decode-data-uri.node.js"}
1
+ {"version":3,"file":"decode-data-uri.node.js","names":["isArrayBuffer","x","ArrayBuffer","isBuffer","Buffer","decodeDataUri","uri","dataIndex","indexOf","buffer","mimeType","slice","from","trim","decodeURIComponent","startsWith","arrayBuffer","toArrayBuffer","data","typedArray","Uint8Array","isView","text","uint8Array","TextEncoder","encode","_toArrayBuffer","Error","JSON","stringify"],"sources":["../../../../../src/node/fetch/utils/decode-data-uri.node.ts"],"sourcesContent":["// Based on binary-gltf-utils under MIT license: Copyright (c) 2016-17 Karl Cheng\n\nconst isArrayBuffer = (x) => x && x instanceof ArrayBuffer;\nconst isBuffer = (x) => x && x instanceof Buffer;\n\n/**\n * Parses a data URI into a buffer, as well as retrieving its declared MIME type.\n *\n * @param {string} uri - a data URI (assumed to be valid)\n * @returns {Object} { buffer, mimeType }\n */\nexport function decodeDataUri(uri: string): {arrayBuffer: ArrayBuffer; mimeType: string} {\n const dataIndex = uri.indexOf(',');\n\n let buffer;\n let mimeType;\n if (uri.slice(dataIndex - 7, dataIndex) === ';base64') {\n buffer = Buffer.from(uri.slice(dataIndex + 1), 'base64');\n mimeType = uri.slice(5, dataIndex - 7).trim();\n } else {\n buffer = Buffer.from(decodeURIComponent(uri.slice(dataIndex + 1)));\n mimeType = uri.slice(5, dataIndex).trim();\n }\n\n if (!mimeType) {\n mimeType = 'text/plain;charset=US-ASCII';\n } else if (mimeType.startsWith(';')) {\n mimeType = `text/plain${mimeType}`;\n }\n\n return {arrayBuffer: toArrayBuffer(buffer), mimeType};\n}\n\n/**\n * @param data\n * @todo Duplicate of core\n */\nexport function toArrayBuffer(data: unknown): ArrayBuffer {\n if (isArrayBuffer(data)) {\n return data as ArrayBuffer;\n }\n\n // TODO - per docs we should just be able to call buffer.buffer, but there are issues\n if (isBuffer(data)) {\n // @ts-expect-error\n const typedArray = new Uint8Array(data);\n return typedArray.buffer;\n }\n\n // Careful - Node Buffers will look like ArrayBuffers (keep after isBuffer)\n if (ArrayBuffer.isView(data)) {\n return data.buffer;\n }\n\n if (typeof data === 'string') {\n const text = data;\n const uint8Array = new TextEncoder().encode(text);\n return uint8Array.buffer;\n }\n\n // HACK to support Blob polyfill\n // @ts-expect-error\n if (data && typeof data === 'object' && data._toArrayBuffer) {\n // @ts-expect-error\n return data._toArrayBuffer();\n }\n\n throw new Error(`toArrayBuffer(${JSON.stringify(data, null, 2).slice(10)})`);\n}\n"],"mappings":";;AAEA,MAAMA,aAAa,GAAIC,CAAC,IAAKA,CAAC,IAAIA,CAAC,YAAYC,WAAW;AAC1D,MAAMC,QAAQ,GAAIF,CAAC,IAAKA,CAAC,IAAIA,CAAC,YAAYG,MAAM;;AAQhD,OAAO,SAASC,aAAa,CAACC,GAAW,EAAgD;EACvF,MAAMC,SAAS,GAAGD,GAAG,CAACE,OAAO,CAAC,GAAG,CAAC;EAElC,IAAIC,MAAM;EACV,IAAIC,QAAQ;EACZ,IAAIJ,GAAG,CAACK,KAAK,CAACJ,SAAS,GAAG,CAAC,EAAEA,SAAS,CAAC,KAAK,SAAS,EAAE;IACrDE,MAAM,GAAGL,MAAM,CAACQ,IAAI,CAACN,GAAG,CAACK,KAAK,CAACJ,SAAS,GAAG,CAAC,CAAC,EAAE,QAAQ,CAAC;IACxDG,QAAQ,GAAGJ,GAAG,CAACK,KAAK,CAAC,CAAC,EAAEJ,SAAS,GAAG,CAAC,CAAC,CAACM,IAAI,EAAE;EAC/C,CAAC,MAAM;IACLJ,MAAM,GAAGL,MAAM,CAACQ,IAAI,CAACE,kBAAkB,CAACR,GAAG,CAACK,KAAK,CAACJ,SAAS,GAAG,CAAC,CAAC,CAAC,CAAC;IAClEG,QAAQ,GAAGJ,GAAG,CAACK,KAAK,CAAC,CAAC,EAAEJ,SAAS,CAAC,CAACM,IAAI,EAAE;EAC3C;EAEA,IAAI,CAACH,QAAQ,EAAE;IACbA,QAAQ,GAAG,6BAA6B;EAC1C,CAAC,MAAM,IAAIA,QAAQ,CAACK,UAAU,CAAC,GAAG,CAAC,EAAE;IACnCL,QAAQ,uBAAgBA,QAAQ,CAAE;EACpC;EAEA,OAAO;IAACM,WAAW,EAAEC,aAAa,CAACR,MAAM,CAAC;IAAEC;EAAQ,CAAC;AACvD;;AAMA,OAAO,SAASO,aAAa,CAACC,IAAa,EAAe;EACxD,IAAIlB,aAAa,CAACkB,IAAI,CAAC,EAAE;IACvB,OAAOA,IAAI;EACb;;EAGA,IAAIf,QAAQ,CAACe,IAAI,CAAC,EAAE;IAElB,MAAMC,UAAU,GAAG,IAAIC,UAAU,CAACF,IAAI,CAAC;IACvC,OAAOC,UAAU,CAACV,MAAM;EAC1B;;EAGA,IAAIP,WAAW,CAACmB,MAAM,CAACH,IAAI,CAAC,EAAE;IAC5B,OAAOA,IAAI,CAACT,MAAM;EACpB;EAEA,IAAI,OAAOS,IAAI,KAAK,QAAQ,EAAE;IAC5B,MAAMI,IAAI,GAAGJ,IAAI;IACjB,MAAMK,UAAU,GAAG,IAAIC,WAAW,EAAE,CAACC,MAAM,CAACH,IAAI,CAAC;IACjD,OAAOC,UAAU,CAACd,MAAM;EAC1B;;EAIA,IAAIS,IAAI,IAAI,OAAOA,IAAI,KAAK,QAAQ,IAAIA,IAAI,CAACQ,cAAc,EAAE;IAE3D,OAAOR,IAAI,CAACQ,cAAc,EAAE;EAC9B;EAEA,MAAM,IAAIC,KAAK,yBAAkBC,IAAI,CAACC,SAAS,CAACX,IAAI,EAAE,IAAI,EAAE,CAAC,CAAC,CAACP,KAAK,CAAC,EAAE,CAAC,OAAI;AAC9E"}
@@ -3,7 +3,6 @@ import http from 'http';
3
3
  import https from 'https';
4
4
  import zlib from 'zlib';
5
5
  import { toArrayBuffer } from './decode-data-uri.node';
6
-
7
6
  function isRequestURL(url) {
8
7
  return url.startsWith('http:') || url.startsWith('https:');
9
8
  }
@@ -32,13 +31,10 @@ export function decompressReadStream(readStream, headers) {
32
31
  switch (headers.get('content-encoding')) {
33
32
  case 'br':
34
33
  return readStream.pipe(zlib.createBrotliDecompress());
35
-
36
34
  case 'gzip':
37
35
  return readStream.pipe(zlib.createGunzip());
38
-
39
36
  case 'deflate':
40
37
  return readStream.pipe(zlib.createDeflate());
41
-
42
38
  default:
43
39
  return readStream;
44
40
  }
@@ -47,12 +43,12 @@ export async function concatenateReadStream(readStream) {
47
43
  const arrayBufferChunks = [];
48
44
  return await new Promise((resolve, reject) => {
49
45
  readStream.on('error', error => reject(error));
46
+
50
47
  readStream.on('readable', () => readStream.read());
51
48
  readStream.on('data', chunk => {
52
49
  if (typeof chunk === 'string') {
53
50
  reject(new Error('Read stream not binary'));
54
51
  }
55
-
56
52
  arrayBufferChunks.push(toArrayBuffer(chunk));
57
53
  });
58
54
  readStream.on('end', () => {
@@ -65,7 +61,6 @@ export async function concatenateReadStream(readStream) {
65
61
  function getRequestOptions(url, options) {
66
62
  const originalHeaders = (options === null || options === void 0 ? void 0 : options.headers) || {};
67
63
  const headers = {};
68
-
69
64
  for (const key of Object.keys(originalHeaders)) {
70
65
  headers[key.toLowerCase()] = originalHeaders[key];
71
66
  }
@@ -85,10 +80,12 @@ function getRequestOptions(url, options) {
85
80
 
86
81
  export function concatenateArrayBuffers(sources) {
87
82
  const sourceArrays = sources.map(source2 => source2 instanceof ArrayBuffer ? new Uint8Array(source2) : source2);
83
+
88
84
  const byteLength = sourceArrays.reduce((length, typedArray) => length + typedArray.byteLength, 0);
85
+
89
86
  const result = new Uint8Array(byteLength);
90
- let offset = 0;
91
87
 
88
+ let offset = 0;
92
89
  for (const sourceArray of sourceArrays) {
93
90
  result.set(sourceArray, offset);
94
91
  offset += sourceArray.byteLength;
@@ -1 +1 @@
1
- {"version":3,"sources":["../../../../../src/node/fetch/utils/stream-utils.node.ts"],"names":["fs","http","https","zlib","toArrayBuffer","isRequestURL","url","startsWith","createReadStream","options","noqueryUrl","split","Promise","resolve","reject","stream","encoding","once","on","error","requestFunction","request","requestOptions","getRequestOptions","req","res","end","decompressReadStream","readStream","headers","get","pipe","createBrotliDecompress","createGunzip","createDeflate","concatenateReadStream","arrayBufferChunks","read","chunk","Error","push","arrayBuffer","concatenateArrayBuffers","originalHeaders","key","Object","keys","toLowerCase","urlObject","URL","hostname","path","pathname","method","fetch","port","sources","sourceArrays","map","source2","ArrayBuffer","Uint8Array","byteLength","reduce","length","typedArray","result","offset","sourceArray","set","buffer"],"mappings":"AAAA,OAAOA,EAAP,MAAe,IAAf;AACA,OAAOC,IAAP,MAAiB,MAAjB;AACA,OAAOC,KAAP,MAAkB,OAAlB;AACA,OAAOC,IAAP,MAAiB,MAAjB;AAEA,SAAQC,aAAR,QAA4B,wBAA5B;;AAEA,SAASC,YAAT,CAAsBC,GAAtB,EAA4C;AAC1C,SAAOA,GAAG,CAACC,UAAJ,CAAe,OAAf,KAA2BD,GAAG,CAACC,UAAJ,CAAe,QAAf,CAAlC;AACD;;AAGD,OAAO,eAAeC,gBAAf,CAAgCF,GAAhC,EAAqCG,OAArC,EAA4D;AAEjE,MAAI,CAACJ,YAAY,CAACC,GAAD,CAAjB,EAAwB;AACtB,UAAMI,UAAU,GAAGJ,GAAG,CAACK,KAAJ,CAAU,GAAV,EAAe,CAAf,CAAnB;AAEA,WAAO,MAAM,IAAIC,OAAJ,CAAY,CAACC,OAAD,EAAUC,MAAV,KAAqB;AAE5C,YAAMC,MAAM,GAAGf,EAAE,CAACQ,gBAAH,CAAoBE,UAApB,EAAgC;AAACM,QAAAA,QAAQ,EAAE;AAAX,OAAhC,CAAf;AACAD,MAAAA,MAAM,CAACE,IAAP,CAAY,UAAZ,EAAwB,MAAMJ,OAAO,CAACE,MAAD,CAArC;AACAA,MAAAA,MAAM,CAACG,EAAP,CAAU,OAAV,EAAoBC,KAAD,IAAWL,MAAM,CAACK,KAAD,CAApC;AACD,KALY,CAAb;AAMD;;AAID,SAAO,MAAM,IAAIP,OAAJ,CAAY,CAACC,OAAD,EAAUC,MAAV,KAAqB;AAC5C,UAAMM,eAAe,GAAGd,GAAG,CAACC,UAAJ,CAAe,QAAf,IAA2BL,KAAK,CAACmB,OAAjC,GAA2CpB,IAAI,CAACoB,OAAxE;AACA,UAAMC,cAAc,GAAGC,iBAAiB,CAACjB,GAAD,EAAMG,OAAN,CAAxC;AACA,UAAMe,GAAG,GAAGJ,eAAe,CAACE,cAAD,EAAkBG,GAAD,IAASZ,OAAO,CAACY,GAAD,CAAjC,CAA3B;AACAD,IAAAA,GAAG,CAACN,EAAJ,CAAO,OAAP,EAAiBC,KAAD,IAAWL,MAAM,CAACK,KAAD,CAAjC;AACAK,IAAAA,GAAG,CAACE,GAAJ;AACD,GANY,CAAb;AAOD;AAED,OAAO,SAASC,oBAAT,CAA8BC,UAA9B,EAA0CC,OAA1C,EAAmD;AACxD,UAAQA,OAAO,CAACC,GAAR,CAAY,kBAAZ,CAAR;AACE,SAAK,IAAL;AACE,aAAOF,UAAU,CAACG,IAAX,CAAgB5B,IAAI,CAAC6B,sBAAL,EAAhB,CAAP;;AACF,SAAK,MAAL;AACE,aAAOJ,UAAU,CAACG,IAAX,CAAgB5B,IAAI,CAAC8B,YAAL,EAAhB,CAAP;;AACF,SAAK,SAAL;AACE,aAAOL,UAAU,CAACG,IAAX,CAAgB5B,IAAI,CAAC+B,aAAL,EAAhB,CAAP;;AACF;AAEE,aAAON,UAAP;AATJ;AAWD;AAED,OAAO,eAAeO,qBAAf,CAAqCP,UAArC,EAAuE;AAC5E,QAAMQ,iBAAgC,GAAG,EAAzC;AAEA,SAAO,MAAM,IAAIxB,OAAJ,CAAY,CAACC,OAAD,EAAUC,MAAV,KAAqB;AAC5Cc,IAAAA,UAAU,CAACV,EAAX,CAAc,OAAd,EAAwBC,KAAD,IAAWL,MAAM,CAACK,KAAD,CAAxC;AAIAS,IAAAA,UAAU,CAACV,EAAX,CAAc,UAAd,EAA0B,MAAMU,UAAU,CAACS,IAAX,EAAhC;AAEAT,IAAAA,UAAU,CAACV,EAAX,CAAc,MAAd,EAAuBoB,KAAD,IAAW;AAC/B,UAAI,OAAOA,KAAP,KAAiB,QAArB,EAA+B;AAC7BxB,QAAAA,MAAM,CAAC,IAAIyB,KAAJ,CAAU,wBAAV,CAAD,CAAN;AACD;;AACDH,MAAAA,iBAAiB,CAACI,IAAlB,CAAuBpC,aAAa,CAACkC,KAAD,CAApC;AACD,KALD;AAOAV,IAAAA,UAAU,CAACV,EAAX,CAAc,KAAd,EAAqB,MAAM;AACzB,YAAMuB,WAAW,GAAGC,uBAAuB,CAACN,iBAAD,CAA3C;AACAvB,MAAAA,OAAO,CAAC4B,WAAD,CAAP;AACD,KAHD;AAID,GAlBY,CAAb;AAmBD;;AAID,SAASlB,iBAAT,CAA2BjB,GAA3B,EAAwCG,OAAxC,EAAoF;AAElF,QAAMkC,eAAe,GAAG,CAAAlC,OAAO,SAAP,IAAAA,OAAO,WAAP,YAAAA,OAAO,CAAEoB,OAAT,KAAoB,EAA5C;AACA,QAAMA,OAAO,GAAG,EAAhB;;AACA,OAAK,MAAMe,GAAX,IAAkBC,MAAM,CAACC,IAAP,CAAYH,eAAZ,CAAlB,EAAgD;AAC9Cd,IAAAA,OAAO,CAACe,GAAG,CAACG,WAAJ,EAAD,CAAP,GAA6BJ,eAAe,CAACC,GAAD,CAA5C;AACD;;AAGDf,EAAAA,OAAO,CAAC,iBAAD,CAAP,GAA6BA,OAAO,CAAC,iBAAD,CAAP,IAA8B,iBAA3D;AAEA,QAAMmB,SAAS,GAAG,IAAIC,GAAJ,CAAQ3C,GAAR,CAAlB;AACA,SAAO;AACL4C,IAAAA,QAAQ,EAAEF,SAAS,CAACE,QADf;AAELC,IAAAA,IAAI,EAAEH,SAAS,CAACI,QAFX;AAGLC,IAAAA,MAAM,EAAE,KAHH;AAKL,OAAG5C,OALE;AAML,QAAGA,OAAH,aAAGA,OAAH,uBAAGA,OAAO,CAAE6C,KAAZ,CANK;AAQLzB,IAAAA,OARK;AASL0B,IAAAA,IAAI,EAAEP,SAAS,CAACO;AATX,GAAP;AAWD;;AAOD,OAAO,SAASb,uBAAT,CAAiCc,OAAjC,EAAqF;AAE1F,QAAMC,YAAY,GAAGD,OAAO,CAACE,GAAR,CAAaC,OAAD,IAC/BA,OAAO,YAAYC,WAAnB,GAAiC,IAAIC,UAAJ,CAAeF,OAAf,CAAjC,GAA2DA,OADxC,CAArB;AAKA,QAAMG,UAAU,GAAGL,YAAY,CAACM,MAAb,CAAoB,CAACC,MAAD,EAASC,UAAT,KAAwBD,MAAM,GAAGC,UAAU,CAACH,UAAhE,EAA4E,CAA5E,CAAnB;AAGA,QAAMI,MAAM,GAAG,IAAIL,UAAJ,CAAeC,UAAf,CAAf;AAGA,MAAIK,MAAM,GAAG,CAAb;;AACA,OAAK,MAAMC,WAAX,IAA0BX,YAA1B,EAAwC;AACtCS,IAAAA,MAAM,CAACG,GAAP,CAAWD,WAAX,EAAwBD,MAAxB;AACAA,IAAAA,MAAM,IAAIC,WAAW,CAACN,UAAtB;AACD;;AAGD,SAAOI,MAAM,CAACI,MAAd;AACD","sourcesContent":["import fs from 'fs'; // `fs` will be empty object in browsers (see package.json \"browser\" field).\nimport http from 'http';\nimport https from 'https';\nimport zlib from 'zlib';\n\nimport {toArrayBuffer} from './decode-data-uri.node';\n\nfunction isRequestURL(url: string): boolean {\n return url.startsWith('http:') || url.startsWith('https:');\n}\n\n// Returns a promise that resolves to a readable stream\nexport async function createReadStream(url, options): Promise<any> {\n // Handle file streams in node\n if (!isRequestURL(url)) {\n const noqueryUrl = url.split('?')[0];\n // Now open the stream\n return await new Promise((resolve, reject) => {\n // @ts-ignore\n const stream = fs.createReadStream(noqueryUrl, {encoding: null});\n stream.once('readable', () => resolve(stream));\n stream.on('error', (error) => reject(error));\n });\n }\n\n // HANDLE HTTP/HTTPS REQUESTS IN NODE\n // TODO: THIS IS BAD SINCE WE RETURN A PROMISE INSTEAD OF A STREAM\n return await new Promise((resolve, reject) => {\n const requestFunction = url.startsWith('https:') ? https.request : http.request;\n const requestOptions = getRequestOptions(url, options);\n const req = requestFunction(requestOptions, (res) => resolve(res));\n req.on('error', (error) => reject(error));\n req.end();\n });\n}\n\nexport function decompressReadStream(readStream, headers) {\n switch (headers.get('content-encoding')) {\n case 'br':\n return readStream.pipe(zlib.createBrotliDecompress());\n case 'gzip':\n return readStream.pipe(zlib.createGunzip());\n case 'deflate':\n return readStream.pipe(zlib.createDeflate());\n default:\n // No compression or an unknown one, just return it as is\n return readStream;\n }\n}\n\nexport async function concatenateReadStream(readStream): Promise<ArrayBuffer> {\n const arrayBufferChunks: ArrayBuffer[] = [];\n\n return await new Promise((resolve, reject) => {\n readStream.on('error', (error) => reject(error));\n\n // Once the readable callback has been added, stream switches to \"flowing mode\"\n // In Node 10 (but not 12 and 14) this causes `data` and `end` to never be called unless we read data here\n readStream.on('readable', () => readStream.read());\n\n readStream.on('data', (chunk) => {\n if (typeof chunk === 'string') {\n reject(new Error('Read stream not binary'));\n }\n arrayBufferChunks.push(toArrayBuffer(chunk));\n });\n\n readStream.on('end', () => {\n const arrayBuffer = concatenateArrayBuffers(arrayBufferChunks);\n resolve(arrayBuffer);\n });\n });\n}\n\n// HELPERS\n\nfunction getRequestOptions(url: string, options?: {fetch?: typeof fetch; headers?}) {\n // Ensure header keys are lower case so that we can merge without duplicates\n const originalHeaders = options?.headers || {};\n const headers = {};\n for (const key of Object.keys(originalHeaders)) {\n headers[key.toLowerCase()] = originalHeaders[key];\n }\n\n // Add default accept-encoding to headers\n headers['accept-encoding'] = headers['accept-encoding'] || 'gzip,br,deflate';\n\n const urlObject = new URL(url);\n return {\n hostname: urlObject.hostname,\n path: urlObject.pathname,\n method: 'GET',\n // Add options and user provided 'options.fetch' overrides if available\n ...options,\n ...options?.fetch,\n // Override with updated headers with accepted encodings:\n headers,\n port: urlObject.port\n };\n}\n\n/**\n * Concatenate a sequence of ArrayBuffers\n * @return A concatenated ArrayBuffer\n * @note duplicates loader-utils since polyfills should be independent\n */\nexport function concatenateArrayBuffers(sources: (ArrayBuffer | Uint8Array)[]): ArrayBuffer {\n // Make sure all inputs are wrapped in typed arrays\n const sourceArrays = sources.map((source2) =>\n source2 instanceof ArrayBuffer ? new Uint8Array(source2) : source2\n );\n\n // Get length of all inputs\n const byteLength = sourceArrays.reduce((length, typedArray) => length + typedArray.byteLength, 0);\n\n // Allocate array with space for all inputs\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let offset = 0;\n for (const sourceArray of sourceArrays) {\n result.set(sourceArray, offset);\n offset += sourceArray.byteLength;\n }\n\n // We work with ArrayBuffers, discard the typed array wrapper\n return result.buffer;\n}\n"],"file":"stream-utils.node.js"}
1
+ {"version":3,"file":"stream-utils.node.js","names":["fs","http","https","zlib","toArrayBuffer","isRequestURL","url","startsWith","createReadStream","options","noqueryUrl","split","Promise","resolve","reject","stream","encoding","once","on","error","requestFunction","request","requestOptions","getRequestOptions","req","res","end","decompressReadStream","readStream","headers","get","pipe","createBrotliDecompress","createGunzip","createDeflate","concatenateReadStream","arrayBufferChunks","read","chunk","Error","push","arrayBuffer","concatenateArrayBuffers","originalHeaders","key","Object","keys","toLowerCase","urlObject","URL","hostname","path","pathname","method","fetch","port","sources","sourceArrays","map","source2","ArrayBuffer","Uint8Array","byteLength","reduce","length","typedArray","result","offset","sourceArray","set","buffer"],"sources":["../../../../../src/node/fetch/utils/stream-utils.node.ts"],"sourcesContent":["import fs from 'fs'; // `fs` will be empty object in browsers (see package.json \"browser\" field).\nimport http from 'http';\nimport https from 'https';\nimport zlib from 'zlib';\n\nimport {toArrayBuffer} from './decode-data-uri.node';\n\nfunction isRequestURL(url: string): boolean {\n return url.startsWith('http:') || url.startsWith('https:');\n}\n\n// Returns a promise that resolves to a readable stream\nexport async function createReadStream(url, options): Promise<any> {\n // Handle file streams in node\n if (!isRequestURL(url)) {\n const noqueryUrl = url.split('?')[0];\n // Now open the stream\n return await new Promise((resolve, reject) => {\n // @ts-ignore\n const stream = fs.createReadStream(noqueryUrl, {encoding: null});\n stream.once('readable', () => resolve(stream));\n stream.on('error', (error) => reject(error));\n });\n }\n\n // HANDLE HTTP/HTTPS REQUESTS IN NODE\n // TODO: THIS IS BAD SINCE WE RETURN A PROMISE INSTEAD OF A STREAM\n return await new Promise((resolve, reject) => {\n const requestFunction = url.startsWith('https:') ? https.request : http.request;\n const requestOptions = getRequestOptions(url, options);\n const req = requestFunction(requestOptions, (res) => resolve(res));\n req.on('error', (error) => reject(error));\n req.end();\n });\n}\n\nexport function decompressReadStream(readStream, headers) {\n switch (headers.get('content-encoding')) {\n case 'br':\n return readStream.pipe(zlib.createBrotliDecompress());\n case 'gzip':\n return readStream.pipe(zlib.createGunzip());\n case 'deflate':\n return readStream.pipe(zlib.createDeflate());\n default:\n // No compression or an unknown one, just return it as is\n return readStream;\n }\n}\n\nexport async function concatenateReadStream(readStream): Promise<ArrayBuffer> {\n const arrayBufferChunks: ArrayBuffer[] = [];\n\n return await new Promise((resolve, reject) => {\n readStream.on('error', (error) => reject(error));\n\n // Once the readable callback has been added, stream switches to \"flowing mode\"\n // In Node 10 (but not 12 and 14) this causes `data` and `end` to never be called unless we read data here\n readStream.on('readable', () => readStream.read());\n\n readStream.on('data', (chunk) => {\n if (typeof chunk === 'string') {\n reject(new Error('Read stream not binary'));\n }\n arrayBufferChunks.push(toArrayBuffer(chunk));\n });\n\n readStream.on('end', () => {\n const arrayBuffer = concatenateArrayBuffers(arrayBufferChunks);\n resolve(arrayBuffer);\n });\n });\n}\n\n// HELPERS\n\nfunction getRequestOptions(url: string, options?: {fetch?: typeof fetch; headers?}) {\n // Ensure header keys are lower case so that we can merge without duplicates\n const originalHeaders = options?.headers || {};\n const headers = {};\n for (const key of Object.keys(originalHeaders)) {\n headers[key.toLowerCase()] = originalHeaders[key];\n }\n\n // Add default accept-encoding to headers\n headers['accept-encoding'] = headers['accept-encoding'] || 'gzip,br,deflate';\n\n const urlObject = new URL(url);\n return {\n hostname: urlObject.hostname,\n path: urlObject.pathname,\n method: 'GET',\n // Add options and user provided 'options.fetch' overrides if available\n ...options,\n ...options?.fetch,\n // Override with updated headers with accepted encodings:\n headers,\n port: urlObject.port\n };\n}\n\n/**\n * Concatenate a sequence of ArrayBuffers\n * @return A concatenated ArrayBuffer\n * @note duplicates loader-utils since polyfills should be independent\n */\nexport function concatenateArrayBuffers(sources: (ArrayBuffer | Uint8Array)[]): ArrayBuffer {\n // Make sure all inputs are wrapped in typed arrays\n const sourceArrays = sources.map((source2) =>\n source2 instanceof ArrayBuffer ? new Uint8Array(source2) : source2\n );\n\n // Get length of all inputs\n const byteLength = sourceArrays.reduce((length, typedArray) => length + typedArray.byteLength, 0);\n\n // Allocate array with space for all inputs\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let offset = 0;\n for (const sourceArray of sourceArrays) {\n result.set(sourceArray, offset);\n offset += sourceArray.byteLength;\n }\n\n // We work with ArrayBuffers, discard the typed array wrapper\n return result.buffer;\n}\n"],"mappings":"AAAA,OAAOA,EAAE,MAAM,IAAI;AACnB,OAAOC,IAAI,MAAM,MAAM;AACvB,OAAOC,KAAK,MAAM,OAAO;AACzB,OAAOC,IAAI,MAAM,MAAM;AAEvB,SAAQC,aAAa,QAAO,wBAAwB;AAEpD,SAASC,YAAY,CAACC,GAAW,EAAW;EAC1C,OAAOA,GAAG,CAACC,UAAU,CAAC,OAAO,CAAC,IAAID,GAAG,CAACC,UAAU,CAAC,QAAQ,CAAC;AAC5D;;AAGA,OAAO,eAAeC,gBAAgB,CAACF,GAAG,EAAEG,OAAO,EAAgB;EAEjE,IAAI,CAACJ,YAAY,CAACC,GAAG,CAAC,EAAE;IACtB,MAAMI,UAAU,GAAGJ,GAAG,CAACK,KAAK,CAAC,GAAG,CAAC,CAAC,CAAC,CAAC;IAEpC,OAAO,MAAM,IAAIC,OAAO,CAAC,CAACC,OAAO,EAAEC,MAAM,KAAK;MAE5C,MAAMC,MAAM,GAAGf,EAAE,CAACQ,gBAAgB,CAACE,UAAU,EAAE;QAACM,QAAQ,EAAE;MAAI,CAAC,CAAC;MAChED,MAAM,CAACE,IAAI,CAAC,UAAU,EAAE,MAAMJ,OAAO,CAACE,MAAM,CAAC,CAAC;MAC9CA,MAAM,CAACG,EAAE,CAAC,OAAO,EAAGC,KAAK,IAAKL,MAAM,CAACK,KAAK,CAAC,CAAC;IAC9C,CAAC,CAAC;EACJ;;EAIA,OAAO,MAAM,IAAIP,OAAO,CAAC,CAACC,OAAO,EAAEC,MAAM,KAAK;IAC5C,MAAMM,eAAe,GAAGd,GAAG,CAACC,UAAU,CAAC,QAAQ,CAAC,GAAGL,KAAK,CAACmB,OAAO,GAAGpB,IAAI,CAACoB,OAAO;IAC/E,MAAMC,cAAc,GAAGC,iBAAiB,CAACjB,GAAG,EAAEG,OAAO,CAAC;IACtD,MAAMe,GAAG,GAAGJ,eAAe,CAACE,cAAc,EAAGG,GAAG,IAAKZ,OAAO,CAACY,GAAG,CAAC,CAAC;IAClED,GAAG,CAACN,EAAE,CAAC,OAAO,EAAGC,KAAK,IAAKL,MAAM,CAACK,KAAK,CAAC,CAAC;IACzCK,GAAG,CAACE,GAAG,EAAE;EACX,CAAC,CAAC;AACJ;AAEA,OAAO,SAASC,oBAAoB,CAACC,UAAU,EAAEC,OAAO,EAAE;EACxD,QAAQA,OAAO,CAACC,GAAG,CAAC,kBAAkB,CAAC;IACrC,KAAK,IAAI;MACP,OAAOF,UAAU,CAACG,IAAI,CAAC5B,IAAI,CAAC6B,sBAAsB,EAAE,CAAC;IACvD,KAAK,MAAM;MACT,OAAOJ,UAAU,CAACG,IAAI,CAAC5B,IAAI,CAAC8B,YAAY,EAAE,CAAC;IAC7C,KAAK,SAAS;MACZ,OAAOL,UAAU,CAACG,IAAI,CAAC5B,IAAI,CAAC+B,aAAa,EAAE,CAAC;IAC9C;MAEE,OAAON,UAAU;EAAC;AAExB;AAEA,OAAO,eAAeO,qBAAqB,CAACP,UAAU,EAAwB;EAC5E,MAAMQ,iBAAgC,GAAG,EAAE;EAE3C,OAAO,MAAM,IAAIxB,OAAO,CAAC,CAACC,OAAO,EAAEC,MAAM,KAAK;IAC5Cc,UAAU,CAACV,EAAE,CAAC,OAAO,EAAGC,KAAK,IAAKL,MAAM,CAACK,KAAK,CAAC,CAAC;;IAIhDS,UAAU,CAACV,EAAE,CAAC,UAAU,EAAE,MAAMU,UAAU,CAACS,IAAI,EAAE,CAAC;IAElDT,UAAU,CAACV,EAAE,CAAC,MAAM,EAAGoB,KAAK,IAAK;MAC/B,IAAI,OAAOA,KAAK,KAAK,QAAQ,EAAE;QAC7BxB,MAAM,CAAC,IAAIyB,KAAK,CAAC,wBAAwB,CAAC,CAAC;MAC7C;MACAH,iBAAiB,CAACI,IAAI,CAACpC,aAAa,CAACkC,KAAK,CAAC,CAAC;IAC9C,CAAC,CAAC;IAEFV,UAAU,CAACV,EAAE,CAAC,KAAK,EAAE,MAAM;MACzB,MAAMuB,WAAW,GAAGC,uBAAuB,CAACN,iBAAiB,CAAC;MAC9DvB,OAAO,CAAC4B,WAAW,CAAC;IACtB,CAAC,CAAC;EACJ,CAAC,CAAC;AACJ;;AAIA,SAASlB,iBAAiB,CAACjB,GAAW,EAAEG,OAA0C,EAAE;EAElF,MAAMkC,eAAe,GAAG,CAAAlC,OAAO,aAAPA,OAAO,uBAAPA,OAAO,CAAEoB,OAAO,KAAI,CAAC,CAAC;EAC9C,MAAMA,OAAO,GAAG,CAAC,CAAC;EAClB,KAAK,MAAMe,GAAG,IAAIC,MAAM,CAACC,IAAI,CAACH,eAAe,CAAC,EAAE;IAC9Cd,OAAO,CAACe,GAAG,CAACG,WAAW,EAAE,CAAC,GAAGJ,eAAe,CAACC,GAAG,CAAC;EACnD;;EAGAf,OAAO,CAAC,iBAAiB,CAAC,GAAGA,OAAO,CAAC,iBAAiB,CAAC,IAAI,iBAAiB;EAE5E,MAAMmB,SAAS,GAAG,IAAIC,GAAG,CAAC3C,GAAG,CAAC;EAC9B,OAAO;IACL4C,QAAQ,EAAEF,SAAS,CAACE,QAAQ;IAC5BC,IAAI,EAAEH,SAAS,CAACI,QAAQ;IACxBC,MAAM,EAAE,KAAK;IAEb,GAAG5C,OAAO;IACV,IAAGA,OAAO,aAAPA,OAAO,uBAAPA,OAAO,CAAE6C,KAAK;IAEjBzB,OAAO;IACP0B,IAAI,EAAEP,SAAS,CAACO;EAClB,CAAC;AACH;;AAOA,OAAO,SAASb,uBAAuB,CAACc,OAAqC,EAAe;EAE1F,MAAMC,YAAY,GAAGD,OAAO,CAACE,GAAG,CAAEC,OAAO,IACvCA,OAAO,YAAYC,WAAW,GAAG,IAAIC,UAAU,CAACF,OAAO,CAAC,GAAGA,OAAO,CACnE;;EAGD,MAAMG,UAAU,GAAGL,YAAY,CAACM,MAAM,CAAC,CAACC,MAAM,EAAEC,UAAU,KAAKD,MAAM,GAAGC,UAAU,CAACH,UAAU,EAAE,CAAC,CAAC;;EAGjG,MAAMI,MAAM,GAAG,IAAIL,UAAU,CAACC,UAAU,CAAC;;EAGzC,IAAIK,MAAM,GAAG,CAAC;EACd,KAAK,MAAMC,WAAW,IAAIX,YAAY,EAAE;IACtCS,MAAM,CAACG,GAAG,CAACD,WAAW,EAAED,MAAM,CAAC;IAC/BA,MAAM,IAAIC,WAAW,CAACN,UAAU;EAClC;;EAGA,OAAOI,MAAM,CAACI,MAAM;AACtB"}
@@ -2,11 +2,8 @@ import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
2
2
  export class BlobStreamController {
3
3
  constructor(chunks) {
4
4
  _defineProperty(this, "chunks", void 0);
5
-
6
5
  _defineProperty(this, "isWorking", false);
7
-
8
6
  _defineProperty(this, "isCancelled", false);
9
-
10
7
  this.chunks = chunks;
11
8
  }
12
9
 
@@ -19,17 +16,14 @@ export class BlobStreamController {
19
16
  chunks
20
17
  } = this;
21
18
  this.isWorking = true;
22
-
23
19
  while (!this.isCancelled && (controller.desiredSize || 0) > 0) {
24
20
  let next;
25
-
26
21
  try {
27
22
  next = chunks.next();
28
23
  } catch (error) {
29
24
  controller.error(error);
30
25
  break;
31
26
  }
32
-
33
27
  if (next) {
34
28
  if (!next.done && !this.isCancelled) {
35
29
  controller.enqueue(next.value);
@@ -38,7 +32,6 @@ export class BlobStreamController {
38
32
  }
39
33
  }
40
34
  }
41
-
42
35
  this.isWorking = false;
43
36
  }
44
37
 
@@ -51,6 +44,5 @@ export class BlobStreamController {
51
44
  cancel() {
52
45
  this.isCancelled = true;
53
46
  }
54
-
55
47
  }
56
48
  //# sourceMappingURL=blob-stream-controller.js.map