@loaders.gl/polyfills 3.3.0-alpha.8 → 3.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/dist.min.js +64 -15
- package/dist/es5/index.js +17 -11
- package/dist/es5/index.js.map +1 -1
- package/dist/es5/node/fetch/fetch-file.node.js +88 -0
- package/dist/es5/node/fetch/fetch-file.node.js.map +1 -0
- package/dist/es5/node/fetch/fetch.node.js +90 -35
- package/dist/es5/node/fetch/fetch.node.js.map +1 -1
- package/dist/es5/node/fetch/headers.node.js +2 -2
- package/dist/es5/node/fetch/headers.node.js.map +1 -1
- package/dist/es5/node/fetch/response.node.js +5 -4
- package/dist/es5/node/fetch/response.node.js.map +1 -1
- package/dist/es5/node/fetch/utils/stream-utils.node.js +9 -93
- package/dist/es5/node/fetch/utils/stream-utils.node.js.map +1 -1
- package/dist/es5/node/images/parse-image.node.js +30 -18
- package/dist/es5/node/images/parse-image.node.js.map +1 -1
- package/dist/esm/index.js +5 -4
- package/dist/esm/index.js.map +1 -1
- package/dist/esm/node/fetch/fetch-file.node.js +54 -0
- package/dist/esm/node/fetch/fetch-file.node.js.map +1 -0
- package/dist/esm/node/fetch/fetch.node.js +43 -18
- package/dist/esm/node/fetch/fetch.node.js.map +1 -1
- package/dist/esm/node/fetch/headers.node.js +1 -1
- package/dist/esm/node/fetch/headers.node.js.map +1 -1
- package/dist/esm/node/fetch/response.node.js +3 -2
- package/dist/esm/node/fetch/response.node.js.map +1 -1
- package/dist/esm/node/fetch/utils/stream-utils.node.js +3 -46
- package/dist/esm/node/fetch/utils/stream-utils.node.js.map +1 -1
- package/dist/esm/node/images/parse-image.node.js +23 -16
- package/dist/esm/node/images/parse-image.node.js.map +1 -1
- package/dist/index.d.ts +2 -1
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +13 -14
- package/dist/lib/encoding.js +1 -1
- package/dist/node/fetch/fetch-file.node.d.ts +4 -0
- package/dist/node/fetch/fetch-file.node.d.ts.map +1 -0
- package/dist/node/fetch/fetch-file.node.js +51 -0
- package/dist/node/fetch/fetch.node.d.ts +6 -1
- package/dist/node/fetch/fetch.node.d.ts.map +1 -1
- package/dist/node/fetch/fetch.node.js +57 -31
- package/dist/node/fetch/headers.node.d.ts +1 -1
- package/dist/node/fetch/headers.node.d.ts.map +1 -1
- package/dist/node/fetch/headers.node.js +2 -1
- package/dist/node/fetch/response.node.d.ts +2 -2
- package/dist/node/fetch/response.node.d.ts.map +1 -1
- package/dist/node/fetch/response.node.js +5 -6
- package/dist/node/fetch/utils/stream-utils.node.d.ts +8 -1
- package/dist/node/fetch/utils/stream-utils.node.d.ts.map +1 -1
- package/dist/node/fetch/utils/stream-utils.node.js +10 -54
- package/dist/node/images/parse-image.node.d.ts.map +1 -1
- package/dist/node/images/parse-image.node.js +26 -18
- package/package.json +3 -2
- package/src/index.ts +5 -4
- package/src/node/fetch/fetch-file.node.ts +51 -0
- package/src/node/fetch/fetch.node.ts +64 -30
- package/src/node/fetch/headers.node.ts +1 -1
- package/src/node/fetch/response.node.ts +4 -2
- package/src/node/fetch/utils/stream-utils.node.ts +10 -59
- package/src/node/images/parse-image.node.ts +31 -20
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
|
|
2
2
|
let _Symbol$iterator;
|
|
3
3
|
_Symbol$iterator = Symbol.iterator;
|
|
4
|
-
export
|
|
4
|
+
export class Headers {
|
|
5
5
|
constructor(headers) {
|
|
6
6
|
_defineProperty(this, "map", void 0);
|
|
7
7
|
this.map = {};
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"headers.node.js","names":["Symbol","iterator","Headers","constructor","headers","map","forEach","value","name","append","Array","isArray","header","Object","getOwnPropertyNames","normalizeName","normalizeValue","oldValue","delete","get","has","hasOwnProperty","set","visitor","thisArg","call","keys","items","push","iteratorFor","values","entries","String","test","TypeError","toLowerCase","next","shift","done","undefined"],"sources":["../../../../src/node/fetch/headers.node.ts"],"sourcesContent":["/**\n * Polyfill for Browser Headers\n * Based on https://github.com/github/fetch under MIT license\n */\nexport
|
|
1
|
+
{"version":3,"file":"headers.node.js","names":["Symbol","iterator","Headers","constructor","headers","map","forEach","value","name","append","Array","isArray","header","Object","getOwnPropertyNames","normalizeName","normalizeValue","oldValue","delete","get","has","hasOwnProperty","set","visitor","thisArg","call","keys","items","push","iteratorFor","values","entries","String","test","TypeError","toLowerCase","next","shift","done","undefined"],"sources":["../../../../src/node/fetch/headers.node.ts"],"sourcesContent":["/**\n * Polyfill for Browser Headers\n * Based on https://github.com/github/fetch under MIT license\n */\nexport class Headers {\n map: {};\n\n constructor(headers) {\n this.map = {};\n\n if (headers instanceof Headers) {\n headers.forEach((value, name) => this.append(name, value));\n } else if (Array.isArray(headers)) {\n headers.forEach((header) => this.append(header[0], header[1]));\n } else if (headers) {\n Object.getOwnPropertyNames(headers).forEach((name) => this.append(name, headers[name]));\n }\n }\n\n append(name, value) {\n name = normalizeName(name);\n value = normalizeValue(value);\n const oldValue = this.map[name];\n this.map[name] = oldValue ? `${oldValue}, ${value}` : value;\n }\n\n delete(name) {\n delete this.map[normalizeName(name)];\n }\n\n get(name) {\n name = normalizeName(name);\n return this.has(name) ? this.map[name] : null;\n }\n\n has(name) {\n return this.map.hasOwnProperty(normalizeName(name));\n }\n\n set(name, value) {\n this.map[normalizeName(name)] = normalizeValue(value);\n }\n\n forEach(visitor, thisArg = null) {\n for (const name in this.map) {\n if (this.map.hasOwnProperty(name)) {\n if (thisArg) {\n visitor.call(thisArg, this.map[name], name, this);\n } else {\n visitor(this.map[name], name, this);\n }\n }\n }\n }\n\n keys() {\n const items: any[] = [];\n this.forEach(function (value, name) {\n items.push(name);\n });\n return iteratorFor(items);\n }\n\n values() {\n const items: any[] = [];\n this.forEach(function (value) {\n items.push(value);\n });\n return iteratorFor(items);\n }\n\n entries() {\n const items: any[] = [];\n this.forEach(function (value, name) {\n items.push([name, value]);\n });\n return iteratorFor(items);\n }\n\n *[Symbol.iterator]() {\n // @ts-ignore must have a '[Symbol.iterator]()' method that returns an iterator.\n yield* this.entries();\n }\n}\n\nfunction normalizeName(name) {\n if (typeof name !== 'string') {\n name = String(name);\n }\n if (/[^a-z0-9\\-#$%&'*+.^_`|~]/i.test(name) || name === '') {\n throw new TypeError('Invalid character in header field name');\n }\n return name.toLowerCase();\n}\n\nfunction normalizeValue(value) {\n if (typeof value !== 'string') {\n value = String(value);\n }\n return value;\n}\n\n// Build a destructive iterator for the value list\nfunction iteratorFor(items) {\n const iterator = {\n next() {\n const value = items.shift();\n return {done: value === undefined, value};\n }\n };\n\n iterator[Symbol.iterator] = function () {\n return iterator;\n };\n\n return iterator;\n}\n"],"mappings":";;mBA+EIA,MAAM,CAACC,QAAQ;AA3EnB,OAAO,MAAMC,OAAO,CAAC;EAGnBC,WAAW,CAACC,OAAO,EAAE;IAAA;IACnB,IAAI,CAACC,GAAG,GAAG,CAAC,CAAC;IAEb,IAAID,OAAO,YAAYF,OAAO,EAAE;MAC9BE,OAAO,CAACE,OAAO,CAAC,CAACC,KAAK,EAAEC,IAAI,KAAK,IAAI,CAACC,MAAM,CAACD,IAAI,EAAED,KAAK,CAAC,CAAC;IAC5D,CAAC,MAAM,IAAIG,KAAK,CAACC,OAAO,CAACP,OAAO,CAAC,EAAE;MACjCA,OAAO,CAACE,OAAO,CAAEM,MAAM,IAAK,IAAI,CAACH,MAAM,CAACG,MAAM,CAAC,CAAC,CAAC,EAAEA,MAAM,CAAC,CAAC,CAAC,CAAC,CAAC;IAChE,CAAC,MAAM,IAAIR,OAAO,EAAE;MAClBS,MAAM,CAACC,mBAAmB,CAACV,OAAO,CAAC,CAACE,OAAO,CAAEE,IAAI,IAAK,IAAI,CAACC,MAAM,CAACD,IAAI,EAAEJ,OAAO,CAACI,IAAI,CAAC,CAAC,CAAC;IACzF;EACF;EAEAC,MAAM,CAACD,IAAI,EAAED,KAAK,EAAE;IAClBC,IAAI,GAAGO,aAAa,CAACP,IAAI,CAAC;IAC1BD,KAAK,GAAGS,cAAc,CAACT,KAAK,CAAC;IAC7B,MAAMU,QAAQ,GAAG,IAAI,CAACZ,GAAG,CAACG,IAAI,CAAC;IAC/B,IAAI,CAACH,GAAG,CAACG,IAAI,CAAC,GAAGS,QAAQ,aAAMA,QAAQ,eAAKV,KAAK,IAAKA,KAAK;EAC7D;EAEAW,MAAM,CAACV,IAAI,EAAE;IACX,OAAO,IAAI,CAACH,GAAG,CAACU,aAAa,CAACP,IAAI,CAAC,CAAC;EACtC;EAEAW,GAAG,CAACX,IAAI,EAAE;IACRA,IAAI,GAAGO,aAAa,CAACP,IAAI,CAAC;IAC1B,OAAO,IAAI,CAACY,GAAG,CAACZ,IAAI,CAAC,GAAG,IAAI,CAACH,GAAG,CAACG,IAAI,CAAC,GAAG,IAAI;EAC/C;EAEAY,GAAG,CAACZ,IAAI,EAAE;IACR,OAAO,IAAI,CAACH,GAAG,CAACgB,cAAc,CAACN,aAAa,CAACP,IAAI,CAAC,CAAC;EACrD;EAEAc,GAAG,CAACd,IAAI,EAAED,KAAK,EAAE;IACf,IAAI,CAACF,GAAG,CAACU,aAAa,CAACP,IAAI,CAAC,CAAC,GAAGQ,cAAc,CAACT,KAAK,CAAC;EACvD;EAEAD,OAAO,CAACiB,OAAO,EAAkB;IAAA,IAAhBC,OAAO,uEAAG,IAAI;IAC7B,KAAK,MAAMhB,IAAI,IAAI,IAAI,CAACH,GAAG,EAAE;MAC3B,IAAI,IAAI,CAACA,GAAG,CAACgB,cAAc,CAACb,IAAI,CAAC,EAAE;QACjC,IAAIgB,OAAO,EAAE;UACXD,OAAO,CAACE,IAAI,CAACD,OAAO,EAAE,IAAI,CAACnB,GAAG,CAACG,IAAI,CAAC,EAAEA,IAAI,EAAE,IAAI,CAAC;QACnD,CAAC,MAAM;UACLe,OAAO,CAAC,IAAI,CAAClB,GAAG,CAACG,IAAI,CAAC,EAAEA,IAAI,EAAE,IAAI,CAAC;QACrC;MACF;IACF;EACF;EAEAkB,IAAI,GAAG;IACL,MAAMC,KAAY,GAAG,EAAE;IACvB,IAAI,CAACrB,OAAO,CAAC,UAAUC,KAAK,EAAEC,IAAI,EAAE;MAClCmB,KAAK,CAACC,IAAI,CAACpB,IAAI,CAAC;IAClB,CAAC,CAAC;IACF,OAAOqB,WAAW,CAACF,KAAK,CAAC;EAC3B;EAEAG,MAAM,GAAG;IACP,MAAMH,KAAY,GAAG,EAAE;IACvB,IAAI,CAACrB,OAAO,CAAC,UAAUC,KAAK,EAAE;MAC5BoB,KAAK,CAACC,IAAI,CAACrB,KAAK,CAAC;IACnB,CAAC,CAAC;IACF,OAAOsB,WAAW,CAACF,KAAK,CAAC;EAC3B;EAEAI,OAAO,GAAG;IACR,MAAMJ,KAAY,GAAG,EAAE;IACvB,IAAI,CAACrB,OAAO,CAAC,UAAUC,KAAK,EAAEC,IAAI,EAAE;MAClCmB,KAAK,CAACC,IAAI,CAAC,CAACpB,IAAI,EAAED,KAAK,CAAC,CAAC;IAC3B,CAAC,CAAC;IACF,OAAOsB,WAAW,CAACF,KAAK,CAAC;EAC3B;EAEA,sBAAqB;IAEnB,OAAO,IAAI,CAACI,OAAO,EAAE;EACvB;AACF;AAEA,SAAShB,aAAa,CAACP,IAAI,EAAE;EAC3B,IAAI,OAAOA,IAAI,KAAK,QAAQ,EAAE;IAC5BA,IAAI,GAAGwB,MAAM,CAACxB,IAAI,CAAC;EACrB;EACA,IAAI,2BAA2B,CAACyB,IAAI,CAACzB,IAAI,CAAC,IAAIA,IAAI,KAAK,EAAE,EAAE;IACzD,MAAM,IAAI0B,SAAS,CAAC,wCAAwC,CAAC;EAC/D;EACA,OAAO1B,IAAI,CAAC2B,WAAW,EAAE;AAC3B;AAEA,SAASnB,cAAc,CAACT,KAAK,EAAE;EAC7B,IAAI,OAAOA,KAAK,KAAK,QAAQ,EAAE;IAC7BA,KAAK,GAAGyB,MAAM,CAACzB,KAAK,CAAC;EACvB;EACA,OAAOA,KAAK;AACd;;AAGA,SAASsB,WAAW,CAACF,KAAK,EAAE;EAC1B,MAAM1B,QAAQ,GAAG;IACfmC,IAAI,GAAG;MACL,MAAM7B,KAAK,GAAGoB,KAAK,CAACU,KAAK,EAAE;MAC3B,OAAO;QAACC,IAAI,EAAE/B,KAAK,KAAKgC,SAAS;QAAEhC;MAAK,CAAC;IAC3C;EACF,CAAC;EAEDN,QAAQ,CAACD,MAAM,CAACC,QAAQ,CAAC,GAAG,YAAY;IACtC,OAAOA,QAAQ;EACjB,CAAC;EAED,OAAOA,QAAQ;AACjB"}
|
|
@@ -1,14 +1,15 @@
|
|
|
1
1
|
import _defineProperty from "@babel/runtime/helpers/esm/defineProperty";
|
|
2
|
+
|
|
2
3
|
import { assert } from '../../utils/assert';
|
|
3
4
|
import { decompressReadStream, concatenateReadStream } from './utils/stream-utils.node';
|
|
4
|
-
import Headers from './headers.node';
|
|
5
|
+
import { Headers } from './headers.node';
|
|
5
6
|
const isBoolean = x => typeof x === 'boolean';
|
|
6
7
|
const isFunction = x => typeof x === 'function';
|
|
7
8
|
const isObject = x => x !== null && typeof x === 'object';
|
|
8
9
|
const isReadableNodeStream = x => isObject(x) && isFunction(x.read) && isFunction(x.pipe) && isBoolean(x.readable);
|
|
9
10
|
|
|
10
11
|
import { Readable } from 'stream';
|
|
11
|
-
export
|
|
12
|
+
export class Response {
|
|
12
13
|
constructor(body, options) {
|
|
13
14
|
_defineProperty(this, "ok", void 0);
|
|
14
15
|
_defineProperty(this, "status", void 0);
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"response.node.js","names":["assert","decompressReadStream","concatenateReadStream","Headers","isBoolean","x","isFunction","isObject","isReadableNodeStream","read","pipe","readable","Readable","Response","constructor","body","options","headers","status","statusText","url","ok","_body","from","TextEncoder","encode","ArrayBuffer","bodyUsed","arrayBuffer","data","text","textDecoder","TextDecoder","decode","json","JSON","parse","blob","Blob","Error"],"sources":["../../../../src/node/fetch/response.node.ts"],"sourcesContent":["
|
|
1
|
+
{"version":3,"file":"response.node.js","names":["assert","decompressReadStream","concatenateReadStream","Headers","isBoolean","x","isFunction","isObject","isReadableNodeStream","read","pipe","readable","Readable","Response","constructor","body","options","headers","status","statusText","url","ok","_body","from","TextEncoder","encode","ArrayBuffer","bodyUsed","arrayBuffer","data","text","textDecoder","TextDecoder","decode","json","JSON","parse","blob","Blob","Error"],"sources":["../../../../src/node/fetch/response.node.ts"],"sourcesContent":["// loaders.gl, MIT license\n\nimport {assert} from '../../utils/assert';\nimport {decompressReadStream, concatenateReadStream} from './utils/stream-utils.node';\nimport {Headers} from './headers.node';\n\nconst isBoolean = (x) => typeof x === 'boolean';\nconst isFunction = (x) => typeof x === 'function';\nconst isObject = (x) => x !== null && typeof x === 'object';\nconst isReadableNodeStream = (x) =>\n isObject(x) && isFunction(x.read) && isFunction(x.pipe) && isBoolean(x.readable);\n\n/**\n * Polyfill for Browser Response\n *\n * Under Node.js we return a mock \"fetch response object\"\n * so that apps can use the same API as in the browser.\n *\n * Note: This is intended to be a \"lightweight\" implementation and will have limitations.\n *\n * See https://developer.mozilla.org/en-US/docs/Web/API/Response\n */\nimport {Readable} from 'stream';\n\nexport class Response {\n readonly ok: boolean;\n readonly status: number;\n readonly statusText: string;\n readonly headers: Headers;\n readonly url: string;\n bodyUsed: boolean = false;\n private readonly _body;\n\n // TODO - handle ArrayBuffer, ArrayBufferView, Buffer\n constructor(\n body,\n options: {\n headers?;\n status?: number;\n statusText?: string;\n url: string;\n }\n ) {\n const {headers, status = 200, statusText = 'OK', url} = options || {};\n\n this.url = url;\n this.ok = status === 200;\n this.status = status; // TODO - handle errors and set status\n this.statusText = statusText;\n this.headers = new Headers(options?.headers || {});\n\n // Check for content-encoding and create a decompression stream\n if (isReadableNodeStream(body)) {\n this._body = decompressReadStream(body, headers);\n } else if (typeof body === 'string') {\n this._body = Readable.from([new TextEncoder().encode(body)]);\n } else {\n this._body = Readable.from([body || new ArrayBuffer(0)]);\n }\n }\n\n // Subset of Properties\n\n // Returns a readable stream to the \"body\" of the response (or file)\n get body() {\n assert(!this.bodyUsed);\n assert(isReadableNodeStream(this._body)); // Not implemented: conversion of ArrayBuffer etc to stream\n this.bodyUsed = true;\n return this._body;\n }\n\n // Subset of Methods\n\n async arrayBuffer() {\n if (!isReadableNodeStream(this._body)) {\n return this._body || new ArrayBuffer(0);\n }\n const data = await concatenateReadStream(this._body);\n return data;\n }\n\n async text() {\n const arrayBuffer = await this.arrayBuffer();\n const textDecoder = new TextDecoder();\n return textDecoder.decode(arrayBuffer);\n }\n\n async json() {\n const text = await this.text();\n return JSON.parse(text);\n }\n\n async blob() {\n if (typeof Blob === 'undefined') {\n throw new Error('Blob polyfill not installed');\n }\n return new Blob([await this.arrayBuffer()]);\n }\n}\n"],"mappings":";;AAEA,SAAQA,MAAM,QAAO,oBAAoB;AACzC,SAAQC,oBAAoB,EAAEC,qBAAqB,QAAO,2BAA2B;AACrF,SAAQC,OAAO,QAAO,gBAAgB;AAEtC,MAAMC,SAAS,GAAIC,CAAC,IAAK,OAAOA,CAAC,KAAK,SAAS;AAC/C,MAAMC,UAAU,GAAID,CAAC,IAAK,OAAOA,CAAC,KAAK,UAAU;AACjD,MAAME,QAAQ,GAAIF,CAAC,IAAKA,CAAC,KAAK,IAAI,IAAI,OAAOA,CAAC,KAAK,QAAQ;AAC3D,MAAMG,oBAAoB,GAAIH,CAAC,IAC7BE,QAAQ,CAACF,CAAC,CAAC,IAAIC,UAAU,CAACD,CAAC,CAACI,IAAI,CAAC,IAAIH,UAAU,CAACD,CAAC,CAACK,IAAI,CAAC,IAAIN,SAAS,CAACC,CAAC,CAACM,QAAQ,CAAC;;AAYlF,SAAQC,QAAQ,QAAO,QAAQ;AAE/B,OAAO,MAAMC,QAAQ,CAAC;EAUpBC,WAAW,CACTC,IAAI,EACJC,OAKC,EACD;IAAA;IAAA;IAAA;IAAA;IAAA;IAAA,kCAZkB,KAAK;IAAA;IAavB,MAAM;MAACC,OAAO;MAAEC,MAAM,GAAG,GAAG;MAAEC,UAAU,GAAG,IAAI;MAAEC;IAAG,CAAC,GAAGJ,OAAO,IAAI,CAAC,CAAC;IAErE,IAAI,CAACI,GAAG,GAAGA,GAAG;IACd,IAAI,CAACC,EAAE,GAAGH,MAAM,KAAK,GAAG;IACxB,IAAI,CAACA,MAAM,GAAGA,MAAM;IACpB,IAAI,CAACC,UAAU,GAAGA,UAAU;IAC5B,IAAI,CAACF,OAAO,GAAG,IAAId,OAAO,CAAC,CAAAa,OAAO,aAAPA,OAAO,uBAAPA,OAAO,CAAEC,OAAO,KAAI,CAAC,CAAC,CAAC;;IAGlD,IAAIT,oBAAoB,CAACO,IAAI,CAAC,EAAE;MAC9B,IAAI,CAACO,KAAK,GAAGrB,oBAAoB,CAACc,IAAI,EAAEE,OAAO,CAAC;IAClD,CAAC,MAAM,IAAI,OAAOF,IAAI,KAAK,QAAQ,EAAE;MACnC,IAAI,CAACO,KAAK,GAAGV,QAAQ,CAACW,IAAI,CAAC,CAAC,IAAIC,WAAW,EAAE,CAACC,MAAM,CAACV,IAAI,CAAC,CAAC,CAAC;IAC9D,CAAC,MAAM;MACL,IAAI,CAACO,KAAK,GAAGV,QAAQ,CAACW,IAAI,CAAC,CAACR,IAAI,IAAI,IAAIW,WAAW,CAAC,CAAC,CAAC,CAAC,CAAC;IAC1D;EACF;;EAKA,IAAIX,IAAI,GAAG;IACTf,MAAM,CAAC,CAAC,IAAI,CAAC2B,QAAQ,CAAC;IACtB3B,MAAM,CAACQ,oBAAoB,CAAC,IAAI,CAACc,KAAK,CAAC,CAAC;IACxC,IAAI,CAACK,QAAQ,GAAG,IAAI;IACpB,OAAO,IAAI,CAACL,KAAK;EACnB;;EAIA,MAAMM,WAAW,GAAG;IAClB,IAAI,CAACpB,oBAAoB,CAAC,IAAI,CAACc,KAAK,CAAC,EAAE;MACrC,OAAO,IAAI,CAACA,KAAK,IAAI,IAAII,WAAW,CAAC,CAAC,CAAC;IACzC;IACA,MAAMG,IAAI,GAAG,MAAM3B,qBAAqB,CAAC,IAAI,CAACoB,KAAK,CAAC;IACpD,OAAOO,IAAI;EACb;EAEA,MAAMC,IAAI,GAAG;IACX,MAAMF,WAAW,GAAG,MAAM,IAAI,CAACA,WAAW,EAAE;IAC5C,MAAMG,WAAW,GAAG,IAAIC,WAAW,EAAE;IACrC,OAAOD,WAAW,CAACE,MAAM,CAACL,WAAW,CAAC;EACxC;EAEA,MAAMM,IAAI,GAAG;IACX,MAAMJ,IAAI,GAAG,MAAM,IAAI,CAACA,IAAI,EAAE;IAC9B,OAAOK,IAAI,CAACC,KAAK,CAACN,IAAI,CAAC;EACzB;EAEA,MAAMO,IAAI,GAAG;IACX,IAAI,OAAOC,IAAI,KAAK,WAAW,EAAE;MAC/B,MAAM,IAAIC,KAAK,CAAC,6BAA6B,CAAC;IAChD;IACA,OAAO,IAAID,IAAI,CAAC,CAAC,MAAM,IAAI,CAACV,WAAW,EAAE,CAAC,CAAC;EAC7C;AACF"}
|
|
@@ -1,32 +1,8 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
import https from 'https';
|
|
1
|
+
|
|
2
|
+
|
|
4
3
|
import zlib from 'zlib';
|
|
5
4
|
import { toArrayBuffer } from './decode-data-uri.node';
|
|
6
|
-
function isRequestURL(url) {
|
|
7
|
-
return url.startsWith('http:') || url.startsWith('https:');
|
|
8
|
-
}
|
|
9
5
|
|
|
10
|
-
export async function createReadStream(url, options) {
|
|
11
|
-
if (!isRequestURL(url)) {
|
|
12
|
-
const noqueryUrl = url.split('?')[0];
|
|
13
|
-
return await new Promise((resolve, reject) => {
|
|
14
|
-
const stream = fs.createReadStream(noqueryUrl, {
|
|
15
|
-
encoding: null
|
|
16
|
-
});
|
|
17
|
-
stream.once('readable', () => resolve(stream));
|
|
18
|
-
stream.on('error', error => reject(error));
|
|
19
|
-
});
|
|
20
|
-
}
|
|
21
|
-
|
|
22
|
-
return await new Promise((resolve, reject) => {
|
|
23
|
-
const requestFunction = url.startsWith('https:') ? https.request : http.request;
|
|
24
|
-
const requestOptions = getRequestOptions(url, options);
|
|
25
|
-
const req = requestFunction(requestOptions, res => resolve(res));
|
|
26
|
-
req.on('error', error => reject(error));
|
|
27
|
-
req.end();
|
|
28
|
-
});
|
|
29
|
-
}
|
|
30
6
|
export function decompressReadStream(readStream, headers) {
|
|
31
7
|
switch (headers.get('content-encoding')) {
|
|
32
8
|
case 'br':
|
|
@@ -39,6 +15,7 @@ export function decompressReadStream(readStream, headers) {
|
|
|
39
15
|
return readStream;
|
|
40
16
|
}
|
|
41
17
|
}
|
|
18
|
+
|
|
42
19
|
export async function concatenateReadStream(readStream) {
|
|
43
20
|
const arrayBufferChunks = [];
|
|
44
21
|
return await new Promise((resolve, reject) => {
|
|
@@ -58,26 +35,6 @@ export async function concatenateReadStream(readStream) {
|
|
|
58
35
|
});
|
|
59
36
|
}
|
|
60
37
|
|
|
61
|
-
function getRequestOptions(url, options) {
|
|
62
|
-
const originalHeaders = (options === null || options === void 0 ? void 0 : options.headers) || {};
|
|
63
|
-
const headers = {};
|
|
64
|
-
for (const key of Object.keys(originalHeaders)) {
|
|
65
|
-
headers[key.toLowerCase()] = originalHeaders[key];
|
|
66
|
-
}
|
|
67
|
-
|
|
68
|
-
headers['accept-encoding'] = headers['accept-encoding'] || 'gzip,br,deflate';
|
|
69
|
-
const urlObject = new URL(url);
|
|
70
|
-
return {
|
|
71
|
-
hostname: urlObject.hostname,
|
|
72
|
-
path: urlObject.pathname,
|
|
73
|
-
method: 'GET',
|
|
74
|
-
...options,
|
|
75
|
-
...(options === null || options === void 0 ? void 0 : options.fetch),
|
|
76
|
-
headers,
|
|
77
|
-
port: urlObject.port
|
|
78
|
-
};
|
|
79
|
-
}
|
|
80
|
-
|
|
81
38
|
export function concatenateArrayBuffers(sources) {
|
|
82
39
|
const sourceArrays = sources.map(source2 => source2 instanceof ArrayBuffer ? new Uint8Array(source2) : source2);
|
|
83
40
|
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"stream-utils.node.js","names":["
|
|
1
|
+
{"version":3,"file":"stream-utils.node.js","names":["zlib","toArrayBuffer","decompressReadStream","readStream","headers","get","pipe","createBrotliDecompress","createGunzip","createDeflate","concatenateReadStream","arrayBufferChunks","Promise","resolve","reject","on","error","read","chunk","Error","push","arrayBuffer","concatenateArrayBuffers","sources","sourceArrays","map","source2","ArrayBuffer","Uint8Array","byteLength","reduce","length","typedArray","result","offset","sourceArray","set","buffer"],"sources":["../../../../../src/node/fetch/utils/stream-utils.node.ts"],"sourcesContent":["// loaders.gl, MIT license\n\nimport zlib from 'zlib';\n\nimport {toArrayBuffer} from './decode-data-uri.node';\n\n/**\n *\n */\nexport function decompressReadStream(readStream, headers) {\n switch (headers.get('content-encoding')) {\n case 'br':\n return readStream.pipe(zlib.createBrotliDecompress());\n case 'gzip':\n return readStream.pipe(zlib.createGunzip());\n case 'deflate':\n return readStream.pipe(zlib.createDeflate());\n default:\n // No compression or an unknown one, just return it as is\n return readStream;\n }\n}\n\n/**\n *\n * @param readStream\n * @returns\n */\nexport async function concatenateReadStream(readStream): Promise<ArrayBuffer> {\n const arrayBufferChunks: ArrayBuffer[] = [];\n\n return await new Promise((resolve, reject) => {\n readStream.on('error', (error) => reject(error));\n\n // Once the readable callback has been added, stream switches to \"flowing mode\"\n // In Node 10 (but not 12 and 14) this causes `data` and `end` to never be called unless we read data here\n readStream.on('readable', () => readStream.read());\n\n readStream.on('data', (chunk) => {\n if (typeof chunk === 'string') {\n reject(new Error('Read stream not binary'));\n }\n arrayBufferChunks.push(toArrayBuffer(chunk));\n });\n\n readStream.on('end', () => {\n const arrayBuffer = concatenateArrayBuffers(arrayBufferChunks);\n resolve(arrayBuffer);\n });\n });\n}\n\n/**\n * Concatenate a sequence of ArrayBuffers\n * @return A concatenated ArrayBuffer\n * @note duplicates loader-utils since polyfills should be independent\n */\nexport function concatenateArrayBuffers(sources: (ArrayBuffer | Uint8Array)[]): ArrayBuffer {\n // Make sure all inputs are wrapped in typed arrays\n const sourceArrays = sources.map((source2) =>\n source2 instanceof ArrayBuffer ? new Uint8Array(source2) : source2\n );\n\n // Get length of all inputs\n const byteLength = sourceArrays.reduce((length, typedArray) => length + typedArray.byteLength, 0);\n\n // Allocate array with space for all inputs\n const result = new Uint8Array(byteLength);\n\n // Copy the subarrays\n let offset = 0;\n for (const sourceArray of sourceArrays) {\n result.set(sourceArray, offset);\n offset += sourceArray.byteLength;\n }\n\n // We work with ArrayBuffers, discard the typed array wrapper\n return result.buffer;\n}\n"],"mappings":";;AAEA,OAAOA,IAAI,MAAM,MAAM;AAEvB,SAAQC,aAAa,QAAO,wBAAwB;;AAKpD,OAAO,SAASC,oBAAoB,CAACC,UAAU,EAAEC,OAAO,EAAE;EACxD,QAAQA,OAAO,CAACC,GAAG,CAAC,kBAAkB,CAAC;IACrC,KAAK,IAAI;MACP,OAAOF,UAAU,CAACG,IAAI,CAACN,IAAI,CAACO,sBAAsB,EAAE,CAAC;IACvD,KAAK,MAAM;MACT,OAAOJ,UAAU,CAACG,IAAI,CAACN,IAAI,CAACQ,YAAY,EAAE,CAAC;IAC7C,KAAK,SAAS;MACZ,OAAOL,UAAU,CAACG,IAAI,CAACN,IAAI,CAACS,aAAa,EAAE,CAAC;IAC9C;MAEE,OAAON,UAAU;EAAC;AAExB;;AAOA,OAAO,eAAeO,qBAAqB,CAACP,UAAU,EAAwB;EAC5E,MAAMQ,iBAAgC,GAAG,EAAE;EAE3C,OAAO,MAAM,IAAIC,OAAO,CAAC,CAACC,OAAO,EAAEC,MAAM,KAAK;IAC5CX,UAAU,CAACY,EAAE,CAAC,OAAO,EAAGC,KAAK,IAAKF,MAAM,CAACE,KAAK,CAAC,CAAC;;IAIhDb,UAAU,CAACY,EAAE,CAAC,UAAU,EAAE,MAAMZ,UAAU,CAACc,IAAI,EAAE,CAAC;IAElDd,UAAU,CAACY,EAAE,CAAC,MAAM,EAAGG,KAAK,IAAK;MAC/B,IAAI,OAAOA,KAAK,KAAK,QAAQ,EAAE;QAC7BJ,MAAM,CAAC,IAAIK,KAAK,CAAC,wBAAwB,CAAC,CAAC;MAC7C;MACAR,iBAAiB,CAACS,IAAI,CAACnB,aAAa,CAACiB,KAAK,CAAC,CAAC;IAC9C,CAAC,CAAC;IAEFf,UAAU,CAACY,EAAE,CAAC,KAAK,EAAE,MAAM;MACzB,MAAMM,WAAW,GAAGC,uBAAuB,CAACX,iBAAiB,CAAC;MAC9DE,OAAO,CAACQ,WAAW,CAAC;IACtB,CAAC,CAAC;EACJ,CAAC,CAAC;AACJ;;AAOA,OAAO,SAASC,uBAAuB,CAACC,OAAqC,EAAe;EAE1F,MAAMC,YAAY,GAAGD,OAAO,CAACE,GAAG,CAAEC,OAAO,IACvCA,OAAO,YAAYC,WAAW,GAAG,IAAIC,UAAU,CAACF,OAAO,CAAC,GAAGA,OAAO,CACnE;;EAGD,MAAMG,UAAU,GAAGL,YAAY,CAACM,MAAM,CAAC,CAACC,MAAM,EAAEC,UAAU,KAAKD,MAAM,GAAGC,UAAU,CAACH,UAAU,EAAE,CAAC,CAAC;;EAGjG,MAAMI,MAAM,GAAG,IAAIL,UAAU,CAACC,UAAU,CAAC;;EAGzC,IAAIK,MAAM,GAAG,CAAC;EACd,KAAK,MAAMC,WAAW,IAAIX,YAAY,EAAE;IACtCS,MAAM,CAACG,GAAG,CAACD,WAAW,EAAED,MAAM,CAAC;IAC/BA,MAAM,IAAIC,WAAW,CAACN,UAAU;EAClC;;EAGA,OAAOI,MAAM,CAACI,MAAM;AACtB"}
|
|
@@ -1,24 +1,31 @@
|
|
|
1
1
|
import getPixels from 'get-pixels';
|
|
2
|
-
import { assert } from '../../utils/assert';
|
|
3
|
-
import util from 'util';
|
|
4
2
|
|
|
5
3
|
export async function parseImageNode(arrayBuffer, mimeType) {
|
|
6
|
-
|
|
7
|
-
|
|
8
|
-
|
|
4
|
+
if (!mimeType) {
|
|
5
|
+
throw new Error('MIMEType is required to parse image under Node.js');
|
|
6
|
+
}
|
|
9
7
|
const buffer = arrayBuffer instanceof Buffer ? arrayBuffer : Buffer.from(arrayBuffer);
|
|
10
8
|
const ndarray = await getPixelsAsync(buffer, mimeType);
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
9
|
+
return ndarray;
|
|
10
|
+
}
|
|
11
|
+
|
|
12
|
+
function getPixelsAsync(buffer, mimeType) {
|
|
13
|
+
return new Promise(resolve => getPixels(buffer, mimeType, (err, ndarray) => {
|
|
14
|
+
if (err) {
|
|
15
|
+
throw err;
|
|
16
|
+
}
|
|
17
|
+
const shape = [...ndarray.shape];
|
|
18
|
+
const layers = ndarray.shape.length === 4 ? ndarray.shape.shift() : 1;
|
|
19
|
+
const data = ndarray.data instanceof Buffer ? new Uint8Array(ndarray.data) : ndarray.data;
|
|
14
20
|
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
21
|
+
resolve({
|
|
22
|
+
shape,
|
|
23
|
+
data,
|
|
24
|
+
width: ndarray.shape[0],
|
|
25
|
+
height: ndarray.shape[1],
|
|
26
|
+
components: ndarray.shape[2],
|
|
27
|
+
layers: layers ? [layers] : []
|
|
28
|
+
});
|
|
29
|
+
}));
|
|
23
30
|
}
|
|
24
31
|
//# sourceMappingURL=parse-image.node.js.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"parse-image.node.js","names":["getPixels","
|
|
1
|
+
{"version":3,"file":"parse-image.node.js","names":["getPixels","parseImageNode","arrayBuffer","mimeType","Error","buffer","Buffer","from","ndarray","getPixelsAsync","Promise","resolve","err","shape","layers","length","shift","data","Uint8Array","width","height","components"],"sources":["../../../../src/node/images/parse-image.node.ts"],"sourcesContent":["import getPixels from 'get-pixels';\n\n// Note: These types are also defined in @loaders.gl/images and need to be kept in sync\ntype NDArray = {\n shape: number[];\n data: Uint8Array;\n width: number;\n height: number;\n components: number;\n layers: number[];\n};\n\nexport async function parseImageNode(arrayBuffer: ArrayBuffer, mimeType: string): Promise<NDArray> {\n if (!mimeType) {\n throw new Error('MIMEType is required to parse image under Node.js');\n }\n\n const buffer = arrayBuffer instanceof Buffer ? arrayBuffer : Buffer.from(arrayBuffer);\n\n const ndarray = await getPixelsAsync(buffer, mimeType);\n\n return ndarray;\n}\n\n// TODO - check if getPixels callback is asynchronous if provided with buffer input\n// if not, parseImage can be a sync function\nfunction getPixelsAsync(buffer: Buffer, mimeType: string): Promise<NDArray> {\n return new Promise<NDArray>((resolve) =>\n getPixels(buffer, mimeType, (err, ndarray) => {\n if (err) {\n throw err;\n }\n\n const shape = [...ndarray.shape];\n const layers = ndarray.shape.length === 4 ? ndarray.shape.shift() : 1;\n const data = ndarray.data instanceof Buffer ? new Uint8Array(ndarray.data) : ndarray.data;\n\n // extract width/height etc\n resolve({\n shape,\n data,\n width: ndarray.shape[0],\n height: ndarray.shape[1],\n components: ndarray.shape[2],\n // TODO - error\n layers: layers ? [layers] : []\n });\n })\n );\n}\n"],"mappings":"AAAA,OAAOA,SAAS,MAAM,YAAY;;AAYlC,OAAO,eAAeC,cAAc,CAACC,WAAwB,EAAEC,QAAgB,EAAoB;EACjG,IAAI,CAACA,QAAQ,EAAE;IACb,MAAM,IAAIC,KAAK,CAAC,mDAAmD,CAAC;EACtE;EAEA,MAAMC,MAAM,GAAGH,WAAW,YAAYI,MAAM,GAAGJ,WAAW,GAAGI,MAAM,CAACC,IAAI,CAACL,WAAW,CAAC;EAErF,MAAMM,OAAO,GAAG,MAAMC,cAAc,CAACJ,MAAM,EAAEF,QAAQ,CAAC;EAEtD,OAAOK,OAAO;AAChB;;AAIA,SAASC,cAAc,CAACJ,MAAc,EAAEF,QAAgB,EAAoB;EAC1E,OAAO,IAAIO,OAAO,CAAWC,OAAO,IAClCX,SAAS,CAACK,MAAM,EAAEF,QAAQ,EAAE,CAACS,GAAG,EAAEJ,OAAO,KAAK;IAC5C,IAAII,GAAG,EAAE;MACP,MAAMA,GAAG;IACX;IAEA,MAAMC,KAAK,GAAG,CAAC,GAAGL,OAAO,CAACK,KAAK,CAAC;IAChC,MAAMC,MAAM,GAAGN,OAAO,CAACK,KAAK,CAACE,MAAM,KAAK,CAAC,GAAGP,OAAO,CAACK,KAAK,CAACG,KAAK,EAAE,GAAG,CAAC;IACrE,MAAMC,IAAI,GAAGT,OAAO,CAACS,IAAI,YAAYX,MAAM,GAAG,IAAIY,UAAU,CAACV,OAAO,CAACS,IAAI,CAAC,GAAGT,OAAO,CAACS,IAAI;;IAGzFN,OAAO,CAAC;MACNE,KAAK;MACLI,IAAI;MACJE,KAAK,EAAEX,OAAO,CAACK,KAAK,CAAC,CAAC,CAAC;MACvBO,MAAM,EAAEZ,OAAO,CAACK,KAAK,CAAC,CAAC,CAAC;MACxBQ,UAAU,EAAEb,OAAO,CAACK,KAAK,CAAC,CAAC,CAAC;MAE5BC,MAAM,EAAEA,MAAM,GAAG,CAACA,MAAM,CAAC,GAAG;IAC9B,CAAC,CAAC;EACJ,CAAC,CAAC,CACH;AACH"}
|
package/dist/index.d.ts
CHANGED
|
@@ -3,5 +3,6 @@ export { BlobPolyfill } from './node/file/blob';
|
|
|
3
3
|
export { FileReaderPolyfill } from './node/file/file-reader';
|
|
4
4
|
export { FilePolyfill } from './node/file/file';
|
|
5
5
|
export { installFilePolyfills } from './node/file/install-file-polyfills';
|
|
6
|
-
export {
|
|
6
|
+
export { fetchNode as _fetchNode } from './node/fetch/fetch.node';
|
|
7
|
+
export { fetchFileNode as _fetchFileNode } from './node/fetch/fetch-file.node';
|
|
7
8
|
//# sourceMappingURL=index.d.ts.map
|
package/dist/index.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAiBA,OAAO,EAAC,sBAAsB,EAAC,MAAM,6BAA6B,CAAC;AACnE,OAAO,EAAC,YAAY,EAAC,MAAM,kBAAkB,CAAC;AAC9C,OAAO,EAAC,kBAAkB,EAAC,MAAM,yBAAyB,CAAC;AAC3D,OAAO,EAAC,YAAY,EAAC,MAAM,kBAAkB,CAAC;AAC9C,OAAO,EAAC,oBAAoB,EAAC,MAAM,oCAAoC,CAAC;AACxE,OAAO,EAAC,
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":"AAiBA,OAAO,EAAC,sBAAsB,EAAC,MAAM,6BAA6B,CAAC;AACnE,OAAO,EAAC,YAAY,EAAC,MAAM,kBAAkB,CAAC;AAC9C,OAAO,EAAC,kBAAkB,EAAC,MAAM,yBAAyB,CAAC;AAC3D,OAAO,EAAC,YAAY,EAAC,MAAM,kBAAkB,CAAC;AAC9C,OAAO,EAAC,oBAAoB,EAAC,MAAM,oCAAoC,CAAC;AACxE,OAAO,EAAC,SAAS,IAAI,UAAU,EAAC,MAAM,yBAAyB,CAAC;AAChE,OAAO,EAAC,aAAa,IAAI,cAAc,EAAC,MAAM,8BAA8B,CAAC"}
|
package/dist/index.js
CHANGED
|
@@ -22,11 +22,8 @@ var __importStar = (this && this.__importStar) || function (mod) {
|
|
|
22
22
|
__setModuleDefault(result, mod);
|
|
23
23
|
return result;
|
|
24
24
|
};
|
|
25
|
-
var __importDefault = (this && this.__importDefault) || function (mod) {
|
|
26
|
-
return (mod && mod.__esModule) ? mod : { "default": mod };
|
|
27
|
-
};
|
|
28
25
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
29
|
-
exports._fetchNode = exports.installFilePolyfills = exports.FilePolyfill = exports.FileReaderPolyfill = exports.BlobPolyfill = exports.ReadableStreamPolyfill = void 0;
|
|
26
|
+
exports._fetchFileNode = exports._fetchNode = exports.installFilePolyfills = exports.FilePolyfill = exports.FileReaderPolyfill = exports.BlobPolyfill = exports.ReadableStreamPolyfill = void 0;
|
|
30
27
|
/* eslint-disable dot-notation */
|
|
31
28
|
const xmldom_1 = require("@xmldom/xmldom");
|
|
32
29
|
const globals_1 = require("./utils/globals");
|
|
@@ -34,9 +31,9 @@ const encoding_1 = require("./lib/encoding");
|
|
|
34
31
|
const all_settled_1 = require("./promise/all-settled");
|
|
35
32
|
// Node specific
|
|
36
33
|
const base64 = __importStar(require("./node/buffer/btoa.node"));
|
|
37
|
-
const headers_node_1 =
|
|
38
|
-
const response_node_1 =
|
|
39
|
-
const fetch_node_1 =
|
|
34
|
+
const headers_node_1 = require("./node/fetch/headers.node");
|
|
35
|
+
const response_node_1 = require("./node/fetch/response.node");
|
|
36
|
+
const fetch_node_1 = require("./node/fetch/fetch.node");
|
|
40
37
|
const encode_image_node_1 = require("./node/images/encode-image.node");
|
|
41
38
|
const parse_image_node_1 = require("./node/images/parse-image.node");
|
|
42
39
|
var readable_stream_1 = require("./node/file/readable-stream");
|
|
@@ -50,7 +47,9 @@ Object.defineProperty(exports, "FilePolyfill", { enumerable: true, get: function
|
|
|
50
47
|
var install_file_polyfills_1 = require("./node/file/install-file-polyfills");
|
|
51
48
|
Object.defineProperty(exports, "installFilePolyfills", { enumerable: true, get: function () { return install_file_polyfills_1.installFilePolyfills; } });
|
|
52
49
|
var fetch_node_2 = require("./node/fetch/fetch.node");
|
|
53
|
-
Object.defineProperty(exports, "_fetchNode", { enumerable: true, get: function () { return
|
|
50
|
+
Object.defineProperty(exports, "_fetchNode", { enumerable: true, get: function () { return fetch_node_2.fetchNode; } });
|
|
51
|
+
var fetch_file_node_1 = require("./node/fetch/fetch-file.node");
|
|
52
|
+
Object.defineProperty(exports, "_fetchFileNode", { enumerable: true, get: function () { return fetch_file_node_1.fetchFileNode; } });
|
|
54
53
|
// POLYFILLS: TextEncoder, TextDecoder
|
|
55
54
|
// - Recent Node versions have these classes but virtually no encodings unless special build.
|
|
56
55
|
// - Browser: Edge, IE11 do not have these
|
|
@@ -74,14 +73,14 @@ if (!globals_1.isBrowser && !('btoa' in globals_1.global) && base64.btoa) {
|
|
|
74
73
|
// POLYFILL: fetch
|
|
75
74
|
// - Node: Yes
|
|
76
75
|
// - Browser: No. For This polyfill is node only, IE11 etc, install external polyfill
|
|
77
|
-
if (!globals_1.isBrowser && !('Headers' in globals_1.global) && headers_node_1.
|
|
78
|
-
globals_1.global['Headers'] = headers_node_1.
|
|
76
|
+
if (!globals_1.isBrowser && !('Headers' in globals_1.global) && headers_node_1.Headers) {
|
|
77
|
+
globals_1.global['Headers'] = headers_node_1.Headers;
|
|
79
78
|
}
|
|
80
|
-
if (!globals_1.isBrowser && !('Response' in globals_1.global) && response_node_1.
|
|
81
|
-
globals_1.global['Response'] = response_node_1.
|
|
79
|
+
if (!globals_1.isBrowser && !('Response' in globals_1.global) && response_node_1.Response) {
|
|
80
|
+
globals_1.global['Response'] = response_node_1.Response;
|
|
82
81
|
}
|
|
83
|
-
if (!globals_1.isBrowser && !('fetch' in globals_1.global) && fetch_node_1.
|
|
84
|
-
globals_1.global['fetch'] = fetch_node_1.
|
|
82
|
+
if (!globals_1.isBrowser && !('fetch' in globals_1.global) && fetch_node_1.fetchNode) {
|
|
83
|
+
globals_1.global['fetch'] = fetch_node_1.fetchNode;
|
|
85
84
|
}
|
|
86
85
|
// POLYFILL: DOMParser
|
|
87
86
|
// - Node: Yes
|
package/dist/lib/encoding.js
CHANGED
|
@@ -11,7 +11,7 @@ exports.TextDecoder = exports.TextEncoder = void 0;
|
|
|
11
11
|
// See LICENSE.md for more information.
|
|
12
12
|
const encoding_indexes_1 = __importDefault(require("./encoding-indexes"));
|
|
13
13
|
// Note: Aaian character indices add half a megabyte to bundle. Ignore, since we really only want the built-in UTF8...
|
|
14
|
-
// import indexes from './encoding-indexes-asian
|
|
14
|
+
// import indexes from './encoding-indexes-asian';
|
|
15
15
|
global['encoding-indexes'] = encoding_indexes_1.default || {};
|
|
16
16
|
//
|
|
17
17
|
// Utilities
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"fetch-file.node.d.ts","sourceRoot":"","sources":["../../../src/node/fetch/fetch-file.node.ts"],"names":[],"mappings":"AAGA,OAAO,EAAC,QAAQ,EAAC,MAAM,iBAAiB,CAAC;AAGzC,wBAAgB,YAAY,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAEjD;AAED,wBAAsB,aAAa,CAAC,GAAG,EAAE,MAAM,EAAE,OAAO,KAAA,GAAG,OAAO,CAAC,QAAQ,CAAC,CAsB3E"}
|
|
@@ -0,0 +1,51 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
// loaders.gl, MIT license
|
|
3
|
+
var __importDefault = (this && this.__importDefault) || function (mod) {
|
|
4
|
+
return (mod && mod.__esModule) ? mod : { "default": mod };
|
|
5
|
+
};
|
|
6
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
7
|
+
exports.fetchFileNode = exports.isRequestURL = void 0;
|
|
8
|
+
const fs_1 = __importDefault(require("fs")); // `fs` will be empty object in browsers (see package.json "browser" field).
|
|
9
|
+
const response_node_1 = require("./response.node");
|
|
10
|
+
const headers_node_1 = require("./headers.node");
|
|
11
|
+
function isRequestURL(url) {
|
|
12
|
+
return url.startsWith('http:') || url.startsWith('https:');
|
|
13
|
+
}
|
|
14
|
+
exports.isRequestURL = isRequestURL;
|
|
15
|
+
async function fetchFileNode(url, options) {
|
|
16
|
+
const noqueryUrl = url.split('?')[0];
|
|
17
|
+
try {
|
|
18
|
+
// Now open the stream
|
|
19
|
+
const body = await new Promise((resolve, reject) => {
|
|
20
|
+
// @ts-ignore
|
|
21
|
+
const stream = fs_1.default.createReadStream(noqueryUrl, { encoding: null });
|
|
22
|
+
stream.once('readable', () => resolve(stream));
|
|
23
|
+
stream.on('error', (error) => reject(error));
|
|
24
|
+
});
|
|
25
|
+
const status = 200;
|
|
26
|
+
const statusText = 'OK';
|
|
27
|
+
const headers = getHeadersForFile(noqueryUrl);
|
|
28
|
+
return new response_node_1.Response(body, { headers, status, statusText, url });
|
|
29
|
+
}
|
|
30
|
+
catch (error) {
|
|
31
|
+
const status = 400;
|
|
32
|
+
const statusText = error.message;
|
|
33
|
+
const headers = {};
|
|
34
|
+
return new response_node_1.Response(error.message, { headers, status, statusText, url });
|
|
35
|
+
}
|
|
36
|
+
}
|
|
37
|
+
exports.fetchFileNode = fetchFileNode;
|
|
38
|
+
function getHeadersForFile(noqueryUrl) {
|
|
39
|
+
const headers = {};
|
|
40
|
+
// Fix up content length if we can for best progress experience
|
|
41
|
+
if (!headers['content-length']) {
|
|
42
|
+
const stats = fs_1.default.statSync(noqueryUrl);
|
|
43
|
+
headers['content-length'] = stats.size;
|
|
44
|
+
}
|
|
45
|
+
// Automatically decompress gzipped files with .gz extension
|
|
46
|
+
if (noqueryUrl.endsWith('.gz')) {
|
|
47
|
+
noqueryUrl = noqueryUrl.slice(0, -3);
|
|
48
|
+
headers['content-encoding'] = 'gzip';
|
|
49
|
+
}
|
|
50
|
+
return new headers_node_1.Headers(headers);
|
|
51
|
+
}
|
|
@@ -1,7 +1,12 @@
|
|
|
1
|
+
/// <reference types="node" />
|
|
2
|
+
import http from 'http';
|
|
3
|
+
import { Response } from './response.node';
|
|
1
4
|
/**
|
|
2
5
|
* Emulation of Browser fetch for Node.js
|
|
3
6
|
* @param url
|
|
4
7
|
* @param options
|
|
5
8
|
*/
|
|
6
|
-
export
|
|
9
|
+
export declare function fetchNode(url: string, options: any): Promise<Response>;
|
|
10
|
+
/** Returns a promise that resolves to a readable stream */
|
|
11
|
+
export declare function createHTTPRequestReadStream(url: string, options: any): Promise<http.IncomingMessage>;
|
|
7
12
|
//# sourceMappingURL=fetch.node.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"fetch.node.d.ts","sourceRoot":"","sources":["../../../src/node/fetch/fetch.node.ts"],"names":[],"mappings":"
|
|
1
|
+
{"version":3,"file":"fetch.node.d.ts","sourceRoot":"","sources":["../../../src/node/fetch/fetch.node.ts"],"names":[],"mappings":";AAEA,OAAO,IAAI,MAAM,MAAM,CAAC;AAExB,OAAO,EAAC,QAAQ,EAAC,MAAM,iBAAiB,CAAC;AASzC;;;;GAIG;AAEH,wBAAsB,SAAS,CAAC,GAAG,EAAE,MAAM,EAAE,OAAO,KAAA,GAAG,OAAO,CAAC,QAAQ,CAAC,CA6CvE;AAED,2DAA2D;AAC3D,wBAAsB,2BAA2B,CAC/C,GAAG,EAAE,MAAM,EACX,OAAO,KAAA,GACN,OAAO,CAAC,IAAI,CAAC,eAAe,CAAC,CAW/B"}
|
|
@@ -1,13 +1,16 @@
|
|
|
1
1
|
"use strict";
|
|
2
|
+
// loaders.gl, MIT license
|
|
2
3
|
var __importDefault = (this && this.__importDefault) || function (mod) {
|
|
3
4
|
return (mod && mod.__esModule) ? mod : { "default": mod };
|
|
4
5
|
};
|
|
5
6
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
6
|
-
|
|
7
|
-
const
|
|
8
|
-
const
|
|
7
|
+
exports.createHTTPRequestReadStream = exports.fetchNode = void 0;
|
|
8
|
+
const http_1 = __importDefault(require("http"));
|
|
9
|
+
const https_1 = __importDefault(require("https"));
|
|
10
|
+
const response_node_1 = require("./response.node");
|
|
11
|
+
const headers_node_1 = require("./headers.node");
|
|
9
12
|
const decode_data_uri_node_1 = require("./utils/decode-data-uri.node");
|
|
10
|
-
const
|
|
13
|
+
const fetch_file_node_1 = require("./fetch-file.node");
|
|
11
14
|
const isDataURL = (url) => url.startsWith('data:');
|
|
12
15
|
const isRequestURL = (url) => url.startsWith('http:') || url.startsWith('https:');
|
|
13
16
|
/**
|
|
@@ -15,13 +18,18 @@ const isRequestURL = (url) => url.startsWith('http:') || url.startsWith('https:'
|
|
|
15
18
|
* @param url
|
|
16
19
|
* @param options
|
|
17
20
|
*/
|
|
21
|
+
// eslint-disable-next-line complexity
|
|
18
22
|
async function fetchNode(url, options) {
|
|
19
23
|
try {
|
|
24
|
+
// Handle file streams in node
|
|
25
|
+
if (!isRequestURL(url) && !isDataURL(url)) {
|
|
26
|
+
return await (0, fetch_file_node_1.fetchFileNode)(url, options);
|
|
27
|
+
}
|
|
20
28
|
// Handle data urls in node, to match `fetch``
|
|
21
29
|
// Note - this loses the MIME type, data URIs are handled directly in fetch
|
|
22
30
|
if (isDataURL(url)) {
|
|
23
31
|
const { arrayBuffer, mimeType } = (0, decode_data_uri_node_1.decodeDataUri)(url);
|
|
24
|
-
const response = new response_node_1.
|
|
32
|
+
const response = new response_node_1.Response(arrayBuffer, {
|
|
25
33
|
headers: { 'content-type': mimeType },
|
|
26
34
|
url
|
|
27
35
|
});
|
|
@@ -35,7 +43,7 @@ async function fetchNode(url, options) {
|
|
|
35
43
|
syntheticResponseHeaders['content-encoding'] = 'gzip';
|
|
36
44
|
}
|
|
37
45
|
// Need to create the stream in advance since Response constructor needs to be sync
|
|
38
|
-
const body = await (
|
|
46
|
+
const body = await createHTTPRequestReadStream(originalUrl, options);
|
|
39
47
|
const headers = getHeaders(url, body, syntheticResponseHeaders);
|
|
40
48
|
const { status, statusText } = getStatus(body);
|
|
41
49
|
const followRedirect = !options || options.followRedirect || options.followRedirect === undefined;
|
|
@@ -44,14 +52,28 @@ async function fetchNode(url, options) {
|
|
|
44
52
|
// Redirect
|
|
45
53
|
return await fetchNode(redirectUrl, options);
|
|
46
54
|
}
|
|
47
|
-
return new response_node_1.
|
|
55
|
+
return new response_node_1.Response(body, { headers, status, statusText, url });
|
|
48
56
|
}
|
|
49
57
|
catch (error) {
|
|
50
58
|
// TODO - what error code to use here?
|
|
51
|
-
return new response_node_1.
|
|
59
|
+
return new response_node_1.Response(null, { status: 400, statusText: String(error), url });
|
|
52
60
|
}
|
|
53
61
|
}
|
|
54
|
-
exports.
|
|
62
|
+
exports.fetchNode = fetchNode;
|
|
63
|
+
/** Returns a promise that resolves to a readable stream */
|
|
64
|
+
async function createHTTPRequestReadStream(url, options) {
|
|
65
|
+
// HANDLE HTTP/HTTPS REQUESTS IN NODE
|
|
66
|
+
// TODO: THIS IS BAD SINCE WE RETURN A PROMISE INSTEAD OF A STREAM
|
|
67
|
+
return await new Promise((resolve, reject) => {
|
|
68
|
+
const requestOptions = getRequestOptions(url, options);
|
|
69
|
+
const req = url.startsWith('https:')
|
|
70
|
+
? https_1.default.request(requestOptions, (res) => resolve(res))
|
|
71
|
+
: http_1.default.request(requestOptions, (res) => resolve(res));
|
|
72
|
+
req.on('error', (error) => reject(error));
|
|
73
|
+
req.end();
|
|
74
|
+
});
|
|
75
|
+
}
|
|
76
|
+
exports.createHTTPRequestReadStream = createHTTPRequestReadStream;
|
|
55
77
|
/**
|
|
56
78
|
* Generate redirect url from location without origin and protocol.
|
|
57
79
|
* @param originalUrl
|
|
@@ -67,7 +89,28 @@ function generateRedirectUrl(originalUrl, location) {
|
|
|
67
89
|
return url.href;
|
|
68
90
|
}
|
|
69
91
|
// HELPER FUNCTIONS
|
|
70
|
-
|
|
92
|
+
function getRequestOptions(url, options) {
|
|
93
|
+
// Ensure header keys are lower case so that we can merge without duplicates
|
|
94
|
+
const originalHeaders = options?.headers || {};
|
|
95
|
+
const headers = {};
|
|
96
|
+
for (const key of Object.keys(originalHeaders)) {
|
|
97
|
+
headers[key.toLowerCase()] = originalHeaders[key];
|
|
98
|
+
}
|
|
99
|
+
// Add default accept-encoding to headers
|
|
100
|
+
headers['accept-encoding'] = headers['accept-encoding'] || 'gzip,br,deflate';
|
|
101
|
+
const urlObject = new URL(url);
|
|
102
|
+
return {
|
|
103
|
+
hostname: urlObject.hostname,
|
|
104
|
+
path: urlObject.pathname,
|
|
105
|
+
method: 'GET',
|
|
106
|
+
// Add options and user provided 'options.fetch' overrides if available
|
|
107
|
+
...options,
|
|
108
|
+
...options?.fetch,
|
|
109
|
+
// Override with updated headers with accepted encodings:
|
|
110
|
+
headers,
|
|
111
|
+
port: urlObject.port
|
|
112
|
+
};
|
|
113
|
+
}
|
|
71
114
|
function getStatus(httpResponse) {
|
|
72
115
|
if (httpResponse.statusCode) {
|
|
73
116
|
return { status: httpResponse.statusCode, statusText: httpResponse.statusMessage || 'NA' };
|
|
@@ -91,27 +134,10 @@ function getHeaders(url, httpResponse, additionalHeaders = {}) {
|
|
|
91
134
|
}
|
|
92
135
|
}
|
|
93
136
|
Object.assign(headers, additionalHeaders);
|
|
94
|
-
return new headers_node_1.
|
|
137
|
+
return new headers_node_1.Headers(headers);
|
|
95
138
|
}
|
|
139
|
+
/** Needs to be read from actual headers */
|
|
96
140
|
function getContentLength(url) {
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
return null;
|
|
100
|
-
}
|
|
101
|
-
else if (isDataURL(url)) {
|
|
102
|
-
// TODO - remove media type etc
|
|
103
|
-
return url.length - 'data:'.length;
|
|
104
|
-
}
|
|
105
|
-
// File URL
|
|
106
|
-
// TODO - how to handle non-existing file, this presumably just throws
|
|
107
|
-
try {
|
|
108
|
-
// strip query params from URL
|
|
109
|
-
const noqueryUrl = url.split('?')[0];
|
|
110
|
-
const stats = fs_1.default.statSync(noqueryUrl);
|
|
111
|
-
return stats.size;
|
|
112
|
-
}
|
|
113
|
-
catch (error) {
|
|
114
|
-
// ignore for now
|
|
115
|
-
}
|
|
116
|
-
return null;
|
|
141
|
+
// TODO - remove media type etc
|
|
142
|
+
return isDataURL(url) ? url.length - 'data:'.length : null;
|
|
117
143
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"headers.node.d.ts","sourceRoot":"","sources":["../../../src/node/fetch/headers.node.ts"],"names":[],"mappings":"AAAA;;;GAGG;AACH,
|
|
1
|
+
{"version":3,"file":"headers.node.d.ts","sourceRoot":"","sources":["../../../src/node/fetch/headers.node.ts"],"names":[],"mappings":"AAAA;;;GAGG;AACH,qBAAa,OAAO;IAClB,GAAG,EAAE,EAAE,CAAC;gBAEI,OAAO,KAAA;IAYnB,MAAM,CAAC,IAAI,KAAA,EAAE,KAAK,KAAA;IAOlB,MAAM,CAAC,IAAI,KAAA;IAIX,GAAG,CAAC,IAAI,KAAA;IAKR,GAAG,CAAC,IAAI,KAAA;IAIR,GAAG,CAAC,IAAI,KAAA,EAAE,KAAK,KAAA;IAIf,OAAO,CAAC,OAAO,KAAA,EAAE,OAAO,OAAO;IAY/B,IAAI;;;;;;IAQJ,MAAM;;;;;;IAQN,OAAO;;;;;;IAQN,CAAC,MAAM,CAAC,QAAQ,CAAC;CAInB"}
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
"use strict";
|
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.Headers = void 0;
|
|
3
4
|
/**
|
|
4
5
|
* Polyfill for Browser Headers
|
|
5
6
|
* Based on https://github.com/github/fetch under MIT license
|
|
@@ -74,7 +75,7 @@ class Headers {
|
|
|
74
75
|
yield* this.entries();
|
|
75
76
|
}
|
|
76
77
|
}
|
|
77
|
-
exports.
|
|
78
|
+
exports.Headers = Headers;
|
|
78
79
|
function normalizeName(name) {
|
|
79
80
|
if (typeof name !== 'string') {
|
|
80
81
|
name = String(name);
|