@expofp/loader 1.0.58 → 1.0.60

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (32) hide show
  1. package/dist/bundle/bundle.js +156 -156
  2. package/dist/bundle/bundle.js.map +1 -1
  3. package/dist/bundle/{downloadZip-BbVnryS9.js → downloadZip-CohB2rQN.js} +326 -324
  4. package/dist/bundle/downloadZip-CohB2rQN.js.map +1 -0
  5. package/dist/bundle/makeOffline-DuRXI_Dd.js.map +1 -1
  6. package/dist/bundle/makeOfflineBundle-XjvmvDQE.js.map +1 -1
  7. package/dist/esm/importJson.js +2 -1
  8. package/dist/esm/index.js +3 -3
  9. package/dist/esm/loadAndWaitGlobal.js +1 -1
  10. package/dist/esm/logger.d.ts +1 -2
  11. package/dist/esm/logger.js +4 -1
  12. package/dist/esm/offline/downloadZip.js +2 -0
  13. package/dist/esm/offline/makeOffline.js +2 -1
  14. package/dist/esm/offline/makeOfflineBundle.js +1 -1
  15. package/dist/esm/resolve.js +2 -1
  16. package/dist/esm/resolvers/_OLD_expoResolver.js +1 -1
  17. package/dist/esm/resolvers/assetResolver.js +1 -1
  18. package/dist/esm/resolvers/bundleAssetsResolver.js +1 -1
  19. package/dist/esm/resolvers/expoRuntimeBranchResolver.js +1 -1
  20. package/dist/esm/resolvers/expoRuntimeGetBranchResolver.js +1 -1
  21. package/dist/esm/resolvers/expoRuntimeResolver.js +2 -1
  22. package/dist/esm/resolvers/httpResolver.js +1 -1
  23. package/dist/esm/resolvers/legacyAssetUrlsResolver.js +9 -10
  24. package/dist/esm/resolvers/legacyDataResolver.js +1 -1
  25. package/dist/esm/shared.d.ts +0 -1
  26. package/dist/esm/shared.js +0 -6
  27. package/package.json +1 -1
  28. package/dist/bundle/downloadZip-BbVnryS9.js.map +0 -1
  29. package/dist/esm/offline/downloadZip copy.d.ts +0 -1
  30. package/dist/esm/offline/downloadZip copy.js +0 -89
  31. package/dist/esm/offline/makeOfflineBundle copy.d.ts +0 -1
  32. package/dist/esm/offline/makeOfflineBundle copy.js +0 -92
@@ -1 +1 @@
1
- {"version":3,"file":"makeOffline-DuRXI_Dd.js","sources":["../../src/offline/hashString.ts","../../src/offline/slugify.ts","../../src/offline/tools.ts","../../src/offline/makeOffline.ts"],"sourcesContent":["// 32-bit murmur-ish hash with configurable seed\nfunction murmur32WithSeed(str: string, seed: number): string {\n let h = seed >>> 0;\n\n for (let i = 0; i < str.length; i++) {\n h ^= str.charCodeAt(i);\n h = Math.imul(h, 0x01000193); // FNV-ish prime, good mixer\n }\n\n return (h >>> 0).toString(16).padStart(8, '0'); // 8 hex chars\n}\n\n// 128-bit (MD5-width) hash: 4 × 32-bit parts concatenated\nexport function hashString(str: string): string {\n return (\n murmur32WithSeed(str, 0x811c9dc5) + // your original seed\n murmur32WithSeed(str, 0x21f0aaad) +\n murmur32WithSeed(str, 0x1b873593) +\n murmur32WithSeed(str, 0x85ebca6b)\n );\n}\n","// Module-level caches\nconst inputToSlug = new Map<string, string>();\nconst slugToInput = new Map<string, string>();\n\nconst WINDOWS_RESERVED = new Set([\n \"con\", \"prn\", \"aux\", \"nul\",\n \"com1\", \"com2\", \"com3\", \"com4\", \"com5\", \"com6\", \"com7\", \"com8\", \"com9\",\n \"lpt1\", \"lpt2\", \"lpt3\", \"lpt4\", \"lpt5\", \"lpt6\", \"lpt7\", \"lpt8\", \"lpt9\",\n]);\n\nfunction makeBaseSlug(input: string): string {\n let slug = input\n .normalize(\"NFKD\")\n // All non letters/numbers → \"-\"\n .replace(/[^\\p{Letter}\\p{Number}]+/gu, \"-\")\n // Collapse multiple \"-\"\n .replace(/-+/g, \"-\")\n // Trim \"-\" from start/end\n .replace(/^-|-$/g, \"\")\n .toLowerCase();\n\n // Strip forbidden Windows characters just in case\n slug = slug.replace(/[<>:\"/\\\\|?*]/g, \"\");\n\n // Windows forbids trailing space/period\n slug = slug.replace(/[. ]+$/g, \"\");\n\n // Fallback if everything was stripped\n if (!slug) slug = \"file\";\n\n // Avoid bare reserved device names\n if (WINDOWS_RESERVED.has(slug)) {\n slug += \"-file\";\n }\n\n return slug;\n}\n\nexport function slugifyFsUnique(input: string): string {\n // If we've seen this exact input before, return the same slug\n const existing = inputToSlug.get(input);\n if (existing) return existing;\n\n const base = makeBaseSlug(input);\n let candidate = base;\n let counter = 2;\n\n while (true) {\n const existingInput = slugToInput.get(candidate);\n\n if (!existingInput) {\n // Free slug → claim it for this input\n slugToInput.set(candidate, input);\n inputToSlug.set(input, candidate);\n return candidate;\n }\n\n if (existingInput === input) {\n // Same input somehow (super defensive)\n inputToSlug.set(input, candidate);\n return candidate;\n }\n\n // Collision: same slug already used by different input → add suffix\n candidate = `${base}-${counter++}`;\n }\n}\n\n// // Optional: to reset between runs/tests\n// export function resetSlugCache() {\n// inputToSlug.clear();\n// slugToInput.clear();\n// }","import { hashString } from './hashString';\nimport { slugifyFsUnique } from './slugify';\n\nexport function makeUniqueJsonTargetPathFromString(str: string, namespace: string = ''): string {\n // const hash = hashString(str);\n let result = slugifyFsUnique(str); // + '-' + hash;\n if (namespace) {\n result = `${slugifyFsUnique(namespace)}$${result}`;\n }\n\n if (result.endsWith('/')) {\n result += 'index.json';\n } else if (!result.endsWith('.json')) {\n result += '.json';\n }\n return './' + result;\n // handle directory case\n}\n\nexport function makeTargetPathFromUrl(url: string, prefix: string = ''): string {\n // https://example.com/dir1/dir2/a.js => \"{prefix}{origin-slug}/dir1/dir2/a.js\";\n // https://example.com/dir1/dir2/a.js?params => \"{prefix}{origin-slug}/dir1/dir2/a{paramsmd5hash}.js\";\n // use slugify.ts\n\n try {\n new URL(url);\n } catch {\n debugger;\n }\n const urlObj = new URL(url);\n const origin = `${urlObj.protocol}//${urlObj.host}`;\n const originSlug = slugifyFsUnique(origin);\n\n let pathname = urlObj.pathname;\n let search = urlObj.search;\n\n // if path doesn't end with extension, throw\n if (!pathname.match(/\\.[^\\/]+$/)) {\n throw new Error(`Cannot make target path from URL without file extension: ${url}`);\n }\n\n const extension = pathname.substring(pathname.lastIndexOf('.'));\n let pathnameWithoutExtension = pathname.substring(0, pathname.lastIndexOf('.'));\n\n // check the pathname contains only valid fs characters\n const invalidPathnameChars = pathnameWithoutExtension.match(/[^a-zA-Z0-9\\-._\\/]/g);\n if (invalidPathnameChars) {\n const fixedPathnameWithoutExtension = slugifyFsUnique(pathnameWithoutExtension);\n console.warn(\n `Warning: pathname contains invalid filesystem characters (${[\n ...new Set(invalidPathnameChars),\n ].join(\n ', '\n )}), slugifying it: ${pathnameWithoutExtension}${extension} => ${fixedPathnameWithoutExtension}${extension}`\n );\n pathnameWithoutExtension = fixedPathnameWithoutExtension;\n }\n\n pathname = pathnameWithoutExtension + extension;\n\n if (pathname.length > 120) {\n console.warn(\n `Warning: pathname is too long (${pathname.length} characters), truncating to 150 characters: ${pathname}`\n );\n pathname = pathname.substring(0, 120 - extension.length) + extension;\n }\n\n if (search) {\n // create a hash from search params\n const hash = hashString(search);\n const dotIndex = pathname.lastIndexOf('.');\n if (dotIndex !== -1) {\n pathname = `${pathname.slice(0, dotIndex)}.${hash}${pathname.slice(dotIndex)}`;\n } else {\n pathname = `${pathname}${hash}`;\n }\n }\n\n // // handle directory case\n // if (pathname.endsWith('/')) {\n // pathname += '__index.json';\n // }\n\n return `./${prefix}${originSlug}${pathname}`;\n}\n","import {\n canResolve,\n parseRefValue,\n resolverResolve,\n type ResolveContextInternal,\n} from '../resolve';\nimport { resolvers } from '../resolvers';\nimport { createMergedObjectWithOverridenNonRefProps, deepClone, log } from '../shared';\nimport type { LocalFile, MakeOfflineResult, ManifestData, Resolver } from '../types';\nimport { makeTargetPathFromUrl, makeUniqueJsonTargetPathFromString } from './tools';\n\nexport async function makeOffline(manifest: any): Promise<MakeOfflineResult> {\n log('makeOffline', manifest);\n // if (typeof manifest !== 'object' || manifest === null) {\n // throw new Error('Manifest must be an object');\n // }\n\n const res = await makeOfflineInternal(manifest);\n let done = false;\n let resultManifest: any;\n async function* files() {\n resultManifest = yield* res;\n done = true;\n }\n\n return {\n get manifest(): ManifestData {\n if (!done) throw new Error('Iterate over files before getting manifest');\n return resultManifest;\n },\n files: files() as AsyncGenerator<LocalFile, void, void>,\n };\n}\n\nasync function* makeOfflineInternal(manifest: any): AsyncGenerator<LocalFile, ManifestData, void> {\n log('makeOfflineInternal', manifest);\n const resolveContext: ResolveContextInternal = {\n refCache: new Map<string, Promise<any>>(),\n forceFetch: true,\n signal: null,\n };\n const parent = { manifest };\n yield* walk(parent, 'manifest');\n return parent.manifest;\n\n async function* walk(node: any, key: string | number): AsyncGenerator<LocalFile> {\n // console.log('walk', node, key, node[key]);\n // let node[key] = node[key];\n if (typeof node[key] !== 'object' || node[key] === null) {\n return;\n }\n // let resolversToUse1 = resolvers.filter((r) => r.canResolve(node[key].$ref));\n // const usedResolvers = new Set<Resolver>();\n // do {\n if ('$ref' in node[key]) {\n const resolversToUse = resolvers.filter((r) => canResolve(r, node[key].$ref));\n if (resolversToUse.length === 0) {\n throw new Error(`No resolver found for ref: ${node[key].$ref}`);\n }\n if (resolversToUse.length > 1) {\n throw new Error(`Multiple resolvers can make offline ref: ${node[key].$ref}`);\n }\n\n const resolver = resolversToUse[0];\n\n let func: typeof offlineLocalizeRef;\n switch (resolver.offlineMethod) {\n case 'localizeRef':\n func = offlineLocalizeRef;\n break;\n case 'resolveRef':\n func = offlineResolveRef;\n break;\n default:\n throw new Error(`Unknown offlineMethod: ${resolver.offlineMethod}`);\n }\n\n const mergeRef = yield* func(resolver, node[key].$ref, resolveContext);\n\n if (Object.isFrozen(node[key])) throw new Error('Unexpected frozen node during makeOffline');\n\n node[key] = createMergedObjectWithOverridenNonRefProps({ $ref: mergeRef }, node[key]);\n }\n\n // recurse\n if (Array.isArray(node[key])) {\n for (const [index] of node[key].entries()) {\n yield* walk(node[key], index);\n }\n } else if (typeof node[key] === 'object' && node[key] !== null) {\n for (const key1 of Object.keys(node[key])) {\n // debugger;\n yield* walk(node[key], key1);\n }\n }\n }\n}\n\n// function canResolve(resolver: Resolver, ref: string): boolean {\n// if (resolver.canResolve) {\n// return resolver.canResolve(ref);\n// }\n// if (resolver.schema) {\n// return canResolveRefSchema(ref, resolver.schema);\n// }\n// throw new Error('Resolver is missing canResolve method and schema property');\n// }\n\nasync function* offlineLocalizeRef(\n resolver: Resolver,\n ref: string,\n context: ResolveContextInternal\n): AsyncGenerator<LocalFile, string, void> {\n const refData = deepClone(await resolverResolve(resolver, ref, context));\n // emit assets\n yield* makeOfflineInternal(refData);\n const url = parseRefValue(ref);\n const targetFilePath = makeTargetPathFromUrl(url);\n // TODO: handle CSS and other text files that may reference other assets\n yield { url, targetFilePath };\n const schema = ref.substring(0, ref.length - url.length - 1);\n return `${schema}:${targetFilePath}`;\n}\n\nasync function* offlineResolveRef(\n resolver: Resolver,\n ref: string,\n context: ResolveContextInternal\n): AsyncGenerator<LocalFile, string, void> {\n const refData = deepClone(await resolverResolve(resolver, ref, context));\n const targetFilePath = makeUniqueJsonTargetPathFromString(ref);\n const data = yield* makeOfflineInternal(refData);\n yield { data, targetFilePath };\n return targetFilePath;\n}\n\n// export function createMakeOfflineResult(\n// manifest: any,\n// walk: (node: any, key: string | number) => AsyncGenerator<LocalFile>\n// ): MakeOfflineResult {\n// log('createMakeOfflineResult', manifest);\n// if (typeof manifest !== 'object' || manifest === null) {\n// throw new Error('Manifest must be an object');\n// }\n// const parent = { manifest };\n// let done = false;\n// async function* files() {\n// yield* walk(parent, 'manifest');\n// done = true;\n// }\n\n// return {\n// get manifest() {\n// if (!done) {\n// throw new Error('Cannot access manifest before all files are generated');\n// }\n// return parent.manifest;\n// },\n// files: files(),\n// };\n// }\n"],"names":["murmur32WithSeed","str","seed","h","i","hashString","inputToSlug","slugToInput","WINDOWS_RESERVED","makeBaseSlug","input","slug","slugifyFsUnique","existing","base","candidate","counter","existingInput","makeUniqueJsonTargetPathFromString","namespace","result","makeTargetPathFromUrl","url","prefix","urlObj","origin","originSlug","pathname","search","extension","pathnameWithoutExtension","invalidPathnameChars","fixedPathnameWithoutExtension","hash","dotIndex","makeOffline","manifest","log","res","makeOfflineInternal","done","resultManifest","files","resolveContext","parent","walk","node","key","resolversToUse","resolvers","r","canResolve","resolver","func","offlineLocalizeRef","offlineResolveRef","mergeRef","createMergedObjectWithOverridenNonRefProps","index","key1","ref","context","refData","deepClone","resolverResolve","parseRefValue","targetFilePath"],"mappings":";AACA,SAASA,EAAiBC,GAAaC,GAAsB;AAC3D,MAAIC,IAAID,MAAS;AAEjB,WAASE,IAAI,GAAGA,IAAIH,EAAI,QAAQG;AAC9B,IAAAD,KAAKF,EAAI,WAAWG,CAAC,GACrBD,IAAI,KAAK,KAAKA,GAAG,QAAU;AAG7B,UAAQA,MAAM,GAAG,SAAS,EAAE,EAAE,SAAS,GAAG,GAAG;AAC/C;AAGO,SAASE,EAAWJ,GAAqB;AAC9C,SACED,EAAiBC,GAAK,UAAU;AAAA,EAChCD,EAAiBC,GAAK,SAAU,IAChCD,EAAiBC,GAAK,SAAU,IAChCD,EAAiBC,GAAK,UAAU;AAEpC;ACnBA,MAAMK,wBAAkB,IAAA,GAClBC,wBAAkB,IAAA,GAElBC,wBAAuB,IAAI;AAAA,EAC/B;AAAA,EAAO;AAAA,EAAO;AAAA,EAAO;AAAA,EACrB;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAChE;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAClE,CAAC;AAED,SAASC,EAAaC,GAAuB;AAC3C,MAAIC,IAAOD,EACR,UAAU,MAAM,EAEhB,QAAQ,8BAA8B,GAAG,EAEzC,QAAQ,OAAO,GAAG,EAElB,QAAQ,UAAU,EAAE,EACpB,YAAA;AAGH,SAAAC,IAAOA,EAAK,QAAQ,iBAAiB,EAAE,GAGvCA,IAAOA,EAAK,QAAQ,WAAW,EAAE,GAG5BA,MAAMA,IAAO,SAGdH,EAAiB,IAAIG,CAAI,MAC3BA,KAAQ,UAGHA;AACT;AAEO,SAASC,EAAgBF,GAAuB;AAErD,QAAMG,IAAWP,EAAY,IAAII,CAAK;AACtC,MAAIG,EAAU,QAAOA;AAErB,QAAMC,IAAOL,EAAaC,CAAK;AAC/B,MAAIK,IAAYD,GACZE,IAAU;AAEd,aAAa;AACX,UAAMC,IAAgBV,EAAY,IAAIQ,CAAS;AAE/C,QAAI,CAACE;AAEH,aAAAV,EAAY,IAAIQ,GAAWL,CAAK,GAChCJ,EAAY,IAAII,GAAOK,CAAS,GACzBA;AAGT,QAAIE,MAAkBP;AAEpB,aAAAJ,EAAY,IAAII,GAAOK,CAAS,GACzBA;AAIT,IAAAA,IAAY,GAAGD,CAAI,IAAIE,GAAS;AAAA,EAClC;AACF;AC/DO,SAASE,EAAmCjB,GAAakB,IAAoB,IAAY;AAE9F,MAAIC,IAASR,EAAgBX,CAAG;AAChC,SAAIkB,MACFC,IAAS,GAAGR,EAAgBO,CAAS,CAAC,IAAIC,CAAM,KAG9CA,EAAO,SAAS,GAAG,IACrBA,KAAU,eACAA,EAAO,SAAS,OAAO,MACjCA,KAAU,UAEL,OAAOA;AAEhB;AAEO,SAASC,EAAsBC,GAAaC,IAAiB,IAAY;AAK9E,MAAI;AACF,QAAI,IAAID,CAAG;AAAA,EACb,QAAQ;AACN;AAAA,EACF;AACA,QAAME,IAAS,IAAI,IAAIF,CAAG,GACpBG,IAAS,GAAGD,EAAO,QAAQ,KAAKA,EAAO,IAAI,IAC3CE,IAAad,EAAgBa,CAAM;AAEzC,MAAIE,IAAWH,EAAO,UAClBI,IAASJ,EAAO;AAGpB,MAAI,CAACG,EAAS,MAAM,WAAW;AAC7B,UAAM,IAAI,MAAM,4DAA4DL,CAAG,EAAE;AAGnF,QAAMO,IAAYF,EAAS,UAAUA,EAAS,YAAY,GAAG,CAAC;AAC9D,MAAIG,IAA2BH,EAAS,UAAU,GAAGA,EAAS,YAAY,GAAG,CAAC;AAG9E,QAAMI,IAAuBD,EAAyB,MAAM,qBAAqB;AACjF,MAAIC,GAAsB;AACxB,UAAMC,IAAgCpB,EAAgBkB,CAAwB;AAC9E,YAAQ;AAAA,MACN,6DAA6D;AAAA,QAC3D,GAAG,IAAI,IAAIC,CAAoB;AAAA,MAAA,EAC/B;AAAA,QACA;AAAA,MAAA,CACD,qBAAqBD,CAAwB,GAAGD,CAAS,OAAOG,CAA6B,GAAGH,CAAS;AAAA,IAAA,GAE5GC,IAA2BE;AAAA,EAC7B;AAWA,MATAL,IAAWG,IAA2BD,GAElCF,EAAS,SAAS,QACpB,QAAQ;AAAA,IACN,kCAAkCA,EAAS,MAAM,+CAA+CA,CAAQ;AAAA,EAAA,GAE1GA,IAAWA,EAAS,UAAU,GAAG,MAAME,EAAU,MAAM,IAAIA,IAGzDD,GAAQ;AAEV,UAAMK,IAAO5B,EAAWuB,CAAM,GACxBM,IAAWP,EAAS,YAAY,GAAG;AACzC,IAAIO,MAAa,KACfP,IAAW,GAAGA,EAAS,MAAM,GAAGO,CAAQ,CAAC,IAAID,CAAI,GAAGN,EAAS,MAAMO,CAAQ,CAAC,KAE5EP,IAAW,GAAGA,CAAQ,GAAGM,CAAI;AAAA,EAEjC;AAOA,SAAO,KAAKV,CAAM,GAAGG,CAAU,GAAGC,CAAQ;AAC5C;ACzEA,eAAsBQ,EAAYC,GAA2C;AAC3E,EAAAC,EAAI,eAAeD,CAAQ;AAK3B,QAAME,IAAM,MAAMC,EAAoBH,CAAQ;AAC9C,MAAII,IAAO,IACPC;AACJ,kBAAgBC,IAAQ;AACtB,IAAAD,IAAiB,OAAOH,GACxBE,IAAO;AAAA,EACT;AAEA,SAAO;AAAA,IACL,IAAI,WAAyB;AAC3B,UAAI,CAACA,EAAM,OAAM,IAAI,MAAM,4CAA4C;AACvE,aAAOC;AAAA,IACT;AAAA,IACA,OAAOC,EAAA;AAAA,EAAM;AAEjB;AAEA,gBAAgBH,EAAoBH,GAA8D;AAChG,EAAAC,EAAI,uBAAuBD,CAAQ;AACnC,QAAMO,IAAyC;AAAA,IAC7C,8BAAc,IAAA;AAAA,IACd,YAAY;AAAA,IACZ,QAAQ;AAAA,EAAA,GAEJC,IAAS,EAAE,UAAAR,EAAA;AACjB,gBAAOS,EAAKD,GAAQ,UAAU,GACvBA,EAAO;AAEd,kBAAgBC,EAAKC,GAAWC,GAAiD;AAG/E,QAAI,SAAOD,EAAKC,CAAG,KAAM,YAAYD,EAAKC,CAAG,MAAM,OAMnD;AAAA,UAAI,UAAUD,EAAKC,CAAG,GAAG;AACvB,cAAMC,IAAiBC,EAAU,OAAO,CAACC,MAAMC,EAAWD,GAAGJ,EAAKC,CAAG,EAAE,IAAI,CAAC;AAC5E,YAAIC,EAAe,WAAW;AAC5B,gBAAM,IAAI,MAAM,8BAA8BF,EAAKC,CAAG,EAAE,IAAI,EAAE;AAEhE,YAAIC,EAAe,SAAS;AAC1B,gBAAM,IAAI,MAAM,4CAA4CF,EAAKC,CAAG,EAAE,IAAI,EAAE;AAG9E,cAAMK,IAAWJ,EAAe,CAAC;AAEjC,YAAIK;AACJ,gBAAQD,EAAS,eAAA;AAAA,UACf,KAAK;AACH,YAAAC,IAAOC;AACP;AAAA,UACF,KAAK;AACH,YAAAD,IAAOE;AACP;AAAA,UACF;AACE,kBAAM,IAAI,MAAM,0BAA0BH,EAAS,aAAa,EAAE;AAAA,QAAA;AAGtE,cAAMI,IAAW,OAAOH,EAAKD,GAAUN,EAAKC,CAAG,EAAE,MAAMJ,CAAc;AAErE,YAAI,OAAO,SAASG,EAAKC,CAAG,CAAC,EAAG,OAAM,IAAI,MAAM,2CAA2C;AAE3F,QAAAD,EAAKC,CAAG,IAAIU,EAA2C,EAAE,MAAMD,EAAA,GAAYV,EAAKC,CAAG,CAAC;AAAA,MACtF;AAGA,UAAI,MAAM,QAAQD,EAAKC,CAAG,CAAC;AACzB,mBAAW,CAACW,CAAK,KAAKZ,EAAKC,CAAG,EAAE;AAC9B,iBAAOF,EAAKC,EAAKC,CAAG,GAAGW,CAAK;AAAA,eAErB,OAAOZ,EAAKC,CAAG,KAAM,YAAYD,EAAKC,CAAG,MAAM;AACxD,mBAAWY,KAAQ,OAAO,KAAKb,EAAKC,CAAG,CAAC;AAEtC,iBAAOF,EAAKC,EAAKC,CAAG,GAAGY,CAAI;AAAA;AAAA,EAGjC;AACF;AAYA,gBAAgBL,EACdF,GACAQ,GACAC,GACyC;AACzC,QAAMC,IAAUC,EAAU,MAAMC,EAAgBZ,GAAUQ,GAAKC,CAAO,CAAC;AAEvE,SAAOtB,EAAoBuB,CAAO;AAClC,QAAMxC,IAAM2C,EAAcL,CAAG,GACvBM,IAAiB7C,EAAsBC,CAAG;AAEhD,eAAM,EAAE,KAAAA,GAAK,gBAAA4C,EAAA,GAEN,GADQN,EAAI,UAAU,GAAGA,EAAI,SAAStC,EAAI,SAAS,CAAC,CAC3C,IAAI4C,CAAc;AACpC;AAEA,gBAAgBX,EACdH,GACAQ,GACAC,GACyC;AACzC,QAAMC,IAAUC,EAAU,MAAMC,EAAgBZ,GAAUQ,GAAKC,CAAO,CAAC,GACjEK,IAAiBhD,EAAmC0C,CAAG;AAE7D,eAAM,EAAE,MADK,OAAOrB,EAAoBuB,CAAO,GACjC,gBAAAI,EAAA,GACPA;AACT;"}
1
+ {"version":3,"file":"makeOffline-DuRXI_Dd.js","sources":["../../src/offline/hashString.ts","../../src/offline/slugify.ts","../../src/offline/tools.ts","../../src/offline/makeOffline.ts"],"sourcesContent":["// 32-bit murmur-ish hash with configurable seed\nfunction murmur32WithSeed(str: string, seed: number): string {\n let h = seed >>> 0;\n\n for (let i = 0; i < str.length; i++) {\n h ^= str.charCodeAt(i);\n h = Math.imul(h, 0x01000193); // FNV-ish prime, good mixer\n }\n\n return (h >>> 0).toString(16).padStart(8, '0'); // 8 hex chars\n}\n\n// 128-bit (MD5-width) hash: 4 × 32-bit parts concatenated\nexport function hashString(str: string): string {\n return (\n murmur32WithSeed(str, 0x811c9dc5) + // your original seed\n murmur32WithSeed(str, 0x21f0aaad) +\n murmur32WithSeed(str, 0x1b873593) +\n murmur32WithSeed(str, 0x85ebca6b)\n );\n}\n","// Module-level caches\nconst inputToSlug = new Map<string, string>();\nconst slugToInput = new Map<string, string>();\n\nconst WINDOWS_RESERVED = new Set([\n \"con\", \"prn\", \"aux\", \"nul\",\n \"com1\", \"com2\", \"com3\", \"com4\", \"com5\", \"com6\", \"com7\", \"com8\", \"com9\",\n \"lpt1\", \"lpt2\", \"lpt3\", \"lpt4\", \"lpt5\", \"lpt6\", \"lpt7\", \"lpt8\", \"lpt9\",\n]);\n\nfunction makeBaseSlug(input: string): string {\n let slug = input\n .normalize(\"NFKD\")\n // All non letters/numbers → \"-\"\n .replace(/[^\\p{Letter}\\p{Number}]+/gu, \"-\")\n // Collapse multiple \"-\"\n .replace(/-+/g, \"-\")\n // Trim \"-\" from start/end\n .replace(/^-|-$/g, \"\")\n .toLowerCase();\n\n // Strip forbidden Windows characters just in case\n slug = slug.replace(/[<>:\"/\\\\|?*]/g, \"\");\n\n // Windows forbids trailing space/period\n slug = slug.replace(/[. ]+$/g, \"\");\n\n // Fallback if everything was stripped\n if (!slug) slug = \"file\";\n\n // Avoid bare reserved device names\n if (WINDOWS_RESERVED.has(slug)) {\n slug += \"-file\";\n }\n\n return slug;\n}\n\nexport function slugifyFsUnique(input: string): string {\n // If we've seen this exact input before, return the same slug\n const existing = inputToSlug.get(input);\n if (existing) return existing;\n\n const base = makeBaseSlug(input);\n let candidate = base;\n let counter = 2;\n\n while (true) {\n const existingInput = slugToInput.get(candidate);\n\n if (!existingInput) {\n // Free slug → claim it for this input\n slugToInput.set(candidate, input);\n inputToSlug.set(input, candidate);\n return candidate;\n }\n\n if (existingInput === input) {\n // Same input somehow (super defensive)\n inputToSlug.set(input, candidate);\n return candidate;\n }\n\n // Collision: same slug already used by different input → add suffix\n candidate = `${base}-${counter++}`;\n }\n}\n\n// // Optional: to reset between runs/tests\n// export function resetSlugCache() {\n// inputToSlug.clear();\n// slugToInput.clear();\n// }","import { hashString } from './hashString';\nimport { slugifyFsUnique } from './slugify';\n\nexport function makeUniqueJsonTargetPathFromString(str: string, namespace: string = ''): string {\n // const hash = hashString(str);\n let result = slugifyFsUnique(str); // + '-' + hash;\n if (namespace) {\n result = `${slugifyFsUnique(namespace)}$${result}`;\n }\n\n if (result.endsWith('/')) {\n result += 'index.json';\n } else if (!result.endsWith('.json')) {\n result += '.json';\n }\n return './' + result;\n // handle directory case\n}\n\nexport function makeTargetPathFromUrl(url: string, prefix: string = ''): string {\n // https://example.com/dir1/dir2/a.js => \"{prefix}{origin-slug}/dir1/dir2/a.js\";\n // https://example.com/dir1/dir2/a.js?params => \"{prefix}{origin-slug}/dir1/dir2/a{paramsmd5hash}.js\";\n // use slugify.ts\n\n try {\n new URL(url);\n } catch {\n debugger;\n }\n const urlObj = new URL(url);\n const origin = `${urlObj.protocol}//${urlObj.host}`;\n const originSlug = slugifyFsUnique(origin);\n\n let pathname = urlObj.pathname;\n let search = urlObj.search;\n\n // if path doesn't end with extension, throw\n if (!pathname.match(/\\.[^\\/]+$/)) {\n throw new Error(`Cannot make target path from URL without file extension: ${url}`);\n }\n\n const extension = pathname.substring(pathname.lastIndexOf('.'));\n let pathnameWithoutExtension = pathname.substring(0, pathname.lastIndexOf('.'));\n\n // check the pathname contains only valid fs characters\n const invalidPathnameChars = pathnameWithoutExtension.match(/[^a-zA-Z0-9\\-._\\/]/g);\n if (invalidPathnameChars) {\n const fixedPathnameWithoutExtension = slugifyFsUnique(pathnameWithoutExtension);\n console.warn(\n `Warning: pathname contains invalid filesystem characters (${[\n ...new Set(invalidPathnameChars),\n ].join(\n ', '\n )}), slugifying it: ${pathnameWithoutExtension}${extension} => ${fixedPathnameWithoutExtension}${extension}`\n );\n pathnameWithoutExtension = fixedPathnameWithoutExtension;\n }\n\n pathname = pathnameWithoutExtension + extension;\n\n if (pathname.length > 120) {\n console.warn(\n `Warning: pathname is too long (${pathname.length} characters), truncating to 150 characters: ${pathname}`\n );\n pathname = pathname.substring(0, 120 - extension.length) + extension;\n }\n\n if (search) {\n // create a hash from search params\n const hash = hashString(search);\n const dotIndex = pathname.lastIndexOf('.');\n if (dotIndex !== -1) {\n pathname = `${pathname.slice(0, dotIndex)}.${hash}${pathname.slice(dotIndex)}`;\n } else {\n pathname = `${pathname}${hash}`;\n }\n }\n\n // // handle directory case\n // if (pathname.endsWith('/')) {\n // pathname += '__index.json';\n // }\n\n return `./${prefix}${originSlug}${pathname}`;\n}\n","import { log } from '../logger';\nimport {\n canResolve,\n parseRefValue,\n resolverResolve,\n type ResolveContextInternal,\n} from '../resolve';\nimport { resolvers } from '../resolvers';\nimport { createMergedObjectWithOverridenNonRefProps, deepClone } from '../shared';\nimport type { LocalFile, MakeOfflineResult, ManifestData, Resolver } from '../types';\nimport { makeTargetPathFromUrl, makeUniqueJsonTargetPathFromString } from './tools';\n\nexport async function makeOffline(manifest: any): Promise<MakeOfflineResult> {\n log('makeOffline', manifest);\n // if (typeof manifest !== 'object' || manifest === null) {\n // throw new Error('Manifest must be an object');\n // }\n\n const res = await makeOfflineInternal(manifest);\n let done = false;\n let resultManifest: any;\n async function* files() {\n resultManifest = yield* res;\n done = true;\n }\n\n return {\n get manifest(): ManifestData {\n if (!done) throw new Error('Iterate over files before getting manifest');\n return resultManifest;\n },\n files: files() as AsyncGenerator<LocalFile, void, void>,\n };\n}\n\nasync function* makeOfflineInternal(manifest: any): AsyncGenerator<LocalFile, ManifestData, void> {\n log('makeOfflineInternal', manifest);\n const resolveContext: ResolveContextInternal = {\n refCache: new Map<string, Promise<any>>(),\n forceFetch: true,\n signal: null,\n };\n const parent = { manifest };\n yield* walk(parent, 'manifest');\n return parent.manifest;\n\n async function* walk(node: any, key: string | number): AsyncGenerator<LocalFile> {\n // console.log('walk', node, key, node[key]);\n // let node[key] = node[key];\n if (typeof node[key] !== 'object' || node[key] === null) {\n return;\n }\n // let resolversToUse1 = resolvers.filter((r) => r.canResolve(node[key].$ref));\n // const usedResolvers = new Set<Resolver>();\n // do {\n if ('$ref' in node[key]) {\n const resolversToUse = resolvers.filter((r) => canResolve(r, node[key].$ref));\n if (resolversToUse.length === 0) {\n throw new Error(`No resolver found for ref: ${node[key].$ref}`);\n }\n if (resolversToUse.length > 1) {\n throw new Error(`Multiple resolvers can make offline ref: ${node[key].$ref}`);\n }\n\n const resolver = resolversToUse[0];\n\n let func: typeof offlineLocalizeRef;\n switch (resolver.offlineMethod) {\n case 'localizeRef':\n func = offlineLocalizeRef;\n break;\n case 'resolveRef':\n func = offlineResolveRef;\n break;\n default:\n throw new Error(`Unknown offlineMethod: ${resolver.offlineMethod}`);\n }\n\n const mergeRef = yield* func(resolver, node[key].$ref, resolveContext);\n\n if (Object.isFrozen(node[key])) throw new Error('Unexpected frozen node during makeOffline');\n\n node[key] = createMergedObjectWithOverridenNonRefProps({ $ref: mergeRef }, node[key]);\n }\n\n // recurse\n if (Array.isArray(node[key])) {\n for (const [index] of node[key].entries()) {\n yield* walk(node[key], index);\n }\n } else if (typeof node[key] === 'object' && node[key] !== null) {\n for (const key1 of Object.keys(node[key])) {\n // debugger;\n yield* walk(node[key], key1);\n }\n }\n }\n}\n\n// function canResolve(resolver: Resolver, ref: string): boolean {\n// if (resolver.canResolve) {\n// return resolver.canResolve(ref);\n// }\n// if (resolver.schema) {\n// return canResolveRefSchema(ref, resolver.schema);\n// }\n// throw new Error('Resolver is missing canResolve method and schema property');\n// }\n\nasync function* offlineLocalizeRef(\n resolver: Resolver,\n ref: string,\n context: ResolveContextInternal\n): AsyncGenerator<LocalFile, string, void> {\n const refData = deepClone(await resolverResolve(resolver, ref, context));\n // emit assets\n yield* makeOfflineInternal(refData);\n const url = parseRefValue(ref);\n const targetFilePath = makeTargetPathFromUrl(url);\n // TODO: handle CSS and other text files that may reference other assets\n yield { url, targetFilePath };\n const schema = ref.substring(0, ref.length - url.length - 1);\n return `${schema}:${targetFilePath}`;\n}\n\nasync function* offlineResolveRef(\n resolver: Resolver,\n ref: string,\n context: ResolveContextInternal\n): AsyncGenerator<LocalFile, string, void> {\n const refData = deepClone(await resolverResolve(resolver, ref, context));\n const targetFilePath = makeUniqueJsonTargetPathFromString(ref);\n const data = yield* makeOfflineInternal(refData);\n yield { data, targetFilePath };\n return targetFilePath;\n}\n\n// export function createMakeOfflineResult(\n// manifest: any,\n// walk: (node: any, key: string | number) => AsyncGenerator<LocalFile>\n// ): MakeOfflineResult {\n// log('createMakeOfflineResult', manifest);\n// if (typeof manifest !== 'object' || manifest === null) {\n// throw new Error('Manifest must be an object');\n// }\n// const parent = { manifest };\n// let done = false;\n// async function* files() {\n// yield* walk(parent, 'manifest');\n// done = true;\n// }\n\n// return {\n// get manifest() {\n// if (!done) {\n// throw new Error('Cannot access manifest before all files are generated');\n// }\n// return parent.manifest;\n// },\n// files: files(),\n// };\n// }\n"],"names":["murmur32WithSeed","str","seed","h","i","hashString","inputToSlug","slugToInput","WINDOWS_RESERVED","makeBaseSlug","input","slug","slugifyFsUnique","existing","base","candidate","counter","existingInput","makeUniqueJsonTargetPathFromString","namespace","result","makeTargetPathFromUrl","url","prefix","urlObj","origin","originSlug","pathname","search","extension","pathnameWithoutExtension","invalidPathnameChars","fixedPathnameWithoutExtension","hash","dotIndex","makeOffline","manifest","log","res","makeOfflineInternal","done","resultManifest","files","resolveContext","parent","walk","node","key","resolversToUse","resolvers","r","canResolve","resolver","func","offlineLocalizeRef","offlineResolveRef","mergeRef","createMergedObjectWithOverridenNonRefProps","index","key1","ref","context","refData","deepClone","resolverResolve","parseRefValue","targetFilePath"],"mappings":";AACA,SAASA,EAAiBC,GAAaC,GAAsB;AAC3D,MAAIC,IAAID,MAAS;AAEjB,WAASE,IAAI,GAAGA,IAAIH,EAAI,QAAQG;AAC9B,IAAAD,KAAKF,EAAI,WAAWG,CAAC,GACrBD,IAAI,KAAK,KAAKA,GAAG,QAAU;AAG7B,UAAQA,MAAM,GAAG,SAAS,EAAE,EAAE,SAAS,GAAG,GAAG;AAC/C;AAGO,SAASE,EAAWJ,GAAqB;AAC9C,SACED,EAAiBC,GAAK,UAAU;AAAA,EAChCD,EAAiBC,GAAK,SAAU,IAChCD,EAAiBC,GAAK,SAAU,IAChCD,EAAiBC,GAAK,UAAU;AAEpC;ACnBA,MAAMK,wBAAkB,IAAA,GAClBC,wBAAkB,IAAA,GAElBC,wBAAuB,IAAI;AAAA,EAC/B;AAAA,EAAO;AAAA,EAAO;AAAA,EAAO;AAAA,EACrB;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAChE;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAAA,EAAQ;AAClE,CAAC;AAED,SAASC,EAAaC,GAAuB;AAC3C,MAAIC,IAAOD,EACR,UAAU,MAAM,EAEhB,QAAQ,8BAA8B,GAAG,EAEzC,QAAQ,OAAO,GAAG,EAElB,QAAQ,UAAU,EAAE,EACpB,YAAA;AAGH,SAAAC,IAAOA,EAAK,QAAQ,iBAAiB,EAAE,GAGvCA,IAAOA,EAAK,QAAQ,WAAW,EAAE,GAG5BA,MAAMA,IAAO,SAGdH,EAAiB,IAAIG,CAAI,MAC3BA,KAAQ,UAGHA;AACT;AAEO,SAASC,EAAgBF,GAAuB;AAErD,QAAMG,IAAWP,EAAY,IAAII,CAAK;AACtC,MAAIG,EAAU,QAAOA;AAErB,QAAMC,IAAOL,EAAaC,CAAK;AAC/B,MAAIK,IAAYD,GACZE,IAAU;AAEd,aAAa;AACX,UAAMC,IAAgBV,EAAY,IAAIQ,CAAS;AAE/C,QAAI,CAACE;AAEH,aAAAV,EAAY,IAAIQ,GAAWL,CAAK,GAChCJ,EAAY,IAAII,GAAOK,CAAS,GACzBA;AAGT,QAAIE,MAAkBP;AAEpB,aAAAJ,EAAY,IAAII,GAAOK,CAAS,GACzBA;AAIT,IAAAA,IAAY,GAAGD,CAAI,IAAIE,GAAS;AAAA,EAClC;AACF;AC/DO,SAASE,EAAmCjB,GAAakB,IAAoB,IAAY;AAE9F,MAAIC,IAASR,EAAgBX,CAAG;AAChC,SAAIkB,MACFC,IAAS,GAAGR,EAAgBO,CAAS,CAAC,IAAIC,CAAM,KAG9CA,EAAO,SAAS,GAAG,IACrBA,KAAU,eACAA,EAAO,SAAS,OAAO,MACjCA,KAAU,UAEL,OAAOA;AAEhB;AAEO,SAASC,EAAsBC,GAAaC,IAAiB,IAAY;AAK9E,MAAI;AACF,QAAI,IAAID,CAAG;AAAA,EACb,QAAQ;AACN;AAAA,EACF;AACA,QAAME,IAAS,IAAI,IAAIF,CAAG,GACpBG,IAAS,GAAGD,EAAO,QAAQ,KAAKA,EAAO,IAAI,IAC3CE,IAAad,EAAgBa,CAAM;AAEzC,MAAIE,IAAWH,EAAO,UAClBI,IAASJ,EAAO;AAGpB,MAAI,CAACG,EAAS,MAAM,WAAW;AAC7B,UAAM,IAAI,MAAM,4DAA4DL,CAAG,EAAE;AAGnF,QAAMO,IAAYF,EAAS,UAAUA,EAAS,YAAY,GAAG,CAAC;AAC9D,MAAIG,IAA2BH,EAAS,UAAU,GAAGA,EAAS,YAAY,GAAG,CAAC;AAG9E,QAAMI,IAAuBD,EAAyB,MAAM,qBAAqB;AACjF,MAAIC,GAAsB;AACxB,UAAMC,IAAgCpB,EAAgBkB,CAAwB;AAC9E,YAAQ;AAAA,MACN,6DAA6D;AAAA,QAC3D,GAAG,IAAI,IAAIC,CAAoB;AAAA,MAAA,EAC/B;AAAA,QACA;AAAA,MAAA,CACD,qBAAqBD,CAAwB,GAAGD,CAAS,OAAOG,CAA6B,GAAGH,CAAS;AAAA,IAAA,GAE5GC,IAA2BE;AAAA,EAC7B;AAWA,MATAL,IAAWG,IAA2BD,GAElCF,EAAS,SAAS,QACpB,QAAQ;AAAA,IACN,kCAAkCA,EAAS,MAAM,+CAA+CA,CAAQ;AAAA,EAAA,GAE1GA,IAAWA,EAAS,UAAU,GAAG,MAAME,EAAU,MAAM,IAAIA,IAGzDD,GAAQ;AAEV,UAAMK,IAAO5B,EAAWuB,CAAM,GACxBM,IAAWP,EAAS,YAAY,GAAG;AACzC,IAAIO,MAAa,KACfP,IAAW,GAAGA,EAAS,MAAM,GAAGO,CAAQ,CAAC,IAAID,CAAI,GAAGN,EAAS,MAAMO,CAAQ,CAAC,KAE5EP,IAAW,GAAGA,CAAQ,GAAGM,CAAI;AAAA,EAEjC;AAOA,SAAO,KAAKV,CAAM,GAAGG,CAAU,GAAGC,CAAQ;AAC5C;ACxEA,eAAsBQ,EAAYC,GAA2C;AAC3E,EAAAC,EAAI,eAAeD,CAAQ;AAK3B,QAAME,IAAM,MAAMC,EAAoBH,CAAQ;AAC9C,MAAII,IAAO,IACPC;AACJ,kBAAgBC,IAAQ;AACtB,IAAAD,IAAiB,OAAOH,GACxBE,IAAO;AAAA,EACT;AAEA,SAAO;AAAA,IACL,IAAI,WAAyB;AAC3B,UAAI,CAACA,EAAM,OAAM,IAAI,MAAM,4CAA4C;AACvE,aAAOC;AAAA,IACT;AAAA,IACA,OAAOC,EAAA;AAAA,EAAM;AAEjB;AAEA,gBAAgBH,EAAoBH,GAA8D;AAChG,EAAAC,EAAI,uBAAuBD,CAAQ;AACnC,QAAMO,IAAyC;AAAA,IAC7C,8BAAc,IAAA;AAAA,IACd,YAAY;AAAA,IACZ,QAAQ;AAAA,EAAA,GAEJC,IAAS,EAAE,UAAAR,EAAA;AACjB,gBAAOS,EAAKD,GAAQ,UAAU,GACvBA,EAAO;AAEd,kBAAgBC,EAAKC,GAAWC,GAAiD;AAG/E,QAAI,SAAOD,EAAKC,CAAG,KAAM,YAAYD,EAAKC,CAAG,MAAM,OAMnD;AAAA,UAAI,UAAUD,EAAKC,CAAG,GAAG;AACvB,cAAMC,IAAiBC,EAAU,OAAO,CAACC,MAAMC,EAAWD,GAAGJ,EAAKC,CAAG,EAAE,IAAI,CAAC;AAC5E,YAAIC,EAAe,WAAW;AAC5B,gBAAM,IAAI,MAAM,8BAA8BF,EAAKC,CAAG,EAAE,IAAI,EAAE;AAEhE,YAAIC,EAAe,SAAS;AAC1B,gBAAM,IAAI,MAAM,4CAA4CF,EAAKC,CAAG,EAAE,IAAI,EAAE;AAG9E,cAAMK,IAAWJ,EAAe,CAAC;AAEjC,YAAIK;AACJ,gBAAQD,EAAS,eAAA;AAAA,UACf,KAAK;AACH,YAAAC,IAAOC;AACP;AAAA,UACF,KAAK;AACH,YAAAD,IAAOE;AACP;AAAA,UACF;AACE,kBAAM,IAAI,MAAM,0BAA0BH,EAAS,aAAa,EAAE;AAAA,QAAA;AAGtE,cAAMI,IAAW,OAAOH,EAAKD,GAAUN,EAAKC,CAAG,EAAE,MAAMJ,CAAc;AAErE,YAAI,OAAO,SAASG,EAAKC,CAAG,CAAC,EAAG,OAAM,IAAI,MAAM,2CAA2C;AAE3F,QAAAD,EAAKC,CAAG,IAAIU,EAA2C,EAAE,MAAMD,EAAA,GAAYV,EAAKC,CAAG,CAAC;AAAA,MACtF;AAGA,UAAI,MAAM,QAAQD,EAAKC,CAAG,CAAC;AACzB,mBAAW,CAACW,CAAK,KAAKZ,EAAKC,CAAG,EAAE;AAC9B,iBAAOF,EAAKC,EAAKC,CAAG,GAAGW,CAAK;AAAA,eAErB,OAAOZ,EAAKC,CAAG,KAAM,YAAYD,EAAKC,CAAG,MAAM;AACxD,mBAAWY,KAAQ,OAAO,KAAKb,EAAKC,CAAG,CAAC;AAEtC,iBAAOF,EAAKC,EAAKC,CAAG,GAAGY,CAAI;AAAA;AAAA,EAGjC;AACF;AAYA,gBAAgBL,EACdF,GACAQ,GACAC,GACyC;AACzC,QAAMC,IAAUC,EAAU,MAAMC,EAAgBZ,GAAUQ,GAAKC,CAAO,CAAC;AAEvE,SAAOtB,EAAoBuB,CAAO;AAClC,QAAMxC,IAAM2C,EAAcL,CAAG,GACvBM,IAAiB7C,EAAsBC,CAAG;AAEhD,eAAM,EAAE,KAAAA,GAAK,gBAAA4C,EAAA,GAEN,GADQN,EAAI,UAAU,GAAGA,EAAI,SAAStC,EAAI,SAAS,CAAC,CAC3C,IAAI4C,CAAc;AACpC;AAEA,gBAAgBX,EACdH,GACAQ,GACAC,GACyC;AACzC,QAAMC,IAAUC,EAAU,MAAMC,EAAgBZ,GAAUQ,GAAKC,CAAO,CAAC,GACjEK,IAAiBhD,EAAmC0C,CAAG;AAE7D,eAAM,EAAE,MADK,OAAOrB,EAAoBuB,CAAO,GACjC,gBAAAI,EAAA,GACPA;AACT;"}
@@ -1 +1 @@
1
- {"version":3,"file":"makeOfflineBundle-XjvmvDQE.js","sources":["../../src/offline/makeOfflineBundle.ts"],"sourcesContent":["import { log } from '../shared';\nimport { makeOffline } from './makeOffline';\nimport { resolve } from '..';\n\nexport async function* makeOfflineBundle(\n manifest: unknown\n): AsyncGenerator<{ path: string; data: ArrayBuffer }> {\n log('makeOfflineBundle', manifest);\n const offlineData = await makeOffline(manifest);\n for await (const file of offlineData.files) {\n if ('data' in file) {\n const jsonString = JSON.stringify(file.data, null, 2);\n yield {\n path: file.targetFilePath,\n data: new TextEncoder().encode(jsonString).buffer,\n };\n } else if ('url' in file) {\n const blob = await downloadFile(file.url);\n yield {\n path: file.targetFilePath,\n data: await blob.arrayBuffer(),\n };\n }\n }\n // yield* generateJsLoaderFiles();\n const html = await getIndexHtml(offlineData.manifest);\n yield {\n path: 'index.html',\n data: new TextEncoder().encode(html).buffer,\n };\n}\n\nconst MAX_CONCURRENT_DOWNLOADS = 10;\nconst queue = new Set<Promise<Blob>>();\n// have a queue to limit concurrent downloads\nasync function downloadFile(url: string): Promise<Blob> {\n while (queue.size >= MAX_CONCURRENT_DOWNLOADS) {\n await Promise.race(queue);\n }\n const downloadPromise = (async () => {\n log('Fetching file for zip:', url);\n const response = await fetch(url);\n if (!response.ok) {\n throw new Error(`Failed to fetch ${url}: ${response.status} ${response.statusText}`);\n }\n return await response.blob();\n })();\n queue.add(downloadPromise);\n try {\n return await downloadPromise;\n } finally {\n queue.delete(downloadPromise);\n }\n}\n\nasync function getIndexHtml(manifest: unknown) {\n const entryPoint = await resolve(manifest, '/runtime/entry');\n const html = `\n<!DOCTYPE html>\n<script type=\"module\">\n import { load } from ${JSON.stringify(entryPoint)};\n await load(${JSON.stringify(manifest)});\n console.info('🚀 loaded');\n</script> \n`;\n return html;\n}\n"],"names":["makeOfflineBundle","manifest","log","offlineData","makeOffline","file","jsonString","blob","downloadFile","html","getIndexHtml","MAX_CONCURRENT_DOWNLOADS","queue","url","downloadPromise","response","entryPoint","resolve"],"mappings":";;AAIA,gBAAuBA,EACrBC,GACqD;AACrD,EAAAC,EAAI,qBAAqBD,CAAQ;AACjC,QAAME,IAAc,MAAMC,EAAYH,CAAQ;AAC9C,mBAAiBI,KAAQF,EAAY;AACnC,QAAI,UAAUE,GAAM;AAClB,YAAMC,IAAa,KAAK,UAAUD,EAAK,MAAM,MAAM,CAAC;AACpD,YAAM;AAAA,QACJ,MAAMA,EAAK;AAAA,QACX,MAAM,IAAI,YAAA,EAAc,OAAOC,CAAU,EAAE;AAAA,MAAA;AAAA,IAE/C,WAAW,SAASD,GAAM;AACxB,YAAME,IAAO,MAAMC,EAAaH,EAAK,GAAG;AACxC,YAAM;AAAA,QACJ,MAAMA,EAAK;AAAA,QACX,MAAM,MAAME,EAAK,YAAA;AAAA,MAAY;AAAA,IAEjC;AAGF,QAAME,IAAO,MAAMC,EAAaP,EAAY,QAAQ;AACpD,QAAM;AAAA,IACJ,MAAM;AAAA,IACN,MAAM,IAAI,YAAA,EAAc,OAAOM,CAAI,EAAE;AAAA,EAAA;AAEzC;AAEA,MAAME,IAA2B,IAC3BC,wBAAY,IAAA;AAElB,eAAeJ,EAAaK,GAA4B;AACtD,SAAOD,EAAM,QAAQD;AACnB,UAAM,QAAQ,KAAKC,CAAK;AAE1B,QAAME,KAAmB,YAAY;AACnC,IAAAZ,EAAI,0BAA0BW,CAAG;AACjC,UAAME,IAAW,MAAM,MAAMF,CAAG;AAChC,QAAI,CAACE,EAAS;AACZ,YAAM,IAAI,MAAM,mBAAmBF,CAAG,KAAKE,EAAS,MAAM,IAAIA,EAAS,UAAU,EAAE;AAErF,WAAO,MAAMA,EAAS,KAAA;AAAA,EACxB,GAAA;AACA,EAAAH,EAAM,IAAIE,CAAe;AACzB,MAAI;AACF,WAAO,MAAMA;AAAA,EACf,UAAA;AACE,IAAAF,EAAM,OAAOE,CAAe;AAAA,EAC9B;AACF;AAEA,eAAeJ,EAAaT,GAAmB;AAC7C,QAAMe,IAAa,MAAMC,EAAQhB,GAAU,gBAAgB;AAS3D,SARa;AAAA;AAAA;AAAA,yBAGU,KAAK,UAAUe,CAAU,CAAC;AAAA,eACpC,KAAK,UAAUf,CAAQ,CAAC;AAAA;AAAA;AAAA;AAKvC;"}
1
+ {"version":3,"file":"makeOfflineBundle-XjvmvDQE.js","sources":["../../src/offline/makeOfflineBundle.ts"],"sourcesContent":["import { log } from '../logger';\nimport { makeOffline } from './makeOffline';\nimport { resolve } from '..';\n\nexport async function* makeOfflineBundle(\n manifest: unknown\n): AsyncGenerator<{ path: string; data: ArrayBuffer }> {\n log('makeOfflineBundle', manifest);\n const offlineData = await makeOffline(manifest);\n for await (const file of offlineData.files) {\n if ('data' in file) {\n const jsonString = JSON.stringify(file.data, null, 2);\n yield {\n path: file.targetFilePath,\n data: new TextEncoder().encode(jsonString).buffer,\n };\n } else if ('url' in file) {\n const blob = await downloadFile(file.url);\n yield {\n path: file.targetFilePath,\n data: await blob.arrayBuffer(),\n };\n }\n }\n // yield* generateJsLoaderFiles();\n const html = await getIndexHtml(offlineData.manifest);\n yield {\n path: 'index.html',\n data: new TextEncoder().encode(html).buffer,\n };\n}\n\nconst MAX_CONCURRENT_DOWNLOADS = 10;\nconst queue = new Set<Promise<Blob>>();\n// have a queue to limit concurrent downloads\nasync function downloadFile(url: string): Promise<Blob> {\n while (queue.size >= MAX_CONCURRENT_DOWNLOADS) {\n await Promise.race(queue);\n }\n const downloadPromise = (async () => {\n log('Fetching file for zip:', url);\n const response = await fetch(url);\n if (!response.ok) {\n throw new Error(`Failed to fetch ${url}: ${response.status} ${response.statusText}`);\n }\n return await response.blob();\n })();\n queue.add(downloadPromise);\n try {\n return await downloadPromise;\n } finally {\n queue.delete(downloadPromise);\n }\n}\n\nasync function getIndexHtml(manifest: unknown) {\n const entryPoint = await resolve(manifest, '/runtime/entry');\n const html = `\n<!DOCTYPE html>\n<script type=\"module\">\n import { load } from ${JSON.stringify(entryPoint)};\n await load(${JSON.stringify(manifest)});\n console.info('🚀 loaded');\n</script> \n`;\n return html;\n}\n"],"names":["makeOfflineBundle","manifest","log","offlineData","makeOffline","file","jsonString","blob","downloadFile","html","getIndexHtml","MAX_CONCURRENT_DOWNLOADS","queue","url","downloadPromise","response","entryPoint","resolve"],"mappings":";;AAIA,gBAAuBA,EACrBC,GACqD;AACrD,EAAAC,EAAI,qBAAqBD,CAAQ;AACjC,QAAME,IAAc,MAAMC,EAAYH,CAAQ;AAC9C,mBAAiBI,KAAQF,EAAY;AACnC,QAAI,UAAUE,GAAM;AAClB,YAAMC,IAAa,KAAK,UAAUD,EAAK,MAAM,MAAM,CAAC;AACpD,YAAM;AAAA,QACJ,MAAMA,EAAK;AAAA,QACX,MAAM,IAAI,YAAA,EAAc,OAAOC,CAAU,EAAE;AAAA,MAAA;AAAA,IAE/C,WAAW,SAASD,GAAM;AACxB,YAAME,IAAO,MAAMC,EAAaH,EAAK,GAAG;AACxC,YAAM;AAAA,QACJ,MAAMA,EAAK;AAAA,QACX,MAAM,MAAME,EAAK,YAAA;AAAA,MAAY;AAAA,IAEjC;AAGF,QAAME,IAAO,MAAMC,EAAaP,EAAY,QAAQ;AACpD,QAAM;AAAA,IACJ,MAAM;AAAA,IACN,MAAM,IAAI,YAAA,EAAc,OAAOM,CAAI,EAAE;AAAA,EAAA;AAEzC;AAEA,MAAME,IAA2B,IAC3BC,wBAAY,IAAA;AAElB,eAAeJ,EAAaK,GAA4B;AACtD,SAAOD,EAAM,QAAQD;AACnB,UAAM,QAAQ,KAAKC,CAAK;AAE1B,QAAME,KAAmB,YAAY;AACnC,IAAAZ,EAAI,0BAA0BW,CAAG;AACjC,UAAME,IAAW,MAAM,MAAMF,CAAG;AAChC,QAAI,CAACE,EAAS;AACZ,YAAM,IAAI,MAAM,mBAAmBF,CAAG,KAAKE,EAAS,MAAM,IAAIA,EAAS,UAAU,EAAE;AAErF,WAAO,MAAMA,EAAS,KAAA;AAAA,EACxB,GAAA;AACA,EAAAH,EAAM,IAAIE,CAAe;AACzB,MAAI;AACF,WAAO,MAAMA;AAAA,EACf,UAAA;AACE,IAAAF,EAAM,OAAOE,CAAe;AAAA,EAC9B;AACF;AAEA,eAAeJ,EAAaT,GAAmB;AAC7C,QAAMe,IAAa,MAAMC,EAAQhB,GAAU,gBAAgB;AAS3D,SARa;AAAA;AAAA;AAAA,yBAGU,KAAK,UAAUe,CAAU,CAAC;AAAA,eACpC,KAAK,UAAUf,CAAQ,CAAC;AAAA;AAAA;AAAA;AAKvC;"}
@@ -1,4 +1,5 @@
1
- import { deepFreeze, log } from './shared';
1
+ import { log } from './logger';
2
+ import { deepFreeze } from './shared';
2
3
  let importJsonNotAvailable;
3
4
  const jsonFrozen = new WeakSet();
4
5
  // ET: this is a workaround for Vite that analyzes dynamic imports and removes 'with' option
package/dist/esm/index.js CHANGED
@@ -1,11 +1,11 @@
1
+ import { log } from './logger';
1
2
  import { createFunction } from './shared';
2
- import { logger } from './logger';
3
3
  const info = globalThis.__efpLoaderBuildInfo;
4
4
  if (info) {
5
- logger('Loader initialized', { version: info[0], builtAt: info[1] });
5
+ log('Initialized', { version: info[0], builtAt: info[1] });
6
6
  }
7
7
  else {
8
- logger('Loader initialized');
8
+ log('Initialized');
9
9
  }
10
10
  export const mount = createFunction('mount');
11
11
  export const load = createFunction('load');
@@ -1,4 +1,4 @@
1
- //import { log } from './shared';
1
+ //import { log } from './logger';
2
2
  /**
3
3
  * Load a script from `scriptUrl` and optionally wait until `globalVar` appears.
4
4
  * Works in both browser (via <script>) and Node 18+ (via fetch + eval).
@@ -1,2 +1 @@
1
- import debug from 'debug';
2
- export declare const logger: debug.Debugger;
1
+ export declare function log(...args: any[]): void;
@@ -1,2 +1,5 @@
1
1
  import debug from 'debug';
2
- export const logger = debug('efp:loader');
2
+ const logger = debug('efp:loader');
3
+ export function log(...args) {
4
+ logger(...args);
5
+ }
@@ -1,4 +1,5 @@
1
1
  import JSZip from 'jszip';
2
+ import { log } from '../logger';
2
3
  import { makeOfflineBundle } from './makeOfflineBundle';
3
4
  export async function downloadZip(manifest) {
4
5
  const zip = new JSZip();
@@ -9,6 +10,7 @@ export async function downloadZip(manifest) {
9
10
  // Generate archive
10
11
  const blob = await zip.generateAsync({ type: 'blob' });
11
12
  // console.info('Manifest for HTML:', offlineData.manifest);
13
+ log('Generated offline ZIP bundle, size:', blob.size);
12
14
  // Create link & trigger download
13
15
  const a = document.createElement('a');
14
16
  a.href = URL.createObjectURL(blob);
@@ -1,6 +1,7 @@
1
+ import { log } from '../logger';
1
2
  import { canResolve, parseRefValue, resolverResolve, } from '../resolve';
2
3
  import { resolvers } from '../resolvers';
3
- import { createMergedObjectWithOverridenNonRefProps, deepClone, log } from '../shared';
4
+ import { createMergedObjectWithOverridenNonRefProps, deepClone } from '../shared';
4
5
  import { makeTargetPathFromUrl, makeUniqueJsonTargetPathFromString } from './tools';
5
6
  export async function makeOffline(manifest) {
6
7
  log('makeOffline', manifest);
@@ -1,4 +1,4 @@
1
- import { log } from '../shared';
1
+ import { log } from '../logger';
2
2
  import { makeOffline } from './makeOffline';
3
3
  import { resolve } from '..';
4
4
  export async function* makeOfflineBundle(manifest) {
@@ -1,5 +1,6 @@
1
+ import { log } from './logger';
1
2
  import { resolvers } from './resolvers';
2
- import { createMergedObjectWithOverridenNonRefProps, deepFreeze, log, replaceObjectFields, } from './shared';
3
+ import { createMergedObjectWithOverridenNonRefProps, deepFreeze, replaceObjectFields, } from './shared';
3
4
  const globalRefCache = new Map();
4
5
  /**
5
6
  * Resolves a value from an object using a JSON Pointer RFC 6901.
@@ -1,5 +1,5 @@
1
1
  // import { parseRefValue } from '../resolve';
2
- // import { log } from '../shared';
2
+ // import { log } from '../logger';
3
3
  // import type { Resolver } from '../types';
4
4
  // import { httpResolver } from './httpResolver';
5
5
  export {};
@@ -1,5 +1,5 @@
1
1
  import { parseRefValue } from '../resolve';
2
- import { log } from '../shared';
2
+ import { log } from '../logger';
3
3
  export function resolveAssetRefSync(ref) {
4
4
  log('assetResolver resolveRef:', ref);
5
5
  // if (!assetResolver.canResolve(ref)) throw new Error(`Unexpected ref: ${ref}`);
@@ -1,7 +1,7 @@
1
1
  import { importJson } from '../importJson';
2
2
  import { parseRefValue } from '../resolve';
3
3
  import { returnCachedRef } from '../returnCachedRef';
4
- import { log } from '../shared';
4
+ import { log } from '../logger';
5
5
  export const bundleAssetsResolver = {
6
6
  schema: 'bundle-assets',
7
7
  async resolveRef(ref, context) {
@@ -1,7 +1,7 @@
1
1
  import { importJson } from '../importJson';
2
2
  import { parseRefValue } from '../resolve';
3
3
  import { returnCachedRef } from '../returnCachedRef';
4
- import { log } from '../shared';
4
+ import { log } from '../logger';
5
5
  const BASE_URL = 'https://efp-runtime.expofp.com/';
6
6
  export const expoRuntimeBranchResolver = {
7
7
  schema: 'expo-runtime-branch',
@@ -1,6 +1,6 @@
1
1
  import { importJson } from '../importJson';
2
2
  import { parseRefValue } from '../resolve';
3
- import { log } from '../shared';
3
+ import { log } from '../logger';
4
4
  // const SCHEMA = 'expo-runtime-get-branch';
5
5
  export const expoRuntimeGetBranchResolver = {
6
6
  schema: 'expo-runtime-get-branch',
@@ -1,5 +1,6 @@
1
+ import { log } from '../logger';
1
2
  import { parseRefValue } from '../resolve';
2
- import { deepFreeze, log } from '../shared';
3
+ import { deepFreeze } from '../shared';
3
4
  // const SCHEMA = 'expo-runtime';
4
5
  export const expoRuntimeResolver = {
5
6
  schema: 'expo-runtime',
@@ -1,5 +1,5 @@
1
1
  import { importJson } from '../importJson';
2
- import { log } from '../shared';
2
+ import { log } from '../logger';
3
3
  export const httpResolver = {
4
4
  canResolve(ref) {
5
5
  return ref.startsWith('http://') || ref.startsWith('https://') || ref.startsWith('./');
@@ -1,8 +1,7 @@
1
- import { importJson } from '../importJson';
2
1
  import { loadAndWaitGlobal } from '../loadAndWaitGlobal';
3
2
  import { parseRefValue } from '../resolve';
4
3
  import { returnCachedRef } from '../returnCachedRef';
5
- import { log } from '../shared';
4
+ import { log } from '../logger';
6
5
  export const legacyAssetUrlsResolver = {
7
6
  schema: 'legacy-asset-urls',
8
7
  resolveRef(ref, context) {
@@ -24,12 +23,12 @@ export const legacyAssetUrlsResolver = {
24
23
  console.warn(`Could not load version.js at ${versionUrl}, proceeding without it`);
25
24
  }
26
25
  const version = globalThis.__fpDataVersion || (+new Date()).toString();
27
- const filesUrl = `${dataUrlBase}files.json?v=${version}`;
26
+ // const filesUrl = `${dataUrlBase}files.json?v=${version}`;
28
27
  const dataUrl = `${dataUrlBase}data.js?v=${version}`;
29
- const filesPromise = importJson(filesUrl, context).catch(() => {
30
- console.warn(`Could not load files.json at ${filesUrl}, proceeding without it`);
31
- return [];
32
- });
28
+ // const filesPromise = importJson<string[]>(filesUrl, context).catch(() => {
29
+ // console.warn(`Could not load files.json at ${filesUrl}, proceeding without it`);
30
+ // return [];
31
+ // });
33
32
  // data.js
34
33
  const dataLoadPromise = loadAndWaitGlobal(dataUrl, {
35
34
  globalVar: '__data',
@@ -56,16 +55,16 @@ export const legacyAssetUrlsResolver = {
56
55
  signal,
57
56
  });
58
57
  context.importCallback?.(fpSvgUrl, 'script');
59
- await Promise.all([filesPromise, dataLoadPromise, wfDataLoadPromise, fpSvgLoadPromise]);
58
+ await Promise.all([dataLoadPromise, wfDataLoadPromise, fpSvgLoadPromise]);
60
59
  const __fpLayers = (globalThis.__fpLayers ?? []);
61
60
  const fpLayerFiles = __fpLayers.map((x) => `fp.svg.${x.name}.js`);
62
61
  for (const layerFile of fpLayerFiles) {
63
62
  const layerUrl = `${dataUrlBase}${layerFile}?v=${version}`;
64
63
  context.importCallback?.(layerUrl, 'script');
65
64
  }
66
- const files = await filesPromise;
65
+ // const files = await filesPromise;...files,
67
66
  const data = globalThis.__data || {};
68
- const dataFiles = [...files, 'data.js', 'wf.data.js', 'fp.svg.js', ...fpLayerFiles];
67
+ const dataFiles = ['data.js', 'wf.data.js', 'fp.svg.js', ...fpLayerFiles];
69
68
  const dataFilesAssets = dataFiles.reduce((map, file) => {
70
69
  map[file] = { $ref: `asset+${dataUrlBase}${file}?v=${version}` };
71
70
  return map;
@@ -1,6 +1,6 @@
1
1
  import { parseRefValue } from '../resolve';
2
2
  import { returnCachedRef } from '../returnCachedRef';
3
- import { log } from '../shared';
3
+ import { log } from '../logger';
4
4
  export const legacyDataResolver = {
5
5
  schema: 'legacy-data',
6
6
  resolveRef(ref, context) {
@@ -1,5 +1,4 @@
1
1
  export declare function createFunction(name: string): (manifest: unknown, ...args: IArguments[]) => Promise<any>;
2
- export declare function log(...args: any[]): void;
3
2
  export declare function deepFreeze(obj: any): any;
4
3
  export declare function deepClone<T>(obj: T): T;
5
4
  export declare function preloadJson(url: string): void;
@@ -1,4 +1,3 @@
1
- import { logger } from './logger';
2
1
  import { resolve } from './resolve';
3
2
  // export const NOT_APPLICABLE = Symbol('not-applicable');
4
3
  // export const STOP_RESOLVING = Symbol('STOP_RESOLVING');
@@ -16,11 +15,6 @@ export function createFunction(name) {
16
15
  }
17
16
  };
18
17
  }
19
- export function log(...args) {
20
- // if (!(globalThis as any).__EFP_LOADER_DEBUG__ && !import.meta.env.DEV) return;
21
- // console.debug('efp-loader', ...args);
22
- logger(...args);
23
- }
24
18
  export function deepFreeze(obj) {
25
19
  if (obj === null || typeof obj !== 'object')
26
20
  return obj;
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@expofp/loader",
3
- "version": "1.0.58",
3
+ "version": "1.0.60",
4
4
  "description": "ExpoFP JavaScript loader library",
5
5
  "license": "MIT",
6
6
  "homepage": "https://expofp.com",