@xyo-network/diviner-payload-memory 2.81.7 → 2.82.0-rc.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -58,7 +58,7 @@ var MemoryPayloadDiviner = class extends import_diviner_payload_abstract.Payload
58
58
  const parsedOffset = offset || 0;
59
59
  return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
60
60
  const allPairs = await Promise.all(
61
- all.map(async (payload) => [await import_core.PayloadHasher.hashAsync(payload), payload])
61
+ all.map(async (payload) => [await import_core.PayloadHasher.hashAsync(payload, "hex"), payload])
62
62
  );
63
63
  if (hash) {
64
64
  while (allPairs.length > 0 && allPairs[0][0] !== hash) {
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/index.ts","../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["export * from './MemoryPayloadDiviner'\n","import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,aAAS,wBAAS,UAAU,OAAO,yDAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
1
+ {"version":3,"sources":["../../src/index.ts","../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["export * from './MemoryPayloadDiviner'\n","import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload, 'hex'), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,aAAS,wBAAS,UAAU,OAAO,yDAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,SAAS,KAAK,GAAG,OAAO,CAAC;AAAA,QACjH;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
@@ -32,7 +32,7 @@ var MemoryPayloadDiviner = class extends PayloadDiviner {
32
32
  const parsedOffset = offset || 0;
33
33
  return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
34
34
  const allPairs = await Promise.all(
35
- all.map(async (payload) => [await PayloadHasher.hashAsync(payload), payload])
35
+ all.map(async (payload) => [await PayloadHasher.hashAsync(payload, "hex"), payload])
36
36
  );
37
37
  if (hash) {
38
38
  while (allPairs.length > 0 && allPairs[0][0] !== hash) {
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,SAAS,SAAS,UAAU,OAAO,4BAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
1
+ {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload, 'hex'), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,SAAS,SAAS,UAAU,OAAO,4BAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,SAAS,KAAK,GAAG,OAAO,CAAC;AAAA,QACjH;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
@@ -62,7 +62,7 @@ var MemoryPayloadDiviner = class extends import_diviner_payload_abstract.Payload
62
62
  const parsedOffset = offset || 0;
63
63
  return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
64
64
  const allPairs = await Promise.all(
65
- all.map(async (payload) => [await import_core.PayloadHasher.hashAsync(payload), payload])
65
+ all.map(async (payload) => [await import_core.PayloadHasher.hashAsync(payload, "hex"), payload])
66
66
  );
67
67
  if (hash) {
68
68
  while (allPairs.length > 0 && allPairs[0][0] !== hash) {
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/index.ts","../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["export * from './MemoryPayloadDiviner'\n","import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,aAAS,yBAAS,0CAAU,OAAO,+DAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
1
+ {"version":3,"sources":["../../src/index.ts","../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["export * from './MemoryPayloadDiviner'\n","import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload, 'hex'), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,aAAS,yBAAS,0CAAU,OAAO,+DAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,SAAS,KAAK,GAAG,OAAO,CAAC;AAAA,QACjH;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
@@ -36,7 +36,7 @@ var MemoryPayloadDiviner = class extends PayloadDiviner {
36
36
  const parsedOffset = offset || 0;
37
37
  return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
38
38
  const allPairs = await Promise.all(
39
- all.map(async (payload) => [await PayloadHasher.hashAsync(payload), payload])
39
+ all.map(async (payload) => [await PayloadHasher.hashAsync(payload, "hex"), payload])
40
40
  );
41
41
  if (hash) {
42
42
  while (allPairs.length > 0 && allPairs[0][0] !== hash) {
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,SAAS,UAAS,0CAAU,OAAO,kCAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
1
+ {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload, 'hex'), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,SAAS,UAAS,0CAAU,OAAO,kCAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,SAAS,KAAK,GAAG,OAAO,CAAC;AAAA,QACjH;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
package/package.json CHANGED
@@ -10,19 +10,19 @@
10
10
  "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js/issues"
11
11
  },
12
12
  "dependencies": {
13
- "@xylabs/assert": "^2.13.9",
14
- "@xyo-network/core": "~2.81.7",
15
- "@xyo-network/diviner-payload-abstract": "~2.81.7",
16
- "@xyo-network/diviner-payload-model": "~2.81.7",
17
- "@xyo-network/payload-model": "~2.81.7"
13
+ "@xylabs/assert": "^2.13.10",
14
+ "@xyo-network/core": "~2.82.0-rc.1",
15
+ "@xyo-network/diviner-payload-abstract": "~2.82.0-rc.1",
16
+ "@xyo-network/diviner-payload-model": "~2.82.0-rc.1",
17
+ "@xyo-network/payload-model": "~2.82.0-rc.1"
18
18
  },
19
19
  "devDependencies": {
20
- "@xylabs/ts-scripts-yarn3": "^3.2.8",
21
- "@xylabs/tsconfig": "^3.2.8",
22
- "@xyo-network/account": "~2.81.7",
23
- "@xyo-network/memory-archivist": "~2.81.7",
24
- "@xyo-network/node-memory": "~2.81.7",
25
- "@xyo-network/payload-builder": "~2.81.7",
20
+ "@xylabs/ts-scripts-yarn3": "^3.2.10",
21
+ "@xylabs/tsconfig": "^3.2.10",
22
+ "@xyo-network/account": "~2.82.0-rc.1",
23
+ "@xyo-network/memory-archivist": "~2.82.0-rc.1",
24
+ "@xyo-network/node-memory": "~2.82.0-rc.1",
25
+ "@xyo-network/payload-builder": "~2.82.0-rc.1",
26
26
  "typescript": "^5.3.2"
27
27
  },
28
28
  "description": "Primary SDK for using XYO Protocol 2.0",
@@ -65,5 +65,6 @@
65
65
  "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js.git"
66
66
  },
67
67
  "sideEffects": false,
68
- "version": "2.81.7"
68
+ "version": "2.82.0-rc.1",
69
+ "stableVersion": "2.81.7"
69
70
  }
@@ -34,7 +34,7 @@ export class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = Payload
34
34
  ? all.slice(parsedOffset, parsedLimit)
35
35
  : (async () => {
36
36
  const allPairs = await Promise.all(
37
- all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),
37
+ all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload, 'hex'), payload]),
38
38
  )
39
39
  if (hash) {
40
40
  //remove all until found