@xyo-network/diviner-payload-memory 2.75.12 → 2.75.14

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -54,7 +54,7 @@ var MemoryPayloadDiviner = class extends import_diviner_payload_abstract.Payload
54
54
  }
55
55
  const parsedLimit = limit || all.length;
56
56
  const parsedOffset = offset || 0;
57
- return parsedOffset ? all.slice(parsedOffset, parsedLimit) : (async () => {
57
+ return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
58
58
  const allPairs = await Promise.all(
59
59
  all.map(async (payload) => [await import_core.PayloadHasher.hashAsync(payload), payload])
60
60
  );
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return parsedOffset\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,aAAS,wBAAS,UAAU,OAAO,yDAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,eACH,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
1
+ {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,aAAS,wBAAS,UAAU,OAAO,yDAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
@@ -30,7 +30,7 @@ var MemoryPayloadDiviner = class extends PayloadDiviner {
30
30
  }
31
31
  const parsedLimit = limit || all.length;
32
32
  const parsedOffset = offset || 0;
33
- return parsedOffset ? all.slice(parsedOffset, parsedLimit) : (async () => {
33
+ return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
34
34
  const allPairs = await Promise.all(
35
35
  all.map(async (payload) => [await PayloadHasher.hashAsync(payload), payload])
36
36
  );
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return parsedOffset\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,SAAS,SAAS,UAAU,OAAO,4BAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,eACH,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
1
+ {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,SAAS,SAAS,UAAU,OAAO,4BAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
@@ -56,7 +56,7 @@ var MemoryPayloadDiviner = class extends import_diviner_payload_abstract.Payload
56
56
  }
57
57
  const parsedLimit = limit || all.length;
58
58
  const parsedOffset = offset || 0;
59
- return parsedOffset ? all.slice(parsedOffset, parsedLimit) : (async () => {
59
+ return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
60
60
  const allPairs = await Promise.all(
61
61
  all.map(async (payload) => [await import_core.PayloadHasher.hashAsync(payload), payload])
62
62
  );
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/index.ts","../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["export * from './MemoryPayloadDiviner'\n","import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return parsedOffset\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,aAAS,wBAAS,UAAU,OAAO,yDAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,eACH,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
1
+ {"version":3,"sources":["../../src/index.ts","../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["export * from './MemoryPayloadDiviner'\n","import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,aAAS,wBAAS,UAAU,OAAO,yDAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
@@ -30,7 +30,7 @@ var MemoryPayloadDiviner = class extends PayloadDiviner {
30
30
  }
31
31
  const parsedLimit = limit || all.length;
32
32
  const parsedOffset = offset || 0;
33
- return parsedOffset ? all.slice(parsedOffset, parsedLimit) : (async () => {
33
+ return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
34
34
  const allPairs = await Promise.all(
35
35
  all.map(async (payload) => [await PayloadHasher.hashAsync(payload), payload])
36
36
  );
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return parsedOffset\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,SAAS,SAAS,UAAU,OAAO,4BAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,eACH,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
1
+ {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AAC/E,UAAM,SAAS,SAAS,UAAU,OAAO,4BAA4B,GAAG,IAAI,GAAG,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,MAAM,UAAU,MAAM;AAChC,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,SAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAU,UAAU,QAAQ,GAAG,SAAS,KAAK,CAAC,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,YAAY,UAAU,QAAQ,MAAMA,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter"]}
package/dist/docs.json CHANGED
@@ -4216,7 +4216,7 @@
4216
4216
  "fileName": "modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts",
4217
4217
  "line": 8,
4218
4218
  "character": 18,
4219
- "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js/blob/9698850d2c/packages/modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts#L8"
4219
+ "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js/blob/19b8ee0046/packages/modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts#L8"
4220
4220
  }
4221
4221
  ],
4222
4222
  "type": {
@@ -7868,7 +7868,7 @@
7868
7868
  "fileName": "modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts",
7869
7869
  "line": 10,
7870
7870
  "character": 27,
7871
- "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js/blob/9698850d2c/packages/modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts#L10"
7871
+ "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js/blob/19b8ee0046/packages/modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts#L10"
7872
7872
  }
7873
7873
  ],
7874
7874
  "signatures": [
@@ -7883,7 +7883,7 @@
7883
7883
  "fileName": "modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts",
7884
7884
  "line": 10,
7885
7885
  "character": 27,
7886
- "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js/blob/9698850d2c/packages/modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts#L10"
7886
+ "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js/blob/19b8ee0046/packages/modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts#L10"
7887
7887
  }
7888
7888
  ],
7889
7889
  "parameters": [
@@ -17270,7 +17270,7 @@
17270
17270
  "fileName": "modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts",
17271
17271
  "line": 7,
17272
17272
  "character": 13,
17273
- "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js/blob/9698850d2c/packages/modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts#L7"
17273
+ "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js/blob/19b8ee0046/packages/modules/packages/diviner/packages/payload/packages/memory/src/MemoryPayloadDiviner.ts#L7"
17274
17274
  }
17275
17275
  ],
17276
17276
  "typeParameters": [
@@ -58,7 +58,7 @@ var MemoryPayloadDiviner = class extends import_diviner_payload_abstract.Payload
58
58
  }
59
59
  const parsedLimit = limit || all.length;
60
60
  const parsedOffset = offset || 0;
61
- return parsedOffset ? all.slice(parsedOffset, parsedLimit) : (async () => {
61
+ return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
62
62
  const allPairs = await Promise.all(
63
63
  all.map(async (payload) => [await import_core.PayloadHasher.hashAsync(payload), payload])
64
64
  );
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return parsedOffset\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,aAAS,yBAAS,0CAAU,OAAO,+DAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,eACH,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
1
+ {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,aAAS,yBAAS,0CAAU,OAAO,+DAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
@@ -34,7 +34,7 @@ var MemoryPayloadDiviner = class extends PayloadDiviner {
34
34
  }
35
35
  const parsedLimit = limit || all.length;
36
36
  const parsedOffset = offset || 0;
37
- return parsedOffset ? all.slice(parsedOffset, parsedLimit) : (async () => {
37
+ return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
38
38
  const allPairs = await Promise.all(
39
39
  all.map(async (payload) => [await PayloadHasher.hashAsync(payload), payload])
40
40
  );
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return parsedOffset\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,SAAS,UAAS,0CAAU,OAAO,kCAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,eACH,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
1
+ {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,SAAS,UAAS,0CAAU,OAAO,kCAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
@@ -60,7 +60,7 @@ var MemoryPayloadDiviner = class extends import_diviner_payload_abstract.Payload
60
60
  }
61
61
  const parsedLimit = limit || all.length;
62
62
  const parsedOffset = offset || 0;
63
- return parsedOffset ? all.slice(parsedOffset, parsedLimit) : (async () => {
63
+ return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
64
64
  const allPairs = await Promise.all(
65
65
  all.map(async (payload) => [await import_core.PayloadHasher.hashAsync(payload), payload])
66
66
  );
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/index.ts","../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["export * from './MemoryPayloadDiviner'\n","import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return parsedOffset\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,aAAS,yBAAS,0CAAU,OAAO,+DAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,eACH,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
1
+ {"version":3,"sources":["../../src/index.ts","../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["export * from './MemoryPayloadDiviner'\n","import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,oBAAyB;AACzB,kBAA8B;AAC9B,sCAA+B;AAC/B,mCAA+F;AAGxF,IAAM,uBAAN,cAAgG,+CAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,uDAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,aAAS,yBAAS,0CAAU,OAAO,+DAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,gBAAY,wBAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,0BAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
@@ -34,7 +34,7 @@ var MemoryPayloadDiviner = class extends PayloadDiviner {
34
34
  }
35
35
  const parsedLimit = limit || all.length;
36
36
  const parsedOffset = offset || 0;
37
- return parsedOffset ? all.slice(parsedOffset, parsedLimit) : (async () => {
37
+ return offset !== void 0 ? all.slice(parsedOffset, parsedLimit) : (async () => {
38
38
  const allPairs = await Promise.all(
39
39
  all.map(async (payload) => [await PayloadHasher.hashAsync(payload), payload])
40
40
  );
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return parsedOffset\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,SAAS,UAAS,0CAAU,OAAO,kCAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,eACH,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
1
+ {"version":3,"sources":["../../src/MemoryPayloadDiviner.ts"],"sourcesContent":["import { assertEx } from '@xylabs/assert'\nimport { PayloadHasher } from '@xyo-network/core'\nimport { PayloadDiviner } from '@xyo-network/diviner-payload-abstract'\nimport { isPayloadDivinerQueryPayload, PayloadDivinerConfigSchema, PayloadDivinerParams } from '@xyo-network/diviner-payload-model'\nimport { Payload } from '@xyo-network/payload-model'\n\nexport class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = PayloadDivinerParams> extends PayloadDiviner<TParams> {\n static override configSchemas = [PayloadDivinerConfigSchema]\n\n protected override async divineHandler(payloads?: Payload[]): Promise<Payload[]> {\n const filter = assertEx(payloads?.filter(isPayloadDivinerQueryPayload)?.pop(), 'Missing query payload')\n if (!filter) return []\n const archivist = assertEx(await this.readArchivist(), 'Unable to resolve archivist')\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n const { schemas, limit, offset, hash, order, schema, ...props } = filter\n let all = await archivist.all?.()\n if (all) {\n if (order === 'desc') all = all.reverse()\n if (schemas?.length) all = all.filter((payload) => schemas.includes(payload.schema))\n if (Object.keys(props).length > 0) {\n const additionalFilterCriteria = Object.entries(props)\n for (const [prop, filter] of additionalFilterCriteria) {\n const property = prop as keyof Payload\n if (Array.isArray(filter)) {\n all = all.filter((payload) => filter.every((value) => payload?.[property]?.includes(value)))\n } else {\n all = all.filter((payload) => payload?.[property] === filter)\n }\n }\n }\n const parsedLimit = limit || all.length\n const parsedOffset = offset || 0\n return offset !== undefined\n ? all.slice(parsedOffset, parsedLimit)\n : (async () => {\n const allPairs = await Promise.all(\n all.map<Promise<[string, Payload]>>(async (payload) => [await PayloadHasher.hashAsync(payload), payload]),\n )\n if (hash) {\n //remove all until found\n while (allPairs.length > 0 && allPairs[0][0] !== hash) {\n allPairs.shift()\n }\n //remove it if found\n if (allPairs.length > 0 && allPairs[0][0] === hash) {\n allPairs.shift()\n }\n }\n return allPairs.map(([, payload]) => payload)\n })()\n } else {\n throw Error('Archivist does not support \"all\"')\n }\n }\n}\n"],"mappings":";AAAA,SAAS,gBAAgB;AACzB,SAAS,qBAAqB;AAC9B,SAAS,sBAAsB;AAC/B,SAAS,8BAA8B,kCAAwD;AAGxF,IAAM,uBAAN,cAAgG,eAAwB;AAAA,EAC7H,OAAgB,gBAAgB,CAAC,0BAA0B;AAAA,EAE3D,MAAyB,cAAc,UAA0C;AATnF;AAUI,UAAM,SAAS,UAAS,0CAAU,OAAO,kCAAjB,mBAAgD,OAAO,uBAAuB;AACtG,QAAI,CAAC;AAAQ,aAAO,CAAC;AACrB,UAAM,YAAY,SAAS,MAAM,KAAK,cAAc,GAAG,6BAA6B;AAEpF,UAAM,EAAE,SAAS,OAAO,QAAQ,MAAM,OAAO,QAAQ,GAAG,MAAM,IAAI;AAClE,QAAI,MAAM,QAAM,eAAU,QAAV;AAChB,QAAI,KAAK;AACP,UAAI,UAAU;AAAQ,cAAM,IAAI,QAAQ;AACxC,UAAI,mCAAS;AAAQ,cAAM,IAAI,OAAO,CAAC,YAAY,QAAQ,SAAS,QAAQ,MAAM,CAAC;AACnF,UAAI,OAAO,KAAK,KAAK,EAAE,SAAS,GAAG;AACjC,cAAM,2BAA2B,OAAO,QAAQ,KAAK;AACrD,mBAAW,CAAC,MAAMA,OAAM,KAAK,0BAA0B;AACrD,gBAAM,WAAW;AACjB,cAAI,MAAM,QAAQA,OAAM,GAAG;AACzB,kBAAM,IAAI,OAAO,CAAC,YAAYA,QAAO,MAAM,CAAC,UAAO;AAxB/D,kBAAAC;AAwBkE,sBAAAA,MAAA,mCAAU,cAAV,gBAAAA,IAAqB,SAAS;AAAA,aAAM,CAAC;AAAA,UAC7F,OAAO;AACL,kBAAM,IAAI,OAAO,CAAC,aAAY,mCAAU,eAAcD,OAAM;AAAA,UAC9D;AAAA,QACF;AAAA,MACF;AACA,YAAM,cAAc,SAAS,IAAI;AACjC,YAAM,eAAe,UAAU;AAC/B,aAAO,WAAW,SACd,IAAI,MAAM,cAAc,WAAW,KAClC,YAAY;AACX,cAAM,WAAW,MAAM,QAAQ;AAAA,UAC7B,IAAI,IAAgC,OAAO,YAAY,CAAC,MAAM,cAAc,UAAU,OAAO,GAAG,OAAO,CAAC;AAAA,QAC1G;AACA,YAAI,MAAM;AAER,iBAAO,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AACrD,qBAAS,MAAM;AAAA,UACjB;AAEA,cAAI,SAAS,SAAS,KAAK,SAAS,CAAC,EAAE,CAAC,MAAM,MAAM;AAClD,qBAAS,MAAM;AAAA,UACjB;AAAA,QACF;AACA,eAAO,SAAS,IAAI,CAAC,CAAC,EAAE,OAAO,MAAM,OAAO;AAAA,MAC9C,GAAG;AAAA,IACT,OAAO;AACL,YAAM,MAAM,kCAAkC;AAAA,IAChD;AAAA,EACF;AACF;","names":["filter","_a"]}
package/package.json CHANGED
@@ -11,18 +11,18 @@
11
11
  },
12
12
  "dependencies": {
13
13
  "@xylabs/assert": "^2.12.14",
14
- "@xyo-network/core": "~2.75.12",
15
- "@xyo-network/diviner-payload-abstract": "~2.75.12",
16
- "@xyo-network/diviner-payload-model": "~2.75.12",
17
- "@xyo-network/payload-model": "~2.75.12"
14
+ "@xyo-network/core": "~2.75.14",
15
+ "@xyo-network/diviner-payload-abstract": "~2.75.14",
16
+ "@xyo-network/diviner-payload-model": "~2.75.14",
17
+ "@xyo-network/payload-model": "~2.75.14"
18
18
  },
19
19
  "devDependencies": {
20
20
  "@xylabs/ts-scripts-yarn3": "^3.0.88",
21
21
  "@xylabs/tsconfig": "^3.0.88",
22
- "@xyo-network/account": "~2.75.12",
23
- "@xyo-network/memory-archivist": "~2.75.12",
24
- "@xyo-network/node-memory": "~2.75.12",
25
- "@xyo-network/payload-builder": "~2.75.12",
22
+ "@xyo-network/account": "~2.75.14",
23
+ "@xyo-network/memory-archivist": "~2.75.14",
24
+ "@xyo-network/node-memory": "~2.75.14",
25
+ "@xyo-network/payload-builder": "~2.75.14",
26
26
  "typescript": "^5.2.2"
27
27
  },
28
28
  "description": "Primary SDK for using XYO Protocol 2.0",
@@ -68,5 +68,5 @@
68
68
  "url": "https://github.com/XYOracleNetwork/sdk-xyo-client-js.git"
69
69
  },
70
70
  "sideEffects": false,
71
- "version": "2.75.12"
71
+ "version": "2.75.14"
72
72
  }
@@ -30,7 +30,7 @@ export class MemoryPayloadDiviner<TParams extends PayloadDivinerParams = Payload
30
30
  }
31
31
  const parsedLimit = limit || all.length
32
32
  const parsedOffset = offset || 0
33
- return parsedOffset
33
+ return offset !== undefined
34
34
  ? all.slice(parsedOffset, parsedLimit)
35
35
  : (async () => {
36
36
  const allPairs = await Promise.all(