@latticexyz/store-indexer 2.0.0-main-5df1f31b → 2.0.0-main-e48fb3b0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,3 +1,3 @@
1
1
  #!/usr/bin/env node
2
- import{a as T,c as A}from"../chunk-2E5MDUA2.js";import"dotenv/config";import{z as y}from"zod";import Q from"fastify";import{fastifyTRPCPlugin as U}from"@trpc/server/adapters/fastify";import{createAppRouter as q}from"@latticexyz/store-sync/trpc-indexer";import{getAddress as S}from"viem";import{isTableRegistrationLog as v,logToTable as w,storeTables as N}from"@latticexyz/store-sync";import{decodeKey as $,decodeValueArgs as z}from"@latticexyz/protocol-parser";import R from"debug";var k=R("mud:store-indexer");import{tables as o}from"@latticexyz/store-sync/postgres";import{and as D,asc as L,eq as n,or as I}from"drizzle-orm";import{decodeDynamicField as P}from"@latticexyz/protocol-parser";import{bigIntMax as B}from"@latticexyz/common/utils";async function f(a,{chainId:s,address:r,filters:i=[]}){let g=i.length?i.map(e=>D(r!=null?n(o.recordsTable.address,r):void 0,n(o.recordsTable.tableId,e.tableId),e.key0!=null?n(o.recordsTable.key0,e.key0):void 0,e.key1!=null?n(o.recordsTable.key1,e.key1):void 0)):r!=null?[n(o.recordsTable.address,r)]:[],m=(await a.select().from(o.chainTable).where(n(o.chainTable.chainId,s)).limit(1).execute().then(e=>e.find(()=>!0)))?.lastUpdatedBlockNumber??0n,p=await a.select().from(o.recordsTable).where(I(...g)).orderBy(L(o.recordsTable.lastUpdatedBlockNumber)),b=p.reduce((e,u)=>B(e,u.lastUpdatedBlockNumber??0n),m),t=p.filter(e=>!e.isDeleted).map(e=>({address:e.address,eventName:"Store_SetRecord",args:{tableId:e.tableId,keyTuple:P("bytes32[]",e.keyBytes),staticData:e.staticData??"0x",encodedLengths:e.encodedLengths??"0x",dynamicData:e.dynamicData??"0x"}}));return{blockNumber:b,logs:t}}import{groupBy as O}from"@latticexyz/common/utils";async function x(a){return{async getLogs(r){return f(a,r)},async findAll(r){let i=r.filters??[],{blockNumber:g,logs:l}=await f(a,{...r,filters:i.length>0?[...i,{tableId:N.Tables.tableId}]:[]}),m=l.filter(v).map(w),p=O(l,t=>`${S(t.address)}:${t.args.tableId}`),b=m.map(t=>{let u=(p.get(`${S(t.address)}:${t.tableId}`)??[]).map(h=>({key:$(t.keySchema,h.args.keyTuple),value:z(t.valueSchema,h.args)}));return{...t,records:u}});return k("findAll: decoded %d logs across %d tables",l.length,m.length),{blockNumber:g,tables:b}}}}import{drizzle as E}from"drizzle-orm/postgres-js";import H from"postgres";var c=A(y.intersection(T,y.object({DATABASE_URL:y.string()}))),_=E(H(c.DATABASE_URL)),d=Q({maxParamLength:5e3,logger:!0});await d.register(import("@fastify/compress"));await d.register(import("@fastify/cors"));d.get("/healthz",(a,s)=>s.code(200).send());d.get("/readyz",(a,s)=>s.code(200).send());d.register(U,{prefix:"/trpc",trpcOptions:{router:q(),createContext:async()=>({queryAdapter:await x(_)})}});await d.listen({host:c.HOST,port:c.PORT});console.log(`postgres indexer frontend listening on http://${c.HOST}:${c.PORT}`);
2
+ import{a as h,c as A}from"../chunk-2E5MDUA2.js";import"dotenv/config";import{z as y}from"zod";import Q from"fastify";import{fastifyTRPCPlugin as U}from"@trpc/server/adapters/fastify";import{createAppRouter as q}from"@latticexyz/store-sync/trpc-indexer";import{getAddress as S}from"viem";import{isTableRegistrationLog as v,logToTable as w,storeTables as N}from"@latticexyz/store-sync";import{decodeKey as $,decodeValueArgs as z}from"@latticexyz/protocol-parser";import R from"debug";var k=R("mud:store-indexer");import{tables as o}from"@latticexyz/store-sync/postgres";import{and as D,asc as L,eq as n,or as I}from"drizzle-orm";import{decodeDynamicField as P}from"@latticexyz/protocol-parser";import{bigIntMax as B}from"@latticexyz/common/utils";async function f(a,{chainId:s,address:r,filters:i=[]}){let g=i.length?i.map(e=>D(r!=null?n(o.recordsTable.address,r):void 0,n(o.recordsTable.tableId,e.tableId),e.key0!=null?n(o.recordsTable.key0,e.key0):void 0,e.key1!=null?n(o.recordsTable.key1,e.key1):void 0)):r!=null?[n(o.recordsTable.address,r)]:[],m=(await a.select().from(o.configTable).where(n(o.configTable.chainId,s)).limit(1).execute().then(e=>e.find(()=>!0)))?.lastUpdatedBlockNumber??0n,p=await a.select().from(o.recordsTable).where(I(...g)).orderBy(L(o.recordsTable.lastUpdatedBlockNumber)),b=p.reduce((e,u)=>B(e,u.lastUpdatedBlockNumber??0n),m),t=p.filter(e=>!e.isDeleted).map(e=>({address:e.address,eventName:"Store_SetRecord",args:{tableId:e.tableId,keyTuple:P("bytes32[]",e.keyBytes),staticData:e.staticData??"0x",encodedLengths:e.encodedLengths??"0x",dynamicData:e.dynamicData??"0x"}}));return{blockNumber:b,logs:t}}import{groupBy as O}from"@latticexyz/common/utils";async function x(a){return{async getLogs(r){return f(a,r)},async findAll(r){let i=r.filters??[],{blockNumber:g,logs:l}=await f(a,{...r,filters:i.length>0?[...i,{tableId:N.Tables.tableId}]:[]}),m=l.filter(v).map(w),p=O(l,t=>`${S(t.address)}:${t.args.tableId}`),b=m.map(t=>{let u=(p.get(`${S(t.address)}:${t.tableId}`)??[]).map(T=>({key:$(t.keySchema,T.args.keyTuple),value:z(t.valueSchema,T.args)}));return{...t,records:u}});return k("findAll: decoded %d logs across %d tables",l.length,m.length),{blockNumber:g,tables:b}}}}import{drizzle as E}from"drizzle-orm/postgres-js";import H from"postgres";var c=A(y.intersection(h,y.object({DATABASE_URL:y.string()}))),_=E(H(c.DATABASE_URL)),d=Q({maxParamLength:5e3,logger:!0});await d.register(import("@fastify/compress"));await d.register(import("@fastify/cors"));d.get("/healthz",(a,s)=>s.code(200).send());d.get("/readyz",(a,s)=>s.code(200).send());d.register(U,{prefix:"/trpc",trpcOptions:{router:q(),createContext:async()=>({queryAdapter:await x(_)})}});await d.listen({host:c.HOST,port:c.PORT});console.log(`postgres indexer frontend listening on http://${c.HOST}:${c.PORT}`);
3
3
  //# sourceMappingURL=postgres-frontend.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../bin/postgres-frontend.ts","../../src/postgres/createQueryAdapter.ts","../../src/debug.ts","../../src/postgres/getLogs.ts"],"sourcesContent":["#!/usr/bin/env node\nimport \"dotenv/config\";\nimport { z } from \"zod\";\nimport fastify from \"fastify\";\nimport { fastifyTRPCPlugin } from \"@trpc/server/adapters/fastify\";\nimport { AppRouter, createAppRouter } from \"@latticexyz/store-sync/trpc-indexer\";\nimport { createQueryAdapter } from \"../src/postgres/createQueryAdapter\";\nimport { drizzle } from \"drizzle-orm/postgres-js\";\nimport postgres from \"postgres\";\nimport { frontendEnvSchema, parseEnv } from \"./parseEnv\";\n\nconst env = parseEnv(\n z.intersection(\n frontendEnvSchema,\n z.object({\n DATABASE_URL: z.string(),\n })\n )\n);\n\nconst database = drizzle(postgres(env.DATABASE_URL));\n\n// @see https://fastify.dev/docs/latest/\nconst server = fastify({\n maxParamLength: 5000,\n logger: true,\n});\n\nawait server.register(import(\"@fastify/compress\"));\nawait server.register(import(\"@fastify/cors\"));\n\n// k8s healthchecks\nserver.get(\"/healthz\", (req, res) => res.code(200).send());\nserver.get(\"/readyz\", (req, res) => res.code(200).send());\n\n// @see https://trpc.io/docs/server/adapters/fastify\nserver.register(fastifyTRPCPlugin<AppRouter>, {\n prefix: \"/trpc\",\n trpcOptions: {\n router: createAppRouter(),\n createContext: async () => ({\n queryAdapter: await createQueryAdapter(database),\n }),\n },\n});\n\nawait server.listen({ host: env.HOST, port: env.PORT });\nconsole.log(`postgres indexer frontend listening on http://${env.HOST}:${env.PORT}`);\n","import { getAddress } from \"viem\";\nimport { PgDatabase } from \"drizzle-orm/pg-core\";\nimport { TableWithRecords, isTableRegistrationLog, logToTable, storeTables } from \"@latticexyz/store-sync\";\nimport { decodeKey, decodeValueArgs } from \"@latticexyz/protocol-parser\";\nimport { QueryAdapter } from \"@latticexyz/store-sync/trpc-indexer\";\nimport { debug } from \"../debug\";\nimport { getLogs } from \"./getLogs\";\nimport { groupBy } from \"@latticexyz/common/utils\";\n\n/**\n * Creates a query adapter for the tRPC server/client to query data from Postgres.\n *\n * @param {PgDatabase<any>} database Postgres database object from Drizzle\n * @returns {Promise<QueryAdapter>} A set of methods used by tRPC endpoints.\n */\nexport async function createQueryAdapter(database: PgDatabase<any>): Promise<QueryAdapter> {\n const adapter: QueryAdapter = {\n async getLogs(opts) {\n return getLogs(database, opts);\n },\n async findAll(opts) {\n const filters = opts.filters ?? [];\n const { blockNumber, logs } = await getLogs(database, {\n ...opts,\n // make sure we're always retrieving `store.Tables` table, so we can decode table values\n filters: filters.length > 0 ? [...filters, { tableId: storeTables.Tables.tableId }] : [],\n });\n\n const tables = logs.filter(isTableRegistrationLog).map(logToTable);\n\n const logsByTable = groupBy(logs, (log) => `${getAddress(log.address)}:${log.args.tableId}`);\n\n const tablesWithRecords: TableWithRecords[] = tables.map((table) => {\n const tableLogs = logsByTable.get(`${getAddress(table.address)}:${table.tableId}`) ?? [];\n const records = tableLogs.map((log) => ({\n key: decodeKey(table.keySchema, log.args.keyTuple),\n value: decodeValueArgs(table.valueSchema, log.args),\n }));\n\n return {\n ...table,\n records,\n };\n });\n\n debug(\"findAll: decoded %d logs across %d tables\", logs.length, tables.length);\n\n return {\n blockNumber,\n tables: tablesWithRecords,\n };\n },\n };\n return adapter;\n}\n","import createDebug from \"debug\";\n\nexport const debug = createDebug(\"mud:store-indexer\");\n","import { PgDatabase } from \"drizzle-orm/pg-core\";\nimport { Hex } from \"viem\";\nimport { StorageAdapterLog, SyncFilter } from \"@latticexyz/store-sync\";\nimport { tables } from \"@latticexyz/store-sync/postgres\";\nimport { and, asc, eq, or } from \"drizzle-orm\";\nimport { decodeDynamicField } from \"@latticexyz/protocol-parser\";\nimport { bigIntMax } from \"@latticexyz/common/utils\";\n\nexport async function getLogs(\n database: PgDatabase<any>,\n {\n chainId,\n address,\n filters = [],\n }: {\n readonly chainId: number;\n readonly address?: Hex;\n readonly filters?: readonly SyncFilter[];\n }\n): Promise<{ blockNumber: bigint; logs: (StorageAdapterLog & { eventName: \"Store_SetRecord\" })[] }> {\n const conditions = filters.length\n ? filters.map((filter) =>\n and(\n address != null ? eq(tables.recordsTable.address, address) : undefined,\n eq(tables.recordsTable.tableId, filter.tableId),\n filter.key0 != null ? eq(tables.recordsTable.key0, filter.key0) : undefined,\n filter.key1 != null ? eq(tables.recordsTable.key1, filter.key1) : undefined\n )\n )\n : address != null\n ? [eq(tables.recordsTable.address, address)]\n : [];\n\n // Query for the block number that the indexer (i.e. chain) is at, in case the\n // indexer is further along in the chain than a given store/table's last updated\n // block number. We'll then take the highest block number between the indexer's\n // chain state and all the records in the query (in case the records updated\n // between these queries). Using just the highest block number from the queries\n // could potentially signal to the client an older-than-necessary block number,\n // for stores/tables that haven't seen recent activity.\n // TODO: move the block number query into the records query for atomicity so we don't have to merge them here\n const chainState = await database\n .select()\n .from(tables.chainTable)\n .where(eq(tables.chainTable.chainId, chainId))\n .limit(1)\n .execute()\n // Get the first record in a way that returns a possible `undefined`\n // TODO: move this to `.findFirst` after upgrading drizzle or `rows[0]` after enabling `noUncheckedIndexedAccess: true`\n .then((rows) => rows.find(() => true));\n const indexerBlockNumber = chainState?.lastUpdatedBlockNumber ?? 0n;\n\n const records = await database\n .select()\n .from(tables.recordsTable)\n .where(or(...conditions))\n .orderBy(\n asc(tables.recordsTable.lastUpdatedBlockNumber)\n // TODO: add logIndex (https://github.com/latticexyz/mud/issues/1979)\n );\n\n const blockNumber = records.reduce(\n (max, record) => bigIntMax(max, record.lastUpdatedBlockNumber ?? 0n),\n indexerBlockNumber\n );\n\n const logs = records\n // TODO: add this to the query, assuming we can optimize with an index\n .filter((record) => !record.isDeleted)\n .map(\n (record) =>\n ({\n address: record.address,\n eventName: \"Store_SetRecord\",\n args: {\n tableId: record.tableId,\n keyTuple: decodeDynamicField(\"bytes32[]\", record.keyBytes),\n staticData: record.staticData ?? \"0x\",\n encodedLengths: record.encodedLengths ?? \"0x\",\n dynamicData: record.dynamicData ?? \"0x\",\n },\n } as const)\n );\n\n return { blockNumber, logs };\n}\n"],"mappings":";gDACA,MAAO,gBACP,OAAS,KAAAA,MAAS,MAClB,OAAOC,MAAa,UACpB,OAAS,qBAAAC,MAAyB,gCAClC,OAAoB,mBAAAC,MAAuB,sCCL3C,OAAS,cAAAC,MAAkB,OAE3B,OAA2B,0BAAAC,EAAwB,cAAAC,EAAY,eAAAC,MAAmB,yBAClF,OAAS,aAAAC,EAAW,mBAAAC,MAAuB,8BCH3C,OAAOC,MAAiB,QAEjB,IAAMC,EAAQD,EAAY,mBAAmB,ECCpD,OAAS,UAAAE,MAAc,kCACvB,OAAS,OAAAC,EAAK,OAAAC,EAAK,MAAAC,EAAI,MAAAC,MAAU,cACjC,OAAS,sBAAAC,MAA0B,8BACnC,OAAS,aAAAC,MAAiB,2BAE1B,eAAsBC,EACpBC,EACA,CACE,QAAAC,EACA,QAAAC,EACA,QAAAC,EAAU,CAAC,CACb,EAKkG,CAClG,IAAMC,EAAaD,EAAQ,OACvBA,EAAQ,IAAKE,GACXZ,EACES,GAAW,KAAOP,EAAGH,EAAO,aAAa,QAASU,CAAO,EAAI,OAC7DP,EAAGH,EAAO,aAAa,QAASa,EAAO,OAAO,EAC9CA,EAAO,MAAQ,KAAOV,EAAGH,EAAO,aAAa,KAAMa,EAAO,IAAI,EAAI,OAClEA,EAAO,MAAQ,KAAOV,EAAGH,EAAO,aAAa,KAAMa,EAAO,IAAI,EAAI,MACpE,CACF,EACAH,GAAW,KACX,CAACP,EAAGH,EAAO,aAAa,QAASU,CAAO,CAAC,EACzC,CAAC,EAmBCI,GATa,MAAMN,EACtB,OAAO,EACP,KAAKR,EAAO,UAAU,EACtB,MAAMG,EAAGH,EAAO,WAAW,QAASS,CAAO,CAAC,EAC5C,MAAM,CAAC,EACP,QAAQ,EAGR,KAAMM,GAASA,EAAK,KAAK,IAAM,EAAI,CAAC,IACA,wBAA0B,GAE3DC,EAAU,MAAMR,EACnB,OAAO,EACP,KAAKR,EAAO,YAAY,EACxB,MAAMI,EAAG,GAAGQ,CAAU,CAAC,EACvB,QACCV,EAAIF,EAAO,aAAa,sBAAsB,CAEhD,EAEIiB,EAAcD,EAAQ,OAC1B,CAACE,EAAKC,IAAWb,EAAUY,EAAKC,EAAO,wBAA0B,EAAE,EACnEL,CACF,EAEMM,EAAOJ,EAEV,OAAQG,GAAW,CAACA,EAAO,SAAS,EACpC,IACEA,IACE,CACC,QAASA,EAAO,QAChB,UAAW,kBACX,KAAM,CACJ,QAASA,EAAO,QAChB,SAAUd,EAAmB,YAAac,EAAO,QAAQ,EACzD,WAAYA,EAAO,YAAc,KACjC,eAAgBA,EAAO,gBAAkB,KACzC,YAAaA,EAAO,aAAe,IACrC,CACF,EACJ,EAEF,MAAO,CAAE,YAAAF,EAAa,KAAAG,CAAK,CAC7B,CF9EA,OAAS,WAAAC,MAAe,2BAQxB,eAAsBC,EAAmBC,EAAkD,CAsCzF,MArC8B,CAC5B,MAAM,QAAQC,EAAM,CAClB,OAAOC,EAAQF,EAAUC,CAAI,CAC/B,EACA,MAAM,QAAQA,EAAM,CAClB,IAAME,EAAUF,EAAK,SAAW,CAAC,EAC3B,CAAE,YAAAG,EAAa,KAAAC,CAAK,EAAI,MAAMH,EAAQF,EAAU,CACpD,GAAGC,EAEH,QAASE,EAAQ,OAAS,EAAI,CAAC,GAAGA,EAAS,CAAE,QAASG,EAAY,OAAO,OAAQ,CAAC,EAAI,CAAC,CACzF,CAAC,EAEKC,EAASF,EAAK,OAAOG,CAAsB,EAAE,IAAIC,CAAU,EAE3DC,EAAcZ,EAAQO,EAAOM,GAAQ,GAAGC,EAAWD,EAAI,OAAO,KAAKA,EAAI,KAAK,SAAS,EAErFE,EAAwCN,EAAO,IAAKO,GAAU,CAElE,IAAMC,GADYL,EAAY,IAAI,GAAGE,EAAWE,EAAM,OAAO,KAAKA,EAAM,SAAS,GAAK,CAAC,GAC7D,IAAKH,IAAS,CACtC,IAAKK,EAAUF,EAAM,UAAWH,EAAI,KAAK,QAAQ,EACjD,MAAOM,EAAgBH,EAAM,YAAaH,EAAI,IAAI,CACpD,EAAE,EAEF,MAAO,CACL,GAAGG,EACH,QAAAC,CACF,CACF,CAAC,EAED,OAAAG,EAAM,4CAA6Cb,EAAK,OAAQE,EAAO,MAAM,EAEtE,CACL,YAAAH,EACA,OAAQS,CACV,CACF,CACF,CAEF,CD/CA,OAAS,WAAAM,MAAe,0BACxB,OAAOC,MAAc,WAGrB,IAAMC,EAAMC,EACVC,EAAE,aACAC,EACAD,EAAE,OAAO,CACP,aAAcA,EAAE,OAAO,CACzB,CAAC,CACH,CACF,EAEME,EAAWC,EAAQC,EAASN,EAAI,YAAY,CAAC,EAG7CO,EAASC,EAAQ,CACrB,eAAgB,IAChB,OAAQ,EACV,CAAC,EAED,MAAMD,EAAO,SAAS,OAAO,mBAAmB,CAAC,EACjD,MAAMA,EAAO,SAAS,OAAO,eAAe,CAAC,EAG7CA,EAAO,IAAI,WAAY,CAACE,EAAKC,IAAQA,EAAI,KAAK,GAAG,EAAE,KAAK,CAAC,EACzDH,EAAO,IAAI,UAAW,CAACE,EAAKC,IAAQA,EAAI,KAAK,GAAG,EAAE,KAAK,CAAC,EAGxDH,EAAO,SAASI,EAA8B,CAC5C,OAAQ,QACR,YAAa,CACX,OAAQC,EAAgB,EACxB,cAAe,UAAa,CAC1B,aAAc,MAAMC,EAAmBT,CAAQ,CACjD,EACF,CACF,CAAC,EAED,MAAMG,EAAO,OAAO,CAAE,KAAMP,EAAI,KAAM,KAAMA,EAAI,IAAK,CAAC,EACtD,QAAQ,IAAI,iDAAiDA,EAAI,QAAQA,EAAI,MAAM","names":["z","fastify","fastifyTRPCPlugin","createAppRouter","getAddress","isTableRegistrationLog","logToTable","storeTables","decodeKey","decodeValueArgs","createDebug","debug","tables","and","asc","eq","or","decodeDynamicField","bigIntMax","getLogs","database","chainId","address","filters","conditions","filter","indexerBlockNumber","rows","records","blockNumber","max","record","logs","groupBy","createQueryAdapter","database","opts","getLogs","filters","blockNumber","logs","storeTables","tables","isTableRegistrationLog","logToTable","logsByTable","log","getAddress","tablesWithRecords","table","records","decodeKey","decodeValueArgs","debug","drizzle","postgres","env","parseEnv","z","frontendEnvSchema","database","drizzle","postgres","server","fastify","req","res","fastifyTRPCPlugin","createAppRouter","createQueryAdapter"]}
1
+ {"version":3,"sources":["../../bin/postgres-frontend.ts","../../src/postgres/createQueryAdapter.ts","../../src/debug.ts","../../src/postgres/getLogs.ts"],"sourcesContent":["#!/usr/bin/env node\nimport \"dotenv/config\";\nimport { z } from \"zod\";\nimport fastify from \"fastify\";\nimport { fastifyTRPCPlugin } from \"@trpc/server/adapters/fastify\";\nimport { AppRouter, createAppRouter } from \"@latticexyz/store-sync/trpc-indexer\";\nimport { createQueryAdapter } from \"../src/postgres/createQueryAdapter\";\nimport { drizzle } from \"drizzle-orm/postgres-js\";\nimport postgres from \"postgres\";\nimport { frontendEnvSchema, parseEnv } from \"./parseEnv\";\n\nconst env = parseEnv(\n z.intersection(\n frontendEnvSchema,\n z.object({\n DATABASE_URL: z.string(),\n })\n )\n);\n\nconst database = drizzle(postgres(env.DATABASE_URL));\n\n// @see https://fastify.dev/docs/latest/\nconst server = fastify({\n maxParamLength: 5000,\n logger: true,\n});\n\nawait server.register(import(\"@fastify/compress\"));\nawait server.register(import(\"@fastify/cors\"));\n\n// k8s healthchecks\nserver.get(\"/healthz\", (req, res) => res.code(200).send());\nserver.get(\"/readyz\", (req, res) => res.code(200).send());\n\n// @see https://trpc.io/docs/server/adapters/fastify\nserver.register(fastifyTRPCPlugin<AppRouter>, {\n prefix: \"/trpc\",\n trpcOptions: {\n router: createAppRouter(),\n createContext: async () => ({\n queryAdapter: await createQueryAdapter(database),\n }),\n },\n});\n\nawait server.listen({ host: env.HOST, port: env.PORT });\nconsole.log(`postgres indexer frontend listening on http://${env.HOST}:${env.PORT}`);\n","import { getAddress } from \"viem\";\nimport { PgDatabase } from \"drizzle-orm/pg-core\";\nimport { TableWithRecords, isTableRegistrationLog, logToTable, storeTables } from \"@latticexyz/store-sync\";\nimport { decodeKey, decodeValueArgs } from \"@latticexyz/protocol-parser\";\nimport { QueryAdapter } from \"@latticexyz/store-sync/trpc-indexer\";\nimport { debug } from \"../debug\";\nimport { getLogs } from \"./getLogs\";\nimport { groupBy } from \"@latticexyz/common/utils\";\n\n/**\n * Creates a query adapter for the tRPC server/client to query data from Postgres.\n *\n * @param {PgDatabase<any>} database Postgres database object from Drizzle\n * @returns {Promise<QueryAdapter>} A set of methods used by tRPC endpoints.\n */\nexport async function createQueryAdapter(database: PgDatabase<any>): Promise<QueryAdapter> {\n const adapter: QueryAdapter = {\n async getLogs(opts) {\n return getLogs(database, opts);\n },\n async findAll(opts) {\n const filters = opts.filters ?? [];\n const { blockNumber, logs } = await getLogs(database, {\n ...opts,\n // make sure we're always retrieving `store.Tables` table, so we can decode table values\n filters: filters.length > 0 ? [...filters, { tableId: storeTables.Tables.tableId }] : [],\n });\n\n const tables = logs.filter(isTableRegistrationLog).map(logToTable);\n\n const logsByTable = groupBy(logs, (log) => `${getAddress(log.address)}:${log.args.tableId}`);\n\n const tablesWithRecords: TableWithRecords[] = tables.map((table) => {\n const tableLogs = logsByTable.get(`${getAddress(table.address)}:${table.tableId}`) ?? [];\n const records = tableLogs.map((log) => ({\n key: decodeKey(table.keySchema, log.args.keyTuple),\n value: decodeValueArgs(table.valueSchema, log.args),\n }));\n\n return {\n ...table,\n records,\n };\n });\n\n debug(\"findAll: decoded %d logs across %d tables\", logs.length, tables.length);\n\n return {\n blockNumber,\n tables: tablesWithRecords,\n };\n },\n };\n return adapter;\n}\n","import createDebug from \"debug\";\n\nexport const debug = createDebug(\"mud:store-indexer\");\n","import { PgDatabase } from \"drizzle-orm/pg-core\";\nimport { Hex } from \"viem\";\nimport { StorageAdapterLog, SyncFilter } from \"@latticexyz/store-sync\";\nimport { tables } from \"@latticexyz/store-sync/postgres\";\nimport { and, asc, eq, or } from \"drizzle-orm\";\nimport { decodeDynamicField } from \"@latticexyz/protocol-parser\";\nimport { bigIntMax } from \"@latticexyz/common/utils\";\n\nexport async function getLogs(\n database: PgDatabase<any>,\n {\n chainId,\n address,\n filters = [],\n }: {\n readonly chainId: number;\n readonly address?: Hex;\n readonly filters?: readonly SyncFilter[];\n }\n): Promise<{ blockNumber: bigint; logs: (StorageAdapterLog & { eventName: \"Store_SetRecord\" })[] }> {\n const conditions = filters.length\n ? filters.map((filter) =>\n and(\n address != null ? eq(tables.recordsTable.address, address) : undefined,\n eq(tables.recordsTable.tableId, filter.tableId),\n filter.key0 != null ? eq(tables.recordsTable.key0, filter.key0) : undefined,\n filter.key1 != null ? eq(tables.recordsTable.key1, filter.key1) : undefined\n )\n )\n : address != null\n ? [eq(tables.recordsTable.address, address)]\n : [];\n\n // Query for the block number that the indexer (i.e. chain) is at, in case the\n // indexer is further along in the chain than a given store/table's last updated\n // block number. We'll then take the highest block number between the indexer's\n // chain state and all the records in the query (in case the records updated\n // between these queries). Using just the highest block number from the queries\n // could potentially signal to the client an older-than-necessary block number,\n // for stores/tables that haven't seen recent activity.\n // TODO: move the block number query into the records query for atomicity so we don't have to merge them here\n const chainState = await database\n .select()\n .from(tables.configTable)\n .where(eq(tables.configTable.chainId, chainId))\n .limit(1)\n .execute()\n // Get the first record in a way that returns a possible `undefined`\n // TODO: move this to `.findFirst` after upgrading drizzle or `rows[0]` after enabling `noUncheckedIndexedAccess: true`\n .then((rows) => rows.find(() => true));\n const indexerBlockNumber = chainState?.lastUpdatedBlockNumber ?? 0n;\n\n const records = await database\n .select()\n .from(tables.recordsTable)\n .where(or(...conditions))\n .orderBy(\n asc(tables.recordsTable.lastUpdatedBlockNumber)\n // TODO: add logIndex (https://github.com/latticexyz/mud/issues/1979)\n );\n\n const blockNumber = records.reduce(\n (max, record) => bigIntMax(max, record.lastUpdatedBlockNumber ?? 0n),\n indexerBlockNumber\n );\n\n const logs = records\n // TODO: add this to the query, assuming we can optimize with an index\n .filter((record) => !record.isDeleted)\n .map(\n (record) =>\n ({\n address: record.address,\n eventName: \"Store_SetRecord\",\n args: {\n tableId: record.tableId,\n keyTuple: decodeDynamicField(\"bytes32[]\", record.keyBytes),\n staticData: record.staticData ?? \"0x\",\n encodedLengths: record.encodedLengths ?? \"0x\",\n dynamicData: record.dynamicData ?? \"0x\",\n },\n } as const)\n );\n\n return { blockNumber, logs };\n}\n"],"mappings":";gDACA,MAAO,gBACP,OAAS,KAAAA,MAAS,MAClB,OAAOC,MAAa,UACpB,OAAS,qBAAAC,MAAyB,gCAClC,OAAoB,mBAAAC,MAAuB,sCCL3C,OAAS,cAAAC,MAAkB,OAE3B,OAA2B,0BAAAC,EAAwB,cAAAC,EAAY,eAAAC,MAAmB,yBAClF,OAAS,aAAAC,EAAW,mBAAAC,MAAuB,8BCH3C,OAAOC,MAAiB,QAEjB,IAAMC,EAAQD,EAAY,mBAAmB,ECCpD,OAAS,UAAAE,MAAc,kCACvB,OAAS,OAAAC,EAAK,OAAAC,EAAK,MAAAC,EAAI,MAAAC,MAAU,cACjC,OAAS,sBAAAC,MAA0B,8BACnC,OAAS,aAAAC,MAAiB,2BAE1B,eAAsBC,EACpBC,EACA,CACE,QAAAC,EACA,QAAAC,EACA,QAAAC,EAAU,CAAC,CACb,EAKkG,CAClG,IAAMC,EAAaD,EAAQ,OACvBA,EAAQ,IAAKE,GACXZ,EACES,GAAW,KAAOP,EAAGH,EAAO,aAAa,QAASU,CAAO,EAAI,OAC7DP,EAAGH,EAAO,aAAa,QAASa,EAAO,OAAO,EAC9CA,EAAO,MAAQ,KAAOV,EAAGH,EAAO,aAAa,KAAMa,EAAO,IAAI,EAAI,OAClEA,EAAO,MAAQ,KAAOV,EAAGH,EAAO,aAAa,KAAMa,EAAO,IAAI,EAAI,MACpE,CACF,EACAH,GAAW,KACX,CAACP,EAAGH,EAAO,aAAa,QAASU,CAAO,CAAC,EACzC,CAAC,EAmBCI,GATa,MAAMN,EACtB,OAAO,EACP,KAAKR,EAAO,WAAW,EACvB,MAAMG,EAAGH,EAAO,YAAY,QAASS,CAAO,CAAC,EAC7C,MAAM,CAAC,EACP,QAAQ,EAGR,KAAMM,GAASA,EAAK,KAAK,IAAM,EAAI,CAAC,IACA,wBAA0B,GAE3DC,EAAU,MAAMR,EACnB,OAAO,EACP,KAAKR,EAAO,YAAY,EACxB,MAAMI,EAAG,GAAGQ,CAAU,CAAC,EACvB,QACCV,EAAIF,EAAO,aAAa,sBAAsB,CAEhD,EAEIiB,EAAcD,EAAQ,OAC1B,CAACE,EAAKC,IAAWb,EAAUY,EAAKC,EAAO,wBAA0B,EAAE,EACnEL,CACF,EAEMM,EAAOJ,EAEV,OAAQG,GAAW,CAACA,EAAO,SAAS,EACpC,IACEA,IACE,CACC,QAASA,EAAO,QAChB,UAAW,kBACX,KAAM,CACJ,QAASA,EAAO,QAChB,SAAUd,EAAmB,YAAac,EAAO,QAAQ,EACzD,WAAYA,EAAO,YAAc,KACjC,eAAgBA,EAAO,gBAAkB,KACzC,YAAaA,EAAO,aAAe,IACrC,CACF,EACJ,EAEF,MAAO,CAAE,YAAAF,EAAa,KAAAG,CAAK,CAC7B,CF9EA,OAAS,WAAAC,MAAe,2BAQxB,eAAsBC,EAAmBC,EAAkD,CAsCzF,MArC8B,CAC5B,MAAM,QAAQC,EAAM,CAClB,OAAOC,EAAQF,EAAUC,CAAI,CAC/B,EACA,MAAM,QAAQA,EAAM,CAClB,IAAME,EAAUF,EAAK,SAAW,CAAC,EAC3B,CAAE,YAAAG,EAAa,KAAAC,CAAK,EAAI,MAAMH,EAAQF,EAAU,CACpD,GAAGC,EAEH,QAASE,EAAQ,OAAS,EAAI,CAAC,GAAGA,EAAS,CAAE,QAASG,EAAY,OAAO,OAAQ,CAAC,EAAI,CAAC,CACzF,CAAC,EAEKC,EAASF,EAAK,OAAOG,CAAsB,EAAE,IAAIC,CAAU,EAE3DC,EAAcZ,EAAQO,EAAOM,GAAQ,GAAGC,EAAWD,EAAI,OAAO,KAAKA,EAAI,KAAK,SAAS,EAErFE,EAAwCN,EAAO,IAAKO,GAAU,CAElE,IAAMC,GADYL,EAAY,IAAI,GAAGE,EAAWE,EAAM,OAAO,KAAKA,EAAM,SAAS,GAAK,CAAC,GAC7D,IAAKH,IAAS,CACtC,IAAKK,EAAUF,EAAM,UAAWH,EAAI,KAAK,QAAQ,EACjD,MAAOM,EAAgBH,EAAM,YAAaH,EAAI,IAAI,CACpD,EAAE,EAEF,MAAO,CACL,GAAGG,EACH,QAAAC,CACF,CACF,CAAC,EAED,OAAAG,EAAM,4CAA6Cb,EAAK,OAAQE,EAAO,MAAM,EAEtE,CACL,YAAAH,EACA,OAAQS,CACV,CACF,CACF,CAEF,CD/CA,OAAS,WAAAM,MAAe,0BACxB,OAAOC,MAAc,WAGrB,IAAMC,EAAMC,EACVC,EAAE,aACAC,EACAD,EAAE,OAAO,CACP,aAAcA,EAAE,OAAO,CACzB,CAAC,CACH,CACF,EAEME,EAAWC,EAAQC,EAASN,EAAI,YAAY,CAAC,EAG7CO,EAASC,EAAQ,CACrB,eAAgB,IAChB,OAAQ,EACV,CAAC,EAED,MAAMD,EAAO,SAAS,OAAO,mBAAmB,CAAC,EACjD,MAAMA,EAAO,SAAS,OAAO,eAAe,CAAC,EAG7CA,EAAO,IAAI,WAAY,CAACE,EAAKC,IAAQA,EAAI,KAAK,GAAG,EAAE,KAAK,CAAC,EACzDH,EAAO,IAAI,UAAW,CAACE,EAAKC,IAAQA,EAAI,KAAK,GAAG,EAAE,KAAK,CAAC,EAGxDH,EAAO,SAASI,EAA8B,CAC5C,OAAQ,QACR,YAAa,CACX,OAAQC,EAAgB,EACxB,cAAe,UAAa,CAC1B,aAAc,MAAMC,EAAmBT,CAAQ,CACjD,EACF,CACF,CAAC,EAED,MAAMG,EAAO,OAAO,CAAE,KAAMP,EAAI,KAAM,KAAMA,EAAI,IAAK,CAAC,EACtD,QAAQ,IAAI,iDAAiDA,EAAI,QAAQA,EAAI,MAAM","names":["z","fastify","fastifyTRPCPlugin","createAppRouter","getAddress","isTableRegistrationLog","logToTable","storeTables","decodeKey","decodeValueArgs","createDebug","debug","tables","and","asc","eq","or","decodeDynamicField","bigIntMax","getLogs","database","chainId","address","filters","conditions","filter","indexerBlockNumber","rows","records","blockNumber","max","record","logs","groupBy","createQueryAdapter","database","opts","getLogs","filters","blockNumber","logs","storeTables","tables","isTableRegistrationLog","logToTable","logsByTable","log","getAddress","tablesWithRecords","table","records","decodeKey","decodeValueArgs","debug","drizzle","postgres","env","parseEnv","z","frontendEnvSchema","database","drizzle","postgres","server","fastify","req","res","fastifyTRPCPlugin","createAppRouter","createQueryAdapter"]}
@@ -1,3 +1,3 @@
1
1
  #!/usr/bin/env node
2
- import{b as l,c}from"../chunk-2E5MDUA2.js";import"dotenv/config";import{z as n}from"zod";import{eq as T}from"drizzle-orm";import{createPublicClient as b,fallback as u,webSocket as H,http as C}from"viem";import{isDefined as _}from"@latticexyz/common/utils";import{combineLatest as g,filter as E,first as h}from"rxjs";import{drizzle as A}from"drizzle-orm/postgres-js";import L from"postgres";import{createStorageAdapter as R}from"@latticexyz/store-sync/postgres";import{createStoreSync as S}from"@latticexyz/store-sync";var e=c(n.intersection(l,n.object({DATABASE_URL:n.string(),HEALTHCHECK_HOST:n.string().optional(),HEALTHCHECK_PORT:n.coerce.number().optional()}))),k=[e.RPC_WS_URL?H(e.RPC_WS_URL):void 0,e.RPC_HTTP_URL?C(e.RPC_HTTP_URL):void 0].filter(_),i=b({transport:u(k),pollingInterval:e.POLLING_INTERVAL}),P=await i.getChainId(),p=A(L(e.DATABASE_URL)),{storageAdapter:B,tables:m}=await R({database:p,publicClient:i}),a=e.START_BLOCK;try{let t=await p.select().from(m.chainTable).where(T(m.chainTable.chainId,P)).limit(1).execute().then(r=>r.find(()=>!0));t?.lastUpdatedBlockNumber!=null&&(a=t.lastUpdatedBlockNumber+1n,console.log("resuming from block number",a))}catch{}var{latestBlockNumber$:O,storedBlockLogs$:d}=await S({storageAdapter:B,publicClient:i,startBlock:a,maxBlockRange:e.MAX_BLOCK_RANGE,address:e.STORE_ADDRESS});d.subscribe();var f=!1;g([O,d]).pipe(E(([t,{blockNumber:r}])=>t===r),h()).subscribe(()=>{f=!0,console.log("all caught up")});if(e.HEALTHCHECK_HOST!=null||e.HEALTHCHECK_PORT!=null){let{default:t}=await import("fastify"),r=t();r.get("/healthz",(s,o)=>o.code(200).send()),r.get("/readyz",(s,o)=>f?o.code(200).send("ready"):o.code(424).send("backfilling")),r.listen({host:e.HEALTHCHECK_HOST,port:e.HEALTHCHECK_PORT},(s,o)=>{console.log(`postgres indexer healthcheck server listening on ${o}`)})}
2
+ import{b as c,c as d}from"../chunk-2E5MDUA2.js";import"dotenv/config";import{z as a}from"zod";import{eq as u}from"drizzle-orm";import{createPublicClient as T,fallback as g,webSocket as C,http as H}from"viem";import{isDefined as _}from"@latticexyz/common/utils";import{combineLatest as E,filter as h,first as A}from"rxjs";import{drizzle as L}from"drizzle-orm/postgres-js";import R from"postgres";import{cleanDatabase as S,createStorageAdapter as k,shouldCleanDatabase as P}from"@latticexyz/store-sync/postgres";import{createStoreSync as B}from"@latticexyz/store-sync";var e=d(a.intersection(c,a.object({DATABASE_URL:a.string(),HEALTHCHECK_HOST:a.string().optional(),HEALTHCHECK_PORT:a.coerce.number().optional()}))),O=[e.RPC_WS_URL?C(e.RPC_WS_URL):void 0,e.RPC_HTTP_URL?H(e.RPC_HTTP_URL):void 0].filter(_),s=T({transport:g(O),pollingInterval:e.POLLING_INTERVAL}),p=await s.getChainId(),n=L(R(e.DATABASE_URL));await P(n,p)&&(console.log("outdated database detected, clearing data to start fresh"),S(n));var{storageAdapter:U,tables:m}=await k({database:n,publicClient:s}),i=e.START_BLOCK;try{let t=await n.select().from(m.configTable).where(u(m.configTable.chainId,p)).limit(1).execute().then(o=>o.find(()=>!0));t?.lastUpdatedBlockNumber!=null&&(i=t.lastUpdatedBlockNumber+1n,console.log("resuming from block number",i))}catch{}var{latestBlockNumber$:w,storedBlockLogs$:f}=await B({storageAdapter:U,publicClient:s,startBlock:i,maxBlockRange:e.MAX_BLOCK_RANGE,address:e.STORE_ADDRESS});f.subscribe();var b=!1;E([w,f]).pipe(h(([t,{blockNumber:o}])=>t===o),A()).subscribe(()=>{b=!0,console.log("all caught up")});if(e.HEALTHCHECK_HOST!=null||e.HEALTHCHECK_PORT!=null){let{default:t}=await import("fastify"),o=t();o.get("/healthz",(l,r)=>r.code(200).send()),o.get("/readyz",(l,r)=>b?r.code(200).send("ready"):r.code(424).send("backfilling")),o.listen({host:e.HEALTHCHECK_HOST,port:e.HEALTHCHECK_PORT},(l,r)=>{console.log(`postgres indexer healthcheck server listening on ${r}`)})}
3
3
  //# sourceMappingURL=postgres-indexer.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../bin/postgres-indexer.ts"],"sourcesContent":["#!/usr/bin/env node\nimport \"dotenv/config\";\nimport { z } from \"zod\";\nimport { eq } from \"drizzle-orm\";\nimport { createPublicClient, fallback, webSocket, http, Transport } from \"viem\";\nimport { isDefined } from \"@latticexyz/common/utils\";\nimport { combineLatest, filter, first } from \"rxjs\";\nimport { drizzle } from \"drizzle-orm/postgres-js\";\nimport postgres from \"postgres\";\nimport { createStorageAdapter } from \"@latticexyz/store-sync/postgres\";\nimport { createStoreSync } from \"@latticexyz/store-sync\";\nimport { indexerEnvSchema, parseEnv } from \"./parseEnv\";\n\nconst env = parseEnv(\n z.intersection(\n indexerEnvSchema,\n z.object({\n DATABASE_URL: z.string(),\n HEALTHCHECK_HOST: z.string().optional(),\n HEALTHCHECK_PORT: z.coerce.number().optional(),\n })\n )\n);\n\nconst transports: Transport[] = [\n // prefer WS when specified\n env.RPC_WS_URL ? webSocket(env.RPC_WS_URL) : undefined,\n // otherwise use or fallback to HTTP\n env.RPC_HTTP_URL ? http(env.RPC_HTTP_URL) : undefined,\n].filter(isDefined);\n\nconst publicClient = createPublicClient({\n transport: fallback(transports),\n pollingInterval: env.POLLING_INTERVAL,\n});\n\nconst chainId = await publicClient.getChainId();\nconst database = drizzle(postgres(env.DATABASE_URL));\n\nconst { storageAdapter, tables } = await createStorageAdapter({ database, publicClient });\n\nlet startBlock = env.START_BLOCK;\n\n// Resume from latest block stored in DB. This will throw if the DB doesn't exist yet, so we wrap in a try/catch and ignore the error.\n// TODO: query if the DB exists instead of try/catch\ntry {\n const chainState = await database\n .select()\n .from(tables.chainTable)\n .where(eq(tables.chainTable.chainId, chainId))\n .limit(1)\n .execute()\n // Get the first record in a way that returns a possible `undefined`\n // TODO: move this to `.findFirst` after upgrading drizzle or `rows[0]` after enabling `noUncheckedIndexedAccess: true`\n .then((rows) => rows.find(() => true));\n\n if (chainState?.lastUpdatedBlockNumber != null) {\n startBlock = chainState.lastUpdatedBlockNumber + 1n;\n console.log(\"resuming from block number\", startBlock);\n }\n} catch (error) {\n // ignore errors for now\n}\n\nconst { latestBlockNumber$, storedBlockLogs$ } = await createStoreSync({\n storageAdapter,\n publicClient,\n startBlock,\n maxBlockRange: env.MAX_BLOCK_RANGE,\n address: env.STORE_ADDRESS,\n});\n\nstoredBlockLogs$.subscribe();\n\nlet isCaughtUp = false;\ncombineLatest([latestBlockNumber$, storedBlockLogs$])\n .pipe(\n filter(\n ([latestBlockNumber, { blockNumber: lastBlockNumberProcessed }]) => latestBlockNumber === lastBlockNumberProcessed\n ),\n first()\n )\n .subscribe(() => {\n isCaughtUp = true;\n console.log(\"all caught up\");\n });\n\nif (env.HEALTHCHECK_HOST != null || env.HEALTHCHECK_PORT != null) {\n const { default: fastify } = await import(\"fastify\");\n\n const server = fastify();\n\n // k8s healthchecks\n server.get(\"/healthz\", (req, res) => res.code(200).send());\n server.get(\"/readyz\", (req, res) => (isCaughtUp ? res.code(200).send(\"ready\") : res.code(424).send(\"backfilling\")));\n\n server.listen({ host: env.HEALTHCHECK_HOST, port: env.HEALTHCHECK_PORT }, (error, address) => {\n console.log(`postgres indexer healthcheck server listening on ${address}`);\n });\n}\n"],"mappings":";2CACA,MAAO,gBACP,OAAS,KAAAA,MAAS,MAClB,OAAS,MAAAC,MAAU,cACnB,OAAS,sBAAAC,EAAoB,YAAAC,EAAU,aAAAC,EAAW,QAAAC,MAAuB,OACzE,OAAS,aAAAC,MAAiB,2BAC1B,OAAS,iBAAAC,EAAe,UAAAC,EAAQ,SAAAC,MAAa,OAC7C,OAAS,WAAAC,MAAe,0BACxB,OAAOC,MAAc,WACrB,OAAS,wBAAAC,MAA4B,kCACrC,OAAS,mBAAAC,MAAuB,yBAGhC,IAAMC,EAAMC,EACVC,EAAE,aACAC,EACAD,EAAE,OAAO,CACP,aAAcA,EAAE,OAAO,EACvB,iBAAkBA,EAAE,OAAO,EAAE,SAAS,EACtC,iBAAkBA,EAAE,OAAO,OAAO,EAAE,SAAS,CAC/C,CAAC,CACH,CACF,EAEME,EAA0B,CAE9BJ,EAAI,WAAaK,EAAUL,EAAI,UAAU,EAAI,OAE7CA,EAAI,aAAeM,EAAKN,EAAI,YAAY,EAAI,MAC9C,EAAE,OAAOO,CAAS,EAEZC,EAAeC,EAAmB,CACtC,UAAWC,EAASN,CAAU,EAC9B,gBAAiBJ,EAAI,gBACvB,CAAC,EAEKW,EAAU,MAAMH,EAAa,WAAW,EACxCI,EAAWC,EAAQC,EAASd,EAAI,YAAY,CAAC,EAE7C,CAAE,eAAAe,EAAgB,OAAAC,CAAO,EAAI,MAAMC,EAAqB,CAAE,SAAAL,EAAU,aAAAJ,CAAa,CAAC,EAEpFU,EAAalB,EAAI,YAIrB,GAAI,CACF,IAAMmB,EAAa,MAAMP,EACtB,OAAO,EACP,KAAKI,EAAO,UAAU,EACtB,MAAMI,EAAGJ,EAAO,WAAW,QAASL,CAAO,CAAC,EAC5C,MAAM,CAAC,EACP,QAAQ,EAGR,KAAMU,GAASA,EAAK,KAAK,IAAM,EAAI,CAAC,EAEnCF,GAAY,wBAA0B,OACxCD,EAAaC,EAAW,uBAAyB,GACjD,QAAQ,IAAI,6BAA8BD,CAAU,EAExD,MAAE,CAEF,CAEA,GAAM,CAAE,mBAAAI,EAAoB,iBAAAC,CAAiB,EAAI,MAAMC,EAAgB,CACrE,eAAAT,EACA,aAAAP,EACA,WAAAU,EACA,cAAelB,EAAI,gBACnB,QAASA,EAAI,aACf,CAAC,EAEDuB,EAAiB,UAAU,EAE3B,IAAIE,EAAa,GACjBC,EAAc,CAACJ,EAAoBC,CAAgB,CAAC,EACjD,KACCI,EACE,CAAC,CAACC,EAAmB,CAAE,YAAaC,CAAyB,CAAC,IAAMD,IAAsBC,CAC5F,EACAC,EAAM,CACR,EACC,UAAU,IAAM,CACfL,EAAa,GACb,QAAQ,IAAI,eAAe,CAC7B,CAAC,EAEH,GAAIzB,EAAI,kBAAoB,MAAQA,EAAI,kBAAoB,KAAM,CAChE,GAAM,CAAE,QAAS+B,CAAQ,EAAI,KAAM,QAAO,SAAS,EAE7CC,EAASD,EAAQ,EAGvBC,EAAO,IAAI,WAAY,CAACC,EAAKC,IAAQA,EAAI,KAAK,GAAG,EAAE,KAAK,CAAC,EACzDF,EAAO,IAAI,UAAW,CAACC,EAAKC,IAAST,EAAaS,EAAI,KAAK,GAAG,EAAE,KAAK,OAAO,EAAIA,EAAI,KAAK,GAAG,EAAE,KAAK,aAAa,CAAE,EAElHF,EAAO,OAAO,CAAE,KAAMhC,EAAI,iBAAkB,KAAMA,EAAI,gBAAiB,EAAG,CAACmC,EAAOC,IAAY,CAC5F,QAAQ,IAAI,oDAAoDA,GAAS,CAC3E,CAAC","names":["z","eq","createPublicClient","fallback","webSocket","http","isDefined","combineLatest","filter","first","drizzle","postgres","createStorageAdapter","createStoreSync","env","parseEnv","z","indexerEnvSchema","transports","webSocket","http","isDefined","publicClient","createPublicClient","fallback","chainId","database","drizzle","postgres","storageAdapter","tables","createStorageAdapter","startBlock","chainState","eq","rows","latestBlockNumber$","storedBlockLogs$","createStoreSync","isCaughtUp","combineLatest","filter","latestBlockNumber","lastBlockNumberProcessed","first","fastify","server","req","res","error","address"]}
1
+ {"version":3,"sources":["../../bin/postgres-indexer.ts"],"sourcesContent":["#!/usr/bin/env node\nimport \"dotenv/config\";\nimport { z } from \"zod\";\nimport { eq } from \"drizzle-orm\";\nimport { createPublicClient, fallback, webSocket, http, Transport } from \"viem\";\nimport { isDefined } from \"@latticexyz/common/utils\";\nimport { combineLatest, filter, first } from \"rxjs\";\nimport { drizzle } from \"drizzle-orm/postgres-js\";\nimport postgres from \"postgres\";\nimport { cleanDatabase, createStorageAdapter, shouldCleanDatabase } from \"@latticexyz/store-sync/postgres\";\nimport { createStoreSync } from \"@latticexyz/store-sync\";\nimport { indexerEnvSchema, parseEnv } from \"./parseEnv\";\n\nconst env = parseEnv(\n z.intersection(\n indexerEnvSchema,\n z.object({\n DATABASE_URL: z.string(),\n HEALTHCHECK_HOST: z.string().optional(),\n HEALTHCHECK_PORT: z.coerce.number().optional(),\n })\n )\n);\n\nconst transports: Transport[] = [\n // prefer WS when specified\n env.RPC_WS_URL ? webSocket(env.RPC_WS_URL) : undefined,\n // otherwise use or fallback to HTTP\n env.RPC_HTTP_URL ? http(env.RPC_HTTP_URL) : undefined,\n].filter(isDefined);\n\nconst publicClient = createPublicClient({\n transport: fallback(transports),\n pollingInterval: env.POLLING_INTERVAL,\n});\n\nconst chainId = await publicClient.getChainId();\nconst database = drizzle(postgres(env.DATABASE_URL));\n\nif (await shouldCleanDatabase(database, chainId)) {\n console.log(\"outdated database detected, clearing data to start fresh\");\n cleanDatabase(database);\n}\n\nconst { storageAdapter, tables } = await createStorageAdapter({ database, publicClient });\n\nlet startBlock = env.START_BLOCK;\n\n// Resume from latest block stored in DB. This will throw if the DB doesn't exist yet, so we wrap in a try/catch and ignore the error.\n// TODO: query if the DB exists instead of try/catch\ntry {\n const chainState = await database\n .select()\n .from(tables.configTable)\n .where(eq(tables.configTable.chainId, chainId))\n .limit(1)\n .execute()\n // Get the first record in a way that returns a possible `undefined`\n // TODO: move this to `.findFirst` after upgrading drizzle or `rows[0]` after enabling `noUncheckedIndexedAccess: true`\n .then((rows) => rows.find(() => true));\n\n if (chainState?.lastUpdatedBlockNumber != null) {\n startBlock = chainState.lastUpdatedBlockNumber + 1n;\n console.log(\"resuming from block number\", startBlock);\n }\n} catch (error) {\n // ignore errors for now\n}\n\nconst { latestBlockNumber$, storedBlockLogs$ } = await createStoreSync({\n storageAdapter,\n publicClient,\n startBlock,\n maxBlockRange: env.MAX_BLOCK_RANGE,\n address: env.STORE_ADDRESS,\n});\n\nstoredBlockLogs$.subscribe();\n\nlet isCaughtUp = false;\ncombineLatest([latestBlockNumber$, storedBlockLogs$])\n .pipe(\n filter(\n ([latestBlockNumber, { blockNumber: lastBlockNumberProcessed }]) => latestBlockNumber === lastBlockNumberProcessed\n ),\n first()\n )\n .subscribe(() => {\n isCaughtUp = true;\n console.log(\"all caught up\");\n });\n\nif (env.HEALTHCHECK_HOST != null || env.HEALTHCHECK_PORT != null) {\n const { default: fastify } = await import(\"fastify\");\n\n const server = fastify();\n\n // k8s healthchecks\n server.get(\"/healthz\", (req, res) => res.code(200).send());\n server.get(\"/readyz\", (req, res) => (isCaughtUp ? res.code(200).send(\"ready\") : res.code(424).send(\"backfilling\")));\n\n server.listen({ host: env.HEALTHCHECK_HOST, port: env.HEALTHCHECK_PORT }, (error, address) => {\n console.log(`postgres indexer healthcheck server listening on ${address}`);\n });\n}\n"],"mappings":";gDACA,MAAO,gBACP,OAAS,KAAAA,MAAS,MAClB,OAAS,MAAAC,MAAU,cACnB,OAAS,sBAAAC,EAAoB,YAAAC,EAAU,aAAAC,EAAW,QAAAC,MAAuB,OACzE,OAAS,aAAAC,MAAiB,2BAC1B,OAAS,iBAAAC,EAAe,UAAAC,EAAQ,SAAAC,MAAa,OAC7C,OAAS,WAAAC,MAAe,0BACxB,OAAOC,MAAc,WACrB,OAAS,iBAAAC,EAAe,wBAAAC,EAAsB,uBAAAC,MAA2B,kCACzE,OAAS,mBAAAC,MAAuB,yBAGhC,IAAMC,EAAMC,EACVC,EAAE,aACAC,EACAD,EAAE,OAAO,CACP,aAAcA,EAAE,OAAO,EACvB,iBAAkBA,EAAE,OAAO,EAAE,SAAS,EACtC,iBAAkBA,EAAE,OAAO,OAAO,EAAE,SAAS,CAC/C,CAAC,CACH,CACF,EAEME,EAA0B,CAE9BJ,EAAI,WAAaK,EAAUL,EAAI,UAAU,EAAI,OAE7CA,EAAI,aAAeM,EAAKN,EAAI,YAAY,EAAI,MAC9C,EAAE,OAAOO,CAAS,EAEZC,EAAeC,EAAmB,CACtC,UAAWC,EAASN,CAAU,EAC9B,gBAAiBJ,EAAI,gBACvB,CAAC,EAEKW,EAAU,MAAMH,EAAa,WAAW,EACxCI,EAAWC,EAAQC,EAASd,EAAI,YAAY,CAAC,EAE/C,MAAMe,EAAoBH,EAAUD,CAAO,IAC7C,QAAQ,IAAI,0DAA0D,EACtEK,EAAcJ,CAAQ,GAGxB,GAAM,CAAE,eAAAK,EAAgB,OAAAC,CAAO,EAAI,MAAMC,EAAqB,CAAE,SAAAP,EAAU,aAAAJ,CAAa,CAAC,EAEpFY,EAAapB,EAAI,YAIrB,GAAI,CACF,IAAMqB,EAAa,MAAMT,EACtB,OAAO,EACP,KAAKM,EAAO,WAAW,EACvB,MAAMI,EAAGJ,EAAO,YAAY,QAASP,CAAO,CAAC,EAC7C,MAAM,CAAC,EACP,QAAQ,EAGR,KAAMY,GAASA,EAAK,KAAK,IAAM,EAAI,CAAC,EAEnCF,GAAY,wBAA0B,OACxCD,EAAaC,EAAW,uBAAyB,GACjD,QAAQ,IAAI,6BAA8BD,CAAU,EAExD,MAAE,CAEF,CAEA,GAAM,CAAE,mBAAAI,EAAoB,iBAAAC,CAAiB,EAAI,MAAMC,EAAgB,CACrE,eAAAT,EACA,aAAAT,EACA,WAAAY,EACA,cAAepB,EAAI,gBACnB,QAASA,EAAI,aACf,CAAC,EAEDyB,EAAiB,UAAU,EAE3B,IAAIE,EAAa,GACjBC,EAAc,CAACJ,EAAoBC,CAAgB,CAAC,EACjD,KACCI,EACE,CAAC,CAACC,EAAmB,CAAE,YAAaC,CAAyB,CAAC,IAAMD,IAAsBC,CAC5F,EACAC,EAAM,CACR,EACC,UAAU,IAAM,CACfL,EAAa,GACb,QAAQ,IAAI,eAAe,CAC7B,CAAC,EAEH,GAAI3B,EAAI,kBAAoB,MAAQA,EAAI,kBAAoB,KAAM,CAChE,GAAM,CAAE,QAASiC,CAAQ,EAAI,KAAM,QAAO,SAAS,EAE7CC,EAASD,EAAQ,EAGvBC,EAAO,IAAI,WAAY,CAACC,EAAKC,IAAQA,EAAI,KAAK,GAAG,EAAE,KAAK,CAAC,EACzDF,EAAO,IAAI,UAAW,CAACC,EAAKC,IAAST,EAAaS,EAAI,KAAK,GAAG,EAAE,KAAK,OAAO,EAAIA,EAAI,KAAK,GAAG,EAAE,KAAK,aAAa,CAAE,EAElHF,EAAO,OAAO,CAAE,KAAMlC,EAAI,iBAAkB,KAAMA,EAAI,gBAAiB,EAAG,CAACqC,EAAOC,IAAY,CAC5F,QAAQ,IAAI,oDAAoDA,GAAS,CAC3E,CAAC","names":["z","eq","createPublicClient","fallback","webSocket","http","isDefined","combineLatest","filter","first","drizzle","postgres","cleanDatabase","createStorageAdapter","shouldCleanDatabase","createStoreSync","env","parseEnv","z","indexerEnvSchema","transports","webSocket","http","isDefined","publicClient","createPublicClient","fallback","chainId","database","drizzle","postgres","shouldCleanDatabase","cleanDatabase","storageAdapter","tables","createStorageAdapter","startBlock","chainState","eq","rows","latestBlockNumber$","storedBlockLogs$","createStoreSync","isCaughtUp","combineLatest","filter","latestBlockNumber","lastBlockNumberProcessed","first","fastify","server","req","res","error","address"]}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@latticexyz/store-indexer",
3
- "version": "2.0.0-main-5df1f31b",
3
+ "version": "2.0.0-main-e48fb3b0",
4
4
  "description": "Minimal Typescript indexer for Store",
5
5
  "repository": {
6
6
  "type": "git",
@@ -33,11 +33,11 @@
33
33
  "superjson": "^1.12.4",
34
34
  "viem": "1.14.0",
35
35
  "zod": "^3.21.4",
36
- "@latticexyz/block-logs-stream": "2.0.0-main-5df1f31b",
37
- "@latticexyz/common": "2.0.0-main-5df1f31b",
38
- "@latticexyz/protocol-parser": "2.0.0-main-5df1f31b",
39
- "@latticexyz/store": "2.0.0-main-5df1f31b",
40
- "@latticexyz/store-sync": "2.0.0-main-5df1f31b"
36
+ "@latticexyz/block-logs-stream": "2.0.0-main-e48fb3b0",
37
+ "@latticexyz/common": "2.0.0-main-e48fb3b0",
38
+ "@latticexyz/protocol-parser": "2.0.0-main-e48fb3b0",
39
+ "@latticexyz/store": "2.0.0-main-e48fb3b0",
40
+ "@latticexyz/store-sync": "2.0.0-main-e48fb3b0"
41
41
  },
42
42
  "devDependencies": {
43
43
  "@types/better-sqlite3": "^7.6.4",
@@ -41,8 +41,8 @@ export async function getLogs(
41
41
  // TODO: move the block number query into the records query for atomicity so we don't have to merge them here
42
42
  const chainState = await database
43
43
  .select()
44
- .from(tables.chainTable)
45
- .where(eq(tables.chainTable.chainId, chainId))
44
+ .from(tables.configTable)
45
+ .where(eq(tables.configTable.chainId, chainId))
46
46
  .limit(1)
47
47
  .execute()
48
48
  // Get the first record in a way that returns a possible `undefined`