@lpdjs/firestore-repo-service 2.1.16 → 2.1.17

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,9 +1,9 @@
1
- 'use strict';var u=class{constructor(){this.name="bigquery";}mapType(t){switch(t){case "string":return "STRING";case "number":return "FLOAT64";case "bigint":return "INT64";case "boolean":return "BOOL";case "timestamp":return "TIMESTAMP";case "json":return "JSON";case "text":return "STRING"}}quoteIdentifier(t){return `\`${t}\``}},p=new u,l=class l{constructor(t){this.bigquery=t.bigquery,this.datasetId=t.datasetId;}get dialect(){return p}async tableExists(t){let[e]=await this.dataset.table(t).exists();return e}async getTableColumns(t){let[e]=await this.dataset.table(t).getMetadata();return (e.schema?.fields??[]).map(s=>s.name)}async createTable(t){let e=r=>this.dialect.quoteIdentifier(r),i=t.columns.map(r=>{let a=r.isPrimaryKey?" NOT NULL":"";return ` ${e(r.name)} ${r.sqlType}${a}`}).join(`,
1
+ 'use strict';var o="__sync_version";var d=class{constructor(){this.name="bigquery";}mapType(t){switch(t){case "string":return "STRING";case "number":return "FLOAT64";case "bigint":return "INT64";case "boolean":return "BOOL";case "timestamp":return "TIMESTAMP";case "json":return "JSON";case "text":return "STRING"}}quoteIdentifier(t){return `\`${t}\``}},f=new d,c=class c{constructor(t){this.bigquery=t.bigquery,this.datasetId=t.datasetId;}get dialect(){return f}async tableExists(t){let[e]=await this.dataset.table(t).exists();return e}async getTableColumns(t){let[e]=await this.dataset.table(t).getMetadata();return (e.schema?.fields??[]).map(s=>s.name)}async createTable(t){let e=a=>this.dialect.quoteIdentifier(a),r=t.columns.map(a=>{let n=a.isPrimaryKey?" NOT NULL":"";return ` ${e(a.name)} ${a.sqlType}${n}`}).join(`,
2
2
  `),s=`CREATE TABLE IF NOT EXISTS ${this.fqn(t.tableName)} (
3
- ${i}
4
- );`;await this.bigquery.query({query:s});}async addColumns(t,e){let i=s=>this.dialect.quoteIdentifier(s);for(let s of e){let r=`ALTER TABLE ${this.fqn(t)} ADD COLUMN ${i(s.name)} ${s.sqlType};`;await this.bigquery.query({query:r});}}async insertRows(t,e){e.length!==0&&await this.dataset.table(t).insert(e);}async upsertRows(t,e,i){if(e.length===0)return;let s=Object.keys(e[0]),r=s.filter(n=>n!==i),a=n=>this.dialect.quoteIdentifier(n),g=e.map((n,m)=>`SELECT ${s.map(c=>m===0?`${this.escapeValue(n[c])} AS ${a(c)}`:this.escapeValue(n[c])).join(", ")}`).join(` UNION ALL
5
- `),q=r.map(n=>`T.${a(n)} = S.${a(n)}`).join(", "),y=s.map(n=>a(n)).join(", "),T=s.map(n=>`S.${a(n)}`).join(", "),S=[`MERGE ${this.fqn(t)} AS T`,`USING (
6
- ${g}
7
- ) AS S`,`ON T.${a(i)} = S.${a(i)}`,`WHEN MATCHED THEN UPDATE SET ${q}`,`WHEN NOT MATCHED THEN INSERT (${y}) VALUES (${T});`].join(`
8
- `);await this.bigquery.query({query:S});}async deleteRows(t,e,i){if(i.length===0)return;let s=o=>this.dialect.quoteIdentifier(o),r=i.map(o=>this.escapeValue(o)).join(", "),a=`DELETE FROM ${this.fqn(t)} WHERE ${s(e)} IN (${r});`;await this.bigquery.query({query:a});}async executeRaw(t){await this.bigquery.query({query:t});}get dataset(){return this.bigquery.dataset(this.datasetId)}fqn(t){return `\`${this.datasetId}.${t}\``}escapeValue(t){return t==null?"NULL":typeof t=="boolean"?t?"TRUE":"FALSE":typeof t=="number"||typeof t=="bigint"?String(t):typeof t=="string"?l.ISO_TIMESTAMP_RE.test(t)?`TIMESTAMP('${t}')`:t.startsWith("[")&&t.endsWith("]")||t.startsWith("{")&&t.endsWith("}")?`PARSE_JSON('${t.replace(/'/g,"\\'")}')`:`'${t.replace(/'/g,"\\'")}'`:`PARSE_JSON('${JSON.stringify(t).replace(/'/g,"\\'")}')`}};l.ISO_TIMESTAMP_RE=/^\d{4}-\d{2}-\d{2}T\d{2}:\d{2}:\d{2}/;var d=l;exports.BigQueryAdapter=d;exports.bigqueryDialect=p;//# sourceMappingURL=bigquery.cjs.map
3
+ ${r}
4
+ );`;await this.bigquery.query({query:s});}async addColumns(t,e){let r=s=>this.dialect.quoteIdentifier(s);for(let s of e){let a=`ALTER TABLE ${this.fqn(t)} ADD COLUMN ${r(s.name)} ${s.sqlType};`;await this.bigquery.query({query:a});}}async insertRows(t,e){e.length!==0&&await this.dataset.table(t).insert(e);}async upsertRows(t,e,r){if(e.length===0)return;let s=Object.keys(e[0]),a=s.filter(i=>i!==r),n=i=>this.dialect.quoteIdentifier(i),q=e.map((i,E)=>`SELECT ${s.map(u=>E===0?`${this.escapeValue(i[u])} AS ${n(u)}`:this.escapeValue(i[u])).join(", ")}`).join(` UNION ALL
5
+ `),y=a.map(i=>`T.${n(i)} = S.${n(i)}`).join(", "),S=s.map(i=>n(i)).join(", "),T=s.map(i=>`S.${n(i)}`).join(", "),m=s.includes(o)?` AND (T.${n(o)} IS NULL OR S.${n(o)} > T.${n(o)})`:"",p=[`MERGE ${this.fqn(t)} AS T`,`USING (
6
+ ${q}
7
+ ) AS S`,`ON T.${n(r)} = S.${n(r)}`,`WHEN MATCHED${m} THEN UPDATE SET ${y}`,`WHEN NOT MATCHED THEN INSERT (${S}) VALUES (${T});`].join(`
8
+ `);await this.bigquery.query({query:p});}async deleteRows(t,e,r){if(r.length===0)return;let s=l=>this.dialect.quoteIdentifier(l),a=r.map(l=>this.escapeValue(l)).join(", "),n=`DELETE FROM ${this.fqn(t)} WHERE ${s(e)} IN (${a});`;await this.bigquery.query({query:n});}async executeRaw(t){await this.bigquery.query({query:t});}get dataset(){return this.bigquery.dataset(this.datasetId)}fqn(t){return `\`${this.datasetId}.${t}\``}escapeValue(t){return t==null?"NULL":typeof t=="boolean"?t?"TRUE":"FALSE":typeof t=="number"||typeof t=="bigint"?String(t):typeof t=="string"?c.ISO_TIMESTAMP_RE.test(t)?`TIMESTAMP('${t}')`:t.startsWith("[")&&t.endsWith("]")||t.startsWith("{")&&t.endsWith("}")?`PARSE_JSON('${t.replace(/'/g,"\\'")}')`:`'${t.replace(/'/g,"\\'")}'`:`PARSE_JSON('${JSON.stringify(t).replace(/'/g,"\\'")}')`}};c.ISO_TIMESTAMP_RE=/^\d{4}-\d{2}-\d{2}T\d{2}:\d{2}:\d{2}/;var g=c;exports.BigQueryAdapter=g;exports.bigqueryDialect=f;//# sourceMappingURL=bigquery.cjs.map
9
9
  //# sourceMappingURL=bigquery.cjs.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/sync/adapters/bigquery.ts"],"names":["BigQueryDialect","logical","id","bigqueryDialect","_BigQueryAdapter","options","tableName","exists","metadata","f","table","qi","cols","c","notNull","ddl","columns","stmt","rows","primaryKey","allKeys","nonPkCols","k","source","row","i","updateSet","insertCols","insertVals","query","ids","escaped","v","sql","BigQueryAdapter"],"mappings":"aAaA,IAAMA,EAAN,KAA4C,CAA5C,WAAA,EAAA,CACE,IAAA,CAAS,KAAO,WAAA,CAEhB,OAAA,CAAQC,CAAAA,CAA8B,CACpC,OAAQA,CAAAA,EACN,KAAK,QAAA,CACH,OAAO,QAAA,CACT,KAAK,QAAA,CACH,OAAO,UACT,KAAK,QAAA,CACH,OAAO,OAAA,CACT,KAAK,SAAA,CACH,OAAO,MAAA,CACT,KAAK,YACH,OAAO,WAAA,CACT,KAAK,MAAA,CACH,OAAO,OACT,KAAK,MAAA,CACH,OAAO,QACX,CACF,CAEA,eAAA,CAAgBC,CAAAA,CAAoB,CAClC,OAAO,CAAA,EAAA,EAAKA,CAAE,CAAA,EAAA,CAChB,CACF,EAGaC,CAAAA,CAA8B,IAAIH,EAuBlCI,CAAAA,CAAN,MAAMA,CAAsC,CAIjD,WAAA,CAAYC,CAAAA,CAA+C,CACzD,KAAK,QAAA,CAAWA,CAAAA,CAAQ,QAAA,CACxB,IAAA,CAAK,UAAYA,CAAAA,CAAQ,UAC3B,CAGA,IAAI,SAAsB,CACxB,OAAOF,CACT,CAGA,MAAM,YAAYG,CAAAA,CAAqC,CACrD,GAAM,CAACC,CAAM,CAAA,CAAI,MAAM,IAAA,CAAK,OAAA,CAAQ,MAAMD,CAAS,CAAA,CAAE,MAAA,EAAO,CAC5D,OAAOC,CACT,CAGA,MAAM,eAAA,CAAgBD,EAAsC,CAC1D,GAAM,CAACE,CAAQ,EAAI,MAAM,IAAA,CAAK,OAAA,CAAQ,KAAA,CAAMF,CAAS,CAAA,CAAE,WAAA,EAAY,CAEnE,OAAA,CADwCE,EAAS,MAAA,EAAQ,MAAA,EAAU,EAAC,EACtD,GAAA,CAAKC,GAAMA,CAAAA,CAAE,IAAI,CACjC,CAGA,MAAM,WAAA,CAAYC,CAAAA,CAAmC,CACnD,IAAMC,EAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,eAAA,CAAgBA,CAAE,CAAA,CACpDU,CAAAA,CAAOF,EAAM,OAAA,CAChB,GAAA,CAAKG,GAAM,CACV,IAAMC,CAAAA,CAAUD,CAAAA,CAAE,aAAe,WAAA,CAAc,EAAA,CAC/C,OAAO,CAAA,EAAA,EAAKF,EAAGE,CAAAA,CAAE,IAAI,CAAC,CAAA,CAAA,EAAIA,EAAE,OAAO,CAAA,EAAGC,CAAO,CAAA,CAC/C,CAAC,EACA,IAAA,CAAK,CAAA;AAAA,CAAK,EAEPC,CAAAA,CAAM,CAAA,2BAAA,EAA8B,KAAK,GAAA,CAAIL,CAAAA,CAAM,SAAS,CAAC,CAAA;AAAA,EAAOE,CAAI;AAAA,EAAA,CAAA,CAC9E,MAAM,IAAA,CAAK,QAAA,CAAS,KAAA,CAAM,CAAE,KAAA,CAAOG,CAAI,CAAC,EAC1C,CAGA,MAAM,UAAA,CAAWT,CAAAA,CAAmBU,EAAqC,CACvE,IAAML,CAAAA,CAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,eAAA,CAAgBA,CAAE,EAC1D,IAAA,IAAWW,CAAAA,IAAKG,CAAAA,CAAS,CACvB,IAAMC,CAAAA,CAAO,CAAA,YAAA,EAAe,IAAA,CAAK,IAAIX,CAAS,CAAC,CAAA,YAAA,EAAeK,CAAAA,CAAGE,CAAAA,CAAE,IAAI,CAAC,CAAA,CAAA,EAAIA,EAAE,OAAO,CAAA,CAAA,CAAA,CACrF,MAAM,IAAA,CAAK,QAAA,CAAS,KAAA,CAAM,CAAE,KAAA,CAAOI,CAAK,CAAC,EAC3C,CACF,CAGA,MAAM,UAAA,CACJX,CAAAA,CACAY,CAAAA,CACe,CACXA,CAAAA,CAAK,MAAA,GAAW,CAAA,EACpB,MAAM,IAAA,CAAK,OAAA,CAAQ,KAAA,CAAMZ,CAAS,EAAE,MAAA,CAAOY,CAAI,EACjD,CAQA,MAAM,UAAA,CACJZ,CAAAA,CACAY,CAAAA,CACAC,EACe,CACf,GAAID,CAAAA,CAAK,MAAA,GAAW,CAAA,CAAG,OAEvB,IAAME,CAAAA,CAAU,OAAO,IAAA,CAAKF,CAAAA,CAAK,CAAC,CAAE,CAAA,CAC9BG,CAAAA,CAAYD,CAAAA,CAAQ,MAAA,CAAQE,GAAMA,CAAAA,GAAMH,CAAU,CAAA,CAClDR,CAAAA,CAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,eAAA,CAAgBA,CAAE,CAAA,CAgBpDqB,CAAAA,CAbUL,CAAAA,CAAK,GAAA,CAAI,CAACM,CAAAA,CAAKC,CAAAA,GAUtB,CAAA,OAAA,EATQL,EACZ,GAAA,CAAKE,CAAAA,EAEFG,CAAAA,GAAM,CAAA,CACF,CAAA,EAAG,IAAA,CAAK,WAAA,CAAYD,CAAAA,CAAIF,CAAC,CAAC,CAAC,CAAA,IAAA,EAAOX,CAAAA,CAAGW,CAAC,CAAC,CAAA,CAAA,CACvC,IAAA,CAAK,YAAYE,CAAAA,CAAIF,CAAC,CAAC,CAE9B,EACA,IAAA,CAAK,IAAI,CACW,CAAA,CACxB,EAEsB,IAAA,CAAK,CAAA;AAAA,IAAA,CAAkB,CAAA,CAGxCI,EAAYL,CAAAA,CACf,GAAA,CAAKR,GAAM,CAAA,EAAA,EAAKF,CAAAA,CAAGE,CAAC,CAAC,CAAA,KAAA,EAAQF,CAAAA,CAAGE,CAAC,CAAC,CAAA,CAAE,EACpC,IAAA,CAAK,IAAI,EAGNc,CAAAA,CAAaP,CAAAA,CAAQ,GAAA,CAAKP,CAAAA,EAAMF,CAAAA,CAAGE,CAAC,CAAC,CAAA,CAAE,IAAA,CAAK,IAAI,CAAA,CAChDe,CAAAA,CAAaR,EAAQ,GAAA,CAAKP,CAAAA,EAAM,CAAA,EAAA,EAAKF,CAAAA,CAAGE,CAAC,CAAC,EAAE,CAAA,CAAE,IAAA,CAAK,IAAI,CAAA,CAEvDgB,CAAAA,CAAQ,CACZ,SAAS,IAAA,CAAK,GAAA,CAAIvB,CAAS,CAAC,CAAA,KAAA,CAAA,CAC5B,CAAA;AAAA,IAAA,EAAgBiB,CAAM;AAAA,QAAA,CAAA,CACtB,QAAQZ,CAAAA,CAAGQ,CAAU,CAAC,CAAA,KAAA,EAAQR,CAAAA,CAAGQ,CAAU,CAAC,CAAA,CAAA,CAC5C,CAAA,6BAAA,EAAgCO,CAAS,GACzC,CAAA,8BAAA,EAAiCC,CAAU,aAAaC,CAAU,CAAA,EAAA,CACpE,EAAE,IAAA,CAAK;AAAA,CAAI,EAEX,MAAM,IAAA,CAAK,SAAS,KAAA,CAAM,CAAE,MAAAC,CAAM,CAAC,EACrC,CAGA,MAAM,WACJvB,CAAAA,CACAa,CAAAA,CACAW,EACe,CACf,GAAIA,EAAI,MAAA,GAAW,CAAA,CAAG,OAEtB,IAAMnB,EAAMT,CAAAA,EAAe,IAAA,CAAK,QAAQ,eAAA,CAAgBA,CAAE,EACpD6B,CAAAA,CAAUD,CAAAA,CAAI,IAAKE,CAAAA,EAAM,IAAA,CAAK,YAAYA,CAAC,CAAC,EAAE,IAAA,CAAK,IAAI,EACvDH,CAAAA,CAAQ,CAAA,YAAA,EAAe,IAAA,CAAK,GAAA,CAAIvB,CAAS,CAAC,CAAA,OAAA,EAAUK,EAAGQ,CAAU,CAAC,QAAQY,CAAO,CAAA,EAAA,CAAA,CAEvF,MAAM,IAAA,CAAK,QAAA,CAAS,MAAM,CAAE,KAAA,CAAAF,CAAM,CAAC,EACrC,CAGA,MAAM,UAAA,CAAWI,CAAAA,CAA4B,CAC3C,MAAM,IAAA,CAAK,QAAA,CAAS,MAAM,CAAE,KAAA,CAAOA,CAAI,CAAC,EAC1C,CAOA,IAAY,OAAA,EAAU,CACpB,OAAO,IAAA,CAAK,SAAS,OAAA,CAAQ,IAAA,CAAK,SAAS,CAC7C,CAGQ,GAAA,CAAI3B,CAAAA,CAA2B,CACrC,OAAO,CAAA,EAAA,EAAK,KAAK,SAAS,CAAA,CAAA,EAAIA,CAAS,CAAA,EAAA,CACzC,CAOQ,YAAY0B,CAAAA,CAAoB,CACtC,OAAIA,CAAAA,EAAM,IAAA,CAAgC,OACtC,OAAOA,CAAAA,EAAM,UAAkBA,CAAAA,CAAI,MAAA,CAAS,OAAA,CAC5C,OAAOA,GAAM,QAAA,EAAY,OAAOA,GAAM,QAAA,CAAiB,MAAA,CAAOA,CAAC,CAAA,CAC/D,OAAOA,GAAM,QAAA,CAEX5B,CAAAA,CAAgB,iBAAiB,IAAA,CAAK4B,CAAC,EAClC,CAAA,WAAA,EAAcA,CAAC,KAIrBA,CAAAA,CAAE,UAAA,CAAW,GAAG,CAAA,EAAKA,EAAE,QAAA,CAAS,GAAG,GACnCA,CAAAA,CAAE,UAAA,CAAW,GAAG,CAAA,EAAKA,CAAAA,CAAE,SAAS,GAAG,CAAA,CAE7B,eAAeA,CAAAA,CAAE,OAAA,CAAQ,KAAM,KAAK,CAAC,KAEvC,CAAA,CAAA,EAAIA,CAAAA,CAAE,OAAA,CAAQ,IAAA,CAAM,KAAK,CAAC,CAAA,CAAA,CAAA,CAG5B,eAAe,IAAA,CAAK,SAAA,CAAUA,CAAC,CAAA,CAAE,OAAA,CAAQ,KAAM,KAAK,CAAC,IAC9D,CACF,EA5Ka5B,EAmJa,gBAAA,CACtB,sCAAA,KApJS8B,CAAAA,CAAN9B","file":"bigquery.cjs","sourcesContent":["import type {\n LogicalType,\n SqlAdapter,\n SqlColumn,\n SqlDialect,\n SqlTableDef,\n} from \"../types\";\n\n// ---------------------------------------------------------------------------\n// Dialect (internal — used only by BigQueryAdapter)\n// ---------------------------------------------------------------------------\n\n/** BigQuery SQL dialect mapping. */\nclass BigQueryDialect implements SqlDialect {\n readonly name = \"bigquery\";\n\n mapType(logical: LogicalType): string {\n switch (logical) {\n case \"string\":\n return \"STRING\";\n case \"number\":\n return \"FLOAT64\";\n case \"bigint\":\n return \"INT64\";\n case \"boolean\":\n return \"BOOL\";\n case \"timestamp\":\n return \"TIMESTAMP\";\n case \"json\":\n return \"JSON\";\n case \"text\":\n return \"STRING\";\n }\n }\n\n quoteIdentifier(id: string): string {\n return `\\`${id}\\``;\n }\n}\n\n/** Shared BigQuery dialect singleton. */\nexport const bigqueryDialect: SqlDialect = new BigQueryDialect();\n\n// ---------------------------------------------------------------------------\n// Adapter\n// ---------------------------------------------------------------------------\n\n/**\n * BigQuery implementation of {@link SqlAdapter}.\n *\n * Accepts an already-configured BigQuery client so the library does not pull\n * in `@google-cloud/bigquery` as a hard dependency.\n *\n * @example\n * ```ts\n * import { BigQuery } from \"@google-cloud/bigquery\";\n * import { BigQueryAdapter } from \"./adapters/bigquery\";\n *\n * const adapter = new BigQueryAdapter({\n * bigquery: new BigQuery({ projectId: \"my-project\" }),\n * datasetId: \"my_dataset\",\n * });\n * ```\n */\nexport class BigQueryAdapter implements SqlAdapter {\n private readonly bigquery: any;\n private readonly datasetId: string;\n\n constructor(options: { bigquery: any; datasetId: string }) {\n this.bigquery = options.bigquery;\n this.datasetId = options.datasetId;\n }\n\n /** The BigQuery SQL dialect. */\n get dialect(): SqlDialect {\n return bigqueryDialect;\n }\n\n /** Check whether a table exists in the dataset. */\n async tableExists(tableName: string): Promise<boolean> {\n const [exists] = await this.dataset.table(tableName).exists();\n return exists;\n }\n\n /** Return the column names currently present in the table. */\n async getTableColumns(tableName: string): Promise<string[]> {\n const [metadata] = await this.dataset.table(tableName).getMetadata();\n const fields: Array<{ name: string }> = metadata.schema?.fields ?? [];\n return fields.map((f) => f.name);\n }\n\n /** Create a table using a fully-qualified name. */\n async createTable(table: SqlTableDef): Promise<void> {\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n const cols = table.columns\n .map((c) => {\n const notNull = c.isPrimaryKey ? \" NOT NULL\" : \"\";\n return ` ${qi(c.name)} ${c.sqlType}${notNull}`;\n })\n .join(\",\\n\");\n\n const ddl = `CREATE TABLE IF NOT EXISTS ${this.fqn(table.tableName)} (\\n${cols}\\n);`;\n await this.bigquery.query({ query: ddl });\n }\n\n /** Add columns to an existing table using a fully-qualified name. */\n async addColumns(tableName: string, columns: SqlColumn[]): Promise<void> {\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n for (const c of columns) {\n const stmt = `ALTER TABLE ${this.fqn(tableName)} ADD COLUMN ${qi(c.name)} ${c.sqlType};`;\n await this.bigquery.query({ query: stmt });\n }\n }\n\n /** Append rows via BigQuery streaming insert. */\n async insertRows(\n tableName: string,\n rows: Record<string, unknown>[],\n ): Promise<void> {\n if (rows.length === 0) return;\n await this.dataset.table(tableName).insert(rows);\n }\n\n /**\n * Upsert rows using a MERGE DML statement.\n *\n * Builds a source table from inline SELECT … UNION ALL rows and merges\n * into the target on the given primary key.\n */\n async upsertRows(\n tableName: string,\n rows: Record<string, unknown>[],\n primaryKey: string,\n ): Promise<void> {\n if (rows.length === 0) return;\n\n const allKeys = Object.keys(rows[0]!);\n const nonPkCols = allKeys.filter((k) => k !== primaryKey);\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n\n // Build inline source: SELECT val AS col, … UNION ALL SELECT …\n const selects = rows.map((row, i) => {\n const values = allKeys\n .map((k) => {\n const aliased =\n i === 0\n ? `${this.escapeValue(row[k])} AS ${qi(k)}`\n : this.escapeValue(row[k]);\n return aliased;\n })\n .join(\", \");\n return `SELECT ${values}`;\n });\n\n const source = selects.join(\" UNION ALL\\n \");\n\n // UPDATE SET clause (non-PK columns)\n const updateSet = nonPkCols\n .map((c) => `T.${qi(c)} = S.${qi(c)}`)\n .join(\", \");\n\n // INSERT columns / values\n const insertCols = allKeys.map((c) => qi(c)).join(\", \");\n const insertVals = allKeys.map((c) => `S.${qi(c)}`).join(\", \");\n\n const query = [\n `MERGE ${this.fqn(tableName)} AS T`,\n `USING (\\n ${source}\\n ) AS S`,\n `ON T.${qi(primaryKey)} = S.${qi(primaryKey)}`,\n `WHEN MATCHED THEN UPDATE SET ${updateSet}`,\n `WHEN NOT MATCHED THEN INSERT (${insertCols}) VALUES (${insertVals});`,\n ].join(\"\\n\");\n\n await this.bigquery.query({ query });\n }\n\n /** Delete rows by primary-key values. */\n async deleteRows(\n tableName: string,\n primaryKey: string,\n ids: string[],\n ): Promise<void> {\n if (ids.length === 0) return;\n\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n const escaped = ids.map((v) => this.escapeValue(v)).join(\", \");\n const query = `DELETE FROM ${this.fqn(tableName)} WHERE ${qi(primaryKey)} IN (${escaped});`;\n\n await this.bigquery.query({ query });\n }\n\n /** Execute a raw SQL statement (used by the migration manager). */\n async executeRaw(sql: string): Promise<void> {\n await this.bigquery.query({ query: sql });\n }\n\n // -------------------------------------------------------------------------\n // Private helpers\n // -------------------------------------------------------------------------\n\n /** The BigQuery Dataset handle. */\n private get dataset() {\n return this.bigquery.dataset(this.datasetId);\n }\n\n /** Return the fully-qualified table reference (`` `dataset.table` ``). */\n private fqn(tableName: string): string {\n return `\\`${this.datasetId}.${tableName}\\``;\n }\n\n /** ISO 8601 timestamp pattern (e.g. 2026-03-29T20:59:27.394Z) */\n private static readonly ISO_TIMESTAMP_RE =\n /^\\d{4}-\\d{2}-\\d{2}T\\d{2}:\\d{2}:\\d{2}/;\n\n /** Escape a value for use as a SQL literal. */\n private escapeValue(v: unknown): string {\n if (v === null || v === undefined) return \"NULL\";\n if (typeof v === \"boolean\") return v ? \"TRUE\" : \"FALSE\";\n if (typeof v === \"number\" || typeof v === \"bigint\") return String(v);\n if (typeof v === \"string\") {\n // ISO 8601 timestamps → TIMESTAMP literal (keeps type-safety with BQ TIMESTAMP columns)\n if (BigQueryAdapter.ISO_TIMESTAMP_RE.test(v)) {\n return `TIMESTAMP('${v}')`;\n }\n // Detect JSON strings (arrays/objects) → use PARSE_JSON for native JSON columns\n if (\n (v.startsWith(\"[\") && v.endsWith(\"]\")) ||\n (v.startsWith(\"{\") && v.endsWith(\"}\"))\n ) {\n return `PARSE_JSON('${v.replace(/'/g, \"\\\\'\")}')`;\n }\n return `'${v.replace(/'/g, \"\\\\'\")}'`;\n }\n // Objects / arrays → JSON\n return `PARSE_JSON('${JSON.stringify(v).replace(/'/g, \"\\\\'\")}')`;\n }\n}\n"]}
1
+ {"version":3,"sources":["../../src/sync/constants.ts","../../src/sync/adapters/bigquery.ts"],"names":["SYNC_VERSION_COLUMN","BigQueryDialect","logical","id","bigqueryDialect","_BigQueryAdapter","options","tableName","exists","metadata","f","table","qi","cols","c","notNull","ddl","columns","stmt","rows","primaryKey","allKeys","nonPkCols","k","source","row","i","updateSet","insertCols","insertVals","versionGuard","query","ids","escaped","v","sql","BigQueryAdapter"],"mappings":"aAaO,IAAMA,EAAsB,gBAAA,CCCnC,IAAMC,CAAAA,CAAN,KAA4C,CAA5C,WAAA,EAAA,CACE,IAAA,CAAS,IAAA,CAAO,WAAA,CAEhB,QAAQC,CAAAA,CAA8B,CACpC,OAAQA,CAAAA,EACN,KAAK,QAAA,CACH,OAAO,QAAA,CACT,KAAK,SACH,OAAO,SAAA,CACT,KAAK,QAAA,CACH,OAAO,OAAA,CACT,KAAK,SAAA,CACH,OAAO,OACT,KAAK,WAAA,CACH,OAAO,WAAA,CACT,KAAK,OACH,OAAO,MAAA,CACT,KAAK,MAAA,CACH,OAAO,QACX,CACF,CAEA,eAAA,CAAgBC,EAAoB,CAClC,OAAO,CAAA,EAAA,EAAKA,CAAE,IAChB,CACF,CAAA,CAGaC,EAA8B,IAAIH,CAAAA,CAuBlCI,EAAN,MAAMA,CAAsC,CAIjD,WAAA,CAAYC,EAA+C,CACzD,IAAA,CAAK,QAAA,CAAWA,CAAAA,CAAQ,SACxB,IAAA,CAAK,SAAA,CAAYA,CAAAA,CAAQ,UAC3B,CAGA,IAAI,OAAA,EAAsB,CACxB,OAAOF,CACT,CAGA,MAAM,WAAA,CAAYG,CAAAA,CAAqC,CACrD,GAAM,CAACC,CAAM,CAAA,CAAI,MAAM,KAAK,OAAA,CAAQ,KAAA,CAAMD,CAAS,CAAA,CAAE,QAAO,CAC5D,OAAOC,CACT,CAGA,MAAM,gBAAgBD,CAAAA,CAAsC,CAC1D,GAAM,CAACE,CAAQ,CAAA,CAAI,MAAM,IAAA,CAAK,OAAA,CAAQ,MAAMF,CAAS,CAAA,CAAE,WAAA,EAAY,CAEnE,QADwCE,CAAAA,CAAS,MAAA,EAAQ,QAAU,EAAC,EACtD,IAAKC,CAAAA,EAAMA,CAAAA,CAAE,IAAI,CACjC,CAGA,MAAM,WAAA,CAAYC,CAAAA,CAAmC,CACnD,IAAMC,CAAAA,CAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,gBAAgBA,CAAE,CAAA,CACpDU,EAAOF,CAAAA,CAAM,OAAA,CAChB,IAAKG,CAAAA,EAAM,CACV,IAAMC,CAAAA,CAAUD,EAAE,YAAA,CAAe,WAAA,CAAc,GAC/C,OAAO,CAAA,EAAA,EAAKF,EAAGE,CAAAA,CAAE,IAAI,CAAC,CAAA,CAAA,EAAIA,EAAE,OAAO,CAAA,EAAGC,CAAO,CAAA,CAC/C,CAAC,EACA,IAAA,CAAK,CAAA;AAAA,CAAK,EAEPC,CAAAA,CAAM,CAAA,2BAAA,EAA8B,KAAK,GAAA,CAAIL,CAAAA,CAAM,SAAS,CAAC,CAAA;AAAA,EAAOE,CAAI;AAAA,EAAA,CAAA,CAC9E,MAAM,IAAA,CAAK,QAAA,CAAS,KAAA,CAAM,CAAE,KAAA,CAAOG,CAAI,CAAC,EAC1C,CAGA,MAAM,UAAA,CAAWT,CAAAA,CAAmBU,EAAqC,CACvE,IAAML,CAAAA,CAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,eAAA,CAAgBA,CAAE,EAC1D,IAAA,IAAWW,CAAAA,IAAKG,CAAAA,CAAS,CACvB,IAAMC,CAAAA,CAAO,CAAA,YAAA,EAAe,IAAA,CAAK,IAAIX,CAAS,CAAC,CAAA,YAAA,EAAeK,CAAAA,CAAGE,CAAAA,CAAE,IAAI,CAAC,CAAA,CAAA,EAAIA,EAAE,OAAO,CAAA,CAAA,CAAA,CACrF,MAAM,IAAA,CAAK,QAAA,CAAS,KAAA,CAAM,CAAE,KAAA,CAAOI,CAAK,CAAC,EAC3C,CACF,CAGA,MAAM,UAAA,CACJX,CAAAA,CACAY,CAAAA,CACe,CACXA,CAAAA,CAAK,MAAA,GAAW,CAAA,EACpB,MAAM,IAAA,CAAK,OAAA,CAAQ,KAAA,CAAMZ,CAAS,EAAE,MAAA,CAAOY,CAAI,EACjD,CAQA,MAAM,UAAA,CACJZ,CAAAA,CACAY,CAAAA,CACAC,EACe,CACf,GAAID,CAAAA,CAAK,MAAA,GAAW,CAAA,CAAG,OAEvB,IAAME,CAAAA,CAAU,OAAO,IAAA,CAAKF,CAAAA,CAAK,CAAC,CAAE,CAAA,CAC9BG,CAAAA,CAAYD,CAAAA,CAAQ,MAAA,CAAQE,GAAMA,CAAAA,GAAMH,CAAU,CAAA,CAClDR,CAAAA,CAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,eAAA,CAAgBA,CAAE,CAAA,CAgBpDqB,CAAAA,CAbUL,CAAAA,CAAK,GAAA,CAAI,CAACM,CAAAA,CAAKC,CAAAA,GAUtB,CAAA,OAAA,EATQL,EACZ,GAAA,CAAKE,CAAAA,EAEFG,CAAAA,GAAM,CAAA,CACF,CAAA,EAAG,IAAA,CAAK,WAAA,CAAYD,CAAAA,CAAIF,CAAC,CAAC,CAAC,CAAA,IAAA,EAAOX,CAAAA,CAAGW,CAAC,CAAC,CAAA,CAAA,CACvC,IAAA,CAAK,YAAYE,CAAAA,CAAIF,CAAC,CAAC,CAE9B,EACA,IAAA,CAAK,IAAI,CACW,CAAA,CACxB,EAEsB,IAAA,CAAK,CAAA;AAAA,IAAA,CAAkB,CAAA,CAKxCI,CAAAA,CAAYL,CAAAA,CACf,GAAA,CAAKR,GAAM,CAAA,EAAA,EAAKF,CAAAA,CAAGE,CAAC,CAAC,CAAA,KAAA,EAAQF,CAAAA,CAAGE,CAAC,CAAC,EAAE,CAAA,CACpC,IAAA,CAAK,IAAI,CAAA,CAGNc,CAAAA,CAAaP,CAAAA,CAAQ,GAAA,CAAKP,CAAAA,EAAMF,EAAGE,CAAC,CAAC,CAAA,CAAE,IAAA,CAAK,IAAI,CAAA,CAChDe,CAAAA,CAAaR,CAAAA,CAAQ,IAAKP,CAAAA,EAAM,CAAA,EAAA,EAAKF,CAAAA,CAAGE,CAAC,CAAC,CAAA,CAAE,CAAA,CAAE,IAAA,CAAK,IAAI,CAAA,CAKvDgB,CAAAA,CAAeT,CAAAA,CAAQ,QAAA,CAASrB,CAAmB,CAAA,CACrD,CAAA,QAAA,EAAWY,CAAAA,CAAGZ,CAAmB,CAAC,CAAA,cAAA,EAAiBY,CAAAA,CAAGZ,CAAmB,CAAC,CAAA,KAAA,EAAQY,CAAAA,CAAGZ,CAAmB,CAAC,CAAA,CAAA,CAAA,CACzG,EAAA,CAEE+B,CAAAA,CAAQ,CACZ,CAAA,MAAA,EAAS,IAAA,CAAK,GAAA,CAAIxB,CAAS,CAAC,CAAA,KAAA,CAAA,CAC5B,CAAA;AAAA,IAAA,EAAgBiB,CAAM;AAAA,QAAA,CAAA,CACtB,CAAA,KAAA,EAAQZ,EAAGQ,CAAU,CAAC,QAAQR,CAAAA,CAAGQ,CAAU,CAAC,CAAA,CAAA,CAC5C,CAAA,YAAA,EAAeU,CAAY,CAAA,iBAAA,EAAoBH,CAAS,GACxD,CAAA,8BAAA,EAAiCC,CAAU,aAAaC,CAAU,CAAA,EAAA,CACpE,EAAE,IAAA,CAAK;AAAA,CAAI,EAEX,MAAM,IAAA,CAAK,SAAS,KAAA,CAAM,CAAE,MAAAE,CAAM,CAAC,EACrC,CAGA,MAAM,WACJxB,CAAAA,CACAa,CAAAA,CACAY,EACe,CACf,GAAIA,EAAI,MAAA,GAAW,CAAA,CAAG,OAEtB,IAAMpB,EAAMT,CAAAA,EAAe,IAAA,CAAK,QAAQ,eAAA,CAAgBA,CAAE,EACpD8B,CAAAA,CAAUD,CAAAA,CAAI,IAAKE,CAAAA,EAAM,IAAA,CAAK,YAAYA,CAAC,CAAC,EAAE,IAAA,CAAK,IAAI,EACvDH,CAAAA,CAAQ,CAAA,YAAA,EAAe,IAAA,CAAK,GAAA,CAAIxB,CAAS,CAAC,CAAA,OAAA,EAAUK,EAAGQ,CAAU,CAAC,QAAQa,CAAO,CAAA,EAAA,CAAA,CAEvF,MAAM,IAAA,CAAK,QAAA,CAAS,MAAM,CAAE,KAAA,CAAAF,CAAM,CAAC,EACrC,CAGA,MAAM,UAAA,CAAWI,CAAAA,CAA4B,CAC3C,MAAM,IAAA,CAAK,QAAA,CAAS,MAAM,CAAE,KAAA,CAAOA,CAAI,CAAC,EAC1C,CAOA,IAAY,OAAA,EAAU,CACpB,OAAO,IAAA,CAAK,SAAS,OAAA,CAAQ,IAAA,CAAK,SAAS,CAC7C,CAGQ,GAAA,CAAI5B,CAAAA,CAA2B,CACrC,OAAO,CAAA,EAAA,EAAK,KAAK,SAAS,CAAA,CAAA,EAAIA,CAAS,CAAA,EAAA,CACzC,CAOQ,YAAY2B,CAAAA,CAAoB,CACtC,OAAIA,CAAAA,EAAM,IAAA,CAAgC,OACtC,OAAOA,CAAAA,EAAM,UAAkBA,CAAAA,CAAI,MAAA,CAAS,OAAA,CAC5C,OAAOA,GAAM,QAAA,EAAY,OAAOA,GAAM,QAAA,CAAiB,MAAA,CAAOA,CAAC,CAAA,CAC/D,OAAOA,GAAM,QAAA,CAEX7B,CAAAA,CAAgB,iBAAiB,IAAA,CAAK6B,CAAC,EAClC,CAAA,WAAA,EAAcA,CAAC,KAIrBA,CAAAA,CAAE,UAAA,CAAW,GAAG,CAAA,EAAKA,EAAE,QAAA,CAAS,GAAG,GACnCA,CAAAA,CAAE,UAAA,CAAW,GAAG,CAAA,EAAKA,CAAAA,CAAE,SAAS,GAAG,CAAA,CAE7B,eAAeA,CAAAA,CAAE,OAAA,CAAQ,KAAM,KAAK,CAAC,KAEvC,CAAA,CAAA,EAAIA,CAAAA,CAAE,OAAA,CAAQ,IAAA,CAAM,KAAK,CAAC,CAAA,CAAA,CAAA,CAG5B,eAAe,IAAA,CAAK,SAAA,CAAUA,CAAC,CAAA,CAAE,OAAA,CAAQ,KAAM,KAAK,CAAC,IAC9D,CACF,EArLa7B,EA4Ja,gBAAA,CACtB,sCAAA,KA7JS+B,CAAAA,CAAN/B","file":"bigquery.cjs","sourcesContent":["/**\n * Internal constants shared between the worker, queue, schema mapper and\n * SQL adapters.\n */\n\n/**\n * Name of the SQL column that stores the publish-time `version` of each\n * sync event. Used by the worker to discard out-of-order PubSub deliveries\n * (the MERGE only updates rows when the incoming version is strictly\n * greater than the stored one).\n *\n * Two underscores prefix avoids collisions with user-defined fields.\n */\nexport const SYNC_VERSION_COLUMN = \"__sync_version\";\n","import { SYNC_VERSION_COLUMN } from \"../constants\";\nimport type {\n LogicalType,\n SqlAdapter,\n SqlColumn,\n SqlDialect,\n SqlTableDef,\n} from \"../types\";\n\n// ---------------------------------------------------------------------------\n// Dialect (internal — used only by BigQueryAdapter)\n// ---------------------------------------------------------------------------\n\n/** BigQuery SQL dialect mapping. */\nclass BigQueryDialect implements SqlDialect {\n readonly name = \"bigquery\";\n\n mapType(logical: LogicalType): string {\n switch (logical) {\n case \"string\":\n return \"STRING\";\n case \"number\":\n return \"FLOAT64\";\n case \"bigint\":\n return \"INT64\";\n case \"boolean\":\n return \"BOOL\";\n case \"timestamp\":\n return \"TIMESTAMP\";\n case \"json\":\n return \"JSON\";\n case \"text\":\n return \"STRING\";\n }\n }\n\n quoteIdentifier(id: string): string {\n return `\\`${id}\\``;\n }\n}\n\n/** Shared BigQuery dialect singleton. */\nexport const bigqueryDialect: SqlDialect = new BigQueryDialect();\n\n// ---------------------------------------------------------------------------\n// Adapter\n// ---------------------------------------------------------------------------\n\n/**\n * BigQuery implementation of {@link SqlAdapter}.\n *\n * Accepts an already-configured BigQuery client so the library does not pull\n * in `@google-cloud/bigquery` as a hard dependency.\n *\n * @example\n * ```ts\n * import { BigQuery } from \"@google-cloud/bigquery\";\n * import { BigQueryAdapter } from \"./adapters/bigquery\";\n *\n * const adapter = new BigQueryAdapter({\n * bigquery: new BigQuery({ projectId: \"my-project\" }),\n * datasetId: \"my_dataset\",\n * });\n * ```\n */\nexport class BigQueryAdapter implements SqlAdapter {\n private readonly bigquery: any;\n private readonly datasetId: string;\n\n constructor(options: { bigquery: any; datasetId: string }) {\n this.bigquery = options.bigquery;\n this.datasetId = options.datasetId;\n }\n\n /** The BigQuery SQL dialect. */\n get dialect(): SqlDialect {\n return bigqueryDialect;\n }\n\n /** Check whether a table exists in the dataset. */\n async tableExists(tableName: string): Promise<boolean> {\n const [exists] = await this.dataset.table(tableName).exists();\n return exists;\n }\n\n /** Return the column names currently present in the table. */\n async getTableColumns(tableName: string): Promise<string[]> {\n const [metadata] = await this.dataset.table(tableName).getMetadata();\n const fields: Array<{ name: string }> = metadata.schema?.fields ?? [];\n return fields.map((f) => f.name);\n }\n\n /** Create a table using a fully-qualified name. */\n async createTable(table: SqlTableDef): Promise<void> {\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n const cols = table.columns\n .map((c) => {\n const notNull = c.isPrimaryKey ? \" NOT NULL\" : \"\";\n return ` ${qi(c.name)} ${c.sqlType}${notNull}`;\n })\n .join(\",\\n\");\n\n const ddl = `CREATE TABLE IF NOT EXISTS ${this.fqn(table.tableName)} (\\n${cols}\\n);`;\n await this.bigquery.query({ query: ddl });\n }\n\n /** Add columns to an existing table using a fully-qualified name. */\n async addColumns(tableName: string, columns: SqlColumn[]): Promise<void> {\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n for (const c of columns) {\n const stmt = `ALTER TABLE ${this.fqn(tableName)} ADD COLUMN ${qi(c.name)} ${c.sqlType};`;\n await this.bigquery.query({ query: stmt });\n }\n }\n\n /** Append rows via BigQuery streaming insert. */\n async insertRows(\n tableName: string,\n rows: Record<string, unknown>[],\n ): Promise<void> {\n if (rows.length === 0) return;\n await this.dataset.table(tableName).insert(rows);\n }\n\n /**\n * Upsert rows using a MERGE DML statement.\n *\n * Builds a source table from inline SELECT … UNION ALL rows and merges\n * into the target on the given primary key.\n */\n async upsertRows(\n tableName: string,\n rows: Record<string, unknown>[],\n primaryKey: string,\n ): Promise<void> {\n if (rows.length === 0) return;\n\n const allKeys = Object.keys(rows[0]!);\n const nonPkCols = allKeys.filter((k) => k !== primaryKey);\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n\n // Build inline source: SELECT val AS col, … UNION ALL SELECT …\n const selects = rows.map((row, i) => {\n const values = allKeys\n .map((k) => {\n const aliased =\n i === 0\n ? `${this.escapeValue(row[k])} AS ${qi(k)}`\n : this.escapeValue(row[k]);\n return aliased;\n })\n .join(\", \");\n return `SELECT ${values}`;\n });\n\n const source = selects.join(\" UNION ALL\\n \");\n\n // UPDATE SET clause (non-PK columns).\n // Note: when __sync_version is present we still update it so the row\n // tracks the latest applied version.\n const updateSet = nonPkCols\n .map((c) => `T.${qi(c)} = S.${qi(c)}`)\n .join(\", \");\n\n // INSERT columns / values\n const insertCols = allKeys.map((c) => qi(c)).join(\", \");\n const insertVals = allKeys.map((c) => `S.${qi(c)}`).join(\", \");\n\n // Out-of-order protection: only UPDATE when the incoming version is\n // strictly greater than the stored one (NULL stored version means the\n // row pre-dates versioning → always update).\n const versionGuard = allKeys.includes(SYNC_VERSION_COLUMN)\n ? ` AND (T.${qi(SYNC_VERSION_COLUMN)} IS NULL OR S.${qi(SYNC_VERSION_COLUMN)} > T.${qi(SYNC_VERSION_COLUMN)})`\n : \"\";\n\n const query = [\n `MERGE ${this.fqn(tableName)} AS T`,\n `USING (\\n ${source}\\n ) AS S`,\n `ON T.${qi(primaryKey)} = S.${qi(primaryKey)}`,\n `WHEN MATCHED${versionGuard} THEN UPDATE SET ${updateSet}`,\n `WHEN NOT MATCHED THEN INSERT (${insertCols}) VALUES (${insertVals});`,\n ].join(\"\\n\");\n\n await this.bigquery.query({ query });\n }\n\n /** Delete rows by primary-key values. */\n async deleteRows(\n tableName: string,\n primaryKey: string,\n ids: string[],\n ): Promise<void> {\n if (ids.length === 0) return;\n\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n const escaped = ids.map((v) => this.escapeValue(v)).join(\", \");\n const query = `DELETE FROM ${this.fqn(tableName)} WHERE ${qi(primaryKey)} IN (${escaped});`;\n\n await this.bigquery.query({ query });\n }\n\n /** Execute a raw SQL statement (used by the migration manager). */\n async executeRaw(sql: string): Promise<void> {\n await this.bigquery.query({ query: sql });\n }\n\n // -------------------------------------------------------------------------\n // Private helpers\n // -------------------------------------------------------------------------\n\n /** The BigQuery Dataset handle. */\n private get dataset() {\n return this.bigquery.dataset(this.datasetId);\n }\n\n /** Return the fully-qualified table reference (`` `dataset.table` ``). */\n private fqn(tableName: string): string {\n return `\\`${this.datasetId}.${tableName}\\``;\n }\n\n /** ISO 8601 timestamp pattern (e.g. 2026-03-29T20:59:27.394Z) */\n private static readonly ISO_TIMESTAMP_RE =\n /^\\d{4}-\\d{2}-\\d{2}T\\d{2}:\\d{2}:\\d{2}/;\n\n /** Escape a value for use as a SQL literal. */\n private escapeValue(v: unknown): string {\n if (v === null || v === undefined) return \"NULL\";\n if (typeof v === \"boolean\") return v ? \"TRUE\" : \"FALSE\";\n if (typeof v === \"number\" || typeof v === \"bigint\") return String(v);\n if (typeof v === \"string\") {\n // ISO 8601 timestamps → TIMESTAMP literal (keeps type-safety with BQ TIMESTAMP columns)\n if (BigQueryAdapter.ISO_TIMESTAMP_RE.test(v)) {\n return `TIMESTAMP('${v}')`;\n }\n // Detect JSON strings (arrays/objects) → use PARSE_JSON for native JSON columns\n if (\n (v.startsWith(\"[\") && v.endsWith(\"]\")) ||\n (v.startsWith(\"{\") && v.endsWith(\"}\"))\n ) {\n return `PARSE_JSON('${v.replace(/'/g, \"\\\\'\")}')`;\n }\n return `'${v.replace(/'/g, \"\\\\'\")}'`;\n }\n // Objects / arrays → JSON\n return `PARSE_JSON('${JSON.stringify(v).replace(/'/g, \"\\\\'\")}')`;\n }\n}\n"]}
@@ -1,4 +1,4 @@
1
- import { c as SqlDialect, S as SqlAdapter, e as SqlTableDef, d as SqlColumn } from '../types-Cc3LHZ4c.cjs';
1
+ import { c as SqlDialect, S as SqlAdapter, e as SqlTableDef, d as SqlColumn } from '../types-CCFcY7YL.cjs';
2
2
 
3
3
  /** Shared BigQuery dialect singleton. */
4
4
  declare const bigqueryDialect: SqlDialect;
@@ -1,4 +1,4 @@
1
- import { c as SqlDialect, S as SqlAdapter, e as SqlTableDef, d as SqlColumn } from '../types-Cc3LHZ4c.js';
1
+ import { c as SqlDialect, S as SqlAdapter, e as SqlTableDef, d as SqlColumn } from '../types-CCFcY7YL.js';
2
2
 
3
3
  /** Shared BigQuery dialect singleton. */
4
4
  declare const bigqueryDialect: SqlDialect;
@@ -1,9 +1,9 @@
1
- var u=class{constructor(){this.name="bigquery";}mapType(t){switch(t){case "string":return "STRING";case "number":return "FLOAT64";case "bigint":return "INT64";case "boolean":return "BOOL";case "timestamp":return "TIMESTAMP";case "json":return "JSON";case "text":return "STRING"}}quoteIdentifier(t){return `\`${t}\``}},p=new u,l=class l{constructor(t){this.bigquery=t.bigquery,this.datasetId=t.datasetId;}get dialect(){return p}async tableExists(t){let[e]=await this.dataset.table(t).exists();return e}async getTableColumns(t){let[e]=await this.dataset.table(t).getMetadata();return (e.schema?.fields??[]).map(s=>s.name)}async createTable(t){let e=r=>this.dialect.quoteIdentifier(r),i=t.columns.map(r=>{let a=r.isPrimaryKey?" NOT NULL":"";return ` ${e(r.name)} ${r.sqlType}${a}`}).join(`,
1
+ var o="__sync_version";var d=class{constructor(){this.name="bigquery";}mapType(t){switch(t){case "string":return "STRING";case "number":return "FLOAT64";case "bigint":return "INT64";case "boolean":return "BOOL";case "timestamp":return "TIMESTAMP";case "json":return "JSON";case "text":return "STRING"}}quoteIdentifier(t){return `\`${t}\``}},f=new d,c=class c{constructor(t){this.bigquery=t.bigquery,this.datasetId=t.datasetId;}get dialect(){return f}async tableExists(t){let[e]=await this.dataset.table(t).exists();return e}async getTableColumns(t){let[e]=await this.dataset.table(t).getMetadata();return (e.schema?.fields??[]).map(s=>s.name)}async createTable(t){let e=a=>this.dialect.quoteIdentifier(a),r=t.columns.map(a=>{let n=a.isPrimaryKey?" NOT NULL":"";return ` ${e(a.name)} ${a.sqlType}${n}`}).join(`,
2
2
  `),s=`CREATE TABLE IF NOT EXISTS ${this.fqn(t.tableName)} (
3
- ${i}
4
- );`;await this.bigquery.query({query:s});}async addColumns(t,e){let i=s=>this.dialect.quoteIdentifier(s);for(let s of e){let r=`ALTER TABLE ${this.fqn(t)} ADD COLUMN ${i(s.name)} ${s.sqlType};`;await this.bigquery.query({query:r});}}async insertRows(t,e){e.length!==0&&await this.dataset.table(t).insert(e);}async upsertRows(t,e,i){if(e.length===0)return;let s=Object.keys(e[0]),r=s.filter(n=>n!==i),a=n=>this.dialect.quoteIdentifier(n),g=e.map((n,m)=>`SELECT ${s.map(c=>m===0?`${this.escapeValue(n[c])} AS ${a(c)}`:this.escapeValue(n[c])).join(", ")}`).join(` UNION ALL
5
- `),q=r.map(n=>`T.${a(n)} = S.${a(n)}`).join(", "),y=s.map(n=>a(n)).join(", "),T=s.map(n=>`S.${a(n)}`).join(", "),S=[`MERGE ${this.fqn(t)} AS T`,`USING (
6
- ${g}
7
- ) AS S`,`ON T.${a(i)} = S.${a(i)}`,`WHEN MATCHED THEN UPDATE SET ${q}`,`WHEN NOT MATCHED THEN INSERT (${y}) VALUES (${T});`].join(`
8
- `);await this.bigquery.query({query:S});}async deleteRows(t,e,i){if(i.length===0)return;let s=o=>this.dialect.quoteIdentifier(o),r=i.map(o=>this.escapeValue(o)).join(", "),a=`DELETE FROM ${this.fqn(t)} WHERE ${s(e)} IN (${r});`;await this.bigquery.query({query:a});}async executeRaw(t){await this.bigquery.query({query:t});}get dataset(){return this.bigquery.dataset(this.datasetId)}fqn(t){return `\`${this.datasetId}.${t}\``}escapeValue(t){return t==null?"NULL":typeof t=="boolean"?t?"TRUE":"FALSE":typeof t=="number"||typeof t=="bigint"?String(t):typeof t=="string"?l.ISO_TIMESTAMP_RE.test(t)?`TIMESTAMP('${t}')`:t.startsWith("[")&&t.endsWith("]")||t.startsWith("{")&&t.endsWith("}")?`PARSE_JSON('${t.replace(/'/g,"\\'")}')`:`'${t.replace(/'/g,"\\'")}'`:`PARSE_JSON('${JSON.stringify(t).replace(/'/g,"\\'")}')`}};l.ISO_TIMESTAMP_RE=/^\d{4}-\d{2}-\d{2}T\d{2}:\d{2}:\d{2}/;var d=l;export{d as BigQueryAdapter,p as bigqueryDialect};//# sourceMappingURL=bigquery.js.map
3
+ ${r}
4
+ );`;await this.bigquery.query({query:s});}async addColumns(t,e){let r=s=>this.dialect.quoteIdentifier(s);for(let s of e){let a=`ALTER TABLE ${this.fqn(t)} ADD COLUMN ${r(s.name)} ${s.sqlType};`;await this.bigquery.query({query:a});}}async insertRows(t,e){e.length!==0&&await this.dataset.table(t).insert(e);}async upsertRows(t,e,r){if(e.length===0)return;let s=Object.keys(e[0]),a=s.filter(i=>i!==r),n=i=>this.dialect.quoteIdentifier(i),q=e.map((i,E)=>`SELECT ${s.map(u=>E===0?`${this.escapeValue(i[u])} AS ${n(u)}`:this.escapeValue(i[u])).join(", ")}`).join(` UNION ALL
5
+ `),y=a.map(i=>`T.${n(i)} = S.${n(i)}`).join(", "),S=s.map(i=>n(i)).join(", "),T=s.map(i=>`S.${n(i)}`).join(", "),m=s.includes(o)?` AND (T.${n(o)} IS NULL OR S.${n(o)} > T.${n(o)})`:"",p=[`MERGE ${this.fqn(t)} AS T`,`USING (
6
+ ${q}
7
+ ) AS S`,`ON T.${n(r)} = S.${n(r)}`,`WHEN MATCHED${m} THEN UPDATE SET ${y}`,`WHEN NOT MATCHED THEN INSERT (${S}) VALUES (${T});`].join(`
8
+ `);await this.bigquery.query({query:p});}async deleteRows(t,e,r){if(r.length===0)return;let s=l=>this.dialect.quoteIdentifier(l),a=r.map(l=>this.escapeValue(l)).join(", "),n=`DELETE FROM ${this.fqn(t)} WHERE ${s(e)} IN (${a});`;await this.bigquery.query({query:n});}async executeRaw(t){await this.bigquery.query({query:t});}get dataset(){return this.bigquery.dataset(this.datasetId)}fqn(t){return `\`${this.datasetId}.${t}\``}escapeValue(t){return t==null?"NULL":typeof t=="boolean"?t?"TRUE":"FALSE":typeof t=="number"||typeof t=="bigint"?String(t):typeof t=="string"?c.ISO_TIMESTAMP_RE.test(t)?`TIMESTAMP('${t}')`:t.startsWith("[")&&t.endsWith("]")||t.startsWith("{")&&t.endsWith("}")?`PARSE_JSON('${t.replace(/'/g,"\\'")}')`:`'${t.replace(/'/g,"\\'")}'`:`PARSE_JSON('${JSON.stringify(t).replace(/'/g,"\\'")}')`}};c.ISO_TIMESTAMP_RE=/^\d{4}-\d{2}-\d{2}T\d{2}:\d{2}:\d{2}/;var g=c;export{g as BigQueryAdapter,f as bigqueryDialect};//# sourceMappingURL=bigquery.js.map
9
9
  //# sourceMappingURL=bigquery.js.map
@@ -1 +1 @@
1
- {"version":3,"sources":["../../src/sync/adapters/bigquery.ts"],"names":["BigQueryDialect","logical","id","bigqueryDialect","_BigQueryAdapter","options","tableName","exists","metadata","f","table","qi","cols","c","notNull","ddl","columns","stmt","rows","primaryKey","allKeys","nonPkCols","k","source","row","i","updateSet","insertCols","insertVals","query","ids","escaped","v","sql","BigQueryAdapter"],"mappings":"AAaA,IAAMA,EAAN,KAA4C,CAA5C,WAAA,EAAA,CACE,IAAA,CAAS,KAAO,WAAA,CAEhB,OAAA,CAAQC,CAAAA,CAA8B,CACpC,OAAQA,CAAAA,EACN,KAAK,QAAA,CACH,OAAO,QAAA,CACT,KAAK,QAAA,CACH,OAAO,UACT,KAAK,QAAA,CACH,OAAO,OAAA,CACT,KAAK,SAAA,CACH,OAAO,MAAA,CACT,KAAK,YACH,OAAO,WAAA,CACT,KAAK,MAAA,CACH,OAAO,OACT,KAAK,MAAA,CACH,OAAO,QACX,CACF,CAEA,eAAA,CAAgBC,CAAAA,CAAoB,CAClC,OAAO,CAAA,EAAA,EAAKA,CAAE,CAAA,EAAA,CAChB,CACF,EAGaC,CAAAA,CAA8B,IAAIH,EAuBlCI,CAAAA,CAAN,MAAMA,CAAsC,CAIjD,WAAA,CAAYC,CAAAA,CAA+C,CACzD,KAAK,QAAA,CAAWA,CAAAA,CAAQ,QAAA,CACxB,IAAA,CAAK,UAAYA,CAAAA,CAAQ,UAC3B,CAGA,IAAI,SAAsB,CACxB,OAAOF,CACT,CAGA,MAAM,YAAYG,CAAAA,CAAqC,CACrD,GAAM,CAACC,CAAM,CAAA,CAAI,MAAM,IAAA,CAAK,OAAA,CAAQ,MAAMD,CAAS,CAAA,CAAE,MAAA,EAAO,CAC5D,OAAOC,CACT,CAGA,MAAM,eAAA,CAAgBD,EAAsC,CAC1D,GAAM,CAACE,CAAQ,EAAI,MAAM,IAAA,CAAK,OAAA,CAAQ,KAAA,CAAMF,CAAS,CAAA,CAAE,WAAA,EAAY,CAEnE,OAAA,CADwCE,EAAS,MAAA,EAAQ,MAAA,EAAU,EAAC,EACtD,GAAA,CAAKC,GAAMA,CAAAA,CAAE,IAAI,CACjC,CAGA,MAAM,WAAA,CAAYC,CAAAA,CAAmC,CACnD,IAAMC,EAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,eAAA,CAAgBA,CAAE,CAAA,CACpDU,CAAAA,CAAOF,EAAM,OAAA,CAChB,GAAA,CAAKG,GAAM,CACV,IAAMC,CAAAA,CAAUD,CAAAA,CAAE,aAAe,WAAA,CAAc,EAAA,CAC/C,OAAO,CAAA,EAAA,EAAKF,EAAGE,CAAAA,CAAE,IAAI,CAAC,CAAA,CAAA,EAAIA,EAAE,OAAO,CAAA,EAAGC,CAAO,CAAA,CAC/C,CAAC,EACA,IAAA,CAAK,CAAA;AAAA,CAAK,EAEPC,CAAAA,CAAM,CAAA,2BAAA,EAA8B,KAAK,GAAA,CAAIL,CAAAA,CAAM,SAAS,CAAC,CAAA;AAAA,EAAOE,CAAI;AAAA,EAAA,CAAA,CAC9E,MAAM,IAAA,CAAK,QAAA,CAAS,KAAA,CAAM,CAAE,KAAA,CAAOG,CAAI,CAAC,EAC1C,CAGA,MAAM,UAAA,CAAWT,CAAAA,CAAmBU,EAAqC,CACvE,IAAML,CAAAA,CAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,eAAA,CAAgBA,CAAE,EAC1D,IAAA,IAAWW,CAAAA,IAAKG,CAAAA,CAAS,CACvB,IAAMC,CAAAA,CAAO,CAAA,YAAA,EAAe,IAAA,CAAK,IAAIX,CAAS,CAAC,CAAA,YAAA,EAAeK,CAAAA,CAAGE,CAAAA,CAAE,IAAI,CAAC,CAAA,CAAA,EAAIA,EAAE,OAAO,CAAA,CAAA,CAAA,CACrF,MAAM,IAAA,CAAK,QAAA,CAAS,KAAA,CAAM,CAAE,KAAA,CAAOI,CAAK,CAAC,EAC3C,CACF,CAGA,MAAM,UAAA,CACJX,CAAAA,CACAY,CAAAA,CACe,CACXA,CAAAA,CAAK,MAAA,GAAW,CAAA,EACpB,MAAM,IAAA,CAAK,OAAA,CAAQ,KAAA,CAAMZ,CAAS,EAAE,MAAA,CAAOY,CAAI,EACjD,CAQA,MAAM,UAAA,CACJZ,CAAAA,CACAY,CAAAA,CACAC,EACe,CACf,GAAID,CAAAA,CAAK,MAAA,GAAW,CAAA,CAAG,OAEvB,IAAME,CAAAA,CAAU,OAAO,IAAA,CAAKF,CAAAA,CAAK,CAAC,CAAE,CAAA,CAC9BG,CAAAA,CAAYD,CAAAA,CAAQ,MAAA,CAAQE,GAAMA,CAAAA,GAAMH,CAAU,CAAA,CAClDR,CAAAA,CAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,eAAA,CAAgBA,CAAE,CAAA,CAgBpDqB,CAAAA,CAbUL,CAAAA,CAAK,GAAA,CAAI,CAACM,CAAAA,CAAKC,CAAAA,GAUtB,CAAA,OAAA,EATQL,EACZ,GAAA,CAAKE,CAAAA,EAEFG,CAAAA,GAAM,CAAA,CACF,CAAA,EAAG,IAAA,CAAK,WAAA,CAAYD,CAAAA,CAAIF,CAAC,CAAC,CAAC,CAAA,IAAA,EAAOX,CAAAA,CAAGW,CAAC,CAAC,CAAA,CAAA,CACvC,IAAA,CAAK,YAAYE,CAAAA,CAAIF,CAAC,CAAC,CAE9B,EACA,IAAA,CAAK,IAAI,CACW,CAAA,CACxB,EAEsB,IAAA,CAAK,CAAA;AAAA,IAAA,CAAkB,CAAA,CAGxCI,EAAYL,CAAAA,CACf,GAAA,CAAKR,GAAM,CAAA,EAAA,EAAKF,CAAAA,CAAGE,CAAC,CAAC,CAAA,KAAA,EAAQF,CAAAA,CAAGE,CAAC,CAAC,CAAA,CAAE,EACpC,IAAA,CAAK,IAAI,EAGNc,CAAAA,CAAaP,CAAAA,CAAQ,GAAA,CAAKP,CAAAA,EAAMF,CAAAA,CAAGE,CAAC,CAAC,CAAA,CAAE,IAAA,CAAK,IAAI,CAAA,CAChDe,CAAAA,CAAaR,EAAQ,GAAA,CAAKP,CAAAA,EAAM,CAAA,EAAA,EAAKF,CAAAA,CAAGE,CAAC,CAAC,EAAE,CAAA,CAAE,IAAA,CAAK,IAAI,CAAA,CAEvDgB,CAAAA,CAAQ,CACZ,SAAS,IAAA,CAAK,GAAA,CAAIvB,CAAS,CAAC,CAAA,KAAA,CAAA,CAC5B,CAAA;AAAA,IAAA,EAAgBiB,CAAM;AAAA,QAAA,CAAA,CACtB,QAAQZ,CAAAA,CAAGQ,CAAU,CAAC,CAAA,KAAA,EAAQR,CAAAA,CAAGQ,CAAU,CAAC,CAAA,CAAA,CAC5C,CAAA,6BAAA,EAAgCO,CAAS,GACzC,CAAA,8BAAA,EAAiCC,CAAU,aAAaC,CAAU,CAAA,EAAA,CACpE,EAAE,IAAA,CAAK;AAAA,CAAI,EAEX,MAAM,IAAA,CAAK,SAAS,KAAA,CAAM,CAAE,MAAAC,CAAM,CAAC,EACrC,CAGA,MAAM,WACJvB,CAAAA,CACAa,CAAAA,CACAW,EACe,CACf,GAAIA,EAAI,MAAA,GAAW,CAAA,CAAG,OAEtB,IAAMnB,EAAMT,CAAAA,EAAe,IAAA,CAAK,QAAQ,eAAA,CAAgBA,CAAE,EACpD6B,CAAAA,CAAUD,CAAAA,CAAI,IAAKE,CAAAA,EAAM,IAAA,CAAK,YAAYA,CAAC,CAAC,EAAE,IAAA,CAAK,IAAI,EACvDH,CAAAA,CAAQ,CAAA,YAAA,EAAe,IAAA,CAAK,GAAA,CAAIvB,CAAS,CAAC,CAAA,OAAA,EAAUK,EAAGQ,CAAU,CAAC,QAAQY,CAAO,CAAA,EAAA,CAAA,CAEvF,MAAM,IAAA,CAAK,QAAA,CAAS,MAAM,CAAE,KAAA,CAAAF,CAAM,CAAC,EACrC,CAGA,MAAM,UAAA,CAAWI,CAAAA,CAA4B,CAC3C,MAAM,IAAA,CAAK,QAAA,CAAS,MAAM,CAAE,KAAA,CAAOA,CAAI,CAAC,EAC1C,CAOA,IAAY,OAAA,EAAU,CACpB,OAAO,IAAA,CAAK,SAAS,OAAA,CAAQ,IAAA,CAAK,SAAS,CAC7C,CAGQ,GAAA,CAAI3B,CAAAA,CAA2B,CACrC,OAAO,CAAA,EAAA,EAAK,KAAK,SAAS,CAAA,CAAA,EAAIA,CAAS,CAAA,EAAA,CACzC,CAOQ,YAAY0B,CAAAA,CAAoB,CACtC,OAAIA,CAAAA,EAAM,IAAA,CAAgC,OACtC,OAAOA,CAAAA,EAAM,UAAkBA,CAAAA,CAAI,MAAA,CAAS,OAAA,CAC5C,OAAOA,GAAM,QAAA,EAAY,OAAOA,GAAM,QAAA,CAAiB,MAAA,CAAOA,CAAC,CAAA,CAC/D,OAAOA,GAAM,QAAA,CAEX5B,CAAAA,CAAgB,iBAAiB,IAAA,CAAK4B,CAAC,EAClC,CAAA,WAAA,EAAcA,CAAC,KAIrBA,CAAAA,CAAE,UAAA,CAAW,GAAG,CAAA,EAAKA,EAAE,QAAA,CAAS,GAAG,GACnCA,CAAAA,CAAE,UAAA,CAAW,GAAG,CAAA,EAAKA,CAAAA,CAAE,SAAS,GAAG,CAAA,CAE7B,eAAeA,CAAAA,CAAE,OAAA,CAAQ,KAAM,KAAK,CAAC,KAEvC,CAAA,CAAA,EAAIA,CAAAA,CAAE,OAAA,CAAQ,IAAA,CAAM,KAAK,CAAC,CAAA,CAAA,CAAA,CAG5B,eAAe,IAAA,CAAK,SAAA,CAAUA,CAAC,CAAA,CAAE,OAAA,CAAQ,KAAM,KAAK,CAAC,IAC9D,CACF,EA5Ka5B,EAmJa,gBAAA,CACtB,sCAAA,KApJS8B,CAAAA,CAAN9B","file":"bigquery.js","sourcesContent":["import type {\n LogicalType,\n SqlAdapter,\n SqlColumn,\n SqlDialect,\n SqlTableDef,\n} from \"../types\";\n\n// ---------------------------------------------------------------------------\n// Dialect (internal — used only by BigQueryAdapter)\n// ---------------------------------------------------------------------------\n\n/** BigQuery SQL dialect mapping. */\nclass BigQueryDialect implements SqlDialect {\n readonly name = \"bigquery\";\n\n mapType(logical: LogicalType): string {\n switch (logical) {\n case \"string\":\n return \"STRING\";\n case \"number\":\n return \"FLOAT64\";\n case \"bigint\":\n return \"INT64\";\n case \"boolean\":\n return \"BOOL\";\n case \"timestamp\":\n return \"TIMESTAMP\";\n case \"json\":\n return \"JSON\";\n case \"text\":\n return \"STRING\";\n }\n }\n\n quoteIdentifier(id: string): string {\n return `\\`${id}\\``;\n }\n}\n\n/** Shared BigQuery dialect singleton. */\nexport const bigqueryDialect: SqlDialect = new BigQueryDialect();\n\n// ---------------------------------------------------------------------------\n// Adapter\n// ---------------------------------------------------------------------------\n\n/**\n * BigQuery implementation of {@link SqlAdapter}.\n *\n * Accepts an already-configured BigQuery client so the library does not pull\n * in `@google-cloud/bigquery` as a hard dependency.\n *\n * @example\n * ```ts\n * import { BigQuery } from \"@google-cloud/bigquery\";\n * import { BigQueryAdapter } from \"./adapters/bigquery\";\n *\n * const adapter = new BigQueryAdapter({\n * bigquery: new BigQuery({ projectId: \"my-project\" }),\n * datasetId: \"my_dataset\",\n * });\n * ```\n */\nexport class BigQueryAdapter implements SqlAdapter {\n private readonly bigquery: any;\n private readonly datasetId: string;\n\n constructor(options: { bigquery: any; datasetId: string }) {\n this.bigquery = options.bigquery;\n this.datasetId = options.datasetId;\n }\n\n /** The BigQuery SQL dialect. */\n get dialect(): SqlDialect {\n return bigqueryDialect;\n }\n\n /** Check whether a table exists in the dataset. */\n async tableExists(tableName: string): Promise<boolean> {\n const [exists] = await this.dataset.table(tableName).exists();\n return exists;\n }\n\n /** Return the column names currently present in the table. */\n async getTableColumns(tableName: string): Promise<string[]> {\n const [metadata] = await this.dataset.table(tableName).getMetadata();\n const fields: Array<{ name: string }> = metadata.schema?.fields ?? [];\n return fields.map((f) => f.name);\n }\n\n /** Create a table using a fully-qualified name. */\n async createTable(table: SqlTableDef): Promise<void> {\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n const cols = table.columns\n .map((c) => {\n const notNull = c.isPrimaryKey ? \" NOT NULL\" : \"\";\n return ` ${qi(c.name)} ${c.sqlType}${notNull}`;\n })\n .join(\",\\n\");\n\n const ddl = `CREATE TABLE IF NOT EXISTS ${this.fqn(table.tableName)} (\\n${cols}\\n);`;\n await this.bigquery.query({ query: ddl });\n }\n\n /** Add columns to an existing table using a fully-qualified name. */\n async addColumns(tableName: string, columns: SqlColumn[]): Promise<void> {\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n for (const c of columns) {\n const stmt = `ALTER TABLE ${this.fqn(tableName)} ADD COLUMN ${qi(c.name)} ${c.sqlType};`;\n await this.bigquery.query({ query: stmt });\n }\n }\n\n /** Append rows via BigQuery streaming insert. */\n async insertRows(\n tableName: string,\n rows: Record<string, unknown>[],\n ): Promise<void> {\n if (rows.length === 0) return;\n await this.dataset.table(tableName).insert(rows);\n }\n\n /**\n * Upsert rows using a MERGE DML statement.\n *\n * Builds a source table from inline SELECT … UNION ALL rows and merges\n * into the target on the given primary key.\n */\n async upsertRows(\n tableName: string,\n rows: Record<string, unknown>[],\n primaryKey: string,\n ): Promise<void> {\n if (rows.length === 0) return;\n\n const allKeys = Object.keys(rows[0]!);\n const nonPkCols = allKeys.filter((k) => k !== primaryKey);\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n\n // Build inline source: SELECT val AS col, … UNION ALL SELECT …\n const selects = rows.map((row, i) => {\n const values = allKeys\n .map((k) => {\n const aliased =\n i === 0\n ? `${this.escapeValue(row[k])} AS ${qi(k)}`\n : this.escapeValue(row[k]);\n return aliased;\n })\n .join(\", \");\n return `SELECT ${values}`;\n });\n\n const source = selects.join(\" UNION ALL\\n \");\n\n // UPDATE SET clause (non-PK columns)\n const updateSet = nonPkCols\n .map((c) => `T.${qi(c)} = S.${qi(c)}`)\n .join(\", \");\n\n // INSERT columns / values\n const insertCols = allKeys.map((c) => qi(c)).join(\", \");\n const insertVals = allKeys.map((c) => `S.${qi(c)}`).join(\", \");\n\n const query = [\n `MERGE ${this.fqn(tableName)} AS T`,\n `USING (\\n ${source}\\n ) AS S`,\n `ON T.${qi(primaryKey)} = S.${qi(primaryKey)}`,\n `WHEN MATCHED THEN UPDATE SET ${updateSet}`,\n `WHEN NOT MATCHED THEN INSERT (${insertCols}) VALUES (${insertVals});`,\n ].join(\"\\n\");\n\n await this.bigquery.query({ query });\n }\n\n /** Delete rows by primary-key values. */\n async deleteRows(\n tableName: string,\n primaryKey: string,\n ids: string[],\n ): Promise<void> {\n if (ids.length === 0) return;\n\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n const escaped = ids.map((v) => this.escapeValue(v)).join(\", \");\n const query = `DELETE FROM ${this.fqn(tableName)} WHERE ${qi(primaryKey)} IN (${escaped});`;\n\n await this.bigquery.query({ query });\n }\n\n /** Execute a raw SQL statement (used by the migration manager). */\n async executeRaw(sql: string): Promise<void> {\n await this.bigquery.query({ query: sql });\n }\n\n // -------------------------------------------------------------------------\n // Private helpers\n // -------------------------------------------------------------------------\n\n /** The BigQuery Dataset handle. */\n private get dataset() {\n return this.bigquery.dataset(this.datasetId);\n }\n\n /** Return the fully-qualified table reference (`` `dataset.table` ``). */\n private fqn(tableName: string): string {\n return `\\`${this.datasetId}.${tableName}\\``;\n }\n\n /** ISO 8601 timestamp pattern (e.g. 2026-03-29T20:59:27.394Z) */\n private static readonly ISO_TIMESTAMP_RE =\n /^\\d{4}-\\d{2}-\\d{2}T\\d{2}:\\d{2}:\\d{2}/;\n\n /** Escape a value for use as a SQL literal. */\n private escapeValue(v: unknown): string {\n if (v === null || v === undefined) return \"NULL\";\n if (typeof v === \"boolean\") return v ? \"TRUE\" : \"FALSE\";\n if (typeof v === \"number\" || typeof v === \"bigint\") return String(v);\n if (typeof v === \"string\") {\n // ISO 8601 timestamps → TIMESTAMP literal (keeps type-safety with BQ TIMESTAMP columns)\n if (BigQueryAdapter.ISO_TIMESTAMP_RE.test(v)) {\n return `TIMESTAMP('${v}')`;\n }\n // Detect JSON strings (arrays/objects) → use PARSE_JSON for native JSON columns\n if (\n (v.startsWith(\"[\") && v.endsWith(\"]\")) ||\n (v.startsWith(\"{\") && v.endsWith(\"}\"))\n ) {\n return `PARSE_JSON('${v.replace(/'/g, \"\\\\'\")}')`;\n }\n return `'${v.replace(/'/g, \"\\\\'\")}'`;\n }\n // Objects / arrays → JSON\n return `PARSE_JSON('${JSON.stringify(v).replace(/'/g, \"\\\\'\")}')`;\n }\n}\n"]}
1
+ {"version":3,"sources":["../../src/sync/constants.ts","../../src/sync/adapters/bigquery.ts"],"names":["SYNC_VERSION_COLUMN","BigQueryDialect","logical","id","bigqueryDialect","_BigQueryAdapter","options","tableName","exists","metadata","f","table","qi","cols","c","notNull","ddl","columns","stmt","rows","primaryKey","allKeys","nonPkCols","k","source","row","i","updateSet","insertCols","insertVals","versionGuard","query","ids","escaped","v","sql","BigQueryAdapter"],"mappings":"AAaO,IAAMA,EAAsB,gBAAA,CCCnC,IAAMC,CAAAA,CAAN,KAA4C,CAA5C,WAAA,EAAA,CACE,IAAA,CAAS,IAAA,CAAO,WAAA,CAEhB,QAAQC,CAAAA,CAA8B,CACpC,OAAQA,CAAAA,EACN,KAAK,QAAA,CACH,OAAO,QAAA,CACT,KAAK,SACH,OAAO,SAAA,CACT,KAAK,QAAA,CACH,OAAO,OAAA,CACT,KAAK,SAAA,CACH,OAAO,OACT,KAAK,WAAA,CACH,OAAO,WAAA,CACT,KAAK,OACH,OAAO,MAAA,CACT,KAAK,MAAA,CACH,OAAO,QACX,CACF,CAEA,eAAA,CAAgBC,EAAoB,CAClC,OAAO,CAAA,EAAA,EAAKA,CAAE,IAChB,CACF,CAAA,CAGaC,EAA8B,IAAIH,CAAAA,CAuBlCI,EAAN,MAAMA,CAAsC,CAIjD,WAAA,CAAYC,EAA+C,CACzD,IAAA,CAAK,QAAA,CAAWA,CAAAA,CAAQ,SACxB,IAAA,CAAK,SAAA,CAAYA,CAAAA,CAAQ,UAC3B,CAGA,IAAI,OAAA,EAAsB,CACxB,OAAOF,CACT,CAGA,MAAM,WAAA,CAAYG,CAAAA,CAAqC,CACrD,GAAM,CAACC,CAAM,CAAA,CAAI,MAAM,KAAK,OAAA,CAAQ,KAAA,CAAMD,CAAS,CAAA,CAAE,QAAO,CAC5D,OAAOC,CACT,CAGA,MAAM,gBAAgBD,CAAAA,CAAsC,CAC1D,GAAM,CAACE,CAAQ,CAAA,CAAI,MAAM,IAAA,CAAK,OAAA,CAAQ,MAAMF,CAAS,CAAA,CAAE,WAAA,EAAY,CAEnE,QADwCE,CAAAA,CAAS,MAAA,EAAQ,QAAU,EAAC,EACtD,IAAKC,CAAAA,EAAMA,CAAAA,CAAE,IAAI,CACjC,CAGA,MAAM,WAAA,CAAYC,CAAAA,CAAmC,CACnD,IAAMC,CAAAA,CAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,gBAAgBA,CAAE,CAAA,CACpDU,EAAOF,CAAAA,CAAM,OAAA,CAChB,IAAKG,CAAAA,EAAM,CACV,IAAMC,CAAAA,CAAUD,EAAE,YAAA,CAAe,WAAA,CAAc,GAC/C,OAAO,CAAA,EAAA,EAAKF,EAAGE,CAAAA,CAAE,IAAI,CAAC,CAAA,CAAA,EAAIA,EAAE,OAAO,CAAA,EAAGC,CAAO,CAAA,CAC/C,CAAC,EACA,IAAA,CAAK,CAAA;AAAA,CAAK,EAEPC,CAAAA,CAAM,CAAA,2BAAA,EAA8B,KAAK,GAAA,CAAIL,CAAAA,CAAM,SAAS,CAAC,CAAA;AAAA,EAAOE,CAAI;AAAA,EAAA,CAAA,CAC9E,MAAM,IAAA,CAAK,QAAA,CAAS,KAAA,CAAM,CAAE,KAAA,CAAOG,CAAI,CAAC,EAC1C,CAGA,MAAM,UAAA,CAAWT,CAAAA,CAAmBU,EAAqC,CACvE,IAAML,CAAAA,CAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,eAAA,CAAgBA,CAAE,EAC1D,IAAA,IAAWW,CAAAA,IAAKG,CAAAA,CAAS,CACvB,IAAMC,CAAAA,CAAO,CAAA,YAAA,EAAe,IAAA,CAAK,IAAIX,CAAS,CAAC,CAAA,YAAA,EAAeK,CAAAA,CAAGE,CAAAA,CAAE,IAAI,CAAC,CAAA,CAAA,EAAIA,EAAE,OAAO,CAAA,CAAA,CAAA,CACrF,MAAM,IAAA,CAAK,QAAA,CAAS,KAAA,CAAM,CAAE,KAAA,CAAOI,CAAK,CAAC,EAC3C,CACF,CAGA,MAAM,UAAA,CACJX,CAAAA,CACAY,CAAAA,CACe,CACXA,CAAAA,CAAK,MAAA,GAAW,CAAA,EACpB,MAAM,IAAA,CAAK,OAAA,CAAQ,KAAA,CAAMZ,CAAS,EAAE,MAAA,CAAOY,CAAI,EACjD,CAQA,MAAM,UAAA,CACJZ,CAAAA,CACAY,CAAAA,CACAC,EACe,CACf,GAAID,CAAAA,CAAK,MAAA,GAAW,CAAA,CAAG,OAEvB,IAAME,CAAAA,CAAU,OAAO,IAAA,CAAKF,CAAAA,CAAK,CAAC,CAAE,CAAA,CAC9BG,CAAAA,CAAYD,CAAAA,CAAQ,MAAA,CAAQE,GAAMA,CAAAA,GAAMH,CAAU,CAAA,CAClDR,CAAAA,CAAMT,CAAAA,EAAe,IAAA,CAAK,OAAA,CAAQ,eAAA,CAAgBA,CAAE,CAAA,CAgBpDqB,CAAAA,CAbUL,CAAAA,CAAK,GAAA,CAAI,CAACM,CAAAA,CAAKC,CAAAA,GAUtB,CAAA,OAAA,EATQL,EACZ,GAAA,CAAKE,CAAAA,EAEFG,CAAAA,GAAM,CAAA,CACF,CAAA,EAAG,IAAA,CAAK,WAAA,CAAYD,CAAAA,CAAIF,CAAC,CAAC,CAAC,CAAA,IAAA,EAAOX,CAAAA,CAAGW,CAAC,CAAC,CAAA,CAAA,CACvC,IAAA,CAAK,YAAYE,CAAAA,CAAIF,CAAC,CAAC,CAE9B,EACA,IAAA,CAAK,IAAI,CACW,CAAA,CACxB,EAEsB,IAAA,CAAK,CAAA;AAAA,IAAA,CAAkB,CAAA,CAKxCI,CAAAA,CAAYL,CAAAA,CACf,GAAA,CAAKR,GAAM,CAAA,EAAA,EAAKF,CAAAA,CAAGE,CAAC,CAAC,CAAA,KAAA,EAAQF,CAAAA,CAAGE,CAAC,CAAC,EAAE,CAAA,CACpC,IAAA,CAAK,IAAI,CAAA,CAGNc,CAAAA,CAAaP,CAAAA,CAAQ,GAAA,CAAKP,CAAAA,EAAMF,EAAGE,CAAC,CAAC,CAAA,CAAE,IAAA,CAAK,IAAI,CAAA,CAChDe,CAAAA,CAAaR,CAAAA,CAAQ,IAAKP,CAAAA,EAAM,CAAA,EAAA,EAAKF,CAAAA,CAAGE,CAAC,CAAC,CAAA,CAAE,CAAA,CAAE,IAAA,CAAK,IAAI,CAAA,CAKvDgB,CAAAA,CAAeT,CAAAA,CAAQ,QAAA,CAASrB,CAAmB,CAAA,CACrD,CAAA,QAAA,EAAWY,CAAAA,CAAGZ,CAAmB,CAAC,CAAA,cAAA,EAAiBY,CAAAA,CAAGZ,CAAmB,CAAC,CAAA,KAAA,EAAQY,CAAAA,CAAGZ,CAAmB,CAAC,CAAA,CAAA,CAAA,CACzG,EAAA,CAEE+B,CAAAA,CAAQ,CACZ,CAAA,MAAA,EAAS,IAAA,CAAK,GAAA,CAAIxB,CAAS,CAAC,CAAA,KAAA,CAAA,CAC5B,CAAA;AAAA,IAAA,EAAgBiB,CAAM;AAAA,QAAA,CAAA,CACtB,CAAA,KAAA,EAAQZ,EAAGQ,CAAU,CAAC,QAAQR,CAAAA,CAAGQ,CAAU,CAAC,CAAA,CAAA,CAC5C,CAAA,YAAA,EAAeU,CAAY,CAAA,iBAAA,EAAoBH,CAAS,GACxD,CAAA,8BAAA,EAAiCC,CAAU,aAAaC,CAAU,CAAA,EAAA,CACpE,EAAE,IAAA,CAAK;AAAA,CAAI,EAEX,MAAM,IAAA,CAAK,SAAS,KAAA,CAAM,CAAE,MAAAE,CAAM,CAAC,EACrC,CAGA,MAAM,WACJxB,CAAAA,CACAa,CAAAA,CACAY,EACe,CACf,GAAIA,EAAI,MAAA,GAAW,CAAA,CAAG,OAEtB,IAAMpB,EAAMT,CAAAA,EAAe,IAAA,CAAK,QAAQ,eAAA,CAAgBA,CAAE,EACpD8B,CAAAA,CAAUD,CAAAA,CAAI,IAAKE,CAAAA,EAAM,IAAA,CAAK,YAAYA,CAAC,CAAC,EAAE,IAAA,CAAK,IAAI,EACvDH,CAAAA,CAAQ,CAAA,YAAA,EAAe,IAAA,CAAK,GAAA,CAAIxB,CAAS,CAAC,CAAA,OAAA,EAAUK,EAAGQ,CAAU,CAAC,QAAQa,CAAO,CAAA,EAAA,CAAA,CAEvF,MAAM,IAAA,CAAK,QAAA,CAAS,MAAM,CAAE,KAAA,CAAAF,CAAM,CAAC,EACrC,CAGA,MAAM,UAAA,CAAWI,CAAAA,CAA4B,CAC3C,MAAM,IAAA,CAAK,QAAA,CAAS,MAAM,CAAE,KAAA,CAAOA,CAAI,CAAC,EAC1C,CAOA,IAAY,OAAA,EAAU,CACpB,OAAO,IAAA,CAAK,SAAS,OAAA,CAAQ,IAAA,CAAK,SAAS,CAC7C,CAGQ,GAAA,CAAI5B,CAAAA,CAA2B,CACrC,OAAO,CAAA,EAAA,EAAK,KAAK,SAAS,CAAA,CAAA,EAAIA,CAAS,CAAA,EAAA,CACzC,CAOQ,YAAY2B,CAAAA,CAAoB,CACtC,OAAIA,CAAAA,EAAM,IAAA,CAAgC,OACtC,OAAOA,CAAAA,EAAM,UAAkBA,CAAAA,CAAI,MAAA,CAAS,OAAA,CAC5C,OAAOA,GAAM,QAAA,EAAY,OAAOA,GAAM,QAAA,CAAiB,MAAA,CAAOA,CAAC,CAAA,CAC/D,OAAOA,GAAM,QAAA,CAEX7B,CAAAA,CAAgB,iBAAiB,IAAA,CAAK6B,CAAC,EAClC,CAAA,WAAA,EAAcA,CAAC,KAIrBA,CAAAA,CAAE,UAAA,CAAW,GAAG,CAAA,EAAKA,EAAE,QAAA,CAAS,GAAG,GACnCA,CAAAA,CAAE,UAAA,CAAW,GAAG,CAAA,EAAKA,CAAAA,CAAE,SAAS,GAAG,CAAA,CAE7B,eAAeA,CAAAA,CAAE,OAAA,CAAQ,KAAM,KAAK,CAAC,KAEvC,CAAA,CAAA,EAAIA,CAAAA,CAAE,OAAA,CAAQ,IAAA,CAAM,KAAK,CAAC,CAAA,CAAA,CAAA,CAG5B,eAAe,IAAA,CAAK,SAAA,CAAUA,CAAC,CAAA,CAAE,OAAA,CAAQ,KAAM,KAAK,CAAC,IAC9D,CACF,EArLa7B,EA4Ja,gBAAA,CACtB,sCAAA,KA7JS+B,CAAAA,CAAN/B","file":"bigquery.js","sourcesContent":["/**\n * Internal constants shared between the worker, queue, schema mapper and\n * SQL adapters.\n */\n\n/**\n * Name of the SQL column that stores the publish-time `version` of each\n * sync event. Used by the worker to discard out-of-order PubSub deliveries\n * (the MERGE only updates rows when the incoming version is strictly\n * greater than the stored one).\n *\n * Two underscores prefix avoids collisions with user-defined fields.\n */\nexport const SYNC_VERSION_COLUMN = \"__sync_version\";\n","import { SYNC_VERSION_COLUMN } from \"../constants\";\nimport type {\n LogicalType,\n SqlAdapter,\n SqlColumn,\n SqlDialect,\n SqlTableDef,\n} from \"../types\";\n\n// ---------------------------------------------------------------------------\n// Dialect (internal — used only by BigQueryAdapter)\n// ---------------------------------------------------------------------------\n\n/** BigQuery SQL dialect mapping. */\nclass BigQueryDialect implements SqlDialect {\n readonly name = \"bigquery\";\n\n mapType(logical: LogicalType): string {\n switch (logical) {\n case \"string\":\n return \"STRING\";\n case \"number\":\n return \"FLOAT64\";\n case \"bigint\":\n return \"INT64\";\n case \"boolean\":\n return \"BOOL\";\n case \"timestamp\":\n return \"TIMESTAMP\";\n case \"json\":\n return \"JSON\";\n case \"text\":\n return \"STRING\";\n }\n }\n\n quoteIdentifier(id: string): string {\n return `\\`${id}\\``;\n }\n}\n\n/** Shared BigQuery dialect singleton. */\nexport const bigqueryDialect: SqlDialect = new BigQueryDialect();\n\n// ---------------------------------------------------------------------------\n// Adapter\n// ---------------------------------------------------------------------------\n\n/**\n * BigQuery implementation of {@link SqlAdapter}.\n *\n * Accepts an already-configured BigQuery client so the library does not pull\n * in `@google-cloud/bigquery` as a hard dependency.\n *\n * @example\n * ```ts\n * import { BigQuery } from \"@google-cloud/bigquery\";\n * import { BigQueryAdapter } from \"./adapters/bigquery\";\n *\n * const adapter = new BigQueryAdapter({\n * bigquery: new BigQuery({ projectId: \"my-project\" }),\n * datasetId: \"my_dataset\",\n * });\n * ```\n */\nexport class BigQueryAdapter implements SqlAdapter {\n private readonly bigquery: any;\n private readonly datasetId: string;\n\n constructor(options: { bigquery: any; datasetId: string }) {\n this.bigquery = options.bigquery;\n this.datasetId = options.datasetId;\n }\n\n /** The BigQuery SQL dialect. */\n get dialect(): SqlDialect {\n return bigqueryDialect;\n }\n\n /** Check whether a table exists in the dataset. */\n async tableExists(tableName: string): Promise<boolean> {\n const [exists] = await this.dataset.table(tableName).exists();\n return exists;\n }\n\n /** Return the column names currently present in the table. */\n async getTableColumns(tableName: string): Promise<string[]> {\n const [metadata] = await this.dataset.table(tableName).getMetadata();\n const fields: Array<{ name: string }> = metadata.schema?.fields ?? [];\n return fields.map((f) => f.name);\n }\n\n /** Create a table using a fully-qualified name. */\n async createTable(table: SqlTableDef): Promise<void> {\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n const cols = table.columns\n .map((c) => {\n const notNull = c.isPrimaryKey ? \" NOT NULL\" : \"\";\n return ` ${qi(c.name)} ${c.sqlType}${notNull}`;\n })\n .join(\",\\n\");\n\n const ddl = `CREATE TABLE IF NOT EXISTS ${this.fqn(table.tableName)} (\\n${cols}\\n);`;\n await this.bigquery.query({ query: ddl });\n }\n\n /** Add columns to an existing table using a fully-qualified name. */\n async addColumns(tableName: string, columns: SqlColumn[]): Promise<void> {\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n for (const c of columns) {\n const stmt = `ALTER TABLE ${this.fqn(tableName)} ADD COLUMN ${qi(c.name)} ${c.sqlType};`;\n await this.bigquery.query({ query: stmt });\n }\n }\n\n /** Append rows via BigQuery streaming insert. */\n async insertRows(\n tableName: string,\n rows: Record<string, unknown>[],\n ): Promise<void> {\n if (rows.length === 0) return;\n await this.dataset.table(tableName).insert(rows);\n }\n\n /**\n * Upsert rows using a MERGE DML statement.\n *\n * Builds a source table from inline SELECT … UNION ALL rows and merges\n * into the target on the given primary key.\n */\n async upsertRows(\n tableName: string,\n rows: Record<string, unknown>[],\n primaryKey: string,\n ): Promise<void> {\n if (rows.length === 0) return;\n\n const allKeys = Object.keys(rows[0]!);\n const nonPkCols = allKeys.filter((k) => k !== primaryKey);\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n\n // Build inline source: SELECT val AS col, … UNION ALL SELECT …\n const selects = rows.map((row, i) => {\n const values = allKeys\n .map((k) => {\n const aliased =\n i === 0\n ? `${this.escapeValue(row[k])} AS ${qi(k)}`\n : this.escapeValue(row[k]);\n return aliased;\n })\n .join(\", \");\n return `SELECT ${values}`;\n });\n\n const source = selects.join(\" UNION ALL\\n \");\n\n // UPDATE SET clause (non-PK columns).\n // Note: when __sync_version is present we still update it so the row\n // tracks the latest applied version.\n const updateSet = nonPkCols\n .map((c) => `T.${qi(c)} = S.${qi(c)}`)\n .join(\", \");\n\n // INSERT columns / values\n const insertCols = allKeys.map((c) => qi(c)).join(\", \");\n const insertVals = allKeys.map((c) => `S.${qi(c)}`).join(\", \");\n\n // Out-of-order protection: only UPDATE when the incoming version is\n // strictly greater than the stored one (NULL stored version means the\n // row pre-dates versioning → always update).\n const versionGuard = allKeys.includes(SYNC_VERSION_COLUMN)\n ? ` AND (T.${qi(SYNC_VERSION_COLUMN)} IS NULL OR S.${qi(SYNC_VERSION_COLUMN)} > T.${qi(SYNC_VERSION_COLUMN)})`\n : \"\";\n\n const query = [\n `MERGE ${this.fqn(tableName)} AS T`,\n `USING (\\n ${source}\\n ) AS S`,\n `ON T.${qi(primaryKey)} = S.${qi(primaryKey)}`,\n `WHEN MATCHED${versionGuard} THEN UPDATE SET ${updateSet}`,\n `WHEN NOT MATCHED THEN INSERT (${insertCols}) VALUES (${insertVals});`,\n ].join(\"\\n\");\n\n await this.bigquery.query({ query });\n }\n\n /** Delete rows by primary-key values. */\n async deleteRows(\n tableName: string,\n primaryKey: string,\n ids: string[],\n ): Promise<void> {\n if (ids.length === 0) return;\n\n const qi = (id: string) => this.dialect.quoteIdentifier(id);\n const escaped = ids.map((v) => this.escapeValue(v)).join(\", \");\n const query = `DELETE FROM ${this.fqn(tableName)} WHERE ${qi(primaryKey)} IN (${escaped});`;\n\n await this.bigquery.query({ query });\n }\n\n /** Execute a raw SQL statement (used by the migration manager). */\n async executeRaw(sql: string): Promise<void> {\n await this.bigquery.query({ query: sql });\n }\n\n // -------------------------------------------------------------------------\n // Private helpers\n // -------------------------------------------------------------------------\n\n /** The BigQuery Dataset handle. */\n private get dataset() {\n return this.bigquery.dataset(this.datasetId);\n }\n\n /** Return the fully-qualified table reference (`` `dataset.table` ``). */\n private fqn(tableName: string): string {\n return `\\`${this.datasetId}.${tableName}\\``;\n }\n\n /** ISO 8601 timestamp pattern (e.g. 2026-03-29T20:59:27.394Z) */\n private static readonly ISO_TIMESTAMP_RE =\n /^\\d{4}-\\d{2}-\\d{2}T\\d{2}:\\d{2}:\\d{2}/;\n\n /** Escape a value for use as a SQL literal. */\n private escapeValue(v: unknown): string {\n if (v === null || v === undefined) return \"NULL\";\n if (typeof v === \"boolean\") return v ? \"TRUE\" : \"FALSE\";\n if (typeof v === \"number\" || typeof v === \"bigint\") return String(v);\n if (typeof v === \"string\") {\n // ISO 8601 timestamps → TIMESTAMP literal (keeps type-safety with BQ TIMESTAMP columns)\n if (BigQueryAdapter.ISO_TIMESTAMP_RE.test(v)) {\n return `TIMESTAMP('${v}')`;\n }\n // Detect JSON strings (arrays/objects) → use PARSE_JSON for native JSON columns\n if (\n (v.startsWith(\"[\") && v.endsWith(\"]\")) ||\n (v.startsWith(\"{\") && v.endsWith(\"}\"))\n ) {\n return `PARSE_JSON('${v.replace(/'/g, \"\\\\'\")}')`;\n }\n return `'${v.replace(/'/g, \"\\\\'\")}'`;\n }\n // Objects / arrays → JSON\n return `PARSE_JSON('${JSON.stringify(v).replace(/'/g, \"\\\\'\")}')`;\n }\n}\n"]}
@@ -1,7 +1,7 @@
1
- 'use strict';function oe(r){let e=[],n=r.replace(/[.*+?^${}()|[\]\\]/g,a=>a===":"?a:`\\${a}`).replace(/:([a-zA-Z_][a-zA-Z0-9_]*)/g,(a,s)=>(e.push(s),"([^/]+)"));return {pattern:new RegExp(`^${n}$`),paramNames:e}}function se(r){let e=r.path??r.url??"/",n=e.indexOf("?");return n===-1?e:e.slice(0,n)}var L=class{constructor(){this.routes=[];this.middlewares=[];this.notFoundHandler=(e,n)=>{n.status(404).send("Not Found");};this.errorHandler=(e,n,a)=>{console.error("[MiniRouter]",e),a.status(500).send("Internal Server Error");};}use(e){return this.middlewares.push(e),this}get(e,n){return this.addRoute("GET",e,n)}post(e,n){return this.addRoute("POST",e,n)}put(e,n){return this.addRoute("PUT",e,n)}patch(e,n){return this.addRoute("PATCH",e,n)}delete(e,n){return this.addRoute("DELETE",e,n)}onNotFound(e){return this.notFoundHandler=e,this}onError(e){return this.errorHandler=e,this}addRoute(e,n,a){let{pattern:s,paramNames:g}=oe(n);return this.routes.push({method:e.toUpperCase(),pattern:s,paramNames:g,handler:a}),this}async handle(e,n){let a=(e.method??"GET").toUpperCase(),s=se(e),g=null,y={};for(let h of this.routes){if(h.method!==a)continue;let S=s.match(h.pattern);if(S){g=h,y={},h.paramNames.forEach((b,l)=>{y[b]=decodeURIComponent(S[l+1]??"");});break}}let R=Object.assign(e,{params:y}),u=g?g.handler:this.notFoundHandler;try{await this.runMiddlewareChain(R,n,u);}catch(h){this.errorHandler(h,e,n);}}async runMiddlewareChain(e,n,a){let s=0,g=async()=>{if(s<this.middlewares.length){let y=this.middlewares[s++];await y(e,n,g);}else await a(e,n);};await g();}};var ae={string:"ZodString",number:"ZodNumber",bigint:"ZodBigInt",boolean:"ZodBoolean",date:"ZodDate",enum:"ZodEnum",nativeEnum:"ZodNativeEnum",literal:"ZodLiteral",object:"ZodObject",array:"ZodArray",optional:"ZodOptional",nullable:"ZodNullable",default:"ZodDefault",coerce:"ZodCoerce",union:"ZodUnion",undefined:"ZodUndefined",unknown:"ZodUnknown",any:"ZodAny",record:"ZodRecord"};function F(r){let e=r,n=e._zod?.def?.type;if(n)return ae[n]??`Zod${n.charAt(0).toUpperCase()}${n.slice(1)}`;let a=e._def?.typeName;return a||""}function W(r){let e=r;if(e._zod?.def?.innerType)return e._zod.def.innerType;if(e._def?.innerType)return e._def.innerType}function K(r){let e=r;return e.shape&&typeof e.shape=="object"?e.shape:e._zod?.def?.shape&&typeof e._zod.def.shape=="object"?e._zod.def.shape:e._def?.shape?typeof e._def.shape=="function"?e._def.shape():e._def.shape:{}}var ie=new Set(["ZodOptional","ZodNullable","ZodDefault"]);function J(r){let e=r,n=false;for(;;){let a=F(e);if(!ie.has(a))break;(a==="ZodOptional"||a==="ZodNullable")&&(n=true);let s=W(e);if(!s)break;e=s;}return {inner:e,nullable:n}}var V={ZodString:"string",ZodNumber:"number",ZodBigInt:"bigint",ZodBoolean:"boolean",ZodDate:"timestamp",ZodEnum:"string",ZodNativeEnum:"string",ZodLiteral:"string"};function ce(r){let{inner:e}=J(r);return V[F(e)]??"json"}function Y(r,e,n,a,s,g,y,R){for(let[u,h]of Object.entries(r)){let S=n?`${n}__${u}`:u;if(s.has(u)||s.has(S))continue;let{inner:b,nullable:l}=J(h),c=F(b),o=a||l;if(c==="ZodObject"){let i=K(b);Y(i,e,S,o,s,g,y,R);continue}let t=V[c]??"json",d=S===y||u===y,f=g[S]??g[u]??S;R.push({name:f,sqlType:e.mapType(t),nullable:d?false:o,isPrimaryKey:d});}}function q(r,e,n={}){let{primaryKey:a,exclude:s=[],columnMap:g={}}=n,y=new Set(s),R=K(r),u=[];return Y(R,e,"",false,y,g,a,u),u}function Q(r){if(r==null)return null;if(typeof r=="object"&&typeof r.toDate=="function")return r.toDate().toISOString();if(r instanceof Date)return r.toISOString();if(Buffer.isBuffer(r))return r.toString("base64");if(r instanceof Uint8Array)return Buffer.from(r).toString("base64");if(typeof r=="object"&&"latitude"in r&&"longitude"in r){let e=r;return JSON.stringify({lat:e.latitude,lng:e.longitude})}return Array.isArray(r)?JSON.stringify(r.map(Q)):r}function X(r,e,n){for(let[a,s]of Object.entries(r)){let g=e?`${e}__${a}`:a;s!=null&&typeof s=="object"&&!Array.isArray(s)&&!(s instanceof Date)&&!Buffer.isBuffer(s)&&!(s instanceof Uint8Array)&&typeof s.toDate!="function"&&!("latitude"in s&&"longitude"in s)?X(s,g,n):n[g]=Q(s);}}function z(r,e){let n=new Set(e?.exclude),a=e?.columnMap??{},s={};X(r,"",s);let g={};for(let[y,R]of Object.entries(s)){if(n.has(y))continue;let u=y.split("__")[0];if(u!==y&&n.has(u))continue;let h=a[y]??(y.includes("__")?a[y.split("__").pop()]:void 0)??y;g[h]=R;}return g}async function U(r,e){let{pubsub:n,topicPrefix:a="firestore-sync",ordering:s=true,subscriptionSuffix:g="sync-sub",includeDLQ:y=true,ackDeadlineSeconds:R=60,messageRetentionDuration:u}=e,h={topics:[],subscriptions:[]};for(let S of Object.keys(r)){let b=`${a}-${S}`,l=`${a}-${S}-${g}`,c=n.topic(b),o=false,[t]=await c.exists();t||(await c.create(),o=true,console.info(`[ensureSyncInfra] Created topic "${b}"`)),h.topics.push({name:b,created:o});let d=c.subscription(l),[f]=await d.exists();if(!f)await d.create({enableMessageOrdering:s,ackDeadlineSeconds:R,...u?{messageRetentionDuration:u}:{}}),console.info(`[ensureSyncInfra] Created subscription "${l}" (ordering=${s})`),h.subscriptions.push({name:l,topic:b,created:true,orderingEnabled:s});else {let i,p=s;try{let[C]=await d.getMetadata();p=!!C?.enableMessageOrdering,p!==s&&(i=`Subscription "${l}" exists with enableMessageOrdering=${p}, but ordering=${s} was requested. This setting is immutable; delete and recreate the subscription to change it.`,console.warn(`[ensureSyncInfra] ${i}`));}catch{}h.subscriptions.push({name:l,topic:b,created:false,orderingEnabled:p,...i?{warning:i}:{}});}if(y){let i=`${a}-${S}-dlq`,p=n.topic(i),[C]=await p.exists(),w=false;C||(await p.create(),w=true,console.info(`[ensureSyncInfra] Created DLQ topic "${i}"`)),h.topics.push({name:i,created:w});}}return h}function I(r,e){if(process.env.FUNCTIONS_EMULATOR==="true"){let s=process.env.GCLOUD_PROJECT??process.env.GOOGLE_CLOUD_PROJECT??"demo-project",g=process.env.FUNCTION_REGION??"us-central1",y=(process.env.FUNCTION_TARGET??"").replace(/\./g,"-");return `/${s}/${g}/${y}${e}`}let n=process.env.K_SERVICE,a=r.hostname??r.headers?.host??"";return n&&a.includes("cloudfunctions.net")?`/${n.toLowerCase()}${e}`:e}function N(r,e,n){return `<!DOCTYPE html>
1
+ 'use strict';function re(o){let e=[],t=o.replace(/[.*+?^${}()|[\]\\]/g,a=>a===":"?a:`\\${a}`).replace(/:([a-zA-Z_][a-zA-Z0-9_]*)/g,(a,s)=>(e.push(s),"([^/]+)"));return {pattern:new RegExp(`^${t}$`),paramNames:e}}function se(o){let e=o.path??o.url??"/",t=e.indexOf("?");return t===-1?e:e.slice(0,t)}var L=class{constructor(){this.routes=[];this.middlewares=[];this.notFoundHandler=(e,t)=>{t.status(404).send("Not Found");};this.errorHandler=(e,t,a)=>{console.error("[MiniRouter]",e),a.status(500).send("Internal Server Error");};}use(e){return this.middlewares.push(e),this}get(e,t){return this.addRoute("GET",e,t)}post(e,t){return this.addRoute("POST",e,t)}put(e,t){return this.addRoute("PUT",e,t)}patch(e,t){return this.addRoute("PATCH",e,t)}delete(e,t){return this.addRoute("DELETE",e,t)}onNotFound(e){return this.notFoundHandler=e,this}onError(e){return this.errorHandler=e,this}addRoute(e,t,a){let{pattern:s,paramNames:l}=re(t);return this.routes.push({method:e.toUpperCase(),pattern:s,paramNames:l,handler:a}),this}async handle(e,t){let a=(e.method??"GET").toUpperCase(),s=se(e),l=null,m={};for(let b of this.routes){if(b.method!==a)continue;let R=s.match(b.pattern);if(R){l=b,m={},b.paramNames.forEach((p,u)=>{m[p]=decodeURIComponent(R[u+1]??"");});break}}let w=Object.assign(e,{params:m}),d=l?l.handler:this.notFoundHandler;try{await this.runMiddlewareChain(w,t,d);}catch(b){this.errorHandler(b,e,t);}}async runMiddlewareChain(e,t,a){let s=0,l=async()=>{if(s<this.middlewares.length){let m=this.middlewares[s++];await m(e,t,l);}else await a(e,t);};await l();}};var ae={string:"ZodString",number:"ZodNumber",bigint:"ZodBigInt",boolean:"ZodBoolean",date:"ZodDate",enum:"ZodEnum",nativeEnum:"ZodNativeEnum",literal:"ZodLiteral",object:"ZodObject",array:"ZodArray",optional:"ZodOptional",nullable:"ZodNullable",default:"ZodDefault",coerce:"ZodCoerce",union:"ZodUnion",undefined:"ZodUndefined",unknown:"ZodUnknown",any:"ZodAny",record:"ZodRecord"};function F(o){let e=o,t=e._zod?.def?.type;if(t)return ae[t]??`Zod${t.charAt(0).toUpperCase()}${t.slice(1)}`;let a=e._def?.typeName;return a||""}function V(o){let e=o;if(e._zod?.def?.innerType)return e._zod.def.innerType;if(e._def?.innerType)return e._def.innerType}function K(o){let e=o;return e.shape&&typeof e.shape=="object"?e.shape:e._zod?.def?.shape&&typeof e._zod.def.shape=="object"?e._zod.def.shape:e._def?.shape?typeof e._def.shape=="function"?e._def.shape():e._def.shape:{}}var A="__sync_version";var ie=new Set(["ZodOptional","ZodNullable","ZodDefault"]);function W(o){let e=o,t=false;for(;;){let a=F(e);if(!ie.has(a))break;(a==="ZodOptional"||a==="ZodNullable")&&(t=true);let s=V(e);if(!s)break;e=s;}return {inner:e,nullable:t}}var J={ZodString:"string",ZodNumber:"number",ZodBigInt:"bigint",ZodBoolean:"boolean",ZodDate:"timestamp",ZodEnum:"string",ZodNativeEnum:"string",ZodLiteral:"string"};function ce(o){let{inner:e}=W(o);return J[F(e)]??"json"}function Y(o,e,t,a,s,l,m,w){for(let[d,b]of Object.entries(o)){let R=t?`${t}__${d}`:d;if(s.has(d)||s.has(R))continue;let{inner:p,nullable:u}=W(b),r=F(p),c=a||u;if(r==="ZodObject"){let i=K(p);Y(i,e,R,c,s,l,m,w);continue}let n=J[r]??"json",f=R===m||d===m,S=l[R]??l[d]??R;w.push({name:S,sqlType:e.mapType(n),nullable:f?false:c,isPrimaryKey:f});}}function E(o,e,t={}){let{primaryKey:a,exclude:s=[],columnMap:l={}}=t,m=new Set(s),w=K(o),d=[];return Y(w,e,"",false,m,l,a,d),d.some(b=>b.name===A)||d.push({name:A,sqlType:e.mapType("bigint"),nullable:true,isPrimaryKey:false,description:"Monotonic publish version (Date.now() ms). Internal."}),d}function Q(o){if(o==null)return null;if(typeof o=="object"&&typeof o.toDate=="function")return o.toDate().toISOString();if(o instanceof Date)return o.toISOString();if(Buffer.isBuffer(o))return o.toString("base64");if(o instanceof Uint8Array)return Buffer.from(o).toString("base64");if(typeof o=="object"&&"latitude"in o&&"longitude"in o){let e=o;return JSON.stringify({lat:e.latitude,lng:e.longitude})}return Array.isArray(o)?JSON.stringify(o.map(Q)):o}function X(o,e,t){for(let[a,s]of Object.entries(o)){let l=e?`${e}__${a}`:a;s!=null&&typeof s=="object"&&!Array.isArray(s)&&!(s instanceof Date)&&!Buffer.isBuffer(s)&&!(s instanceof Uint8Array)&&typeof s.toDate!="function"&&!("latitude"in s&&"longitude"in s)?X(s,l,t):t[l]=Q(s);}}function Z(o,e){let t=new Set(e?.exclude),a=e?.columnMap??{},s={};X(o,"",s);let l={};for(let[m,w]of Object.entries(s)){if(t.has(m))continue;let d=m.split("__")[0];if(d!==m&&t.has(d))continue;let b=a[m]??(m.includes("__")?a[m.split("__").pop()]:void 0)??m;l[b]=w;}return l}function q(o,e){if(process.env.FUNCTIONS_EMULATOR==="true"){let s=process.env.GCLOUD_PROJECT??process.env.GOOGLE_CLOUD_PROJECT??"demo-project",l=process.env.FUNCTION_REGION??"us-central1",m=(process.env.FUNCTION_TARGET??"").replace(/\./g,"-");return `/${s}/${l}/${m}${e}`}let t=process.env.K_SERVICE,a=o.hostname??o.headers?.host??"";return t&&a.includes("cloudfunctions.net")?`/${t.toLowerCase()}${e}`:e}function P(o,e,t){return `<!DOCTYPE html>
2
2
  <html lang="en"><head>
3
3
  <meta charset="utf-8"><meta name="viewport" content="width=device-width,initial-scale=1">
4
- <title>${r} \u2014 Sync Admin</title>
4
+ <title>${o} \u2014 Sync Admin</title>
5
5
  <style>
6
6
  *{box-sizing:border-box;margin:0;padding:0}
7
7
  body{font-family:system-ui,-apple-system,sans-serif;background:#f5f5f5;color:#1a1a1a;padding:2rem}
@@ -26,102 +26,73 @@
26
26
  </style>
27
27
  </head><body>
28
28
  <nav><a href="${e}/">\u2190 Dashboard</a></nav>
29
- <h1>${r}</h1>
30
- ${n}
31
- </body></html>`}function O(r,e,n=200){r.status(n).set("Content-Type","text/html; charset=utf-8").send(e);}function _(r,e,n=200){r.status(n).set("Content-Type","application/json").send(JSON.stringify(e,null,2));}function j(r){return (r.headers?.accept??"").includes("application/json")}function B(r,e,n,a,s,g,y,R){let u=(s.basePath??"/").replace(/\/$/,"")||"",h=s.featuresFlag??{},S=[];for(let[l,c]of Object.entries(r)){let o=g[l];S.push({name:l,schema:c.schema??null,documentKey:c._systemKeys?.[0]??c.documentKey??"docId",tableName:o?.tableName??l,isGroup:!!c._isGroup,repoCfg:o,repo:c});}let b=new L;if(s.auth)if(typeof s.auth=="function")b.use(s.auth);else {let l=s.auth.realm??"Sync Admin",c="Basic "+Buffer.from(`${s.auth.username}:${s.auth.password}`).toString("base64");b.use((o,t,d)=>{if((o.headers?.authorization??"")!==c){t.status(401).set("WWW-Authenticate",`Basic realm="${l}"`).set("Content-Type","text/plain").send("Unauthorized");return}d();});}return b.get(`${u}/`,(l,c)=>{let o=I(l,u),t=S.map(p=>{let C=[];return h.healthCheck&&C.push(`<a class="btn" href="${o}/${p.name}/health">Health</a>`),h.manualSync&&C.push(`<a class="btn btn-primary" href="${o}/${p.name}/force-sync">Force Sync</a>`),`<tr>
32
- <td><strong>${p.name}</strong></td>
33
- <td>${p.tableName}</td>
34
- <td>${p.isGroup?'<span class="badge badge-warn">group</span>':'<span class="badge badge-ok">collection</span>'}</td>
35
- <td>${p.schema?"\u2713":"\u2717"}</td>
36
- <td>${C.join(" ")}</td>
29
+ <h1>${o}</h1>
30
+ ${t}
31
+ </body></html>`}function D(o,e,t=200){o.status(t).set("Content-Type","text/html; charset=utf-8").send(e);}function z(o,e,t=200){o.status(t).set("Content-Type","application/json").send(JSON.stringify(e,null,2));}function I(o){return (o.headers?.accept??"").includes("application/json")}function B(o,e,t,a,s,l,m,w){let d=(s.basePath??"/").replace(/\/$/,"")||"",b=s.featuresFlag??{},R=[];for(let[u,r]of Object.entries(o)){let c=l[u];R.push({name:u,schema:r.schema??null,documentKey:r._systemKeys?.[0]??r.documentKey??"docId",tableName:c?.tableName??u,isGroup:!!r._isGroup,repoCfg:c,repo:r});}let p=new L;if(s.auth)if(typeof s.auth=="function")p.use(s.auth);else {let u=s.auth.realm??"Sync Admin",r="Basic "+Buffer.from(`${s.auth.username}:${s.auth.password}`).toString("base64");p.use((c,n,f)=>{if((c.headers?.authorization??"")!==r){n.status(401).set("WWW-Authenticate",`Basic realm="${u}"`).set("Content-Type","text/plain").send("Unauthorized");return}f();});}return p.get(`${d}/`,(u,r)=>{let c=q(u,d),n=R.map(y=>{let v=[];return b.healthCheck&&v.push(`<a class="btn" href="${c}/${y.name}/health">Health</a>`),b.manualSync&&v.push(`<a class="btn btn-primary" href="${c}/${y.name}/force-sync">Force Sync</a>`),`<tr>
32
+ <td><strong>${y.name}</strong></td>
33
+ <td>${y.tableName}</td>
34
+ <td>${y.isGroup?'<span class="badge badge-warn">group</span>':'<span class="badge badge-ok">collection</span>'}</td>
35
+ <td>${y.schema?"\u2713":"\u2717"}</td>
36
+ <td>${v.join(" ")}</td>
37
37
  </tr>`}).join(`
38
- `),d=h.viewQueue?`<p><a class="btn" href="${o}/queues">View Queues</a></p>`:"",f=h.configCheck?`<p style="margin-top:.5rem"><a class="btn" href="${o}/config-check">\u2699 Config Check</a></p>`:"",i=N("Sync Dashboard",o,`<div class="card">
38
+ `),f=b.viewQueue?`<p><a class="btn" href="${c}/queues">View Queues</a></p>`:"",S=b.configCheck?`<p style="margin-top:.5rem"><a class="btn" href="${c}/config-check">\u2699 Config Check</a></p>`:"",i=P("Sync Dashboard",c,`<div class="card">
39
39
  <table>
40
40
  <thead><tr><th>Repository</th><th>Table</th><th>Type</th><th>Schema</th><th>Actions</th></tr></thead>
41
- <tbody>${t}</tbody>
41
+ <tbody>${n}</tbody>
42
42
  </table>
43
- ${d}
44
43
  ${f}
45
- </div>`);O(c,i);}),b.get(`${u}`,(l,c)=>{let o=I(l,u);c.status(302).set("Location",`${o}/`).send("");}),h.healthCheck&&b.get(`${u}/:repoName/health`,async(l,c)=>{let o=I(l,u),t=S.find(m=>m.name===l.params.repoName);if(!t){O(c,N("Not Found",o,`<p>Unknown repo: ${l.params.repoName}</p>`),404);return}if(!t.schema){O(c,N("Health Check",o,`<p class="badge badge-warn">No Zod schema attached to "${t.name}"</p>`));return}let d=q(t.schema,e.dialect,{primaryKey:t.documentKey,exclude:t.repoCfg?.exclude,columnMap:t.repoCfg?.columnMap}),f=[],i=false,p=null;try{i=await e.tableExists(t.tableName),i&&(f=await e.getTableColumns(t.tableName));}catch(m){p=m?.message??String(m);}let C=new Set(f),w=new Set(d.map(m=>m.name)),T=d.filter(m=>!C.has(m.name)),E=f.filter(m=>!w.has(m)),D=d.filter(m=>C.has(m.name)),P=i&&T.length===0&&!p;if(j(l)){_(c,{repo:t.name,table:t.tableName,tableExists:i,healthy:P,error:p,columns:{expected:d.map(m=>({name:m.name,type:m.sqlType,nullable:m.nullable,isPrimaryKey:m.isPrimaryKey})),actual:f,matched:D.map(m=>m.name),missing:T.map(m=>({name:m.name,type:m.sqlType})),extra:E}});return}let x=P?'<span class="badge badge-ok">Healthy</span>':'<span class="badge badge-err">Unhealthy</span>',$=d.map(m=>{let A=C.has(m.name)?'<span class="badge badge-ok">OK</span>':'<span class="badge badge-err">MISSING</span>';return `<tr><td>${m.name}</td><td>${m.sqlType}</td><td>${m.nullable?"Yes":"No"}</td><td>${m.isPrimaryKey?"\u2713":""}</td><td>${A}</td></tr>`}).join(`
46
- `),k=E.map(m=>`<tr><td>${m}</td><td colspan="3" class="muted">not in schema</td><td><span class="badge badge-warn">EXTRA</span></td></tr>`).join(`
47
- `),v=N(`Health: ${t.name}`,o,`<div class="card">
48
- <p>Table: <code>${t.tableName}</code> ${i?x:'<span class="badge badge-err">NOT FOUND</span>'}</p>
49
- ${p?`<p class="badge badge-err">Error: ${p}</p>`:""}
44
+ ${S}
45
+ </div>`);D(r,i);}),p.get(`${d}`,(u,r)=>{let c=q(u,d);r.status(302).set("Location",`${c}/`).send("");}),b.healthCheck&&p.get(`${d}/:repoName/health`,async(u,r)=>{let c=q(u,d),n=R.find(g=>g.name===u.params.repoName);if(!n){D(r,P("Not Found",c,`<p>Unknown repo: ${u.params.repoName}</p>`),404);return}if(!n.schema){D(r,P("Health Check",c,`<p class="badge badge-warn">No Zod schema attached to "${n.name}"</p>`));return}let f=E(n.schema,e.dialect,{primaryKey:n.documentKey,exclude:n.repoCfg?.exclude,columnMap:n.repoCfg?.columnMap}),S=[],i=false,y=null;try{i=await e.tableExists(n.tableName),i&&(S=await e.getTableColumns(n.tableName));}catch(g){y=g?.message??String(g);}let v=new Set(S),x=new Set(f.map(g=>g.name)),k=f.filter(g=>!v.has(g.name)),_=S.filter(g=>!x.has(g)),O=f.filter(g=>v.has(g.name)),h=i&&k.length===0&&!y;if(I(u)){z(r,{repo:n.name,table:n.tableName,tableExists:i,healthy:h,error:y,columns:{expected:f.map(g=>({name:g.name,type:g.sqlType,nullable:g.nullable,isPrimaryKey:g.isPrimaryKey})),actual:S,matched:O.map(g=>g.name),missing:k.map(g=>({name:g.name,type:g.sqlType})),extra:_}});return}let $=h?'<span class="badge badge-ok">Healthy</span>':'<span class="badge badge-err">Unhealthy</span>',T=f.map(g=>{let j=v.has(g.name)?'<span class="badge badge-ok">OK</span>':'<span class="badge badge-err">MISSING</span>';return `<tr><td>${g.name}</td><td>${g.sqlType}</td><td>${g.nullable?"Yes":"No"}</td><td>${g.isPrimaryKey?"\u2713":""}</td><td>${j}</td></tr>`}).join(`
46
+ `),C=_.map(g=>`<tr><td>${g}</td><td colspan="3" class="muted">not in schema</td><td><span class="badge badge-warn">EXTRA</span></td></tr>`).join(`
47
+ `),N=P(`Health: ${n.name}`,c,`<div class="card">
48
+ <p>Table: <code>${n.tableName}</code> ${i?$:'<span class="badge badge-err">NOT FOUND</span>'}</p>
49
+ ${y?`<p class="badge badge-err">Error: ${y}</p>`:""}
50
50
  <h2>Columns</h2>
51
51
  <table>
52
52
  <thead><tr><th>Column</th><th>SQL Type</th><th>Nullable</th><th>PK</th><th>Status</th></tr></thead>
53
- <tbody>${$}${k}</tbody>
53
+ <tbody>${T}${C}</tbody>
54
54
  </table>
55
- </div>`);O(c,v);}),h.manualSync&&(b.get(`${u}/:repoName/force-sync`,(l,c)=>{let o=I(l,u),t=S.find(f=>f.name===l.params.repoName);if(!t){O(c,N("Not Found",o,`<p>Unknown repo: ${l.params.repoName}</p>`),404);return}let d=N(`Force Sync: ${t.name}`,o,`<div class="card">
56
- <p>This will read <strong>all</strong> documents from the <code>${t.name}</code> Firestore collection
57
- and upsert them into the <code>${t.tableName}</code> SQL table.</p>
55
+ </div>`);D(r,N);}),b.manualSync&&(p.get(`${d}/:repoName/force-sync`,(u,r)=>{let c=q(u,d),n=R.find(S=>S.name===u.params.repoName);if(!n){D(r,P("Not Found",c,`<p>Unknown repo: ${u.params.repoName}</p>`),404);return}let f=P(`Force Sync: ${n.name}`,c,`<div class="card">
56
+ <p>This will read <strong>all</strong> documents from the <code>${n.name}</code> Firestore collection
57
+ and upsert them into the <code>${n.tableName}</code> SQL table.</p>
58
58
  <p class="muted" style="margin:.75rem 0">This may take a while for large collections.</p>
59
- <form method="POST" action="${o}/${t.name}/force-sync">
59
+ <form method="POST" action="${c}/${n.name}/force-sync">
60
60
  <button type="submit" class="btn btn-primary">Start Force Sync</button>
61
61
  </form>
62
- </div>`);O(c,d);}),b.post(`${u}/:repoName/force-sync`,async(l,c)=>{let o=I(l,u),t=S.find(P=>P.name===l.params.repoName);if(!t){_(c,{error:`Unknown repo: ${l.params.repoName}`},404);return}let d=t.repo.ref;if(!d){_(c,{error:`No collection reference for "${t.name}"`},400);return}let f=0,i=0,p=[],C=500,w=d.limit(C),T=null;try{for(;;){let $=await(T?w.startAfter(T):w).get();if($.empty)break;for(let k of $.docs){let v=k.data(),m=String(v[t.documentKey]??k.id),A=z(v,{exclude:t.repoCfg?.exclude,columnMap:t.repoCfg?.columnMap});try{await a({operation:"UPSERT",repoName:t.name,docId:m,data:A,timestamp:new Date().toISOString()}),f++;}catch(Z){i++;let re=Z?.message??String(Z);console.error(`[ForceSync:${t.name}] doc=${m} failed:`,Z),p.length<5&&p.push(`${m}: ${re}`);}}if(T=$.docs[$.docs.length-1],$.docs.length<C)break}let P=n.get(t.name);P&&await P.flush();}catch(P){if(j(l)){_(c,{error:P?.message??String(P),synced:f,errors:i},500);return}O(c,N(`Force Sync: ${t.name}`,o,`<div class="card">
63
- <p class="badge badge-err">Error: ${P?.message??String(P)}</p>
64
- <p>Synced ${f} docs before failure (${i} errors).</p>
65
- </div>`),500);return}if(j(l)){_(c,{repo:t.name,table:t.tableName,synced:f,errors:i,...p.length>0&&{errorSamples:p}});return}let E=p.length>0?`<details style="margin-top:1rem"><summary>First ${p.length} error(s)</summary>
66
- <pre style="white-space:pre-wrap">${p.map(P=>P.replace(/[<>&]/g,x=>`&#${x.charCodeAt(0)};`)).join(`
62
+ </div>`);D(r,f);}),p.post(`${d}/:repoName/force-sync`,async(u,r)=>{let c=q(u,d),n=R.find(h=>h.name===u.params.repoName);if(!n){z(r,{error:`Unknown repo: ${u.params.repoName}`},404);return}let f=n.repo.ref;if(!f){z(r,{error:`No collection reference for "${n.name}"`},400);return}let S=0,i=0,y=[],v=500,x=f.limit(v),k=null;try{for(;;){let T=await(k?x.startAfter(k):x).get();if(T.empty)break;for(let C of T.docs){let N=C.data(),g=String(N[n.documentKey]??C.id),j=Z(N,{exclude:n.repoCfg?.exclude,columnMap:n.repoCfg?.columnMap});try{await a({operation:"UPSERT",repoName:n.name,docId:g,data:j,timestamp:new Date().toISOString()}),S++;}catch(U){i++;let oe=U?.message??String(U);console.error(`[ForceSync:${n.name}] doc=${g} failed:`,U),y.length<5&&y.push(`${g}: ${oe}`);}}if(k=T.docs[T.docs.length-1],T.docs.length<v)break}let h=t.get(n.name);h&&await h.flush();}catch(h){if(I(u)){z(r,{error:h?.message??String(h),synced:S,errors:i},500);return}D(r,P(`Force Sync: ${n.name}`,c,`<div class="card">
63
+ <p class="badge badge-err">Error: ${h?.message??String(h)}</p>
64
+ <p>Synced ${S} docs before failure (${i} errors).</p>
65
+ </div>`),500);return}if(I(u)){z(r,{repo:n.name,table:n.tableName,synced:S,errors:i,...y.length>0&&{errorSamples:y}});return}let _=y.length>0?`<details style="margin-top:1rem"><summary>First ${y.length} error(s)</summary>
66
+ <pre style="white-space:pre-wrap">${y.map(h=>h.replace(/[<>&]/g,$=>`&#${$.charCodeAt(0)};`)).join(`
67
67
 
68
- `)}</pre></details>`:"",D=N(`Force Sync: ${t.name}`,o,`<div class="card">
68
+ `)}</pre></details>`:"",O=P(`Force Sync: ${n.name}`,c,`<div class="card">
69
69
  <p class="badge ${i>0?"badge-warn":"badge-ok"}">${i>0?"Completed with errors":"Complete"}</p>
70
- <p>Synced <strong>${f}</strong> documents to <code>${t.tableName}</code>.</p>
70
+ <p>Synced <strong>${S}</strong> documents to <code>${n.tableName}</code>.</p>
71
71
  ${i>0?`<p class="badge badge-warn">${i} error(s)</p>`:""}
72
- ${E}
73
- </div>`);O(c,D);})),h.viewQueue&&b.get(`${u}/queues`,(l,c)=>{let o=I(l,u),t=[];for(let i of S){let p=n.get(i.name);t.push({repo:i.name,table:i.tableName,pending:p?p.size:0});}if(j(l)){_(c,{queues:t});return}let d=t.map(i=>`<tr><td>${i.repo}</td><td>${i.table}</td><td>${i.pending===0?'<span class="badge badge-ok">0</span>':`<span class="badge badge-warn">${i.pending}</span>`}</td></tr>`).join(`
74
- `),f=N("Sync Queues",o,`<div class="card">
72
+ ${_}
73
+ </div>`);D(r,O);})),b.viewQueue&&p.get(`${d}/queues`,(u,r)=>{let c=q(u,d),n=[];for(let i of R){let y=t.get(i.name);n.push({repo:i.name,table:i.tableName,pending:y?y.size:0});}if(I(u)){z(r,{queues:n});return}let f=n.map(i=>`<tr><td>${i.repo}</td><td>${i.table}</td><td>${i.pending===0?'<span class="badge badge-ok">0</span>':`<span class="badge badge-warn">${i.pending}</span>`}</td></tr>`).join(`
74
+ `),S=P("Sync Queues",c,`<div class="card">
75
75
  <table>
76
76
  <thead><tr><th>Repository</th><th>Table</th><th>Pending</th></tr></thead>
77
- <tbody>${d}</tbody>
77
+ <tbody>${f}</tbody>
78
78
  </table>
79
- </div>`);O(c,f);}),h.configCheck&&(b.get(`${u}/config-check`,async(l,c)=>{let o=I(l,u),t=process.env.GCLOUD_PROJECT??process.env.GOOGLE_CLOUD_PROJECT??process.env.GCP_PROJECT??"unknown",d="https://console.cloud.google.com",f=R??"firestore-sync",i=[];try{await e.tableExists("__nonexistent_health_check__"),i.push({name:"BigQuery API",category:"bigquery",status:"ok",message:"BigQuery API is reachable"});}catch(x){let $=x?.message??String(x),k=$.toLowerCase(),v=k.includes("disabled")||k.includes("has not been used")||k.includes("accessnotconfigured"),m=k.includes("permission")||$.includes("403")||k.includes("access denied"),A=k.includes("project")&&k.includes("not found"),Z=k.includes("not found")||$.includes("404");v?i.push({name:"BigQuery API",category:"bigquery",status:"error",message:"BigQuery API is not enabled",fix:{gcloud:`gcloud services enable bigquery.googleapis.com --project=${t}`,console:`${d}/apis/library/bigquery.googleapis.com?project=${t}`}}):A?i.push({name:"BigQuery Project",category:"bigquery",status:"error",message:$,fix:{hint:"The GCP project does not exist or the credentials don't have access to it. In the Firebase emulator, GCLOUD_PROJECT may override the configured projectId. Ensure you pass the correct projectId to the BigQuery constructor and have valid credentials.",console:`${d}/home/dashboard`}}):m?i.push({name:"BigQuery API",category:"bigquery",status:"error",message:`Permission denied: ${$}`,fix:{hint:"Grant the service account BigQuery Data Editor + BigQuery Job User roles",gcloud:[`SA=$(gcloud run services describe YOUR_SERVICE --region=YOUR_REGION --format="value(spec.template.spec.serviceAccountName)" --project=${t})`,`gcloud projects add-iam-policy-binding ${t} --member="serviceAccount:$SA" --role="roles/bigquery.dataEditor"`,`gcloud projects add-iam-policy-binding ${t} --member="serviceAccount:$SA" --role="roles/bigquery.jobUser"`].join(`
80
- `),console:`${d}/iam-admin/iam?project=${t}`}}):Z?i.push({name:"BigQuery Dataset",category:"bigquery",status:"error",message:`Dataset not found: ${$}`,fix:{hint:"Create the dataset first",gcloud:`bq mk --dataset ${t}:YOUR_DATASET_ID`,console:`${d}/bigquery?project=${t}`}}):i.push({name:"BigQuery API",category:"bigquery",status:"ok",message:"BigQuery API is reachable (table lookup returned expected error)"});}for(let x of S)try{let $=await e.tableExists(x.tableName);i.push({name:`Table: ${x.tableName}`,category:"bigquery",status:$?"ok":"warn",message:$?`Table \`${x.tableName}\` exists`:`Table \`${x.tableName}\` does not exist yet`,...!$&&{fix:{hint:"Table will be auto-created on first sync if autoMigrate is enabled. Or create it manually."}}});}catch($){i.push({name:`Table: ${x.tableName}`,category:"bigquery",status:"error",message:$?.message??String($)});}if(y)for(let x of S){let $=`${f}-${x.name}`;try{let k=y.topic($);if(typeof k.exists=="function"){let[v]=await k.exists();i.push({name:`Topic: ${$}`,category:"pubsub",status:v?"ok":"error",message:v?`Topic \`${$}\` exists`:`Topic \`${$}\` does not exist`,...!v&&{fix:{gcloud:`gcloud pubsub topics create ${$} --project=${t}`,console:`${d}/cloudpubsub/topic/list?project=${t}`}}});}else i.push({name:`Topic: ${$}`,category:"pubsub",status:"warn",message:"Cannot verify topic existence (PubSub client doesn't expose .exists())",fix:{gcloud:`gcloud pubsub topics create ${$} --project=${t}`,console:`${d}/cloudpubsub/topic/list?project=${t}`,hint:"Ensure the topic exists. It is auto-created by the Firebase emulator but must exist in production."}});}catch(k){let v=k?.message??String(k),m=v.includes("disabled")||v.includes("has not been used");if(i.push({name:m?"Pub/Sub API":`Topic: ${$}`,category:"pubsub",status:"error",message:m?"Pub/Sub API is not enabled":v,fix:m?{gcloud:`gcloud services enable pubsub.googleapis.com --project=${t}`,console:`${d}/apis/library/pubsub.googleapis.com?project=${t}`}:{gcloud:`gcloud pubsub topics create ${$} --project=${t}`,console:`${d}/cloudpubsub/topic/list?project=${t}`}}),m)break}}else i.push({name:"Pub/Sub Client",category:"pubsub",status:"warn",message:"PubSub client not available for config check"});if(j(l)){let x=i.every($=>$.status==="ok");_(c,{project:t,healthy:x,checks:i});return}let p=x=>x==="ok"?'<span class="badge badge-ok">OK</span>':x==="warn"?'<span class="badge badge-warn">WARN</span>':'<span class="badge badge-err">ERROR</span>',C={bigquery:i.filter(x=>x.category==="bigquery"),pubsub:i.filter(x=>x.category==="pubsub"),firestore:i.filter(x=>x.category==="firestore")},w=(x,$)=>{if($.length===0)return "";let k=$.map(v=>{let m="";if(v.fix){let A=[];v.fix.hint&&A.push(`<p class="muted">${v.fix.hint}</p>`),v.fix.gcloud&&A.push(`<pre>$ ${v.fix.gcloud}</pre>`),v.fix.console&&A.push(`<p><a href="${v.fix.console}" target="_blank">Open GCP Console \u2192</a></p>`),m=`<div style="margin-top:.5rem">${A.join("")}</div>`;}return `<tr>
81
- <td>${p(v.status)}</td>
82
- <td><strong>${v.name}</strong><br><span class="muted">${v.message}</span>${m}</td>
79
+ </div>`);D(r,S);}),b.configCheck&&p.get(`${d}/config-check`,async(u,r)=>{let c=q(u,d),n=process.env.GCLOUD_PROJECT??process.env.GOOGLE_CLOUD_PROJECT??process.env.GCP_PROJECT??"unknown",f="https://console.cloud.google.com",S=w??"firestore-sync",i=[];try{await e.tableExists("__nonexistent_health_check__"),i.push({name:"BigQuery API",category:"bigquery",status:"ok",message:"BigQuery API is reachable"});}catch(h){let $=h?.message??String(h),T=$.toLowerCase(),C=T.includes("disabled")||T.includes("has not been used")||T.includes("accessnotconfigured"),N=T.includes("permission")||$.includes("403")||T.includes("access denied"),g=T.includes("project")&&T.includes("not found"),j=T.includes("not found")||$.includes("404");C?i.push({name:"BigQuery API",category:"bigquery",status:"error",message:"BigQuery API is not enabled",fix:{gcloud:`gcloud services enable bigquery.googleapis.com --project=${n}`,console:`${f}/apis/library/bigquery.googleapis.com?project=${n}`}}):g?i.push({name:"BigQuery Project",category:"bigquery",status:"error",message:$,fix:{hint:"The GCP project does not exist or the credentials don't have access to it. In the Firebase emulator, GCLOUD_PROJECT may override the configured projectId. Ensure you pass the correct projectId to the BigQuery constructor and have valid credentials.",console:`${f}/home/dashboard`}}):N?i.push({name:"BigQuery API",category:"bigquery",status:"error",message:`Permission denied: ${$}`,fix:{hint:"Grant the service account BigQuery Data Editor + BigQuery Job User roles",gcloud:[`SA=$(gcloud run services describe YOUR_SERVICE --region=YOUR_REGION --format="value(spec.template.spec.serviceAccountName)" --project=${n})`,`gcloud projects add-iam-policy-binding ${n} --member="serviceAccount:$SA" --role="roles/bigquery.dataEditor"`,`gcloud projects add-iam-policy-binding ${n} --member="serviceAccount:$SA" --role="roles/bigquery.jobUser"`].join(`
80
+ `),console:`${f}/iam-admin/iam?project=${n}`}}):j?i.push({name:"BigQuery Dataset",category:"bigquery",status:"error",message:`Dataset not found: ${$}`,fix:{hint:"Create the dataset first",gcloud:`bq mk --dataset ${n}:YOUR_DATASET_ID`,console:`${f}/bigquery?project=${n}`}}):i.push({name:"BigQuery API",category:"bigquery",status:"ok",message:"BigQuery API is reachable (table lookup returned expected error)"});}for(let h of R)try{let $=await e.tableExists(h.tableName);i.push({name:`Table: ${h.tableName}`,category:"bigquery",status:$?"ok":"warn",message:$?`Table \`${h.tableName}\` exists`:`Table \`${h.tableName}\` does not exist yet`,...!$&&{fix:{hint:"Table will be auto-created on first sync if autoMigrate is enabled. Or create it manually."}}});}catch($){i.push({name:`Table: ${h.tableName}`,category:"bigquery",status:"error",message:$?.message??String($)});}if(m)for(let h of R){let $=`${S}-${h.name}`;try{let T=m.topic($);if(typeof T.exists=="function"){let[C]=await T.exists();i.push({name:`Topic: ${$}`,category:"pubsub",status:C?"ok":"error",message:C?`Topic \`${$}\` exists`:`Topic \`${$}\` does not exist`,...!C&&{fix:{gcloud:`gcloud pubsub topics create ${$} --project=${n}`,console:`${f}/cloudpubsub/topic/list?project=${n}`}}});}else i.push({name:`Topic: ${$}`,category:"pubsub",status:"warn",message:"Cannot verify topic existence (PubSub client doesn't expose .exists())",fix:{gcloud:`gcloud pubsub topics create ${$} --project=${n}`,console:`${f}/cloudpubsub/topic/list?project=${n}`,hint:"Ensure the topic exists. It is auto-created by the Firebase emulator but must exist in production."}});}catch(T){let C=T?.message??String(T),N=C.includes("disabled")||C.includes("has not been used");if(i.push({name:N?"Pub/Sub API":`Topic: ${$}`,category:"pubsub",status:"error",message:N?"Pub/Sub API is not enabled":C,fix:N?{gcloud:`gcloud services enable pubsub.googleapis.com --project=${n}`,console:`${f}/apis/library/pubsub.googleapis.com?project=${n}`}:{gcloud:`gcloud pubsub topics create ${$} --project=${n}`,console:`${f}/cloudpubsub/topic/list?project=${n}`}}),N)break}}else i.push({name:"Pub/Sub Client",category:"pubsub",status:"warn",message:"PubSub client not available for config check"});if(I(u)){let h=i.every($=>$.status==="ok");z(r,{project:n,healthy:h,checks:i});return}let y=h=>h==="ok"?'<span class="badge badge-ok">OK</span>':h==="warn"?'<span class="badge badge-warn">WARN</span>':'<span class="badge badge-err">ERROR</span>',v={bigquery:i.filter(h=>h.category==="bigquery"),pubsub:i.filter(h=>h.category==="pubsub"),firestore:i.filter(h=>h.category==="firestore")},x=(h,$)=>{if($.length===0)return "";let T=$.map(C=>{let N="";if(C.fix){let g=[];C.fix.hint&&g.push(`<p class="muted">${C.fix.hint}</p>`),C.fix.gcloud&&g.push(`<pre>$ ${C.fix.gcloud}</pre>`),C.fix.console&&g.push(`<p><a href="${C.fix.console}" target="_blank">Open GCP Console \u2192</a></p>`),N=`<div style="margin-top:.5rem">${g.join("")}</div>`;}return `<tr>
81
+ <td>${y(C.status)}</td>
82
+ <td><strong>${C.name}</strong><br><span class="muted">${C.message}</span>${N}</td>
83
83
  </tr>`}).join(`
84
- `);return `<h2>${x}</h2>
84
+ `);return `<h2>${h}</h2>
85
85
  <table><thead><tr><th style="width:80px">Status</th><th>Check</th></tr></thead>
86
- <tbody>${k}</tbody></table>`},E=i.every(x=>x.status==="ok")?'<span class="badge badge-ok">All checks passed</span>':'<span class="badge badge-warn">Some issues found</span>',D=y?`<form method="POST" action="${o}/config-check/setup-pubsub" style="display:inline">
87
- <button type="submit" class="btn btn-primary">\u2699 Setup Pub/Sub (topics + subscriptions)</button>
88
- </form>
89
- <p class="muted" style="margin-top:.5rem">
90
- Idempotent. Creates missing topics and subscriptions with
91
- <code>enableMessageOrdering=${s.pubsubSetup?.ordering??true}</code>.
92
- Existing subscriptions are kept as-is (the ordering flag is immutable).
93
- </p>`:"",P=N("Config Check",o,`<div class="card">
94
- <p>Project: <code>${t}</code> ${E}</p>
95
- ${w("BigQuery",C.bigquery)}
96
- ${w("Pub/Sub",C.pubsub)}
97
- ${w("Firestore",C.firestore)}
98
- ${D?`<hr style="margin:1.5rem 0"><h2>Actions</h2>${D}`:""}
99
- </div>`);O(c,P);}),y&&b.post(`${u}/config-check/setup-pubsub`,async(l,c)=>{let o=I(l,u),t=s.pubsubSetup??{};try{let d=await U(r,{pubsub:y,topicPrefix:R??"firestore-sync",ordering:t.ordering??!0,subscriptionSuffix:t.subscriptionSuffix??"sync-sub",includeDLQ:t.includeDLQ??!0,ackDeadlineSeconds:t.ackDeadlineSeconds??60,...t.messageRetentionDuration&&{messageRetentionDuration:t.messageRetentionDuration}});if(j(l)){_(c,{ok:!0,...d});return}let f=d.topics.map(p=>`<tr><td><code>${p.name}</code></td><td>${p.created?'<span class="badge badge-ok">created</span>':'<span class="badge">already exists</span>'}</td></tr>`).join(`
100
- `),i=d.subscriptions.map(p=>`<tr>
101
- <td><code>${p.name}</code></td>
102
- <td><code>${p.topic}</code></td>
103
- <td>${p.created?'<span class="badge badge-ok">created</span>':'<span class="badge">already exists</span>'}</td>
104
- <td>${p.orderingEnabled?"\u2713":"\u2717"}</td>
105
- <td>${p.warning?`<span class="badge badge-warn">${p.warning}</span>`:""}</td>
106
- </tr>`).join(`
107
- `);O(c,N("Pub/Sub Setup",o,`<div class="card">
108
- <p><span class="badge badge-ok">Setup complete</span></p>
109
- <h2>Topics</h2>
110
- <table><thead><tr><th>Name</th><th>Status</th></tr></thead>
111
- <tbody>${f}</tbody></table>
112
- <h2>Subscriptions</h2>
113
- <table><thead><tr><th>Name</th><th>Topic</th><th>Status</th><th>Ordering</th><th>Notes</th></tr></thead>
114
- <tbody>${i}</tbody></table>
115
- <p style="margin-top:1rem"><a class="btn" href="${o}/config-check">\u2190 Back to Config Check</a></p>
116
- </div>`));}catch(d){let f=d?.message??String(d);if(j(l)){_(c,{ok:false,error:f},500);return}O(c,N("Pub/Sub Setup \u2014 Error",o,`<div class="card">
117
- <p><span class="badge badge-err">Setup failed</span></p>
118
- <pre>${f}</pre>
119
- <p><a class="btn" href="${o}/config-check">\u2190 Back</a></p>
120
- </div>`),500);}})),async(l,c)=>{await b.handle(l,c);}}var de="firestore-sync";function ue(r,e){let n=e.ref?.path??void 0;return n?`${n}/{docId}`:(console.warn(`[SyncTriggers] Cannot determine collection path for "${r}". Skipping.`),null)}function H(r,e){let{onDocumentCreated:n,onDocumentUpdated:a,onDocumentDeleted:s}=e.deps.firestoreTriggers,g=e.deps.pubsub,y=e?.topicPrefix??de,R={},u=e?.ordering,h=!!u,S=typeof u=="function"?u:u===true?o=>o.docId:null,b=new Map;function l(o){let t=b.get(o);return t||(t=h?g.topic(o,{messageOrdering:true}):g.topic(o),b.set(o,t),t)}async function c(o,t){let d=l(o),f=S?S(t):void 0;try{await d.publishMessage(f!==void 0?{json:t,orderingKey:f}:{json:t});}catch(i){throw f!==void 0&&typeof d.resumePublishing=="function"&&d.resumePublishing(f),i}}for(let[o,t]of Object.entries(r)){let d=e?.repos?.[o],f;if(t._isGroup){if(!d?.triggerPath){console.warn(`[SyncTriggers] Skipping collection-group repo "${o}". Provide a triggerPath in the sync repos config for group collections.`);continue}f=d.triggerPath;}else f=d?.triggerPath??ue(o,t);if(!f)continue;let i=t._systemKeys?.[0]??"docId",p=`${y}-${o}`;R[`${o}_onCreate`]=n(f,async C=>{let w=C.data;if(!w)return;let T=w.data();if(!T)return;let E=String(T[i]??w.id),D=z(T,{exclude:d?.exclude,columnMap:d?.columnMap}),P={operation:"INSERT",repoName:o,docId:E,data:D,timestamp:new Date().toISOString()};await c(p,P);}),R[`${o}_onUpdate`]=a(f,async C=>{let w=C.data?.after;if(!w)return;let T=w.data();if(!T)return;let E=String(T[i]??w.id),D=z(T,{exclude:d?.exclude,columnMap:d?.columnMap}),P={operation:"UPSERT",repoName:o,docId:E,data:D,timestamp:new Date().toISOString()};await c(p,P);}),R[`${o}_onDelete`]=s(f,async C=>{let w=C.data;if(!w)return;let T=w.data(),E=String(T?.[i]??w.id),D={operation:"DELETE",repoName:o,docId:E,data:null,timestamp:new Date().toISOString()};await c(p,D);});}return R}var M=class{constructor(e){this.buffer=[];this.flushing=false;this.timer=null;this.adapter=e.adapter,this.tableName=e.tableName,this.primaryKey=e.primaryKey,this.batchSize=e.batchSize??100,this.onFlushError=e.onFlushError;let n=e.flushIntervalMs??5e3;n>0&&(this.timer=setInterval(()=>{this.flush();},n),typeof this.timer=="object"&&"unref"in this.timer&&this.timer.unref());}get size(){return this.buffer.length}enqueue(...e){this.buffer.push(...e),this.buffer.length>=this.batchSize&&this.flush();}async flush(){if(this.flushing||this.buffer.length===0)return;this.flushing=true;let e=this.buffer.splice(0,this.batchSize);try{let n=[],a=[];for(let s of e)s.operation==="DELETE"?a.push(s.docId):s.data&&n.push(s.data);n.length>0&&await this.adapter.upsertRows(this.tableName,n,this.primaryKey),a.length>0&&await this.adapter.deleteRows(this.tableName,this.primaryKey,a);}catch(n){this.onFlushError?await this.onFlushError(e,n).catch(a=>{console.error(`[SyncQueue] Flush error for ${this.tableName}:`,n),console.error("[SyncQueue] Error handler also failed:",a);}):(this.buffer.unshift(...e),console.error(`[SyncQueue] Flush failed for ${this.tableName}:`,n));}finally{this.flushing=false;}}async shutdown(){this.timer&&(clearInterval(this.timer),this.timer=null),await this.flush();}};var ee=new Set;async function le(r,e,n,a,s,g,y){if(ee.has(r))return;let R=q(n,e.dialect,{primaryKey:s,exclude:g,columnMap:y});if(!await e.tableExists(a))await e.createTable({tableName:a,columns:R});else {let h=new Set(await e.getTableColumns(a)),S=R.filter(b=>!h.has(b.name));S.length>0&&await e.addColumns(a,S);}ee.add(r);}function G(r,e){let{deps:n,adapter:a,batchSize:s=100,flushIntervalMs:g=5e3,autoMigrate:y=false,topicPrefix:R="firestore-sync",repos:u={}}=e,h=new Map;function S(c,o){let t=h.get(c);if(t)return t;let f=u[c]?.tableName??c,i=async(p,C)=>{console.error(`[SyncWorker] Flush failed for "${c}" (${p.length} events):`,C);try{let w=`${R}-${c}-dlq`,T=n.pubsub.topic(w),[E]=await T.exists();E||(await T.create(),console.info(`[SyncWorker] Created DLQ topic "${w}"`));for(let D of p)await T.publishMessage({json:D});}catch(w){console.error(`[SyncWorker] Dead-letter publish also failed for ${c}:`,w);}};return t=new M({adapter:a,tableName:f,primaryKey:o,batchSize:s,flushIntervalMs:g,onFlushError:i}),h.set(c,t),t}async function b(c){let{repoName:o}=c,t=r[o];if(!t){console.warn(`[SyncWorker] Unknown repo "${o}", skipping event`);return}let d=t._systemKeys?.[0]??t.documentKey??"docId",f=u[o],i=f?.columnMap,p=i?.[d]??d;if(y){let w=t.schema??void 0;if(w){let T=f?.tableName??o;await le(o,a,w,T,d,f?.exclude,i);}}S(o,p).enqueue(c);}function l(c){return n.pubsubHandler.onMessagePublished(c,async o=>{let t=o.data?.message?.json??o.data?.json;if(!t){console.warn("[SyncWorker] Received empty PubSub message");return}await b(t);let d=h.get(t.repoName);d&&await d.flush();})}return {handleMessage:b,createHandler:l,queues:h,async shutdown(){let c=[];for(let o of h.values())c.push(o.shutdown());await Promise.all(c);}}}var pe="firestore-sync";function te(r){if(typeof r!="function")return r;let e=r,n;return new Proxy({},{get(a,s){return n||(n=e()),n[s]},has(a,s){return n||(n=e()),s in n}})}function fe(r,e){let{deps:n,adapter:a,topicPrefix:s=pe,batchSize:g,flushIntervalMs:y,autoMigrate:R,admin:u,repos:h,ordering:S}=e,b=te(n.pubsub),l=te(a),c=H(r,{deps:{firestoreTriggers:n.firestoreTriggers,pubsub:b},topicPrefix:s,repos:h,...S!==void 0&&{ordering:S}}),o=G(r,{deps:{pubsubHandler:n.pubsubHandler,pubsub:b},adapter:l,batchSize:g,flushIntervalMs:y,autoMigrate:R,topicPrefix:s,repos:h}),t={};for(let i of Object.keys(r))t[`sync_${i}`]=o.createHandler(`${s}-${i}`);let d=null;u&&(d=B(r,l,o.queues,o.handleMessage,u,h??{},b,s),t.adminsync=u.onRequest?u.httpsOptions?u.onRequest(u.httpsOptions,d):u.onRequest(d):d);let f={functions:{...c,...t},adminHandler:d,handleMessage:o.handleMessage,queues:o.queues,shutdown:o.shutdown};for(let i of ["adminHandler","handleMessage","queues","shutdown"])Object.defineProperty(f,i,{enumerable:false});return f}function ne(r,e){let n=e.columns.map(a=>{let s=a.isPrimaryKey?" NOT NULL":"";return ` ${r.quoteIdentifier(a.name)} ${a.sqlType}${s}`}).join(`,
121
- `);return `CREATE TABLE IF NOT EXISTS ${r.quoteIdentifier(e.tableName)} (
122
- ${n}
123
- );`}function ge(r,e,n){return n.map(a=>`ALTER TABLE ${r.quoteIdentifier(e)} ADD COLUMN ${r.quoteIdentifier(a.name)} ${a.sqlType};`).join(`
124
- `)}function me(r,e,n){let a=[];for(let[s,g]of Object.entries(r)){let y=g.schema??g._schema??void 0;if(!y)continue;let R=n?.repos?.[s],u=R?.tableName??s,h=g._systemKeys?.[0]??g.documentKey??"docId",S=q(y,e,{primaryKey:h,exclude:R?.exclude,columnMap:R?.columnMap}),b={tableName:u,columns:S};a.push(ne(e,b));}return a.join(`
86
+ <tbody>${T}</tbody></table>`},_=i.every(h=>h.status==="ok")?'<span class="badge badge-ok">All checks passed</span>':'<span class="badge badge-warn">Some issues found</span>',O=P("Config Check",c,`<div class="card">
87
+ <p>Project: <code>${n}</code> ${_}</p>
88
+ ${x("BigQuery",v.bigquery)}
89
+ ${x("Pub/Sub",v.pubsub)}
90
+ ${x("Firestore",v.firestore)}
91
+ </div>`);D(r,O);}),async(u,r)=>{await p.handle(u,r);}}var de="firestore-sync";function ue(o,e){let t=e.ref?.path??void 0;return t?`${t}/{docId}`:(console.warn(`[SyncTriggers] Cannot determine collection path for "${o}". Skipping.`),null)}function H(o,e){let{onDocumentCreated:t,onDocumentUpdated:a,onDocumentDeleted:s}=e.deps.firestoreTriggers,l=e.deps.pubsub,m=e?.topicPrefix??de,w={},d=new Map;function b(p){let u=d.get(p);return u||(u=l.topic(p),d.set(p,u),u)}async function R(p,u){await b(p).publishMessage({json:u});}for(let[p,u]of Object.entries(o)){let r=e?.repos?.[p],c;if(u._isGroup){if(!r?.triggerPath){console.warn(`[SyncTriggers] Skipping collection-group repo "${p}". Provide a triggerPath in the sync repos config for group collections.`);continue}c=r.triggerPath;}else c=r?.triggerPath??ue(p,u);if(!c)continue;let n=u._systemKeys?.[0]??"docId",f=`${m}-${p}`;w[`${p}_onCreate`]=t(c,async S=>{let i=S.data;if(!i)return;let y=i.data();if(!y)return;let v=String(y[n]??i.id),x=Z(y,{exclude:r?.exclude,columnMap:r?.columnMap}),k={operation:"INSERT",repoName:p,docId:v,data:x,timestamp:new Date().toISOString(),version:Date.now()};await R(f,k);}),w[`${p}_onUpdate`]=a(c,async S=>{let i=S.data?.after;if(!i)return;let y=i.data();if(!y)return;let v=String(y[n]??i.id),x=Z(y,{exclude:r?.exclude,columnMap:r?.columnMap}),k={operation:"UPSERT",repoName:p,docId:v,data:x,timestamp:new Date().toISOString(),version:Date.now()};await R(f,k);}),w[`${p}_onDelete`]=s(c,async S=>{let i=S.data;if(!i)return;let y=i.data(),v=String(y?.[n]??i.id),x={operation:"DELETE",repoName:p,docId:v,data:null,timestamp:new Date().toISOString(),version:Date.now()};await R(f,x);});}return w}var M=class{constructor(e){this.buffer=[];this.flushing=false;this.timer=null;this.adapter=e.adapter,this.tableName=e.tableName,this.primaryKey=e.primaryKey,this.batchSize=e.batchSize??100,this.onFlushError=e.onFlushError;let t=e.flushIntervalMs??5e3;t>0&&(this.timer=setInterval(()=>{this.flush();},t),typeof this.timer=="object"&&"unref"in this.timer&&this.timer.unref());}get size(){return this.buffer.length}enqueue(...e){this.buffer.push(...e),this.buffer.length>=this.batchSize&&this.flush();}async flush(){if(this.flushing||this.buffer.length===0)return;this.flushing=true;let e=this.buffer.splice(0,this.batchSize);try{let t=new Map,a=[];for(let l of e)if(l.operation==="DELETE")a.push(l.docId),t.delete(l.docId);else if(l.data){let m=t.get(l.docId);if(!m)t.set(l.docId,l.data);else {let w=Number(m[A]??0);Number(l.data[A]??0)>=w&&t.set(l.docId,l.data);}}let s=Array.from(t.values());s.length>0&&await this.adapter.upsertRows(this.tableName,s,this.primaryKey),a.length>0&&await this.adapter.deleteRows(this.tableName,this.primaryKey,a);}catch(t){this.onFlushError?await this.onFlushError(e,t).catch(a=>{console.error(`[SyncQueue] Flush error for ${this.tableName}:`,t),console.error("[SyncQueue] Error handler also failed:",a);}):(this.buffer.unshift(...e),console.error(`[SyncQueue] Flush failed for ${this.tableName}:`,t));}finally{this.flushing=false;}}async shutdown(){this.timer&&(clearInterval(this.timer),this.timer=null),await this.flush();}};var ee=new Set;async function le(o,e,t,a,s,l,m){if(ee.has(o))return;let w=E(t,e.dialect,{primaryKey:s,exclude:l,columnMap:m});if(!await e.tableExists(a))await e.createTable({tableName:a,columns:w});else {let b=new Set(await e.getTableColumns(a)),R=w.filter(p=>!b.has(p.name));R.length>0&&await e.addColumns(a,R);}ee.add(o);}function G(o,e){let{deps:t,adapter:a,batchSize:s=100,flushIntervalMs:l=5e3,autoMigrate:m=false,topicPrefix:w="firestore-sync",repos:d={}}=e,b=new Map;function R(r,c){let n=b.get(r);if(n)return n;let S=d[r]?.tableName??r,i=async(y,v)=>{console.error(`[SyncWorker] Flush failed for "${r}" (${y.length} events):`,v);try{let x=`${w}-${r}-dlq`,k=t.pubsub.topic(x),[_]=await k.exists();_||(await k.create(),console.info(`[SyncWorker] Created DLQ topic "${x}"`));for(let O of y)await k.publishMessage({json:O});}catch(x){console.error(`[SyncWorker] Dead-letter publish also failed for ${r}:`,x);}};return n=new M({adapter:a,tableName:S,primaryKey:c,batchSize:s,flushIntervalMs:l,onFlushError:i}),b.set(r,n),n}async function p(r){let{repoName:c}=r,n=o[c];if(!n){console.warn(`[SyncWorker] Unknown repo "${c}", skipping event`);return}let f=n._systemKeys?.[0]??n.documentKey??"docId",S=d[c],i=S?.columnMap,y=i?.[f]??f;if(m){let x=n.schema??void 0;if(x){let k=S?.tableName??c;await le(c,a,x,k,f,S?.exclude,i);}}let v=R(c,y);r.data&&(r.data[A]=r.version??Date.now()),v.enqueue(r);}function u(r){return t.pubsubHandler.onMessagePublished(r,async c=>{let n=c.data?.message?.json??c.data?.json;if(!n){console.warn("[SyncWorker] Received empty PubSub message");return}await p(n);let f=b.get(n.repoName);f&&await f.flush();})}return {handleMessage:p,createHandler:u,queues:b,async shutdown(){let r=[];for(let c of b.values())r.push(c.shutdown());await Promise.all(r);}}}var pe="firestore-sync";function te(o){if(typeof o!="function")return o;let e=o,t;return new Proxy({},{get(a,s){return t||(t=e()),t[s]},has(a,s){return t||(t=e()),s in t}})}function fe(o,e){let{deps:t,adapter:a,topicPrefix:s=pe,batchSize:l,flushIntervalMs:m,autoMigrate:w,admin:d,repos:b}=e,R=te(t.pubsub),p=te(a),u=H(o,{deps:{firestoreTriggers:t.firestoreTriggers,pubsub:R},topicPrefix:s,repos:b}),r=G(o,{deps:{pubsubHandler:t.pubsubHandler,pubsub:R},adapter:p,batchSize:l,flushIntervalMs:m,autoMigrate:w,topicPrefix:s,repos:b}),c={};for(let S of Object.keys(o))c[`sync_${S}`]=r.createHandler(`${s}-${S}`);let n=null;d&&(n=B(o,p,r.queues,r.handleMessage,d,b??{},R,s),c.adminsync=d.onRequest?d.httpsOptions?d.onRequest(d.httpsOptions,n):d.onRequest(n):n);let f={functions:{...u,...c},adminHandler:n,handleMessage:r.handleMessage,queues:r.queues,shutdown:r.shutdown};for(let S of ["adminHandler","handleMessage","queues","shutdown"])Object.defineProperty(f,S,{enumerable:false});return f}function ne(o,e){let t=e.columns.map(a=>{let s=a.isPrimaryKey?" NOT NULL":"";return ` ${o.quoteIdentifier(a.name)} ${a.sqlType}${s}`}).join(`,
92
+ `);return `CREATE TABLE IF NOT EXISTS ${o.quoteIdentifier(e.tableName)} (
93
+ ${t}
94
+ );`}function me(o,e,t){return t.map(a=>`ALTER TABLE ${o.quoteIdentifier(e)} ADD COLUMN ${o.quoteIdentifier(a.name)} ${a.sqlType};`).join(`
95
+ `)}function ge(o,e,t){let a=[];for(let[s,l]of Object.entries(o)){let m=l.schema??l._schema??void 0;if(!m)continue;let w=t?.repos?.[s],d=w?.tableName??s,b=l._systemKeys?.[0]??l.documentKey??"docId",R=E(m,e,{primaryKey:b,exclude:w?.exclude,columnMap:w?.columnMap}),p={tableName:d,columns:R};a.push(ne(e,p));}return a.join(`
125
96
 
126
- `)}async function ye(r,e,n){let a={created:[],altered:[],upToDate:[],skipped:[]};for(let[s,g]of Object.entries(r)){let y=g.schema??void 0;if(!y){a.skipped.push(s);continue}let R=n?.repos?.[s],u=R?.tableName??s,h=g._systemKeys?.[0]??g.documentKey??"docId",S=q(y,e.dialect,{primaryKey:h,exclude:R?.exclude,columnMap:R?.columnMap}),b={tableName:u,columns:S};if(!await e.tableExists(u))await e.createTable(b),a.created.push(u);else {let c=new Set(await e.getTableColumns(u)),o=S.filter(t=>!c.has(t.name));o.length>0?(await e.addColumns(u,o),a.altered.push(u)):a.upToDate.push(u);}}return a}exports.SyncQueue=M;exports.addColumnsDDL=ge;exports.autoMigrate=ye;exports.createFirestoreSync=fe;exports.createSyncTriggers=H;exports.createSyncWorker=G;exports.createTableDDL=ne;exports.createadminsyncServer=B;exports.ensureSyncInfra=U;exports.generateDDL=me;exports.serializeDocument=z;exports.serializeValue=Q;exports.zodSchemaToColumns=q;exports.zodTypeToLogical=ce;//# sourceMappingURL=index.cjs.map
97
+ `)}async function ye(o,e,t){let a={created:[],altered:[],upToDate:[],skipped:[]};for(let[s,l]of Object.entries(o)){let m=l.schema??void 0;if(!m){a.skipped.push(s);continue}let w=t?.repos?.[s],d=w?.tableName??s,b=l._systemKeys?.[0]??l.documentKey??"docId",R=E(m,e.dialect,{primaryKey:b,exclude:w?.exclude,columnMap:w?.columnMap}),p={tableName:d,columns:R};if(!await e.tableExists(d))await e.createTable(p),a.created.push(d);else {let r=new Set(await e.getTableColumns(d)),c=R.filter(n=>!r.has(n.name));c.length>0?(await e.addColumns(d,c),a.altered.push(d)):a.upToDate.push(d);}}return a}exports.SyncQueue=M;exports.addColumnsDDL=me;exports.autoMigrate=ye;exports.createFirestoreSync=fe;exports.createSyncTriggers=H;exports.createSyncWorker=G;exports.createTableDDL=ne;exports.createadminsyncServer=B;exports.generateDDL=ge;exports.serializeDocument=Z;exports.serializeValue=Q;exports.zodSchemaToColumns=E;exports.zodTypeToLogical=ce;//# sourceMappingURL=index.cjs.map
127
98
  //# sourceMappingURL=index.cjs.map