@nicnocquee/dataqueue 1.32.0 → 1.33.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/cli.js.map CHANGED
@@ -1 +1 @@
1
- {"version":3,"sources":["../src/cli.ts"],"names":[],"mappings":";;;;;AAKA,IAAM,UAAA,GAAa,aAAA,CAAc,MAAA,CAAA,IAAA,CAAY,GAAG,CAAA;AAChD,IAAM,SAAA,GAAY,IAAA,CAAK,OAAA,CAAQ,UAAU,CAAA;AAUlC,SAAS,OACd,IAAA,EACA;AAAA,EACE,MAAM,OAAA,CAAQ,GAAA;AAAA,EACd,IAAA,GAAO,CAAC,IAAA,KAAiB,OAAA,CAAQ,KAAK,IAAI,CAAA;AAAA,EAC1C,aAAA,GAAgB,SAAA;AAAA,EAChB,aAAA,GAAgB,IAAA,CAAK,IAAA,CAAK,SAAA,EAAW,eAAe;AACtD,CAAA,GAAa,EAAC,EACR;AACN,EAAA,MAAM,KAAK,OAAA,EAAS,GAAG,QAAQ,CAAA,GAAI,IAAA;AAEnC,EAAA,SAAS,UAAA,GAAa;AACpB,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA,CAAI,EAAE,CAAA;AACN,IAAA,GAAA,CAAI,UAAU,CAAA;AACd,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA,CAAI,EAAE,CAAA;AACN,IAAA,GAAA,CAAI,QAAQ,CAAA;AACZ,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,IAAA,CAAK,CAAC,CAAA;AAAA;AAGR,EAAA,IAAI,YAAY,SAAA,EAAW;AAEzB,IAAA,IAAI,YAAsB,EAAC;AAC3B,IAAA,MAAM,MAAA,GAAS,QAAA,CAAS,OAAA,CAAQ,IAAI,CAAA;AACpC,IAAA,MAAM,WAAA,GAAc,QAAA,CAAS,OAAA,CAAQ,UAAU,CAAA;AAC/C,IAAA,IAAI,MAAA,KAAW,EAAA,IAAM,QAAA,CAAS,MAAA,GAAS,CAAC,CAAA,EAAG;AACzC,MAAA,SAAA,GAAY,CAAC,IAAA,EAAM,QAAA,CAAS,MAAA,GAAS,CAAC,GAAG,iBAAiB,CAAA;AAC1D,MAAA,QAAA,CAAS,MAAA,CAAO,QAAQ,CAAC,CAAA;AAAA,eAChB,WAAA,KAAgB,EAAA,IAAM,QAAA,CAAS,WAAA,GAAc,CAAC,CAAA,EAAG;AAC1D,MAAA,SAAA,GAAY,CAAC,IAAA,EAAM,QAAA,CAAS,WAAA,GAAc,CAAC,GAAG,iBAAiB,CAAA;AAC/D,MAAA,QAAA,CAAS,MAAA,CAAO,aAAa,CAAC,CAAA;AAAA;AAIhC,IAAA,IAAI,aAAuB,EAAC;AAC5B,IAAA,MAAM,YAAA,GAAe,QAAA,CAAS,OAAA,CAAQ,WAAW,CAAA;AACjD,IAAA,IAAI,YAAA,KAAiB,EAAA,IAAM,QAAA,CAAS,YAAA,GAAe,CAAC,CAAA,EAAG;AACrD,MAAA,UAAA,GAAa,CAAC,WAAA,EAAa,QAAA,CAAS,YAAA,GAAe,CAAC,CAAC,CAAA;AAAA;AAGvD,IAAA,MAAM,MAAA,GAAgC,aAAA;AAAA,MACpC,KAAA;AAAA,MACA;AAAA,QACE,iBAAA;AAAA,QACA,IAAA;AAAA,QACA,IAAA;AAAA,QACA,yBAAA;AAAA,QACA,IAAA;AAAA,QACA,uBAAA;AAAA,QACA,IAAA;AAAA,QACA,aAAA;AAAA,QACA,GAAG,SAAA;AAAA,QACH,GAAG,UAAA;AAAA,QACH,GAAG;AAAA,OACL;AAAA,MACA,EAAE,OAAO,SAAA;AAAU,KACrB;AACA,IAAA,IAAA,CAAK,MAAA,CAAO,UAAU,CAAC,CAAA;AAAA,GACzB,MAAO;AACL,IAAA,UAAA,EAAW;AAAA;AAEf","file":"cli.js","sourcesContent":["// Testable CLI logic for dataqueue\nimport { spawnSync, SpawnSyncReturns } from 'child_process';\nimport path from 'path';\nimport { fileURLToPath } from 'url';\n\nconst __filename = fileURLToPath(import.meta.url);\nconst __dirname = path.dirname(__filename);\n\nexport interface CliDeps {\n log?: (...args: any[]) => void;\n error?: (...args: any[]) => void;\n exit?: (code: number) => void;\n spawnSyncImpl?: (...args: any[]) => SpawnSyncReturns<any>;\n migrationsDir?: string;\n}\n\nexport function runCli(\n argv: string[],\n {\n log = console.log,\n exit = (code: number) => process.exit(code),\n spawnSyncImpl = spawnSync,\n migrationsDir = path.join(__dirname, '../migrations'),\n }: CliDeps = {},\n): void {\n const [, , command, ...restArgs] = argv;\n\n function printUsage() {\n log(\n 'Usage: dataqueue-cli migrate [--envPath <path>] [-s <schema> | --schema <schema>]',\n );\n log('');\n log('Options:');\n log(\n ' --envPath <path> Path to a .env file to load environment variables (passed to node-pg-migrate)',\n );\n log(\n ' -s, --schema <schema> Set the schema to use (passed to node-pg-migrate)',\n );\n log('');\n log('Notes:');\n log(\n ' - The PG_DATAQUEUE_DATABASE environment variable must be set to your Postgres connection string.',\n );\n log(\n ' - For managed Postgres (e.g., DigitalOcean) with SSL, set PGSSLMODE=require and PGSSLROOTCERT to your CA .crt file.',\n );\n log(\n ' Example: PGSSLMODE=require NODE_EXTRA_CA_CERTS=/absolute/path/to/ca.crt PG_DATAQUEUE_DATABASE=... npx dataqueue-cli migrate',\n );\n exit(1);\n }\n\n if (command === 'migrate') {\n // Support for -s or --schema argument\n let schemaArg: string[] = [];\n const sIndex = restArgs.indexOf('-s');\n const schemaIndex = restArgs.indexOf('--schema');\n if (sIndex !== -1 && restArgs[sIndex + 1]) {\n schemaArg = ['-s', restArgs[sIndex + 1], '--create-schema'];\n restArgs.splice(sIndex, 2);\n } else if (schemaIndex !== -1 && restArgs[schemaIndex + 1]) {\n schemaArg = ['-s', restArgs[schemaIndex + 1], '--create-schema'];\n restArgs.splice(schemaIndex, 2);\n }\n\n // Support for --envPath argument\n let envPathArg: string[] = [];\n const envPathIndex = restArgs.indexOf('--envPath');\n if (envPathIndex !== -1 && restArgs[envPathIndex + 1]) {\n envPathArg = ['--envPath', restArgs[envPathIndex + 1]];\n }\n\n const result: SpawnSyncReturns<any> = spawnSyncImpl(\n 'npx',\n [\n 'node-pg-migrate',\n 'up',\n '-t',\n 'dataqueuedev_migrations',\n '-d',\n 'PG_DATAQUEUE_DATABASE',\n '-m',\n migrationsDir,\n ...schemaArg,\n ...envPathArg,\n ...restArgs,\n ],\n { stdio: 'inherit' },\n );\n exit(result.status ?? 1);\n } else {\n printUsage();\n }\n}\n"]}
1
+ {"version":3,"sources":["../src/init-command.ts","../src/cli.ts"],"names":["path"],"mappings":";;;;;;AAYA,IAAM,mBAAA,GAAsB;AAAA,EAC1B,uBAAA;AAAA,EACA,iCAAA;AAAA,EACA;AACF,CAAA;AAEA,IAAM,uBAAA,GAA0B;AAAA,EAC9B,YAAA;AAAA,EACA,SAAA;AAAA,EACA;AACF,CAAA;AAEA,IAAM,cAAA,GAAiB;AAAA,EACrB,IAAA,EAAM,cAAA;AAAA,EACN,mBAAA,EAAqB;AACvB,CAAA;AAKO,IAAM,yBAAA,GAA4B,CAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,CAAA;AAgIlC,IAAM,2BAAA,GAA8B,CAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,CAAA;AA0GpC,IAAM,gBAAA,GAAmB,CAAA;;AAAA;AAAA;;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA,CAAA;AA+CzB,IAAM,cAAA,GAAiB,CAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,CAAA;AAyDvB,SAAS,OAAA,CAAQ;AAAA,EACtB,MAAM,OAAA,CAAQ,GAAA;AAAA,EACd,QAAQ,OAAA,CAAQ,KAAA;AAAA,EAChB,IAAA,GAAO,CAAC,IAAA,KAAiB,OAAA,CAAQ,KAAK,IAAI,CAAA;AAAA,EAC1C,GAAA,GAAM,QAAQ,GAAA,EAAI;AAAA,EAClB,gBAAA,GAAmB,YAAA;AAAA,EACnB,iBAAA,GAAoB,aAAA;AAAA,EACpB,cAAA,GAAiB,UAAA;AAAA,EACjB,aAAA,GAAgB,SAAA;AAAA,EAChB,aAAA,GAAgB;AAClB,CAAA,GAAc,EAAC,EAAS;AACtB,EAAA,IAAI;AACF,IAAA,GAAA,CAAI,CAAA,2BAAA,EAA8B,GAAG,CAAA,GAAA,CAAK,CAAA;AAC1C,IAAA,GAAA,CAAI,EAAE,CAAA;AAEN,IAAA,MAAM,UAAU,qBAAA,CAAsB;AAAA,MACpC,GAAA;AAAA,MACA,cAAA;AAAA,MACA;AAAA,KACD,CAAA;AAED,IAAA,mBAAA,CAAoB;AAAA,MAClB,OAAA;AAAA,MACA,GAAA;AAAA,MACA,cAAA;AAAA,MACA,aAAA;AAAA,MACA,iBAAA;AAAA,MACA;AAAA,KACD,CAAA;AAED,IAAA,iBAAA,CAAkB;AAAA,MAChB,OAAA;AAAA,MACA,GAAA;AAAA,MACA;AAAA,KACD,CAAA;AAED,IAAA,GAAA,CAAI,EAAE,CAAA;AACN,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,IAAA,CAAK,CAAC,CAAA;AAAA,WACC,KAAA,EAAO;AACd,IAAA,MAAM,UAAU,KAAA,YAAiB,KAAA,GAAQ,KAAA,CAAM,OAAA,GAAU,OAAO,KAAK,CAAA;AACrE,IAAA,KAAA,CAAM,CAAA,WAAA,EAAc,OAAO,CAAA,CAAE,CAAA;AAC7B,IAAA,IAAA,CAAK,CAAC,CAAA;AAAA;AAEV;AAKO,SAAS,qBAAA,CAAsB;AAAA,EACpC,GAAA;AAAA,EACA,cAAA;AAAA,EACA;AACF,CAAA,EAImB;AACjB,EAAA,MAAM,eAAA,GAAkB,IAAA,CAAK,IAAA,CAAK,GAAA,EAAK,cAAc,CAAA;AACrD,EAAA,IAAI,CAAC,cAAA,CAAe,eAAe,CAAA,EAAG;AACpC,IAAA,MAAM,IAAI,MAAM,8CAA8C,CAAA;AAAA;AAGhE,EAAA,MAAM,WAAA,GAAc,gBAAA;AAAA,IAClB,gBAAA,CAAiB,iBAAiB,MAAM,CAAA;AAAA,IACxC;AAAA,GACF;AACA,EAAA,IAAI,CAAC,eAAA,CAAgB,WAAW,CAAA,EAAG;AACjC,IAAA,MAAM,IAAI,KAAA;AAAA,MACR;AAAA,KACF;AAAA;AAGF,EAAA,MAAM,MAAA,GAAS,IAAA,CAAK,IAAA,CAAK,GAAA,EAAK,KAAK,CAAA;AACnC,EAAA,MAAM,OAAA,GAAU,cAAA,CAAe,MAAM,CAAA,GAAI,KAAA,GAAQ,GAAA;AACjD,EAAA,MAAM,MAAA,GAAS,IAAA,CAAK,IAAA,CAAK,GAAA,EAAK,SAAS,KAAK,CAAA;AAC5C,EAAA,MAAM,QAAA,GAAW,IAAA,CAAK,IAAA,CAAK,GAAA,EAAK,SAAS,OAAO,CAAA;AAChD,EAAA,MAAM,SAAA,GAAY,eAAe,MAAM,CAAA;AACvC,EAAA,MAAM,WAAA,GAAc,eAAe,QAAQ,CAAA;AAE3C,EAAA,IAAI,CAAC,SAAA,IAAa,CAAC,WAAA,EAAa;AAC9B,IAAA,MAAM,IAAI,KAAA;AAAA,MACR;AAAA,KACF;AAAA;AAGF,EAAA,MAAM,MAAA,GAAqB,YAAY,KAAA,GAAQ,OAAA;AAC/C,EAAA,OAAO,EAAE,GAAA,EAAK,eAAA,EAAiB,WAAA,EAAa,SAAS,MAAA,EAAO;AAC9D;AAKA,SAAS,iBAAA,CAAkB;AAAA,EACzB,OAAA;AAAA,EACA,GAAA;AAAA,EACA;AACF,CAAA,EAIS;AACP,EAAA,MAAM,cAAc,OAAA,CAAQ,WAAA;AAC5B,EAAA,MAAM,YAAA,GAAe,sBAAA,CAAuB,WAAA,EAAa,cAAc,CAAA;AACvE,EAAA,MAAM,eAAA,GAAkB,sBAAA;AAAA,IACtB,WAAA;AAAA,IACA;AAAA,GACF;AACA,EAAA,MAAM,OAAA,GAAU,sBAAA,CAAuB,WAAA,EAAa,SAAS,CAAA;AAE7D,EAAA,KAAA,MAAW,cAAc,mBAAA,EAAqB;AAC5C,IAAA,IAAI,YAAA,CAAa,UAAU,CAAA,EAAG;AAC5B,MAAA,GAAA,CAAI,CAAA,uBAAA,EAA0B,UAAU,CAAA,iBAAA,CAAmB,CAAA;AAC3D,MAAA;AAAA;AAEF,IAAA,YAAA,CAAa,UAAU,CAAA,GAAI,QAAA;AAC3B,IAAA,GAAA,CAAI,CAAA,uBAAA,EAA0B,UAAU,CAAA,CAAE,CAAA;AAAA;AAG5C,EAAA,KAAA,MAAW,iBAAiB,uBAAA,EAAyB;AACnD,IAAA,IAAI,eAAA,CAAgB,aAAa,CAAA,EAAG;AAClC,MAAA,GAAA,CAAI,CAAA,0BAAA,EAA6B,aAAa,CAAA,iBAAA,CAAmB,CAAA;AACjE,MAAA;AAAA;AAEF,IAAA,eAAA,CAAgB,aAAa,CAAA,GAAI,QAAA;AACjC,IAAA,GAAA,CAAI,CAAA,0BAAA,EAA6B,aAAa,CAAA,CAAE,CAAA;AAAA;AAGlD,EAAA,KAAA,MAAW,CAAC,UAAA,EAAY,WAAW,KAAK,MAAA,CAAO,OAAA,CAAQ,cAAc,CAAA,EAAG;AACtE,IAAA,IAAI,OAAA,CAAQ,UAAU,CAAA,EAAG;AACvB,MAAA,GAAA,CAAI,CAAA,oBAAA,EAAuB,UAAU,CAAA,kBAAA,CAAoB,CAAA;AACzD,MAAA;AAAA;AAEF,IAAA,OAAA,CAAQ,UAAU,CAAA,GAAI,WAAA;AACtB,IAAA,GAAA,CAAI,CAAA,oBAAA,EAAuB,UAAU,CAAA,CAAA,CAAG,CAAA;AAAA;AAG1C,EAAA,iBAAA;AAAA,IACE,OAAA,CAAQ,eAAA;AAAA,IACR,GAAG,IAAA,CAAK,SAAA,CAAU,WAAA,EAAa,IAAA,EAAM,CAAC,CAAC;AAAA;AAAA,GACzC;AACF;AAKA,SAAS,mBAAA,CAAoB;AAAA,EAC3B,OAAA;AAAA,EACA,GAAA;AAAA,EACA,cAAA;AAAA,EACA,aAAA;AAAA,EACA,iBAAA;AAAA,EACA;AACF,CAAA,EAOS;AACP,EAAA,MAAM,eAAe,IAAA,CAAK,IAAA;AAAA,IACxB,OAAA,CAAQ,GAAA;AAAA,IACR,OAAA,CAAQ,OAAA;AAAA,IACR,KAAA;AAAA,IACA,KAAA;AAAA,IACA,WAAA;AAAA,IACA,QAAA;AAAA,IACA,aAAA;AAAA,IACA;AAAA,GACF;AACA,EAAA,MAAM,iBAAiB,IAAA,CAAK,IAAA;AAAA,IAC1B,OAAA,CAAQ,GAAA;AAAA,IACR,OAAA,CAAQ,OAAA;AAAA,IACR,OAAA;AAAA,IACA,KAAA;AAAA,IACA,WAAA;AAAA,IACA,QAAA;AAAA,IACA;AAAA,GACF;AACA,EAAA,MAAM,YAAY,IAAA,CAAK,IAAA;AAAA,IACrB,OAAA,CAAQ,GAAA;AAAA,IACR,OAAA,CAAQ,OAAA;AAAA,IACR,KAAA;AAAA,IACA,WAAA;AAAA,IACA;AAAA,GACF;AACA,EAAA,MAAM,QAAA,GAAW,IAAA,CAAK,IAAA,CAAK,OAAA,CAAQ,KAAK,SAAS,CAAA;AAEjD,EAAA,IAAI,OAAA,CAAQ,WAAW,KAAA,EAAO;AAC5B,IAAA,mBAAA,CAAoB;AAAA,MAClB,YAAA,EAAc,YAAA;AAAA,MACd,OAAA,EAAS,yBAAA;AAAA,MACT,cAAA;AAAA,MACA,aAAA;AAAA,MACA,iBAAA;AAAA,MACA,GAAA;AAAA,MACA,OAAA,EAAS,cAAA,CAAe,OAAA,CAAQ,GAAA,EAAK,YAAY;AAAA,KAClD,CAAA;AACD,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AAAA,GACF,MAAO;AACL,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,mBAAA,CAAoB;AAAA,MAClB,YAAA,EAAc,cAAA;AAAA,MACd,OAAA,EAAS,2BAAA;AAAA,MACT,cAAA;AAAA,MACA,aAAA;AAAA,MACA,iBAAA;AAAA,MACA,GAAA;AAAA,MACA,OAAA,EAAS,cAAA,CAAe,OAAA,CAAQ,GAAA,EAAK,cAAc;AAAA,KACpD,CAAA;AAAA;AAGH,EAAA,mBAAA,CAAoB;AAAA,IAClB,YAAA,EAAc,QAAA;AAAA,IACd,OAAA,EAAS,gBAAA;AAAA,IACT,cAAA;AAAA,IACA,aAAA;AAAA,IACA,iBAAA;AAAA,IACA,GAAA;AAAA,IACA,OAAA,EAAS;AAAA,GACV,CAAA;AACD,EAAA,IAAI,cAAA,CAAe,QAAQ,CAAA,EAAG;AAC5B,IAAA,aAAA,CAAc,UAAU,GAAK,CAAA;AAAA;AAG/B,EAAA,mBAAA,CAAoB;AAAA,IAClB,YAAA,EAAc,SAAA;AAAA,IACd,OAAA,EAAS,cAAA;AAAA,IACT,cAAA;AAAA,IACA,aAAA;AAAA,IACA,iBAAA;AAAA,IACA,GAAA;AAAA,IACA,OAAA,EAAS,cAAA,CAAe,OAAA,CAAQ,GAAA,EAAK,SAAS;AAAA,GAC/C,CAAA;AACH;AAKA,SAAS,mBAAA,CAAoB;AAAA,EAC3B,YAAA;AAAA,EACA,OAAA;AAAA,EACA,cAAA;AAAA,EACA,aAAA;AAAA,EACA,iBAAA;AAAA,EACA,GAAA;AAAA,EACA;AACF,CAAA,EAQS;AACP,EAAA,IAAI,cAAA,CAAe,YAAY,CAAA,EAAG;AAChC,IAAA,GAAA,CAAI,CAAA,YAAA,EAAe,OAAO,CAAA,iBAAA,CAAmB,CAAA;AAC7C,IAAA;AAAA;AAGF,EAAA,aAAA,CAAc,KAAK,OAAA,CAAQ,YAAY,GAAG,EAAE,SAAA,EAAW,MAAM,CAAA;AAC7D,EAAA,iBAAA,CAAkB,cAAc,OAAO,CAAA;AACvC,EAAA,GAAA,CAAI,CAAA,YAAA,EAAe,OAAO,CAAA,CAAE,CAAA;AAC9B;AAKA,SAAS,gBAAA,CAAiB,SAAiB,QAAA,EAA8B;AACvE,EAAA,IAAI;AACF,IAAA,MAAM,MAAA,GAAS,IAAA,CAAK,KAAA,CAAM,OAAO,CAAA;AACjC,IAAA,IAAI,CAAC,UAAU,OAAO,MAAA,KAAW,YAAY,KAAA,CAAM,OAAA,CAAQ,MAAM,CAAA,EAAG;AAClE,MAAA,MAAM,IAAI,MAAM,sCAAsC,CAAA;AAAA;AAExD,IAAA,OAAO,MAAA;AAAA,WACA,KAAA,EAAO;AACd,IAAA,MAAM,IAAI,KAAA;AAAA,MACR,CAAA,gCAAA,EAAmC,QAAQ,CAAA,EAAA,EACzC,KAAA,YAAiB,QAAQ,KAAA,CAAM,OAAA,GAAU,MAAA,CAAO,KAAK,CACvD,CAAA;AAAA,KACF;AAAA;AAEJ;AAKA,SAAS,gBAAgB,WAAA,EAAkC;AACzD,EAAA,MAAM,eAAe,WAAA,CAAY,YAAA;AACjC,EAAA,MAAM,kBAAkB,WAAA,CAAY,eAAA;AAEpC,EAAA,OACE,WAAW,YAAA,EAAc,MAAM,CAAA,IAAK,UAAA,CAAW,iBAAiB,MAAM,CAAA;AAE1E;AAKA,SAAS,UAAA,CAAW,SAAkB,WAAA,EAA8B;AAClE,EAAA,IAAI,CAAC,WAAW,OAAO,OAAA,KAAY,YAAY,KAAA,CAAM,OAAA,CAAQ,OAAO,CAAA,EAAG;AACrE,IAAA,OAAO,KAAA;AAAA;AAET,EAAA,OAAO,OAAA,CAAS,OAAA,CAAoB,WAAW,CAAC,CAAA;AAClD;AAKA,SAAS,sBAAA,CACP,aACA,WAAA,EACS;AACT,EAAA,MAAM,YAAA,GAAe,YAAY,WAAW,CAAA;AAC5C,EAAA,IACE,CAAC,gBACD,OAAO,YAAA,KAAiB,YACxB,KAAA,CAAM,OAAA,CAAQ,YAAY,CAAA,EAC1B;AACA,IAAA,WAAA,CAAY,WAAW,IAAI,EAAC;AAAA;AAE9B,EAAA,OAAO,YAAY,WAAW,CAAA;AAChC;AAKA,SAAS,cAAA,CAAe,KAAa,YAAA,EAA8B;AACjE,EAAA,MAAM,QAAA,GAAW,IAAA,CAAK,QAAA,CAAS,GAAA,EAAK,YAAY,CAAA;AAChD,EAAA,OAAO,QAAA,IAAY,GAAA;AACrB;;;AC9rBA,IAAM,UAAA,GAAa,aAAA,CAAc,MAAA,CAAA,IAAA,CAAY,GAAG,CAAA;AAChD,IAAM,SAAA,GAAYA,IAAAA,CAAK,OAAA,CAAQ,UAAU,CAAA;AAYlC,SAAS,OACd,IAAA,EACA;AAAA,EACE,MAAM,OAAA,CAAQ,GAAA;AAAA,EACd,QAAQ,OAAA,CAAQ,KAAA;AAAA,EAChB,IAAA,GAAO,CAAC,IAAA,KAAiB,OAAA,CAAQ,KAAK,IAAI,CAAA;AAAA,EAC1C,aAAA,GAAgB,SAAA;AAAA,EAChB,aAAA,GAAgBA,IAAAA,CAAK,IAAA,CAAK,SAAA,EAAW,eAAe,CAAA;AAAA,EACpD,QAAA;AAAA,EACA,WAAA,GAAc;AAChB,CAAA,GAAa,EAAC,EACR;AACN,EAAA,MAAM,KAAK,OAAA,EAAS,GAAG,QAAQ,CAAA,GAAI,IAAA;AAKnC,EAAA,SAAS,UAAA,GAAa;AACpB,IAAA,GAAA,CAAI,QAAQ,CAAA;AACZ,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA,CAAI,sBAAsB,CAAA;AAC1B,IAAA,GAAA,CAAI,EAAE,CAAA;AACN,IAAA,GAAA,CAAI,sBAAsB,CAAA;AAC1B,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA,CAAI,EAAE,CAAA;AACN,IAAA,GAAA,CAAI,QAAQ,CAAA;AACZ,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA,CAAI,EAAE,CAAA;AACN,IAAA,GAAA,CAAI,iBAAiB,CAAA;AACrB,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,GAAA;AAAA,MACE;AAAA,KACF;AACA,IAAA,IAAA,CAAK,CAAC,CAAA;AAAA;AAGR,EAAA,IAAI,YAAY,SAAA,EAAW;AAEzB,IAAA,IAAI,YAAsB,EAAC;AAC3B,IAAA,MAAM,MAAA,GAAS,QAAA,CAAS,OAAA,CAAQ,IAAI,CAAA;AACpC,IAAA,MAAM,WAAA,GAAc,QAAA,CAAS,OAAA,CAAQ,UAAU,CAAA;AAC/C,IAAA,IAAI,MAAA,KAAW,EAAA,IAAM,QAAA,CAAS,MAAA,GAAS,CAAC,CAAA,EAAG;AACzC,MAAA,SAAA,GAAY,CAAC,IAAA,EAAM,QAAA,CAAS,MAAA,GAAS,CAAC,GAAG,iBAAiB,CAAA;AAC1D,MAAA,QAAA,CAAS,MAAA,CAAO,QAAQ,CAAC,CAAA;AAAA,eAChB,WAAA,KAAgB,EAAA,IAAM,QAAA,CAAS,WAAA,GAAc,CAAC,CAAA,EAAG;AAC1D,MAAA,SAAA,GAAY,CAAC,IAAA,EAAM,QAAA,CAAS,WAAA,GAAc,CAAC,GAAG,iBAAiB,CAAA;AAC/D,MAAA,QAAA,CAAS,MAAA,CAAO,aAAa,CAAC,CAAA;AAAA;AAIhC,IAAA,IAAI,aAAuB,EAAC;AAC5B,IAAA,MAAM,YAAA,GAAe,QAAA,CAAS,OAAA,CAAQ,WAAW,CAAA;AACjD,IAAA,IAAI,YAAA,KAAiB,EAAA,IAAM,QAAA,CAAS,YAAA,GAAe,CAAC,CAAA,EAAG;AACrD,MAAA,UAAA,GAAa,CAAC,WAAA,EAAa,QAAA,CAAS,YAAA,GAAe,CAAC,CAAC,CAAA;AAAA;AAGvD,IAAA,MAAM,MAAA,GAAgC,aAAA;AAAA,MACpC,KAAA;AAAA,MACA;AAAA,QACE,iBAAA;AAAA,QACA,IAAA;AAAA,QACA,IAAA;AAAA,QACA,yBAAA;AAAA,QACA,IAAA;AAAA,QACA,uBAAA;AAAA,QACA,IAAA;AAAA,QACA,aAAA;AAAA,QACA,GAAG,SAAA;AAAA,QACH,GAAG,UAAA;AAAA,QACH,GAAG;AAAA,OACL;AAAA,MACA,EAAE,OAAO,SAAA;AAAU,KACrB;AACA,IAAA,IAAA,CAAK,MAAA,CAAO,UAAU,CAAC,CAAA;AAAA,GACzB,MAAA,IAAW,YAAY,MAAA,EAAQ;AAC7B,IAAA,WAAA,CAAY;AAAA,MACV,GAAA;AAAA,MACA,KAAA;AAAA,MACA,IAAA;AAAA,MACA,GAAG;AAAA,KACJ,CAAA;AAAA,GACH,MAAO;AACL,IAAA,UAAA,EAAW;AAAA;AAEf","file":"cli.js","sourcesContent":["import {\n chmodSync,\n existsSync,\n mkdirSync,\n readFileSync,\n writeFileSync,\n} from 'fs';\nimport path from 'path';\n\ntype JsonObject = Record<string, unknown>;\ntype JsonMap = Record<string, string>;\n\nconst DEPENDENCIES_TO_ADD = [\n '@nicnocquee/dataqueue',\n '@nicnocquee/dataqueue-dashboard',\n '@nicnocquee/dataqueue-react',\n] as const;\n\nconst DEV_DEPENDENCIES_TO_ADD = [\n 'dotenv-cli',\n 'ts-node',\n 'node-pg-migrate',\n] as const;\n\nconst SCRIPTS_TO_ADD = {\n cron: 'bash cron.sh',\n 'migrate-dataqueue': 'dotenv -e .env.local -- dataqueue-cli migrate',\n} as const;\n\n/**\n * App router endpoint template for queue management.\n */\nexport const APP_ROUTER_ROUTE_TEMPLATE = `/**\n * This end point is used to manage the job queue.\n * It supports the following tasks:\n * - reclaim: Reclaim stuck jobs\n * - cleanup: Cleanup old jobs\n * - process: Process jobs\n *\n * Example usage with default values (reclaim stuck jobs for 10 minutes, cleanup old jobs for 30 days, and process jobs with batch size 3, concurrency 2, and verbose true):\n * curl -X POST http://localhost:3000/api/dataqueue/manage/reclaim -H \"Authorization: Bearer $CRON_SECRET\"\n * curl -X POST http://localhost:3000/api/dataqueue/manage/cleanup -H \"Authorization: Bearer $CRON_SECRET\"\n * curl -X POST http://localhost:3000/api/dataqueue/manage/process -H \"Authorization: Bearer $CRON_SECRET\"\n *\n * Example usage with custom values:\n * curl -X POST http://localhost:3000/api/dataqueue/manage/reclaim -H \"Authorization: Bearer $CRON_SECRET\" -d '{\"maxProcessingTimeMinutes\": 15}' -H \"Content-Type: application/json\"\n * curl -X POST http://localhost:3000/api/dataqueue/manage/cleanup -H \"Authorization: Bearer $CRON_SECRET\" -d '{\"daysToKeep\": 15}' -H \"Content-Type: application/json\"\n * curl -X POST http://localhost:3000/api/dataqueue/manage/process -H \"Authorization: Bearer $CRON_SECRET\" -d '{\"batchSize\": 5, \"concurrency\": 3, \"verbose\": false, \"workerId\": \"custom-worker-id\"}' -H \"Content-Type: application/json\"\n *\n * During development, you can run the following script to run the cron jobs continuously in the background:\n * pnpm cron\n */\nimport { getJobQueue, jobHandlers } from '@/lib/dataqueue/queue';\nimport { NextResponse } from 'next/server';\n\nexport async function POST(\n request: Request,\n { params }: { params: Promise<{ task: string[] }> },\n) {\n const { task } = await params;\n const authHeader = request.headers.get('authorization');\n if (authHeader !== \\`Bearer \\${process.env.CRON_SECRET}\\`) {\n return NextResponse.json({ message: 'Unauthorized' }, { status: 401 });\n }\n\n if (!task || task.length === 0) {\n return NextResponse.json({ message: 'Task is required' }, { status: 400 });\n }\n\n const supportedTasks = ['reclaim', 'cleanup', 'process'];\n const theTask = task[0];\n if (!supportedTasks.includes(theTask)) {\n return NextResponse.json(\n { message: 'Task not supported' },\n { status: 400 },\n );\n }\n\n try {\n const jobQueue = getJobQueue();\n\n if (theTask === 'reclaim') {\n let maxProcessingTimeMinutes = 10;\n try {\n const body = await request.json();\n maxProcessingTimeMinutes = body.maxProcessingTimeMinutes || 10;\n } catch {\n // ignore parsing error and use default value\n }\n const reclaimed = await jobQueue.reclaimStuckJobs(\n maxProcessingTimeMinutes,\n );\n console.log(\\`Reclaimed \\${reclaimed} stuck jobs\\`);\n return NextResponse.json({\n message: \\`Stuck jobs reclaimed: \\${reclaimed} with maxProcessingTimeMinutes: \\${maxProcessingTimeMinutes}\\`,\n reclaimed,\n });\n }\n\n if (theTask === 'cleanup') {\n let daysToKeep = 30;\n try {\n const body = await request.json();\n daysToKeep = body.daysToKeep || 30;\n } catch {\n // ignore parsing error and use default value\n }\n const deleted = await jobQueue.cleanupOldJobs(daysToKeep);\n console.log(\\`Deleted \\${deleted} old jobs\\`);\n return NextResponse.json({\n message: \\`Old jobs cleaned up: \\${deleted} with daysToKeep: \\${daysToKeep}\\`,\n deleted,\n });\n }\n\n if (theTask === 'process') {\n let batchSize = 3;\n let concurrency = 2;\n let verbose = true;\n let workerId = \\`manage-\\${theTask}-\\${Date.now()}\\`;\n try {\n const body = await request.json();\n batchSize = body.batchSize || 3;\n concurrency = body.concurrency || 2;\n verbose = body.verbose || true;\n workerId = body.workerId || \\`manage-\\${theTask}-\\${Date.now()}\\`;\n } catch {\n // ignore parsing error and use default value\n }\n const processor = jobQueue.createProcessor(jobHandlers, {\n workerId,\n batchSize,\n concurrency,\n verbose,\n });\n const processed = await processor.start();\n\n return NextResponse.json({\n message: \\`Jobs processed: \\${processed} with workerId: \\${workerId}, batchSize: \\${batchSize}, concurrency: \\${concurrency}, and verbose: \\${verbose}\\`,\n processed,\n });\n }\n\n return NextResponse.json(\n { message: 'Task not supported' },\n { status: 400 },\n );\n } catch (error) {\n console.error('Error processing jobs:', error);\n return NextResponse.json(\n { message: 'Failed to process jobs' },\n { status: 500 },\n );\n }\n}\n`;\n\n/**\n * Pages router endpoint template for queue management.\n */\nexport const PAGES_ROUTER_ROUTE_TEMPLATE = `/**\n * This end point is used to manage the job queue.\n * It supports the following tasks:\n * - reclaim: Reclaim stuck jobs\n * - cleanup: Cleanup old jobs\n * - process: Process jobs\n *\n * Example usage with default values (reclaim stuck jobs for 10 minutes, cleanup old jobs for 30 days, and process jobs with batch size 3, concurrency 2, and verbose true):\n * curl -X POST http://localhost:3000/api/dataqueue/manage/reclaim -H \"Authorization: Bearer $CRON_SECRET\"\n * curl -X POST http://localhost:3000/api/dataqueue/manage/cleanup -H \"Authorization: Bearer $CRON_SECRET\"\n * curl -X POST http://localhost:3000/api/dataqueue/manage/process -H \"Authorization: Bearer $CRON_SECRET\"\n *\n * Example usage with custom values:\n * curl -X POST http://localhost:3000/api/dataqueue/manage/reclaim -H \"Authorization: Bearer $CRON_SECRET\" -d '{\"maxProcessingTimeMinutes\": 15}' -H \"Content-Type: application/json\"\n * curl -X POST http://localhost:3000/api/dataqueue/manage/cleanup -H \"Authorization: Bearer $CRON_SECRET\" -d '{\"daysToKeep\": 15}' -H \"Content-Type: application/json\"\n * curl -X POST http://localhost:3000/api/dataqueue/manage/process -H \"Authorization: Bearer $CRON_SECRET\" -d '{\"batchSize\": 5, \"concurrency\": 3, \"verbose\": false, \"workerId\": \"custom-worker-id\"}' -H \"Content-Type: application/json\"\n *\n * During development, you can run the following script to run the cron jobs continuously in the background:\n * pnpm cron\n */\nimport type { NextApiRequest, NextApiResponse } from 'next';\nimport { getJobQueue, jobHandlers } from '@/lib/dataqueue/queue';\n\ntype ResponseBody = {\n message: string;\n reclaimed?: number;\n deleted?: number;\n processed?: number;\n};\n\nexport default async function handler(\n req: NextApiRequest,\n res: NextApiResponse<ResponseBody>,\n) {\n if (req.method !== 'POST') {\n res.setHeader('Allow', 'POST');\n return res.status(405).json({ message: 'Method not allowed' });\n }\n\n const authHeader = req.headers.authorization;\n if (authHeader !== \\`Bearer \\${process.env.CRON_SECRET}\\`) {\n return res.status(401).json({ message: 'Unauthorized' });\n }\n\n const task = req.query.task;\n const taskArray = Array.isArray(task) ? task : task ? [task] : [];\n if (!taskArray.length) {\n return res.status(400).json({ message: 'Task is required' });\n }\n\n const supportedTasks = ['reclaim', 'cleanup', 'process'];\n const theTask = taskArray[0];\n if (!supportedTasks.includes(theTask)) {\n return res.status(400).json({ message: 'Task not supported' });\n }\n\n try {\n const jobQueue = getJobQueue();\n const body = typeof req.body === 'object' && req.body ? req.body : {};\n\n if (theTask === 'reclaim') {\n const maxProcessingTimeMinutes = body.maxProcessingTimeMinutes || 10;\n const reclaimed = await jobQueue.reclaimStuckJobs(maxProcessingTimeMinutes);\n console.log(\\`Reclaimed \\${reclaimed} stuck jobs\\`);\n return res.status(200).json({\n message: \\`Stuck jobs reclaimed: \\${reclaimed} with maxProcessingTimeMinutes: \\${maxProcessingTimeMinutes}\\`,\n reclaimed,\n });\n }\n\n if (theTask === 'cleanup') {\n const daysToKeep = body.daysToKeep || 30;\n const deleted = await jobQueue.cleanupOldJobs(daysToKeep);\n console.log(\\`Deleted \\${deleted} old jobs\\`);\n return res.status(200).json({\n message: \\`Old jobs cleaned up: \\${deleted} with daysToKeep: \\${daysToKeep}\\`,\n deleted,\n });\n }\n\n const batchSize = body.batchSize || 3;\n const concurrency = body.concurrency || 2;\n const verbose = body.verbose || true;\n const workerId = body.workerId || \\`manage-\\${theTask}-\\${Date.now()}\\`;\n const processor = jobQueue.createProcessor(jobHandlers, {\n workerId,\n batchSize,\n concurrency,\n verbose,\n });\n const processed = await processor.start();\n\n return res.status(200).json({\n message: \\`Jobs processed: \\${processed} with workerId: \\${workerId}, batchSize: \\${batchSize}, concurrency: \\${concurrency}, and verbose: \\${verbose}\\`,\n processed,\n });\n } catch (error) {\n console.error('Error processing jobs:', error);\n return res.status(500).json({ message: 'Failed to process jobs' });\n }\n}\n`;\n\n/**\n * Cron script template for local queue processing.\n */\nexport const CRON_SH_TEMPLATE = `#!/bin/bash\n\n# This script is used to run the cron jobs for the demo app during development.\n# Run it with \\`pnpm cron\\` from the apps/demo directory.\n\nset -a\nsource \"$(dirname \"$0\")/.env.local\"\nset +a\n\nif [ -z \"$CRON_SECRET\" ]; then\n echo \"Error: CRON_SECRET environment variable is not set in .env.local\"\n exit 1\nfi\n\ncleanup() {\n kill 0\n wait\n}\ntrap cleanup SIGINT SIGTERM\n\nwhile true; do\n echo \"Processing jobs...\"\n curl http://localhost:3000/api/dataqueue/manage/process -X POST -H \"Authorization: Bearer $CRON_SECRET\"\n echo \"\"\n sleep 10 # Process jobs every 10 seconds\ndone &\n\nwhile true; do\n echo \"Reclaiming stuck jobs...\"\n curl http://localhost:3000/api/dataqueue/manage/reclaim -X POST -H \"Authorization: Bearer $CRON_SECRET\"\n echo \"\"\n sleep 20 # Reclaim stuck jobs every 20 seconds\ndone &\n\nwhile true; do\n echo \"Cleaning up old jobs...\"\n curl http://localhost:3000/api/dataqueue/manage/cleanup -X POST -H \"Authorization: Bearer $CRON_SECRET\"\n echo \"\"\n sleep 30 # Cleanup old jobs every 30 seconds\ndone &\n\nwait\n`;\n\n/**\n * Queue placeholder template with a single `send_email` job.\n */\nexport const QUEUE_TEMPLATE = `import { initJobQueue, JobHandlers } from '@nicnocquee/dataqueue';\n\nexport type JobPayloadMap = {\n send_email: {\n to: string;\n subject: string;\n body: string;\n };\n};\n\nlet jobQueue: ReturnType<typeof initJobQueue<JobPayloadMap>> | null = null;\n\nexport const getJobQueue = () => {\n if (!jobQueue) {\n jobQueue = initJobQueue<JobPayloadMap>({\n databaseConfig: {\n connectionString: process.env.PG_DATAQUEUE_DATABASE,\n },\n verbose: process.env.NODE_ENV === 'development',\n });\n }\n return jobQueue;\n};\n\nexport const jobHandlers: JobHandlers<JobPayloadMap> = {\n send_email: async (payload) => {\n const { to, subject, body } = payload;\n console.log('send_email placeholder:', { to, subject, body });\n },\n};\n`;\n\nexport interface InitDeps {\n log?: (...args: any[]) => void;\n error?: (...args: any[]) => void;\n exit?: (code: number) => void;\n cwd?: string;\n readFileSyncImpl?: typeof readFileSync;\n writeFileSyncImpl?: typeof writeFileSync;\n existsSyncImpl?: typeof existsSync;\n mkdirSyncImpl?: typeof mkdirSync;\n chmodSyncImpl?: typeof chmodSync;\n}\n\ntype RouterKind = 'app' | 'pages';\n\ninterface ProjectDetails {\n cwd: string;\n packageJsonPath: string;\n packageJson: JsonObject;\n srcRoot: string;\n router: RouterKind;\n}\n\n/**\n * Runs the `dataqueue-cli init` command.\n */\nexport function runInit({\n log = console.log,\n error = console.error,\n exit = (code: number) => process.exit(code),\n cwd = process.cwd(),\n readFileSyncImpl = readFileSync,\n writeFileSyncImpl = writeFileSync,\n existsSyncImpl = existsSync,\n mkdirSyncImpl = mkdirSync,\n chmodSyncImpl = chmodSync,\n}: InitDeps = {}): void {\n try {\n log(`dataqueue: Initializing in ${cwd}...`);\n log('');\n\n const details = detectNextJsAndRouter({\n cwd,\n existsSyncImpl,\n readFileSyncImpl,\n });\n\n createScaffoldFiles({\n details,\n log,\n existsSyncImpl,\n mkdirSyncImpl,\n writeFileSyncImpl,\n chmodSyncImpl,\n });\n\n updatePackageJson({\n details,\n log,\n writeFileSyncImpl,\n });\n\n log('');\n log(\n \"Done! Run your package manager's install command to install new dependencies.\",\n );\n exit(0);\n } catch (cause) {\n const message = cause instanceof Error ? cause.message : String(cause);\n error(`dataqueue: ${message}`);\n exit(1);\n }\n}\n\n/**\n * Detects that the current directory is a Next.js app and chooses the router.\n */\nexport function detectNextJsAndRouter({\n cwd,\n existsSyncImpl,\n readFileSyncImpl,\n}: {\n cwd: string;\n existsSyncImpl: typeof existsSync;\n readFileSyncImpl: typeof readFileSync;\n}): ProjectDetails {\n const packageJsonPath = path.join(cwd, 'package.json');\n if (!existsSyncImpl(packageJsonPath)) {\n throw new Error('package.json not found in current directory.');\n }\n\n const packageJson = parsePackageJson(\n readFileSyncImpl(packageJsonPath, 'utf8'),\n packageJsonPath,\n );\n if (!isNextJsProject(packageJson)) {\n throw new Error(\n \"Not a Next.js project. Could not find 'next' in package.json dependencies.\",\n );\n }\n\n const srcDir = path.join(cwd, 'src');\n const srcRoot = existsSyncImpl(srcDir) ? 'src' : '.';\n const appDir = path.join(cwd, srcRoot, 'app');\n const pagesDir = path.join(cwd, srcRoot, 'pages');\n const hasAppDir = existsSyncImpl(appDir);\n const hasPagesDir = existsSyncImpl(pagesDir);\n\n if (!hasAppDir && !hasPagesDir) {\n throw new Error(\n 'Could not detect Next.js router. Expected either app/ or pages/ directory.',\n );\n }\n\n const router: RouterKind = hasAppDir ? 'app' : 'pages';\n return { cwd, packageJsonPath, packageJson, srcRoot, router };\n}\n\n/**\n * Updates package.json with required dependencies and scripts.\n */\nfunction updatePackageJson({\n details,\n log,\n writeFileSyncImpl,\n}: {\n details: ProjectDetails;\n log: (...args: any[]) => void;\n writeFileSyncImpl: typeof writeFileSync;\n}): void {\n const packageJson = details.packageJson;\n const dependencies = ensureStringMapSection(packageJson, 'dependencies');\n const devDependencies = ensureStringMapSection(\n packageJson,\n 'devDependencies',\n );\n const scripts = ensureStringMapSection(packageJson, 'scripts');\n\n for (const dependency of DEPENDENCIES_TO_ADD) {\n if (dependencies[dependency]) {\n log(` [skipped] dependency ${dependency} (already exists)`);\n continue;\n }\n dependencies[dependency] = 'latest';\n log(` [added] dependency ${dependency}`);\n }\n\n for (const devDependency of DEV_DEPENDENCIES_TO_ADD) {\n if (devDependencies[devDependency]) {\n log(` [skipped] devDependency ${devDependency} (already exists)`);\n continue;\n }\n devDependencies[devDependency] = 'latest';\n log(` [added] devDependency ${devDependency}`);\n }\n\n for (const [scriptName, scriptValue] of Object.entries(SCRIPTS_TO_ADD)) {\n if (scripts[scriptName]) {\n log(` [skipped] script \"${scriptName}\" (already exists)`);\n continue;\n }\n scripts[scriptName] = scriptValue;\n log(` [added] script \"${scriptName}\"`);\n }\n\n writeFileSyncImpl(\n details.packageJsonPath,\n `${JSON.stringify(packageJson, null, 2)}\\n`,\n );\n}\n\n/**\n * Creates all scaffold files for the detected router without overwriting.\n */\nfunction createScaffoldFiles({\n details,\n log,\n existsSyncImpl,\n mkdirSyncImpl,\n writeFileSyncImpl,\n chmodSyncImpl,\n}: {\n details: ProjectDetails;\n log: (...args: any[]) => void;\n existsSyncImpl: typeof existsSync;\n mkdirSyncImpl: typeof mkdirSync;\n writeFileSyncImpl: typeof writeFileSync;\n chmodSyncImpl: typeof chmodSync;\n}): void {\n const appRoutePath = path.join(\n details.cwd,\n details.srcRoot,\n 'app',\n 'api',\n 'dataqueue',\n 'manage',\n '[[...task]]',\n 'route.ts',\n );\n const pagesRoutePath = path.join(\n details.cwd,\n details.srcRoot,\n 'pages',\n 'api',\n 'dataqueue',\n 'manage',\n '[[...task]].ts',\n );\n const queuePath = path.join(\n details.cwd,\n details.srcRoot,\n 'lib',\n 'dataqueue',\n 'queue.ts',\n );\n const cronPath = path.join(details.cwd, 'cron.sh');\n\n if (details.router === 'app') {\n createFileIfMissing({\n absolutePath: appRoutePath,\n content: APP_ROUTER_ROUTE_TEMPLATE,\n existsSyncImpl,\n mkdirSyncImpl,\n writeFileSyncImpl,\n log,\n logPath: toRelativePath(details.cwd, appRoutePath),\n });\n log(\n ' [skipped] pages/api/dataqueue/manage/[[...task]].ts (router not selected)',\n );\n } else {\n log(\n ' [skipped] app/api/dataqueue/manage/[[...task]]/route.ts (router not selected)',\n );\n createFileIfMissing({\n absolutePath: pagesRoutePath,\n content: PAGES_ROUTER_ROUTE_TEMPLATE,\n existsSyncImpl,\n mkdirSyncImpl,\n writeFileSyncImpl,\n log,\n logPath: toRelativePath(details.cwd, pagesRoutePath),\n });\n }\n\n createFileIfMissing({\n absolutePath: cronPath,\n content: CRON_SH_TEMPLATE,\n existsSyncImpl,\n mkdirSyncImpl,\n writeFileSyncImpl,\n log,\n logPath: 'cron.sh',\n });\n if (existsSyncImpl(cronPath)) {\n chmodSyncImpl(cronPath, 0o755);\n }\n\n createFileIfMissing({\n absolutePath: queuePath,\n content: QUEUE_TEMPLATE,\n existsSyncImpl,\n mkdirSyncImpl,\n writeFileSyncImpl,\n log,\n logPath: toRelativePath(details.cwd, queuePath),\n });\n}\n\n/**\n * Creates a file only if it does not already exist.\n */\nfunction createFileIfMissing({\n absolutePath,\n content,\n existsSyncImpl,\n mkdirSyncImpl,\n writeFileSyncImpl,\n log,\n logPath,\n}: {\n absolutePath: string;\n content: string;\n existsSyncImpl: typeof existsSync;\n mkdirSyncImpl: typeof mkdirSync;\n writeFileSyncImpl: typeof writeFileSync;\n log: (...args: any[]) => void;\n logPath: string;\n}): void {\n if (existsSyncImpl(absolutePath)) {\n log(` [skipped] ${logPath} (already exists)`);\n return;\n }\n\n mkdirSyncImpl(path.dirname(absolutePath), { recursive: true });\n writeFileSyncImpl(absolutePath, content);\n log(` [created] ${logPath}`);\n}\n\n/**\n * Parses package.json content with clear source context.\n */\nfunction parsePackageJson(content: string, filePath: string): JsonObject {\n try {\n const parsed = JSON.parse(content);\n if (!parsed || typeof parsed !== 'object' || Array.isArray(parsed)) {\n throw new Error('package.json must contain an object.');\n }\n return parsed as JsonObject;\n } catch (cause) {\n throw new Error(\n `Failed to parse package.json at ${filePath}: ${\n cause instanceof Error ? cause.message : String(cause)\n }`,\n );\n }\n}\n\n/**\n * Returns true when package.json declares Next.js in deps or devDeps.\n */\nfunction isNextJsProject(packageJson: JsonObject): boolean {\n const dependencies = packageJson.dependencies;\n const devDependencies = packageJson.devDependencies;\n\n return (\n hasPackage(dependencies, 'next') || hasPackage(devDependencies, 'next')\n );\n}\n\n/**\n * Returns true when a package name exists in a dependency section object.\n */\nfunction hasPackage(section: unknown, packageName: string): boolean {\n if (!section || typeof section !== 'object' || Array.isArray(section)) {\n return false;\n }\n return Boolean((section as JsonMap)[packageName]);\n}\n\n/**\n * Ensures package.json has a string map section and returns it.\n */\nfunction ensureStringMapSection(\n packageJson: JsonObject,\n sectionName: 'dependencies' | 'devDependencies' | 'scripts',\n): JsonMap {\n const currentValue = packageJson[sectionName];\n if (\n !currentValue ||\n typeof currentValue !== 'object' ||\n Array.isArray(currentValue)\n ) {\n packageJson[sectionName] = {};\n }\n return packageJson[sectionName] as JsonMap;\n}\n\n/**\n * Converts an absolute path to a stable relative path for log output.\n */\nfunction toRelativePath(cwd: string, absolutePath: string): string {\n const relative = path.relative(cwd, absolutePath);\n return relative || '.';\n}\n","// Testable CLI logic for dataqueue\nimport { spawnSync, SpawnSyncReturns } from 'child_process';\nimport path from 'path';\nimport { fileURLToPath } from 'url';\nimport { InitDeps, runInit } from './init-command.js';\n\nconst __filename = fileURLToPath(import.meta.url);\nconst __dirname = path.dirname(__filename);\n\nexport interface CliDeps {\n log?: (...args: any[]) => void;\n error?: (...args: any[]) => void;\n exit?: (code: number) => void;\n spawnSyncImpl?: (...args: any[]) => SpawnSyncReturns<any>;\n migrationsDir?: string;\n initDeps?: InitDeps;\n runInitImpl?: (deps?: InitDeps) => void;\n}\n\nexport function runCli(\n argv: string[],\n {\n log = console.log,\n error = console.error,\n exit = (code: number) => process.exit(code),\n spawnSyncImpl = spawnSync,\n migrationsDir = path.join(__dirname, '../migrations'),\n initDeps,\n runInitImpl = runInit,\n }: CliDeps = {},\n): void {\n const [, , command, ...restArgs] = argv;\n\n /**\n * Prints CLI usage and exits with non-zero code.\n */\n function printUsage() {\n log('Usage:');\n log(\n ' dataqueue-cli migrate [--envPath <path>] [-s <schema> | --schema <schema>]',\n );\n log(' dataqueue-cli init');\n log('');\n log('Options for migrate:');\n log(\n ' --envPath <path> Path to a .env file to load environment variables (passed to node-pg-migrate)',\n );\n log(\n ' -s, --schema <schema> Set the schema to use (passed to node-pg-migrate)',\n );\n log('');\n log('Notes:');\n log(\n ' - The PG_DATAQUEUE_DATABASE environment variable must be set to your Postgres connection string.',\n );\n log(\n ' - For managed Postgres (e.g., DigitalOcean) with SSL, set PGSSLMODE=require and PGSSLROOTCERT to your CA .crt file.',\n );\n log(\n ' Example: PGSSLMODE=require NODE_EXTRA_CA_CERTS=/absolute/path/to/ca.crt PG_DATAQUEUE_DATABASE=... npx dataqueue-cli migrate',\n );\n log('');\n log('Notes for init:');\n log(\n ' - Supports both Next.js App Router and Pages Router (prefers App Router if both exist).',\n );\n log(\n ' - Scaffolds endpoint, cron.sh, queue placeholder, and package.json entries.',\n );\n exit(1);\n }\n\n if (command === 'migrate') {\n // Support for -s or --schema argument\n let schemaArg: string[] = [];\n const sIndex = restArgs.indexOf('-s');\n const schemaIndex = restArgs.indexOf('--schema');\n if (sIndex !== -1 && restArgs[sIndex + 1]) {\n schemaArg = ['-s', restArgs[sIndex + 1], '--create-schema'];\n restArgs.splice(sIndex, 2);\n } else if (schemaIndex !== -1 && restArgs[schemaIndex + 1]) {\n schemaArg = ['-s', restArgs[schemaIndex + 1], '--create-schema'];\n restArgs.splice(schemaIndex, 2);\n }\n\n // Support for --envPath argument\n let envPathArg: string[] = [];\n const envPathIndex = restArgs.indexOf('--envPath');\n if (envPathIndex !== -1 && restArgs[envPathIndex + 1]) {\n envPathArg = ['--envPath', restArgs[envPathIndex + 1]];\n }\n\n const result: SpawnSyncReturns<any> = spawnSyncImpl(\n 'npx',\n [\n 'node-pg-migrate',\n 'up',\n '-t',\n 'dataqueuedev_migrations',\n '-d',\n 'PG_DATAQUEUE_DATABASE',\n '-m',\n migrationsDir,\n ...schemaArg,\n ...envPathArg,\n ...restArgs,\n ],\n { stdio: 'inherit' },\n );\n exit(result.status ?? 1);\n } else if (command === 'init') {\n runInitImpl({\n log,\n error,\n exit,\n ...initDeps,\n });\n } else {\n printUsage();\n }\n}\n"]}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@nicnocquee/dataqueue",
3
- "version": "1.32.0",
3
+ "version": "1.33.0",
4
4
  "description": "PostgreSQL or Redis-backed job queue for Node.js applications with support for serverless environments",
5
5
  "type": "module",
6
6
  "main": "dist/index.js",
package/src/cli.test.ts CHANGED
@@ -22,6 +22,7 @@ function makeDeps() {
22
22
  exit: vi.fn(),
23
23
  spawnSyncImpl: vi.fn(() => makeSpawnSyncReturns(0)),
24
24
  migrationsDir: '/migrations',
25
+ runInitImpl: vi.fn(),
25
26
  } satisfies CliDeps;
26
27
  }
27
28
 
@@ -34,20 +35,30 @@ describe('runCli', () => {
34
35
 
35
36
  it('prints usage and exits with code 1 for no command', () => {
36
37
  runCli(['node', 'cli.js'], deps);
37
- expect(deps.log).toHaveBeenCalledWith(
38
- 'Usage: dataqueue-cli migrate [--envPath <path>] [-s <schema> | --schema <schema>]',
39
- );
38
+ expect(deps.log).toHaveBeenCalledWith('Usage:');
39
+ expect(deps.log).toHaveBeenCalledWith(' dataqueue-cli init');
40
40
  expect(deps.exit).toHaveBeenCalledWith(1);
41
41
  });
42
42
 
43
43
  it('prints usage and exits with code 1 for unknown command', () => {
44
44
  runCli(['node', 'cli.js', 'unknown'], deps);
45
- expect(deps.log).toHaveBeenCalledWith(
46
- 'Usage: dataqueue-cli migrate [--envPath <path>] [-s <schema> | --schema <schema>]',
47
- );
45
+ expect(deps.log).toHaveBeenCalledWith('Usage:');
46
+ expect(deps.log).toHaveBeenCalledWith(' dataqueue-cli init');
48
47
  expect(deps.exit).toHaveBeenCalledWith(1);
49
48
  });
50
49
 
50
+ it('routes init command to runInitImpl', () => {
51
+ runCli(['node', 'cli.js', 'init'], deps);
52
+ expect(deps.runInitImpl).toHaveBeenCalledWith(
53
+ expect.objectContaining({
54
+ log: deps.log,
55
+ error: deps.error,
56
+ exit: deps.exit,
57
+ }),
58
+ );
59
+ expect(deps.spawnSyncImpl).not.toHaveBeenCalled();
60
+ });
61
+
51
62
  it('calls spawnSyncImpl with correct args for migrate', () => {
52
63
  runCli(['node', 'cli.js', 'migrate'], deps);
53
64
  expect(deps.spawnSyncImpl).toHaveBeenCalledWith(
package/src/cli.ts CHANGED
@@ -2,6 +2,7 @@
2
2
  import { spawnSync, SpawnSyncReturns } from 'child_process';
3
3
  import path from 'path';
4
4
  import { fileURLToPath } from 'url';
5
+ import { InitDeps, runInit } from './init-command.js';
5
6
 
6
7
  const __filename = fileURLToPath(import.meta.url);
7
8
  const __dirname = path.dirname(__filename);
@@ -12,25 +13,35 @@ export interface CliDeps {
12
13
  exit?: (code: number) => void;
13
14
  spawnSyncImpl?: (...args: any[]) => SpawnSyncReturns<any>;
14
15
  migrationsDir?: string;
16
+ initDeps?: InitDeps;
17
+ runInitImpl?: (deps?: InitDeps) => void;
15
18
  }
16
19
 
17
20
  export function runCli(
18
21
  argv: string[],
19
22
  {
20
23
  log = console.log,
24
+ error = console.error,
21
25
  exit = (code: number) => process.exit(code),
22
26
  spawnSyncImpl = spawnSync,
23
27
  migrationsDir = path.join(__dirname, '../migrations'),
28
+ initDeps,
29
+ runInitImpl = runInit,
24
30
  }: CliDeps = {},
25
31
  ): void {
26
32
  const [, , command, ...restArgs] = argv;
27
33
 
34
+ /**
35
+ * Prints CLI usage and exits with non-zero code.
36
+ */
28
37
  function printUsage() {
38
+ log('Usage:');
29
39
  log(
30
- 'Usage: dataqueue-cli migrate [--envPath <path>] [-s <schema> | --schema <schema>]',
40
+ ' dataqueue-cli migrate [--envPath <path>] [-s <schema> | --schema <schema>]',
31
41
  );
42
+ log(' dataqueue-cli init');
32
43
  log('');
33
- log('Options:');
44
+ log('Options for migrate:');
34
45
  log(
35
46
  ' --envPath <path> Path to a .env file to load environment variables (passed to node-pg-migrate)',
36
47
  );
@@ -48,6 +59,14 @@ export function runCli(
48
59
  log(
49
60
  ' Example: PGSSLMODE=require NODE_EXTRA_CA_CERTS=/absolute/path/to/ca.crt PG_DATAQUEUE_DATABASE=... npx dataqueue-cli migrate',
50
61
  );
62
+ log('');
63
+ log('Notes for init:');
64
+ log(
65
+ ' - Supports both Next.js App Router and Pages Router (prefers App Router if both exist).',
66
+ );
67
+ log(
68
+ ' - Scaffolds endpoint, cron.sh, queue placeholder, and package.json entries.',
69
+ );
51
70
  exit(1);
52
71
  }
53
72
 
@@ -89,6 +108,13 @@ export function runCli(
89
108
  { stdio: 'inherit' },
90
109
  );
91
110
  exit(result.status ?? 1);
111
+ } else if (command === 'init') {
112
+ runInitImpl({
113
+ log,
114
+ error,
115
+ exit,
116
+ ...initDeps,
117
+ });
92
118
  } else {
93
119
  printUsage();
94
120
  }
@@ -0,0 +1,449 @@
1
+ import path from 'path';
2
+ import { beforeEach, describe, expect, it, vi } from 'vitest';
3
+ import {
4
+ APP_ROUTER_ROUTE_TEMPLATE,
5
+ CRON_SH_TEMPLATE,
6
+ PAGES_ROUTER_ROUTE_TEMPLATE,
7
+ QUEUE_TEMPLATE,
8
+ detectNextJsAndRouter,
9
+ runInit,
10
+ } from './init-command.js';
11
+
12
+ type VirtualFsState = {
13
+ files: Map<string, string>;
14
+ dirs: Set<string>;
15
+ chmodCalls: Array<{ filePath: string; mode: number }>;
16
+ };
17
+
18
+ /**
19
+ * Builds a fake filesystem API surface compatible with `InitDeps`.
20
+ */
21
+ function createVirtualFs(
22
+ cwd: string,
23
+ initialFiles: Record<string, string> = {},
24
+ initialDirs: string[] = [],
25
+ ) {
26
+ const state: VirtualFsState = {
27
+ files: new Map(),
28
+ dirs: new Set([cwd, ...initialDirs.map((dir) => resolvePath(cwd, dir))]),
29
+ chmodCalls: [],
30
+ };
31
+
32
+ for (const [filePath, content] of Object.entries(initialFiles)) {
33
+ const absolutePath = resolvePath(cwd, filePath);
34
+ state.files.set(absolutePath, content);
35
+ state.dirs.add(path.dirname(absolutePath));
36
+ }
37
+
38
+ return {
39
+ state,
40
+ existsSyncImpl: vi.fn((targetPath: string) => {
41
+ return state.files.has(targetPath) || state.dirs.has(targetPath);
42
+ }),
43
+ mkdirSyncImpl: vi.fn((targetPath: string) => {
44
+ state.dirs.add(targetPath);
45
+ }),
46
+ readFileSyncImpl: vi.fn((targetPath: string) => {
47
+ const content = state.files.get(targetPath);
48
+ if (typeof content !== 'string') {
49
+ throw new Error(`ENOENT: ${targetPath}`);
50
+ }
51
+ return content;
52
+ }),
53
+ writeFileSyncImpl: vi.fn((targetPath: string, content: string) => {
54
+ state.files.set(targetPath, content);
55
+ state.dirs.add(path.dirname(targetPath));
56
+ }),
57
+ chmodSyncImpl: vi.fn((filePath: string, mode: number) => {
58
+ state.chmodCalls.push({ filePath, mode });
59
+ }),
60
+ };
61
+ }
62
+
63
+ /**
64
+ * Resolves a project-relative path to absolute for tests.
65
+ */
66
+ function resolvePath(cwd: string, maybeRelativePath: string): string {
67
+ if (path.isAbsolute(maybeRelativePath)) {
68
+ return maybeRelativePath;
69
+ }
70
+ return path.join(cwd, maybeRelativePath);
71
+ }
72
+
73
+ describe('detectNextJsAndRouter', () => {
74
+ const cwd = '/project';
75
+
76
+ it('throws if package.json is missing', () => {
77
+ const fs = createVirtualFs(cwd);
78
+ expect(() =>
79
+ detectNextJsAndRouter({
80
+ cwd,
81
+ existsSyncImpl: fs.existsSyncImpl as any,
82
+ readFileSyncImpl: fs.readFileSyncImpl as any,
83
+ }),
84
+ ).toThrow('package.json not found in current directory.');
85
+ });
86
+
87
+ it('throws if next dependency is missing', () => {
88
+ const fs = createVirtualFs(cwd, {
89
+ 'package.json': JSON.stringify({ name: 'app' }),
90
+ });
91
+
92
+ expect(() =>
93
+ detectNextJsAndRouter({
94
+ cwd,
95
+ existsSyncImpl: fs.existsSyncImpl as any,
96
+ readFileSyncImpl: fs.readFileSyncImpl as any,
97
+ }),
98
+ ).toThrow(
99
+ "Not a Next.js project. Could not find 'next' in package.json dependencies.",
100
+ );
101
+ });
102
+
103
+ it('detects app router when app exists', () => {
104
+ const fs = createVirtualFs(
105
+ cwd,
106
+ {
107
+ 'package.json': JSON.stringify({
108
+ dependencies: { next: '15.0.0' },
109
+ }),
110
+ },
111
+ ['app'],
112
+ );
113
+
114
+ const result = detectNextJsAndRouter({
115
+ cwd,
116
+ existsSyncImpl: fs.existsSyncImpl as any,
117
+ readFileSyncImpl: fs.readFileSyncImpl as any,
118
+ });
119
+
120
+ expect(result.router).toBe('app');
121
+ expect(result.srcRoot).toBe('.');
122
+ });
123
+
124
+ it('detects pages router when only pages exists', () => {
125
+ const fs = createVirtualFs(
126
+ cwd,
127
+ {
128
+ 'package.json': JSON.stringify({
129
+ devDependencies: { next: '15.0.0' },
130
+ }),
131
+ },
132
+ ['pages'],
133
+ );
134
+
135
+ const result = detectNextJsAndRouter({
136
+ cwd,
137
+ existsSyncImpl: fs.existsSyncImpl as any,
138
+ readFileSyncImpl: fs.readFileSyncImpl as any,
139
+ });
140
+
141
+ expect(result.router).toBe('pages');
142
+ expect(result.srcRoot).toBe('.');
143
+ });
144
+
145
+ it('prefers app router when both app and pages exist', () => {
146
+ const fs = createVirtualFs(
147
+ cwd,
148
+ {
149
+ 'package.json': JSON.stringify({
150
+ dependencies: { next: '15.0.0' },
151
+ }),
152
+ },
153
+ ['app', 'pages'],
154
+ );
155
+
156
+ const result = detectNextJsAndRouter({
157
+ cwd,
158
+ existsSyncImpl: fs.existsSyncImpl as any,
159
+ readFileSyncImpl: fs.readFileSyncImpl as any,
160
+ });
161
+
162
+ expect(result.router).toBe('app');
163
+ });
164
+
165
+ it('uses src as root when src exists', () => {
166
+ const fs = createVirtualFs(
167
+ cwd,
168
+ {
169
+ 'package.json': JSON.stringify({
170
+ dependencies: { next: '15.0.0' },
171
+ }),
172
+ },
173
+ ['src', 'src/pages'],
174
+ );
175
+
176
+ const result = detectNextJsAndRouter({
177
+ cwd,
178
+ existsSyncImpl: fs.existsSyncImpl as any,
179
+ readFileSyncImpl: fs.readFileSyncImpl as any,
180
+ });
181
+
182
+ expect(result.srcRoot).toBe('src');
183
+ expect(result.router).toBe('pages');
184
+ });
185
+
186
+ it('throws when neither app nor pages exists', () => {
187
+ const fs = createVirtualFs(cwd, {
188
+ 'package.json': JSON.stringify({
189
+ dependencies: { next: '15.0.0' },
190
+ }),
191
+ });
192
+
193
+ expect(() =>
194
+ detectNextJsAndRouter({
195
+ cwd,
196
+ existsSyncImpl: fs.existsSyncImpl as any,
197
+ readFileSyncImpl: fs.readFileSyncImpl as any,
198
+ }),
199
+ ).toThrow(
200
+ 'Could not detect Next.js router. Expected either app/ or pages/ directory.',
201
+ );
202
+ });
203
+ });
204
+
205
+ describe('runInit', () => {
206
+ const cwd = '/project';
207
+ let log: ReturnType<typeof vi.fn>;
208
+ let error: ReturnType<typeof vi.fn>;
209
+ let exit: ReturnType<typeof vi.fn>;
210
+
211
+ beforeEach(() => {
212
+ log = vi.fn();
213
+ error = vi.fn();
214
+ exit = vi.fn();
215
+ });
216
+
217
+ it('creates app router files, updates package.json, and exits successfully', () => {
218
+ const fs = createVirtualFs(
219
+ cwd,
220
+ {
221
+ 'package.json': JSON.stringify({
222
+ name: 'app',
223
+ dependencies: { next: '15.0.0' },
224
+ }),
225
+ },
226
+ ['app'],
227
+ );
228
+
229
+ runInit({
230
+ cwd,
231
+ log,
232
+ error,
233
+ exit,
234
+ existsSyncImpl: fs.existsSyncImpl as any,
235
+ mkdirSyncImpl: fs.mkdirSyncImpl as any,
236
+ readFileSyncImpl: fs.readFileSyncImpl as any,
237
+ writeFileSyncImpl: fs.writeFileSyncImpl as any,
238
+ chmodSyncImpl: fs.chmodSyncImpl as any,
239
+ });
240
+
241
+ expect(
242
+ fs.state.files.get(
243
+ resolvePath(cwd, 'app/api/dataqueue/manage/[[...task]]/route.ts'),
244
+ ),
245
+ ).toBe(APP_ROUTER_ROUTE_TEMPLATE);
246
+ expect(fs.state.files.get(resolvePath(cwd, 'lib/dataqueue/queue.ts'))).toBe(
247
+ QUEUE_TEMPLATE,
248
+ );
249
+ expect(fs.state.files.get(resolvePath(cwd, 'cron.sh'))).toBe(
250
+ CRON_SH_TEMPLATE,
251
+ );
252
+ expect(fs.state.chmodCalls).toEqual([
253
+ { filePath: resolvePath(cwd, 'cron.sh'), mode: 0o755 },
254
+ ]);
255
+
256
+ const updatedPackageJson = JSON.parse(
257
+ fs.state.files.get(resolvePath(cwd, 'package.json')) || '{}',
258
+ );
259
+ expect(updatedPackageJson.dependencies['@nicnocquee/dataqueue']).toBe(
260
+ 'latest',
261
+ );
262
+ expect(
263
+ updatedPackageJson.dependencies['@nicnocquee/dataqueue-dashboard'],
264
+ ).toBe('latest');
265
+ expect(updatedPackageJson.dependencies['@nicnocquee/dataqueue-react']).toBe(
266
+ 'latest',
267
+ );
268
+ expect(updatedPackageJson.devDependencies['dotenv-cli']).toBe('latest');
269
+ expect(updatedPackageJson.devDependencies['ts-node']).toBe('latest');
270
+ expect(updatedPackageJson.devDependencies['node-pg-migrate']).toBe(
271
+ 'latest',
272
+ );
273
+ expect(updatedPackageJson.scripts.cron).toBe('bash cron.sh');
274
+ expect(updatedPackageJson.scripts['migrate-dataqueue']).toBe(
275
+ 'dotenv -e .env.local -- dataqueue-cli migrate',
276
+ );
277
+
278
+ expect(log).toHaveBeenCalledWith(
279
+ ' [skipped] pages/api/dataqueue/manage/[[...task]].ts (router not selected)',
280
+ );
281
+ expect(log).toHaveBeenCalledWith(
282
+ "Done! Run your package manager's install command to install new dependencies.",
283
+ );
284
+ expect(error).not.toHaveBeenCalled();
285
+ expect(exit).toHaveBeenCalledWith(0);
286
+ });
287
+
288
+ it('creates pages router file when only pages router exists', () => {
289
+ const fs = createVirtualFs(
290
+ cwd,
291
+ {
292
+ 'package.json': JSON.stringify({
293
+ name: 'app',
294
+ dependencies: { next: '15.0.0' },
295
+ }),
296
+ },
297
+ ['pages'],
298
+ );
299
+
300
+ runInit({
301
+ cwd,
302
+ log,
303
+ error,
304
+ exit,
305
+ existsSyncImpl: fs.existsSyncImpl as any,
306
+ mkdirSyncImpl: fs.mkdirSyncImpl as any,
307
+ readFileSyncImpl: fs.readFileSyncImpl as any,
308
+ writeFileSyncImpl: fs.writeFileSyncImpl as any,
309
+ chmodSyncImpl: fs.chmodSyncImpl as any,
310
+ });
311
+
312
+ expect(
313
+ fs.state.files.get(
314
+ resolvePath(cwd, 'pages/api/dataqueue/manage/[[...task]].ts'),
315
+ ),
316
+ ).toBe(PAGES_ROUTER_ROUTE_TEMPLATE);
317
+ expect(log).toHaveBeenCalledWith(
318
+ ' [skipped] app/api/dataqueue/manage/[[...task]]/route.ts (router not selected)',
319
+ );
320
+ expect(exit).toHaveBeenCalledWith(0);
321
+ });
322
+
323
+ it('skips existing files and existing package entries', () => {
324
+ const existingRoute = '/* existing */';
325
+ const existingCron = '#!/bin/bash\n# existing';
326
+ const existingQueue = '// existing queue';
327
+ const fs = createVirtualFs(
328
+ cwd,
329
+ {
330
+ 'package.json': JSON.stringify({
331
+ dependencies: {
332
+ next: '15.0.0',
333
+ '@nicnocquee/dataqueue': '^1.0.0',
334
+ },
335
+ devDependencies: {
336
+ 'dotenv-cli': '^8.0.0',
337
+ },
338
+ scripts: {
339
+ cron: 'bash cron.sh',
340
+ },
341
+ }),
342
+ 'app/api/dataqueue/manage/[[...task]]/route.ts': existingRoute,
343
+ 'cron.sh': existingCron,
344
+ 'lib/dataqueue/queue.ts': existingQueue,
345
+ },
346
+ ['app'],
347
+ );
348
+
349
+ runInit({
350
+ cwd,
351
+ log,
352
+ error,
353
+ exit,
354
+ existsSyncImpl: fs.existsSyncImpl as any,
355
+ mkdirSyncImpl: fs.mkdirSyncImpl as any,
356
+ readFileSyncImpl: fs.readFileSyncImpl as any,
357
+ writeFileSyncImpl: fs.writeFileSyncImpl as any,
358
+ chmodSyncImpl: fs.chmodSyncImpl as any,
359
+ });
360
+
361
+ expect(
362
+ fs.state.files.get(
363
+ resolvePath(cwd, 'app/api/dataqueue/manage/[[...task]]/route.ts'),
364
+ ),
365
+ ).toBe(existingRoute);
366
+ expect(fs.state.files.get(resolvePath(cwd, 'cron.sh'))).toBe(existingCron);
367
+ expect(fs.state.files.get(resolvePath(cwd, 'lib/dataqueue/queue.ts'))).toBe(
368
+ existingQueue,
369
+ );
370
+
371
+ const updatedPackageJson = JSON.parse(
372
+ fs.state.files.get(resolvePath(cwd, 'package.json')) || '{}',
373
+ );
374
+ expect(updatedPackageJson.dependencies['@nicnocquee/dataqueue']).toBe(
375
+ '^1.0.0',
376
+ );
377
+ expect(updatedPackageJson.scripts.cron).toBe('bash cron.sh');
378
+ expect(updatedPackageJson.scripts['migrate-dataqueue']).toBe(
379
+ 'dotenv -e .env.local -- dataqueue-cli migrate',
380
+ );
381
+ expect(log).toHaveBeenCalledWith(
382
+ ' [skipped] dependency @nicnocquee/dataqueue (already exists)',
383
+ );
384
+ expect(log).toHaveBeenCalledWith(
385
+ ' [skipped] script "cron" (already exists)',
386
+ );
387
+ expect(exit).toHaveBeenCalledWith(0);
388
+ });
389
+
390
+ it('works for a monorepo sub-app by using cwd package.json', () => {
391
+ const subAppCwd = '/repo/apps/web';
392
+ const fs = createVirtualFs(
393
+ subAppCwd,
394
+ {
395
+ 'package.json': JSON.stringify({
396
+ dependencies: { next: '15.0.0' },
397
+ }),
398
+ },
399
+ ['app'],
400
+ );
401
+
402
+ runInit({
403
+ cwd: subAppCwd,
404
+ log,
405
+ error,
406
+ exit,
407
+ existsSyncImpl: fs.existsSyncImpl as any,
408
+ mkdirSyncImpl: fs.mkdirSyncImpl as any,
409
+ readFileSyncImpl: fs.readFileSyncImpl as any,
410
+ writeFileSyncImpl: fs.writeFileSyncImpl as any,
411
+ chmodSyncImpl: fs.chmodSyncImpl as any,
412
+ });
413
+
414
+ expect(
415
+ fs.state.files.has(
416
+ resolvePath(subAppCwd, 'app/api/dataqueue/manage/[[...task]]/route.ts'),
417
+ ),
418
+ ).toBe(true);
419
+ expect(exit).toHaveBeenCalledWith(0);
420
+ });
421
+
422
+ it('logs an error and exits with code 1 for invalid package.json', () => {
423
+ const fs = createVirtualFs(
424
+ cwd,
425
+ {
426
+ 'package.json': '{invalid json',
427
+ },
428
+ ['app'],
429
+ );
430
+
431
+ runInit({
432
+ cwd,
433
+ log,
434
+ error,
435
+ exit,
436
+ existsSyncImpl: fs.existsSyncImpl as any,
437
+ mkdirSyncImpl: fs.mkdirSyncImpl as any,
438
+ readFileSyncImpl: fs.readFileSyncImpl as any,
439
+ writeFileSyncImpl: fs.writeFileSyncImpl as any,
440
+ chmodSyncImpl: fs.chmodSyncImpl as any,
441
+ });
442
+
443
+ expect(error).toHaveBeenCalledTimes(1);
444
+ expect(String(error.mock.calls[0][0])).toContain(
445
+ 'Failed to parse package.json',
446
+ );
447
+ expect(exit).toHaveBeenCalledWith(1);
448
+ });
449
+ });