@danceroutine/tango-testing 0.1.0 → 1.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (88) hide show
  1. package/LICENSE +21 -0
  2. package/README.md +85 -0
  3. package/dist/aDBClient-W6eXsK3X.js +21 -0
  4. package/dist/aDBClient-W6eXsK3X.js.map +1 -0
  5. package/dist/assertions/index.js +1 -1
  6. package/dist/{assertions-CN6KxXhH.js → assertions-CCFZ53Y-.js} +1 -1
  7. package/dist/assertions-CCFZ53Y-.js.map +1 -0
  8. package/dist/express/anExpressRequest.d.ts +24 -0
  9. package/dist/express/anExpressResponse.d.ts +9 -0
  10. package/dist/express/index.d.ts +3 -0
  11. package/dist/express/index.js +3 -0
  12. package/dist/express-Czpfz_Ay.js +68 -0
  13. package/dist/express-Czpfz_Ay.js.map +1 -0
  14. package/dist/factories/ModelDataFactory.d.ts +16 -1
  15. package/dist/factories/index.js +1 -1
  16. package/dist/{factories-CCAZ6E-g.js → factories-Cl_CAzbj.js} +19 -4
  17. package/dist/factories-Cl_CAzbj.js.map +1 -0
  18. package/dist/index.d.ts +5 -3
  19. package/dist/index.js +8 -11
  20. package/dist/integration/HarnessStrategyRegistry.d.ts +15 -0
  21. package/dist/integration/TestHarness.d.ts +23 -2
  22. package/dist/integration/anIntegrationHarness.d.ts +5 -0
  23. package/dist/integration/config.d.ts +4 -0
  24. package/dist/integration/conformance/index.d.ts +1 -0
  25. package/dist/integration/conformance/runDialectConformanceSuite.d.ts +11 -0
  26. package/dist/integration/domain/Dialect.d.ts +5 -4
  27. package/dist/integration/domain/ResetMode.d.ts +6 -5
  28. package/dist/integration/index.d.ts +8 -1
  29. package/dist/integration/index.js +3 -2
  30. package/dist/integration/migrations/ApplyAndVerifyMigrations.d.ts +3 -0
  31. package/dist/integration/migrations/AssertMigrationPlan.d.ts +3 -0
  32. package/dist/integration/migrations/IntrospectSchema.d.ts +3 -0
  33. package/dist/integration/orm/createQuerySetFixture.d.ts +10 -0
  34. package/dist/integration/orm/expectQueryResult.d.ts +4 -0
  35. package/dist/integration/orm/index.d.ts +6 -0
  36. package/dist/integration/orm/seedTable.d.ts +5 -0
  37. package/dist/integration/runtime/aTangoConfig.d.ts +8 -0
  38. package/dist/integration/runtime/index.d.ts +6 -0
  39. package/dist/integration/runtime/setupTestTangoRuntime.d.ts +6 -0
  40. package/dist/integration/smoke/AppProcessHarness.d.ts +83 -0
  41. package/dist/integration/smoke/index.d.ts +4 -0
  42. package/dist/integration/strategies/PostgresHarnessStrategy.d.ts +9 -0
  43. package/dist/integration/strategies/SqliteHarnessStrategy.d.ts +9 -0
  44. package/dist/integration-BrJw6NzG.js +747 -0
  45. package/dist/integration-BrJw6NzG.js.map +1 -0
  46. package/dist/mocks/DBClient.d.ts +1 -9
  47. package/dist/mocks/MockQuerySetResult.d.ts +5 -12
  48. package/dist/mocks/aDBClient.d.ts +21 -0
  49. package/dist/mocks/aManager.d.ts +17 -0
  50. package/dist/mocks/aQueryExecutor.d.ts +14 -0
  51. package/dist/mocks/aQueryResult.d.ts +5 -0
  52. package/dist/mocks/aQuerySet.d.ts +8 -0
  53. package/dist/mocks/aRequestContext.d.ts +22 -0
  54. package/dist/mocks/index.d.ts +9 -4
  55. package/dist/mocks/index.js +4 -6
  56. package/dist/mocks-BkwkXQQt.js +136 -0
  57. package/dist/mocks-BkwkXQQt.js.map +1 -0
  58. package/dist/vitest/index.js +3 -2
  59. package/dist/vitest/registerVitestTango.d.ts +3 -3
  60. package/dist/{vitest-PxMJue7R.js → vitest-37qN8D93.js} +4 -4
  61. package/dist/vitest-37qN8D93.js.map +1 -0
  62. package/package.json +81 -68
  63. package/dist/assertions/assertions.js +0 -8
  64. package/dist/assertions-CN6KxXhH.js.map +0 -1
  65. package/dist/factories/ModelDataFactory.js +0 -33
  66. package/dist/factories-CCAZ6E-g.js.map +0 -1
  67. package/dist/index.js.map +0 -1
  68. package/dist/integration/orm.d.ts +0 -9
  69. package/dist/integration/orm.js +0 -39
  70. package/dist/integration/strategies/PostgresHarnessStrategy.js +0 -95
  71. package/dist/integration-CDdpboYz.js +0 -378
  72. package/dist/integration-CDdpboYz.js.map +0 -1
  73. package/dist/mocks/DBClient.js +0 -1
  74. package/dist/mocks/MockQuerySetResult.js +0 -1
  75. package/dist/mocks/RepositoryLike.d.ts +0 -12
  76. package/dist/mocks/RepositoryLike.js +0 -1
  77. package/dist/mocks/aMockDBClient.d.ts +0 -2
  78. package/dist/mocks/aMockDBClient.js +0 -13
  79. package/dist/mocks/aMockQuerySet.d.ts +0 -2
  80. package/dist/mocks/aMockQuerySet.js +0 -15
  81. package/dist/mocks/aMockRepository.d.ts +0 -2
  82. package/dist/mocks/aMockRepository.js +0 -20
  83. package/dist/mocks/types.d.ts +0 -33
  84. package/dist/mocks-qo-1vCez.js +0 -72
  85. package/dist/mocks-qo-1vCez.js.map +0 -1
  86. package/dist/version.d.ts +0 -1
  87. package/dist/vitest/registerVitestTango.js +0 -90
  88. package/dist/vitest-PxMJue7R.js.map +0 -1
@@ -0,0 +1 @@
1
+ {"version":3,"file":"integration-BrJw6NzG.js","names":["harness: IntegrationHarness","options: AssertMigrationPlanOptions","harness: IntegrationHarness","options: ApplyAndVerifyMigrationsOptions","status: MigrationStatus","harness: IntegrationHarness","strategy: HarnessStrategy","options: {\n createOptions?: HarnessOptions;\n migrationsDir?: string;\n }","options: TestTangoConfigOptions","options: TestTangoConfigOptions","defaultDeps: AppProcessHarnessDeps","ms: number","options: AppProcessHarnessOptions","deps: AppProcessHarnessDeps","chunk: unknown","value: unknown","deps: Partial<AppProcessHarnessDeps>","mergedDeps: AppProcessHarnessDeps","path: string","init?: RequestInit","response: Response","expectedStatus: number","label: string","bodyText: string","timeoutMs: number","current: string","chunk: string","defaultCapabilities: DialectTestCapabilities","overrides: Partial<IntegrationHarness>","value: unknown","strategy: HarnessStrategy","dialect: Dialect | string","value: string | undefined","dialect: Dialect","opts: {\n config?: Partial<AdapterConfig>;\n tangoConfigLoader?: () => unknown;\n sqliteFile?: string;\n }","merged: AdapterConfig","value: unknown","explicitSchema?: string","options: HarnessOptions","client: DBClient | null","harness: IntegrationHarness","migrationsDir: string","value: unknown","client: DBClient","options: HarnessOptions","client: DBClient | null","harness: IntegrationHarness","migrationsDir: string","value: unknown","strategy: HarnessStrategy","args: { dialect: Dialect | string; options?: HarnessOptions }","registry?: HarnessStrategyRegistry","options?: HarnessOptions","harness: IntegrationHarness","table: string","rows: T[]","input: {\n harness: IntegrationHarness;\n meta: TableMeta;\n}","executor: QueryExecutor<TModel>","actual: Promise<T> | T","expected: T"],"sources":["../src/integration/domain/Dialect.ts","../src/integration/domain/ResetMode.ts","../src/integration/domain/index.ts","../src/integration/migrations/AssertMigrationPlan.ts","../src/integration/migrations/ApplyAndVerifyMigrations.ts","../src/integration/migrations/IntrospectSchema.ts","../src/integration/migrations/index.ts","../src/integration/conformance/runDialectConformanceSuite.ts","../src/integration/conformance/index.ts","../src/integration/runtime/aTangoConfig.ts","../src/integration/runtime/setupTestTangoRuntime.ts","../src/integration/runtime/index.ts","../src/integration/smoke/AppProcessHarness.ts","../src/integration/smoke/index.ts","../src/integration/anIntegrationHarness.ts","../src/integration/HarnessStrategyRegistry.ts","../src/integration/config.ts","../src/integration/strategies/PostgresHarnessStrategy.ts","../src/integration/strategies/SqliteHarnessStrategy.ts","../src/integration/TestHarness.ts","../src/integration/orm/seedTable.ts","../src/integration/orm/createQuerySetFixture.ts","../src/integration/orm/expectQueryResult.ts","../src/integration/index.ts"],"sourcesContent":["export const Dialect = {\n Sqlite: 'sqlite',\n Postgres: 'postgres',\n} as const;\n\nexport type Dialect = (typeof Dialect)[keyof typeof Dialect];\n","export const ResetMode = {\n Transaction: 'transaction',\n Truncate: 'truncate',\n DropSchema: 'drop-schema',\n} as const;\n\nexport type ResetMode = (typeof ResetMode)[keyof typeof ResetMode];\n","/**\n * Domain boundary barrel: centralizes this subdomain's public contract.\n */\n\nexport { Dialect } from './Dialect';\nexport type { HarnessOptions, HarnessStrategy } from './HarnessStrategy';\nexport type { DialectTestCapabilities, IntegrationHarness } from './IntegrationHarness';\nexport { ResetMode } from './ResetMode';\n","import type { IntegrationHarness } from '../domain';\n\nexport type AssertMigrationPlanOptions = {\n migrationsDir: string;\n expectSqlContains?: string[];\n};\n\n/**\n * Generate a migration plan through a harness and assert that required SQL fragments appear.\n */\nexport async function assertMigrationPlan(\n harness: IntegrationHarness,\n options: AssertMigrationPlanOptions\n): Promise<string> {\n const runner = harness.migrationRunner(options.migrationsDir);\n const plan = await runner.plan();\n\n for (const snippet of options.expectSqlContains ?? []) {\n if (!plan.includes(snippet)) {\n throw new Error(`Expected migration plan to contain: ${snippet}`);\n }\n }\n\n return plan;\n}\n","import type { IntegrationHarness } from '../domain';\n\nexport type ApplyAndVerifyMigrationsOptions = {\n migrationsDir: string;\n toId?: string;\n expectedAppliedIds?: string[];\n};\n\nexport type MigrationStatus = { id: string; applied: boolean };\n\n/**\n * Apply migrations through a harness and optionally verify that specific ids were applied.\n */\nexport async function applyAndVerifyMigrations(\n harness: IntegrationHarness,\n options: ApplyAndVerifyMigrationsOptions\n): Promise<{ statuses: MigrationStatus[] }> {\n const runner = harness.migrationRunner(options.migrationsDir);\n await runner.apply(options.toId);\n const statuses = await runner.status();\n\n for (const id of options.expectedAppliedIds ?? []) {\n const row = statuses.find((status: MigrationStatus) => status.id === id);\n if (!row || !row.applied) {\n throw new Error(`Expected migration ${id} to be applied`);\n }\n }\n\n return { statuses };\n}\n","import { createDefaultIntrospectorStrategy } from '@danceroutine/tango-migrations';\nimport type { Dialect as MigrationDialect } from '@danceroutine/tango-migrations';\nimport { Dialect, type IntegrationHarness } from '../domain';\n\nconst introspectorStrategy = createDefaultIntrospectorStrategy();\n\n/**\n * Introspect the schema visible to a harness using Tango's built-in introspectors.\n */\nexport async function introspectSchema(harness: IntegrationHarness): Promise<unknown> {\n if (harness.dialect !== Dialect.Postgres && harness.dialect !== Dialect.Sqlite) {\n throw new Error(`No introspector registered for dialect: ${String(harness.dialect)}`);\n }\n const dialect = harness.dialect === Dialect.Postgres ? 'postgres' : 'sqlite';\n return introspectorStrategy.introspect(dialect as unknown as MigrationDialect, harness.dbClient);\n}\n","/**\n * Domain boundary barrel: centralizes this subdomain's public contract.\n */\n\nexport { assertMigrationPlan, type AssertMigrationPlanOptions } from './AssertMigrationPlan';\nexport {\n applyAndVerifyMigrations,\n type ApplyAndVerifyMigrationsOptions,\n type MigrationStatus,\n} from './ApplyAndVerifyMigrations';\nexport { introspectSchema } from './IntrospectSchema';\n","import type { HarnessOptions, HarnessStrategy } from '../domain';\n\n/**\n * Shared conformance checks every harness strategy must satisfy.\n *\n * This is intentionally framework-agnostic so first-party and third-party\n * strategies can run the same lifecycle validation.\n */\nexport async function runDialectConformanceSuite(\n strategy: HarnessStrategy,\n options: {\n createOptions?: HarnessOptions;\n migrationsDir?: string;\n } = {}\n): Promise<void> {\n const harness = await strategy.create(options.createOptions);\n\n if (harness.dialect !== strategy.dialect) {\n throw new Error(\n `Conformance failed: harness dialect '${String(harness.dialect)}' does not match strategy dialect '${String(strategy.dialect)}'`\n );\n }\n\n if (harness.capabilities !== strategy.capabilities) {\n throw new Error('Conformance failed: harness capabilities must be strategy capabilities reference');\n }\n\n let resetBeforeSetupThrew = false;\n try {\n await harness.reset();\n } catch {\n resetBeforeSetupThrew = true;\n }\n if (!resetBeforeSetupThrew) {\n throw new Error('Conformance failed: reset() must throw before setup()');\n }\n\n await harness.setup();\n await harness.reset();\n harness.migrationRunner(options.migrationsDir ?? '/tmp/migrations');\n await harness.teardown();\n\n let dbClientAfterTeardownThrew = false;\n try {\n // Access after teardown should fail to prevent stale client usage.\n\n // oxlint-disable-next-line no-unused-expressions\n harness.dbClient;\n } catch {\n dbClientAfterTeardownThrew = true;\n }\n if (!dbClientAfterTeardownThrew) {\n throw new Error('Conformance failed: dbClient getter must throw after teardown()');\n }\n}\n","export { runDialectConformanceSuite } from './runDialectConformanceSuite';\n","import { defineConfig, type TangoConfig } from '@danceroutine/tango-config';\n\nexport type TestTangoConfigOptions = {\n adapter?: 'sqlite' | 'postgres';\n};\n\n/**\n * Create a stable Tango config fixture for runtime-oriented tests.\n */\nexport function aTangoConfig(options: TestTangoConfigOptions = {}): TangoConfig {\n const adapter = options.adapter ?? 'sqlite';\n\n return defineConfig({\n current: 'test',\n environments: {\n development: {\n name: 'development',\n db:\n adapter === 'sqlite'\n ? { adapter: 'sqlite', filename: ':memory:', maxConnections: 1 }\n : {\n adapter: 'postgres',\n url: 'postgres://postgres:postgres@localhost:5432/tango',\n maxConnections: 1,\n },\n migrations: { dir: 'migrations', online: adapter === 'postgres' },\n },\n test: {\n name: 'test',\n db:\n adapter === 'sqlite'\n ? { adapter: 'sqlite', filename: ':memory:', maxConnections: 1 }\n : {\n adapter: 'postgres',\n url: 'postgres://postgres:postgres@localhost:5432/tango_test',\n maxConnections: 1,\n },\n migrations: { dir: 'migrations', online: adapter === 'postgres' },\n },\n production: {\n name: 'production',\n db:\n adapter === 'sqlite'\n ? { adapter: 'sqlite', filename: ':memory:', maxConnections: 1 }\n : {\n adapter: 'postgres',\n url: 'postgres://postgres:postgres@localhost:5432/tango',\n maxConnections: 1,\n },\n migrations: { dir: 'migrations', online: adapter === 'postgres' },\n },\n },\n });\n}\n","import { initializeTangoRuntime, resetTangoRuntime, type TangoRuntime } from '@danceroutine/tango-orm';\nimport { aTangoConfig, type TestTangoConfigOptions } from './aTangoConfig';\n\n/**\n * Reset and initialize the process-default Tango runtime for tests.\n */\nexport async function setupTestTangoRuntime(options: TestTangoConfigOptions = {}): Promise<TangoRuntime> {\n await resetTangoRuntime();\n return initializeTangoRuntime(() => aTangoConfig(options));\n}\n","/**\n * Domain boundary barrel: centralizes runtime-oriented integration fixtures.\n */\n\nexport { aTangoConfig } from './aTangoConfig';\nexport type { TestTangoConfigOptions } from './aTangoConfig';\nexport { setupTestTangoRuntime } from './setupTestTangoRuntime';\n","import { spawn } from 'node:child_process';\n\nexport interface AppProcessHarnessOptions {\n command: string;\n args?: string[];\n cwd?: string;\n env?: Record<string, string | undefined>;\n baseUrl: string;\n readyPath?: string;\n readyTimeoutMs?: number;\n readyIntervalMs?: number;\n stopTimeoutMs?: number;\n}\n\ntype FetchLike = typeof fetch;\ntype ProcessHarnessOutputStream = {\n on(eventName: 'data', listener: (chunk: unknown) => void): unknown;\n};\ntype ProcessHarnessChild = {\n exitCode: number | null;\n killed: boolean;\n kill(signal?: NodeJS.Signals): boolean;\n off(eventName: 'exit', listener: () => void): unknown;\n once(eventName: 'exit', listener: () => void): unknown;\n stdout?: ProcessHarnessOutputStream | null;\n stderr?: ProcessHarnessOutputStream | null;\n};\n\ninterface AppProcessHarnessDeps {\n spawnProcess: (\n command: string,\n args?: readonly string[],\n options?: Parameters<typeof spawn>[2]\n ) => ProcessHarnessChild;\n fetchImpl: FetchLike;\n sleep: (ms: number) => Promise<void>;\n}\n\nconst DEFAULT_READY_TIMEOUT_MS = 30_000;\nconst DEFAULT_READY_INTERVAL_MS = 250;\nconst DEFAULT_STOP_TIMEOUT_MS = 10_000;\nconst MAX_LOG_BUFFER_CHARS = 20_000;\n\nconst defaultDeps: AppProcessHarnessDeps = {\n spawnProcess: (command, args, options) => spawn(command, args as string[], options as Parameters<typeof spawn>[2]),\n fetchImpl: fetch,\n sleep: (ms: number) =>\n new Promise((resolve) => {\n setTimeout(resolve, ms);\n }),\n};\n\n/**\n * Lightweight process harness for end-to-end smoke tests that need a real app process.\n */\nexport class AppProcessHarness {\n static readonly BRAND = 'tango.testing.app_process_harness' as const;\n readonly __tangoBrand: typeof AppProcessHarness.BRAND = AppProcessHarness.BRAND;\n private readonly child: ProcessHarnessChild;\n private readonly baseUrl: string;\n private readonly readyUrl: string;\n private readonly readyTimeoutMs: number;\n private readonly readyIntervalMs: number;\n private readonly stopTimeoutMs: number;\n private readonly deps: AppProcessHarnessDeps;\n private stopped = false;\n private stdoutBuffer = '';\n private stderrBuffer = '';\n\n private constructor(options: AppProcessHarnessOptions, deps: AppProcessHarnessDeps) {\n this.baseUrl = options.baseUrl.replace(/\\/+$/, '');\n this.readyUrl = `${this.baseUrl}${normalizePath(options.readyPath ?? '/health')}`;\n this.readyTimeoutMs = options.readyTimeoutMs ?? DEFAULT_READY_TIMEOUT_MS;\n this.readyIntervalMs = options.readyIntervalMs ?? DEFAULT_READY_INTERVAL_MS;\n this.stopTimeoutMs = options.stopTimeoutMs ?? DEFAULT_STOP_TIMEOUT_MS;\n this.deps = deps;\n\n this.child = this.deps.spawnProcess(options.command, options.args ?? [], {\n cwd: options.cwd,\n env: { ...process.env, ...options.env },\n stdio: 'pipe',\n });\n this.child.stdout?.on('data', (chunk: unknown) => {\n this.stdoutBuffer = appendBuffer(this.stdoutBuffer, String(chunk));\n });\n this.child.stderr?.on('data', (chunk: unknown) => {\n this.stderrBuffer = appendBuffer(this.stderrBuffer, String(chunk));\n });\n }\n\n /**\n * Narrow an unknown value to the smoke-test harness that owns a child process.\n */\n static isAppProcessHarness(value: unknown): value is AppProcessHarness {\n return (\n typeof value === 'object' &&\n value !== null &&\n (value as { __tangoBrand?: unknown }).__tangoBrand === AppProcessHarness.BRAND\n );\n }\n\n /**\n * Spawn the target process and wait until its readiness endpoint responds successfully.\n */\n static async start(\n options: AppProcessHarnessOptions,\n deps: Partial<AppProcessHarnessDeps> = {}\n ): Promise<AppProcessHarness> {\n const mergedDeps: AppProcessHarnessDeps = {\n ...defaultDeps,\n ...deps,\n };\n const harness = new AppProcessHarness(options, mergedDeps);\n await harness.waitForReady();\n return harness;\n }\n\n /**\n * Return the buffered stdout log for recent process output.\n */\n getStdoutLog(): string {\n return this.stdoutBuffer;\n }\n\n /**\n * Return the buffered stderr log for recent process output.\n */\n getStderrLog(): string {\n return this.stderrBuffer;\n }\n\n /**\n * Return stdout and stderr in a single formatted string for debugging failures.\n */\n getCombinedLog(): string {\n const stdout = this.stdoutBuffer.trim();\n const stderr = this.stderrBuffer.trim();\n if (!stdout && !stderr) {\n return '';\n }\n return [`[stdout]\\n${stdout}`, `[stderr]\\n${stderr}`].join('\\n\\n').trim();\n }\n\n /**\n * Issue an HTTP request against the managed application process.\n */\n async request(path: string, init?: RequestInit): Promise<Response> {\n const target = path.startsWith('http') ? path : `${this.baseUrl}${normalizePath(path)}`;\n return this.deps.fetchImpl(target, init);\n }\n\n /**\n * Assert an HTTP response status and include process logs when it mismatches.\n */\n async assertResponseStatus(response: Response, expectedStatus: number, label: string): Promise<void> {\n if (response.status === expectedStatus) {\n return;\n }\n\n let bodyText: string;\n try {\n bodyText = await response.text();\n } catch (error) {\n bodyText = `failed to read response body: ${String(error)}`;\n }\n\n throw new Error(\n [\n `${label}. expected ${String(expectedStatus)} got ${String(response.status)}`,\n `response body: ${bodyText}`,\n `process logs:\\n${this.getCombinedLog()}`,\n ].join('\\n')\n );\n }\n\n /**\n * Stop the managed process, escalating from SIGTERM to SIGKILL when necessary.\n */\n async stop(): Promise<void> {\n if (this.stopped) {\n return;\n }\n this.stopped = true;\n\n if (this.child.exitCode !== null || this.child.killed) {\n return;\n }\n\n this.child.kill('SIGTERM');\n const exited = await this.waitForExit(this.stopTimeoutMs);\n if (!exited && !this.child.killed) {\n this.child.kill('SIGKILL');\n await this.waitForExit(this.stopTimeoutMs);\n }\n }\n\n private async waitForReady(): Promise<void> {\n const deadline = Date.now() + this.readyTimeoutMs;\n\n while (Date.now() < deadline) {\n if (this.child.exitCode !== null) {\n throw new Error(\n `Process exited before ready check succeeded (exitCode=${String(this.child.exitCode)}).\\n${this.getCombinedLog()}`\n );\n }\n try {\n const response = await this.deps.fetchImpl(this.readyUrl);\n if (response.ok) {\n return;\n }\n } catch {\n // Retry until timeout.\n }\n await this.deps.sleep(this.readyIntervalMs);\n }\n\n await this.stop();\n throw new Error(`Timed out waiting for readiness at ${this.readyUrl}.\\n${this.getCombinedLog()}`);\n }\n\n private async waitForExit(timeoutMs: number): Promise<boolean> {\n if (this.child.exitCode !== null) {\n return true;\n }\n\n return await new Promise<boolean>((resolve) => {\n const timer = setTimeout(() => {\n this.child.off('exit', onExit);\n resolve(false);\n }, timeoutMs);\n\n const onExit = () => {\n clearTimeout(timer);\n resolve(true);\n };\n\n this.child.once('exit', onExit);\n });\n }\n}\n\nfunction appendBuffer(current: string, chunk: string): string {\n const next = current + chunk;\n if (next.length <= MAX_LOG_BUFFER_CHARS) {\n return next;\n }\n return next.slice(next.length - MAX_LOG_BUFFER_CHARS);\n}\n\nfunction normalizePath(path: string): string {\n if (!path) {\n return '/';\n }\n return path.startsWith('/') ? path : `/${path}`;\n}\n","/**\n * Domain boundary barrel: centralizes this subdomain's public contract.\n */\n\nexport * from './AppProcessHarness';\n","import { vi } from 'vitest';\nimport { aDBClient } from '../mocks/aDBClient';\nimport { Dialect } from './domain/Dialect';\nimport { ResetMode } from './domain/ResetMode';\nimport type { DialectTestCapabilities, IntegrationHarness } from './domain/IntegrationHarness';\n\nconst defaultCapabilities: DialectTestCapabilities = {\n transactionalDDL: true,\n supportsSchemas: false,\n supportsConcurrentIndex: false,\n supportsDeferredFkValidation: false,\n supportsJsonb: false,\n};\n\n/**\n * Create an integration-harness fixture with optional overrides.\n */\nexport function anIntegrationHarness(overrides: Partial<IntegrationHarness> = {}): IntegrationHarness {\n return {\n dialect: Dialect.Sqlite,\n capabilities: defaultCapabilities,\n resetMode: ResetMode.DropSchema,\n dbClient: aDBClient(),\n setup: vi.fn(async () => {}),\n reset: vi.fn(async () => {}),\n teardown: vi.fn(async () => {}),\n migrationRunner: vi.fn(() => ({}) as never),\n ...overrides,\n };\n}\n","import type { Dialect, HarnessStrategy } from './domain';\n\n/**\n * Registry of test harness strategies keyed by dialect.\n */\nexport class HarnessStrategyRegistry {\n static readonly BRAND = 'tango.testing.harness_strategy_registry' as const;\n readonly __tangoBrand: typeof HarnessStrategyRegistry.BRAND = HarnessStrategyRegistry.BRAND;\n private readonly strategies = new Map<string, HarnessStrategy>();\n\n /**\n * Narrow an unknown value to `HarnessStrategyRegistry`.\n */\n static isHarnessStrategyRegistry(value: unknown): value is HarnessStrategyRegistry {\n return (\n typeof value === 'object' &&\n value !== null &&\n (value as { __tangoBrand?: unknown }).__tangoBrand === HarnessStrategyRegistry.BRAND\n );\n }\n\n /**\n * Register or replace a dialect strategy.\n */\n register(strategy: HarnessStrategy): this {\n this.strategies.set(String(strategy.dialect), strategy);\n return this;\n }\n\n /**\n * Resolve a strategy for a dialect, or throw if none is registered.\n */\n get(dialect: Dialect | string): HarnessStrategy {\n const strategy = this.strategies.get(String(dialect));\n if (!strategy) {\n throw new Error(`No harness strategy registered for dialect: ${String(dialect)}`);\n }\n return strategy;\n }\n\n /**\n * List all registered strategies.\n */\n list(): readonly HarnessStrategy[] {\n return [...this.strategies.values()];\n }\n}\n","import type { AdapterConfig } from '@danceroutine/tango-orm';\nimport { loadConfig } from '@danceroutine/tango-config';\nimport { Dialect } from './domain';\n\nfunction readNumber(value: string | undefined): number | undefined {\n if (!value) return undefined;\n const parsed = Number(value);\n return Number.isFinite(parsed) ? parsed : undefined;\n}\n\n/**\n * Resolve adapter configuration for a test harness from explicit options,\n * typed Tango config, and environment variables in that order.\n */\nexport function resolveAdapterConfig(\n dialect: Dialect,\n opts: {\n config?: Partial<AdapterConfig>;\n tangoConfigLoader?: () => unknown;\n sqliteFile?: string;\n }\n): AdapterConfig {\n const fromOptions = opts.config ?? {};\n\n if (opts.tangoConfigLoader) {\n const loaded = loadConfig(opts.tangoConfigLoader);\n const current = loaded.current.db;\n const merged: AdapterConfig = {\n url: fromOptions.url ?? current.url,\n host: fromOptions.host ?? current.host,\n port: fromOptions.port ?? current.port,\n database: fromOptions.database ?? current.database,\n user: fromOptions.user ?? current.user,\n password: fromOptions.password ?? current.password,\n filename: fromOptions.filename ?? current.filename,\n maxConnections: fromOptions.maxConnections ?? current.maxConnections,\n };\n if (dialect === Dialect.Sqlite) {\n merged.filename = opts.sqliteFile ?? merged.filename ?? ':memory:';\n }\n return merged;\n }\n\n if (dialect === Dialect.Postgres) {\n return {\n url: fromOptions.url ?? process.env.TANGO_DATABASE_URL ?? process.env.DATABASE_URL,\n host: fromOptions.host ?? process.env.TANGO_DB_HOST,\n port: fromOptions.port ?? readNumber(process.env.TANGO_DB_PORT),\n database: fromOptions.database ?? process.env.TANGO_DB_NAME,\n user: fromOptions.user ?? process.env.TANGO_DB_USER,\n password: fromOptions.password ?? process.env.TANGO_DB_PASSWORD,\n maxConnections: fromOptions.maxConnections ?? 10,\n };\n }\n\n return {\n filename: opts.sqliteFile ?? fromOptions.filename ?? process.env.TANGO_SQLITE_FILENAME ?? ':memory:',\n maxConnections: fromOptions.maxConnections ?? 1,\n };\n}\n","import { quoteSqlIdentifier, validateSqlIdentifier } from '@danceroutine/tango-core';\nimport { MigrationRunner } from '@danceroutine/tango-migrations';\nimport type { Dialect as MigrationDialect } from '@danceroutine/tango-migrations';\nimport { PostgresAdapter } from '@danceroutine/tango-orm/connection';\nimport type { DBClient } from '@danceroutine/tango-orm';\nimport { resolveAdapterConfig } from '../config';\nimport {\n Dialect,\n ResetMode,\n type DialectTestCapabilities,\n type HarnessOptions,\n type HarnessStrategy,\n type IntegrationHarness,\n} from '../domain';\n\n/**\n * Harness strategy for PostgreSQL-backed integration tests.\n */\nexport class PostgresHarnessStrategy implements HarnessStrategy {\n static readonly BRAND = 'tango.testing.postgres_harness_strategy' as const;\n readonly __tangoBrand: typeof PostgresHarnessStrategy.BRAND = PostgresHarnessStrategy.BRAND;\n readonly dialect: Dialect = Dialect.Postgres;\n readonly capabilities: DialectTestCapabilities = {\n transactionalDDL: true,\n supportsSchemas: true,\n supportsConcurrentIndex: true,\n supportsDeferredFkValidation: true,\n supportsJsonb: true,\n };\n\n /**\n * Narrow an unknown value to the PostgreSQL integration harness strategy.\n */\n static isPostgresHarnessStrategy(value: unknown): value is PostgresHarnessStrategy {\n return (\n typeof value === 'object' &&\n value !== null &&\n (value as { __tangoBrand?: unknown }).__tangoBrand === PostgresHarnessStrategy.BRAND\n );\n }\n\n private static buildSchemaName(explicitSchema?: string): string {\n if (explicitSchema) return explicitSchema;\n const random = Math.random().toString(36).slice(2, 8);\n return `tango_test_${Date.now()}_${random}`;\n }\n\n /**\n * Create a configured Postgres integration harness instance.\n */\n async create(options: HarnessOptions = {}): Promise<IntegrationHarness> {\n const config = resolveAdapterConfig(Dialect.Postgres, {\n config: options.config,\n tangoConfigLoader: options.tangoConfigLoader,\n });\n\n const adapter = new PostgresAdapter();\n const schemaName = PostgresHarnessStrategy.buildSchemaName(options.schema);\n const resetMode = options.resetMode ?? ResetMode.DropSchema;\n let client: DBClient | null = null;\n\n const ensureSearchPath = async (): Promise<void> => {\n const dbClient = client as DBClient;\n const schema = quoteSqlIdentifier(validateSqlIdentifier(schemaName, 'schema'), 'postgres');\n await dbClient.query(`CREATE SCHEMA IF NOT EXISTS ${schema}`);\n await dbClient.query(`SET search_path TO ${schema}`);\n };\n\n const recreateSchema = async (): Promise<void> => {\n const dbClient = client as DBClient;\n const schema = quoteSqlIdentifier(validateSqlIdentifier(schemaName, 'schema'), 'postgres');\n await dbClient.query(`DROP SCHEMA IF EXISTS ${schema} CASCADE`);\n await dbClient.query(`CREATE SCHEMA ${schema}`);\n await dbClient.query(`SET search_path TO ${schema}`);\n };\n\n const harness: IntegrationHarness = {\n dialect: Dialect.Postgres,\n capabilities: this.capabilities,\n resetMode,\n get dbClient(): DBClient {\n if (!client) {\n throw new Error('Postgres harness not initialized. Call setup() first.');\n }\n return client;\n },\n async setup(): Promise<void> {\n client = await adapter.connect(config);\n await ensureSearchPath();\n },\n async reset(): Promise<void> {\n if (!client) {\n throw new Error('Postgres harness not initialized. Call setup() first.');\n }\n if (resetMode === ResetMode.DropSchema || resetMode === ResetMode.Transaction) {\n await recreateSchema();\n return;\n }\n\n const { rows } = await client.query<{ table_name: string }>(\n `SELECT table_name FROM information_schema.tables WHERE table_schema = $1 AND table_type = 'BASE TABLE'`,\n [schemaName]\n );\n for (const row of rows) {\n const schema = quoteSqlIdentifier(validateSqlIdentifier(schemaName, 'schema'), 'postgres');\n const table = quoteSqlIdentifier(\n validateSqlIdentifier(String(row.table_name), 'table'),\n 'postgres'\n );\n await client.query(`TRUNCATE TABLE ${schema}.${table} RESTART IDENTITY CASCADE`);\n }\n await client.query(\n `SET search_path TO ${quoteSqlIdentifier(validateSqlIdentifier(schemaName, 'schema'), 'postgres')}`\n );\n },\n async teardown(): Promise<void> {\n if (!client) return;\n try {\n await client.query(\n `DROP SCHEMA IF EXISTS ${quoteSqlIdentifier(validateSqlIdentifier(schemaName, 'schema'), 'postgres')} CASCADE`\n );\n } finally {\n await client.close();\n client = null;\n }\n },\n migrationRunner(migrationsDir: string): MigrationRunner {\n if (!client) {\n throw new Error('Postgres harness not initialized. Call setup() first.');\n }\n return new MigrationRunner(client, 'postgres' as MigrationDialect, migrationsDir);\n },\n };\n\n return harness;\n }\n}\n","import { rm } from 'node:fs/promises';\nimport { quoteSqlIdentifier, validateSqlIdentifier } from '@danceroutine/tango-core';\nimport { MigrationRunner } from '@danceroutine/tango-migrations';\nimport type { Dialect as MigrationDialect } from '@danceroutine/tango-migrations';\nimport { SqliteAdapter } from '@danceroutine/tango-orm/connection';\nimport type { DBClient } from '@danceroutine/tango-orm';\nimport { resolveAdapterConfig } from '../config';\nimport {\n Dialect,\n ResetMode,\n type DialectTestCapabilities,\n type HarnessOptions,\n type HarnessStrategy,\n type IntegrationHarness,\n} from '../domain';\n\n/**\n * Harness strategy for SQLite-backed integration tests.\n */\nexport class SqliteHarnessStrategy implements HarnessStrategy {\n static readonly BRAND = 'tango.testing.sqlite_harness_strategy' as const;\n readonly __tangoBrand: typeof SqliteHarnessStrategy.BRAND = SqliteHarnessStrategy.BRAND;\n readonly dialect: Dialect = Dialect.Sqlite;\n readonly capabilities: DialectTestCapabilities = {\n transactionalDDL: true,\n supportsSchemas: false,\n supportsConcurrentIndex: false,\n supportsDeferredFkValidation: false,\n supportsJsonb: false,\n };\n\n /**\n * Narrow an unknown value to the SQLite integration harness strategy.\n */\n static isSqliteHarnessStrategy(value: unknown): value is SqliteHarnessStrategy {\n return (\n typeof value === 'object' &&\n value !== null &&\n (value as { __tangoBrand?: unknown }).__tangoBrand === SqliteHarnessStrategy.BRAND\n );\n }\n\n private static async dropAllTables(client: DBClient): Promise<void> {\n const { rows } = await client.query<{ name: string }>(\n \"SELECT name FROM sqlite_master WHERE type='table' AND name NOT LIKE 'sqlite_%'\"\n );\n for (const row of rows) {\n const table = quoteSqlIdentifier(validateSqlIdentifier(String(row.name), 'table'), 'sqlite');\n await client.query(`DROP TABLE IF EXISTS ${table}`);\n }\n }\n\n /**\n * Create a configured SQLite integration harness instance.\n */\n async create(options: HarnessOptions = {}): Promise<IntegrationHarness> {\n const config = resolveAdapterConfig(Dialect.Sqlite, {\n config: options.config,\n tangoConfigLoader: options.tangoConfigLoader,\n sqliteFile: options.sqliteFile,\n });\n\n const adapter = new SqliteAdapter();\n const resetMode = options.resetMode ?? ResetMode.DropSchema;\n let client: DBClient | null = null;\n\n const reconnect = async (): Promise<DBClient> => {\n client = await adapter.connect(config);\n return client;\n };\n\n const harness: IntegrationHarness = {\n dialect: Dialect.Sqlite,\n capabilities: this.capabilities,\n resetMode,\n get dbClient(): DBClient {\n if (!client) {\n throw new Error('Sqlite harness not initialized. Call setup() first.');\n }\n return client;\n },\n async setup(): Promise<void> {\n await reconnect();\n },\n async reset(): Promise<void> {\n if (!client) {\n throw new Error('Sqlite harness not initialized. Call setup() first.');\n }\n\n if (resetMode === ResetMode.DropSchema && config.filename && config.filename !== ':memory:') {\n await client.close();\n await rm(config.filename, { force: true });\n await reconnect();\n return;\n }\n\n await SqliteHarnessStrategy.dropAllTables(client);\n },\n async teardown(): Promise<void> {\n if (client) {\n await client.close();\n client = null;\n }\n if (config.filename && config.filename !== ':memory:') {\n await rm(config.filename, { force: true });\n }\n },\n migrationRunner(migrationsDir: string): MigrationRunner {\n if (!client) {\n throw new Error('Sqlite harness not initialized. Call setup() first.');\n }\n return new MigrationRunner(client, 'sqlite' as MigrationDialect, migrationsDir);\n },\n };\n\n return harness;\n }\n}\n","import { HarnessStrategyRegistry } from './HarnessStrategyRegistry';\nimport { Dialect, type HarnessOptions, type HarnessStrategy, type IntegrationHarness } from './domain';\nimport { PostgresHarnessStrategy } from './strategies/PostgresHarnessStrategy';\nimport { SqliteHarnessStrategy } from './strategies/SqliteHarnessStrategy';\n\n/**\n * Facade for creating integration test harnesses by dialect.\n */\nexport class TestHarness {\n static readonly BRAND = 'tango.testing.test_harness' as const;\n private static defaultRegistry: HarnessStrategyRegistry | null = null;\n readonly __tangoBrand: typeof TestHarness.BRAND = TestHarness.BRAND;\n\n /**\n * Narrow an unknown value to `TestHarness`.\n */\n static isTestHarness(value: unknown): value is TestHarness {\n return (\n typeof value === 'object' &&\n value !== null &&\n (value as { __tangoBrand?: unknown }).__tangoBrand === TestHarness.BRAND\n );\n }\n\n /**\n * Register a harness strategy on the shared default registry.\n */\n static registerStrategy(strategy: HarnessStrategy): void {\n this.ensureRegistry().register(strategy);\n }\n\n /**\n * Return the shared harness strategy registry.\n */\n static getRegistry(): HarnessStrategyRegistry {\n return this.ensureRegistry();\n }\n\n /**\n * Create a dialect-specific harness from the registry.\n */\n static async forDialect(\n args: { dialect: Dialect | string; options?: HarnessOptions },\n registry?: HarnessStrategyRegistry\n ): Promise<IntegrationHarness> {\n const selectedRegistry = registry ?? this.ensureRegistry();\n const strategy = selectedRegistry.get(args.dialect);\n return strategy.create(args.options);\n }\n\n /**\n * Convenience helper for a SQLite test harness.\n */\n static async sqlite(options?: HarnessOptions): Promise<IntegrationHarness> {\n return this.forDialect({ dialect: Dialect.Sqlite, options });\n }\n\n /**\n * Convenience helper for a Postgres test harness.\n */\n static async postgres(options?: HarnessOptions): Promise<IntegrationHarness> {\n return this.forDialect({ dialect: Dialect.Postgres, options });\n }\n\n private static ensureRegistry(): HarnessStrategyRegistry {\n if (this.defaultRegistry) return this.defaultRegistry;\n\n const registry = new HarnessStrategyRegistry();\n registry.register(new SqliteHarnessStrategy());\n registry.register(new PostgresHarnessStrategy());\n this.defaultRegistry = registry;\n return registry;\n }\n}\n","import { quoteSqlIdentifier, validateSqlIdentifier, type SqlDialect } from '@danceroutine/tango-core';\nimport { Dialect, type IntegrationHarness } from '../domain/index';\n\n/**\n * Seed rows directly into a table for integration tests that need known fixtures.\n */\nexport async function seedTable<T extends Record<string, unknown>>(\n harness: IntegrationHarness,\n table: string,\n rows: T[]\n): Promise<void> {\n if (rows.length === 0) {\n return;\n }\n\n const columns = Object.keys(rows[0] ?? {});\n if (columns.length === 0) {\n return;\n }\n\n const dialect = harness.dialect as SqlDialect;\n const safeTable = quoteSqlIdentifier(validateSqlIdentifier(table, 'table'), dialect);\n const safeColumns = columns.map((column) =>\n quoteSqlIdentifier(validateSqlIdentifier(column, 'column', columns), dialect)\n );\n\n for (const row of rows) {\n const values = columns.map((column) => {\n const value = row[column];\n if (harness.dialect === Dialect.Sqlite && typeof value === 'boolean') {\n return value ? 1 : 0;\n }\n return value;\n });\n const placeholders =\n harness.dialect === Dialect.Postgres\n ? columns.map((_, index) => `$${index + 1}`).join(', ')\n : columns.map(() => '?').join(', ');\n\n await harness.dbClient.query(\n `INSERT INTO ${safeTable} (${safeColumns.join(', ')}) VALUES (${placeholders})`,\n values as unknown[]\n );\n }\n}\n","import { QuerySet, type QueryExecutor } from '@danceroutine/tango-orm';\nimport type { TableMeta } from '@danceroutine/tango-orm/query';\nimport { Dialect, type IntegrationHarness } from '../domain/index';\n\n/**\n * Create a `QuerySet` fixture backed by a real integration harness and supplied table metadata.\n */\nexport function createQuerySetFixture<TModel extends Record<string, unknown>>(input: {\n harness: IntegrationHarness;\n meta: TableMeta;\n}): QuerySet<TModel> {\n const executor: QueryExecutor<TModel> = {\n meta: input.meta,\n client: input.harness.dbClient,\n dialect: input.harness.dialect as Dialect,\n run: async (compiled) => {\n const result = await input.harness.dbClient.query<TModel>(compiled.sql, compiled.params);\n return result.rows;\n },\n };\n\n return new QuerySet<TModel>(executor);\n}\n","/**\n * Assert that a query result matches an expected value using structural equality.\n */\nexport async function expectQueryResult<T>(actual: Promise<T> | T, expected: T): Promise<void> {\n const resolved = await actual;\n\n if (JSON.stringify(resolved) !== JSON.stringify(expected)) {\n throw new Error(`Expected query result ${JSON.stringify(expected)}, got ${JSON.stringify(resolved)}`);\n }\n}\n","/**\n * Domain boundary barrel: exposes namespaced exports for Django-style drill-down\n * imports and curated flat exports for TS-native ergonomics.\n */\n\nexport * as domain from './domain/index';\nexport * as migrations from './migrations/index';\nexport * as conformance from './conformance/index';\nexport * as runtime from './runtime/index';\nexport * as smoke from './smoke/index';\n\nexport * from './domain/index';\nexport * from './anIntegrationHarness';\nexport * from './HarnessStrategyRegistry';\nexport * from './TestHarness';\nexport * from './migrations/index';\nexport * from './conformance/index';\nexport * from './orm/index';\nexport * from './runtime/index';\nexport * from './smoke/index';\n"],"mappings":";;;;;;;;;;;;MAAa,UAAU;CACnB,QAAQ;CACR,UAAU;AACb;;;;MCHY,YAAY;CACrB,aAAa;CACb,UAAU;CACV,YAAY;AACf;;;;;;;;;;;;ACMM,eAAe,oBAClBA,SACAC,SACe;CACf,MAAM,SAAS,QAAQ,gBAAgB,QAAQ,cAAc;CAC7D,MAAM,OAAO,MAAM,OAAO,MAAM;AAEhC,MAAK,MAAM,WAAW,QAAQ,qBAAqB,CAAE,EACjD,MAAK,KAAK,SAAS,QAAQ,CACvB,OAAM,IAAI,OAAO,sCAAsC,QAAQ;AAIvE,QAAO;AACV;;;;ACXM,eAAe,yBAClBC,SACAC,SACwC;CACxC,MAAM,SAAS,QAAQ,gBAAgB,QAAQ,cAAc;AAC7D,OAAM,OAAO,MAAM,QAAQ,KAAK;CAChC,MAAM,WAAW,MAAM,OAAO,QAAQ;AAEtC,MAAK,MAAM,MAAM,QAAQ,sBAAsB,CAAE,GAAE;EAC/C,MAAM,MAAM,SAAS,KAAK,CAACC,WAA4B,OAAO,OAAO,GAAG;AACxE,OAAK,QAAQ,IAAI,QACb,OAAM,IAAI,OAAO,qBAAqB,GAAG;CAEhD;AAED,QAAO,EAAE,SAAU;AACtB;;;;ACzBD,MAAM,uBAAuB,mCAAmC;AAKzD,eAAe,iBAAiBC,SAA+C;AAClF,KAAI,QAAQ,YAAY,QAAQ,YAAY,QAAQ,YAAY,QAAQ,OACpE,OAAM,IAAI,OAAO,0CAA0C,OAAO,QAAQ,QAAQ,CAAC;CAEvF,MAAM,UAAU,QAAQ,YAAY,QAAQ,WAAW,aAAa;AACpE,QAAO,qBAAqB,WAAW,SAAwC,QAAQ,SAAS;AACnG;;;;;;;;;;;;;ACPM,eAAe,2BAClBC,UACAC,UAGI,CAAE,GACO;CACb,MAAM,UAAU,MAAM,SAAS,OAAO,QAAQ,cAAc;AAE5D,KAAI,QAAQ,YAAY,SAAS,QAC7B,OAAM,IAAI,OACL,uCAAuC,OAAO,QAAQ,QAAQ,CAAC,qCAAqC,OAAO,SAAS,QAAQ,CAAC;AAItI,KAAI,QAAQ,iBAAiB,SAAS,aAClC,OAAM,IAAI,MAAM;CAGpB,IAAI,wBAAwB;AAC5B,KAAI;AACA,QAAM,QAAQ,OAAO;CACxB,QAAO;AACJ,0BAAwB;CAC3B;AACD,MAAK,sBACD,OAAM,IAAI,MAAM;AAGpB,OAAM,QAAQ,OAAO;AACrB,OAAM,QAAQ,OAAO;AACrB,SAAQ,gBAAgB,QAAQ,iBAAiB,kBAAkB;AACnE,OAAM,QAAQ,UAAU;CAExB,IAAI,6BAA6B;AACjC,KAAI;AAIA,UAAQ;CACX,QAAO;AACJ,+BAA6B;CAChC;AACD,MAAK,2BACD,OAAM,IAAI,MAAM;AAEvB;;;;;;;;;AC7CM,SAAS,aAAaC,UAAkC,CAAE,GAAe;CAC5E,MAAM,UAAU,QAAQ,WAAW;AAEnC,QAAO,aAAa;EAChB,SAAS;EACT,cAAc;GACV,aAAa;IACT,MAAM;IACN,IACI,YAAY,WACN;KAAE,SAAS;KAAU,UAAU;KAAY,gBAAgB;IAAG,IAC9D;KACI,SAAS;KACT,KAAK;KACL,gBAAgB;IACnB;IACX,YAAY;KAAE,KAAK;KAAc,QAAQ,YAAY;IAAY;GACpE;GACD,MAAM;IACF,MAAM;IACN,IACI,YAAY,WACN;KAAE,SAAS;KAAU,UAAU;KAAY,gBAAgB;IAAG,IAC9D;KACI,SAAS;KACT,KAAK;KACL,gBAAgB;IACnB;IACX,YAAY;KAAE,KAAK;KAAc,QAAQ,YAAY;IAAY;GACpE;GACD,YAAY;IACR,MAAM;IACN,IACI,YAAY,WACN;KAAE,SAAS;KAAU,UAAU;KAAY,gBAAgB;IAAG,IAC9D;KACI,SAAS;KACT,KAAK;KACL,gBAAgB;IACnB;IACX,YAAY;KAAE,KAAK;KAAc,QAAQ,YAAY;IAAY;GACpE;EACJ;CACJ,EAAC;AACL;;;;AC/CM,eAAe,sBAAsBC,UAAkC,CAAE,GAAyB;AACrG,OAAM,mBAAmB;AACzB,QAAO,uBAAuB,MAAM,aAAa,QAAQ,CAAC;AAC7D;;;;;;;;;;;;AC6BD,MAAM,2BAA2B;AACjC,MAAM,4BAA4B;AAClC,MAAM,0BAA0B;AAChC,MAAM,uBAAuB;AAE7B,MAAMC,cAAqC;CACvC,cAAc,CAAC,SAAS,MAAM,YAAY,MAAM,SAAS,MAAkB,QAAuC;CAClH,WAAW;CACX,OAAO,CAACC,OACJ,IAAI,QAAQ,CAAC,YAAY;AACrB,aAAW,SAAS,GAAG;CAC1B;AACR;IAKY,oBAAN,MAAM,kBAAkB;CAC3B,OAAgB,QAAQ;CACxB,eAAwD,kBAAkB;CAC1E;CACA;CACA;CACA;CACA;CACA;CACA;CACA,UAAkB;CAClB,eAAuB;CACvB,eAAuB;CAEvB,YAAoBC,SAAmCC,MAA6B;AAChF,OAAK,UAAU,QAAQ,QAAQ,QAAQ,QAAQ,GAAG;AAClD,OAAK,YAAY,EAAE,KAAK,QAAQ,EAAE,cAAc,QAAQ,aAAa,UAAU,CAAC;AAChF,OAAK,iBAAiB,QAAQ,kBAAkB;AAChD,OAAK,kBAAkB,QAAQ,mBAAmB;AAClD,OAAK,gBAAgB,QAAQ,iBAAiB;AAC9C,OAAK,OAAO;AAEZ,OAAK,QAAQ,KAAK,KAAK,aAAa,QAAQ,SAAS,QAAQ,QAAQ,CAAE,GAAE;GACrE,KAAK,QAAQ;GACb,KAAK;IAAE,GAAG,QAAQ;IAAK,GAAG,QAAQ;GAAK;GACvC,OAAO;EACV,EAAC;AACF,OAAK,MAAM,QAAQ,GAAG,QAAQ,CAACC,UAAmB;AAC9C,QAAK,eAAe,aAAa,KAAK,cAAc,OAAO,MAAM,CAAC;EACrE,EAAC;AACF,OAAK,MAAM,QAAQ,GAAG,QAAQ,CAACA,UAAmB;AAC9C,QAAK,eAAe,aAAa,KAAK,cAAc,OAAO,MAAM,CAAC;EACrE,EAAC;CACL;;;;CAKD,OAAO,oBAAoBC,OAA4C;AACnE,gBACW,UAAU,YACjB,UAAU,QACT,MAAqC,iBAAiB,kBAAkB;CAEhF;;;;CAKD,aAAa,MACTH,SACAI,OAAuC,CAAE,GACf;EAC1B,MAAMC,aAAoC;GACtC,GAAG;GACH,GAAG;EACN;EACD,MAAM,UAAU,IAAI,kBAAkB,SAAS;AAC/C,QAAM,QAAQ,cAAc;AAC5B,SAAO;CACV;;;;CAKD,eAAuB;AACnB,SAAO,KAAK;CACf;;;;CAKD,eAAuB;AACnB,SAAO,KAAK;CACf;;;;CAKD,iBAAyB;EACrB,MAAM,SAAS,KAAK,aAAa,MAAM;EACvC,MAAM,SAAS,KAAK,aAAa,MAAM;AACvC,OAAK,WAAW,OACZ,QAAO;AAEX,SAAO,EAAE,YAAY,OAAO,IAAI,YAAY,OAAO,CAAE,EAAC,KAAK,OAAO,CAAC,MAAM;CAC5E;;;;CAKD,MAAM,QAAQC,MAAcC,MAAuC;EAC/D,MAAM,SAAS,KAAK,WAAW,OAAO,GAAG,QAAQ,EAAE,KAAK,QAAQ,EAAE,cAAc,KAAK,CAAC;AACtF,SAAO,KAAK,KAAK,UAAU,QAAQ,KAAK;CAC3C;;;;CAKD,MAAM,qBAAqBC,UAAoBC,gBAAwBC,OAA8B;AACjG,MAAI,SAAS,WAAW,eACpB;EAGJ,IAAIC;AACJ,MAAI;AACA,cAAW,MAAM,SAAS,MAAM;EACnC,SAAQ,OAAO;AACZ,eAAY,gCAAgC,OAAO,MAAM,CAAC;EAC7D;AAED,QAAM,IAAI,MACN;IACK,EAAE,MAAM,aAAa,OAAO,eAAe,CAAC,OAAO,OAAO,SAAS,OAAO,CAAC;IAC3E,iBAAiB,SAAS;IAC1B,iBAAiB,KAAK,gBAAgB,CAAC;EAC3C,EAAC,KAAK,KAAK;CAEnB;;;;CAKD,MAAM,OAAsB;AACxB,MAAI,KAAK,QACL;AAEJ,OAAK,UAAU;AAEf,MAAI,KAAK,MAAM,aAAa,QAAQ,KAAK,MAAM,OAC3C;AAGJ,OAAK,MAAM,KAAK,UAAU;EAC1B,MAAM,SAAS,MAAM,KAAK,YAAY,KAAK,cAAc;AACzD,OAAK,WAAW,KAAK,MAAM,QAAQ;AAC/B,QAAK,MAAM,KAAK,UAAU;AAC1B,SAAM,KAAK,YAAY,KAAK,cAAc;EAC7C;CACJ;CAED,MAAc,eAA8B;EACxC,MAAM,WAAW,KAAK,KAAK,GAAG,KAAK;AAEnC,SAAO,KAAK,KAAK,GAAG,UAAU;AAC1B,OAAI,KAAK,MAAM,aAAa,KACxB,OAAM,IAAI,OACL,wDAAwD,OAAO,KAAK,MAAM,SAAS,CAAC,MAAM,KAAK,gBAAgB,CAAC;AAGzH,OAAI;IACA,MAAM,WAAW,MAAM,KAAK,KAAK,UAAU,KAAK,SAAS;AACzD,QAAI,SAAS,GACT;GAEP,QAAO,CAEP;AACD,SAAM,KAAK,KAAK,MAAM,KAAK,gBAAgB;EAC9C;AAED,QAAM,KAAK,MAAM;AACjB,QAAM,IAAI,OAAO,qCAAqC,KAAK,SAAS,KAAK,KAAK,gBAAgB,CAAC;CAClG;CAED,MAAc,YAAYC,WAAqC;AAC3D,MAAI,KAAK,MAAM,aAAa,KACxB,QAAO;AAGX,SAAO,MAAM,IAAI,QAAiB,CAAC,YAAY;GAC3C,MAAM,QAAQ,WAAW,MAAM;AAC3B,SAAK,MAAM,IAAI,QAAQ,OAAO;AAC9B,YAAQ,MAAM;GACjB,GAAE,UAAU;GAEb,MAAM,SAAS,MAAM;AACjB,iBAAa,MAAM;AACnB,YAAQ,KAAK;GAChB;AAED,QAAK,MAAM,KAAK,QAAQ,OAAO;EAClC;CACJ;AACJ;AAED,SAAS,aAAaC,SAAiBC,OAAuB;CAC1D,MAAM,OAAO,UAAU;AACvB,KAAI,KAAK,UAAU,qBACf,QAAO;AAEX,QAAO,KAAK,MAAM,KAAK,SAAS,qBAAqB;AACxD;AAED,SAAS,cAAcR,MAAsB;AACzC,MAAK,KACD,QAAO;AAEX,QAAO,KAAK,WAAW,IAAI,GAAG,QAAQ,GAAG,KAAK;AACjD;;;;;;;;;ACxPD,MAAMS,sBAA+C;CACjD,kBAAkB;CAClB,iBAAiB;CACjB,yBAAyB;CACzB,8BAA8B;CAC9B,eAAe;AAClB;AAKM,SAAS,qBAAqBC,YAAyC,CAAE,GAAsB;AAClG,QAAO;EACH,SAAS,QAAQ;EACjB,cAAc;EACd,WAAW,UAAU;EACrB,UAAU,WAAW;EACrB,OAAO,GAAG,GAAG,YAAY,CAAE,EAAC;EAC5B,OAAO,GAAG,GAAG,YAAY,CAAE,EAAC;EAC5B,UAAU,GAAG,GAAG,YAAY,CAAE,EAAC;EAC/B,iBAAiB,GAAG,GAAG,OAAO,CAAE,GAAW;EAC3C,GAAG;CACN;AACJ;;;;ICxBY,0BAAN,MAAM,wBAAwB;CACjC,OAAgB,QAAQ;CACxB,eAA8D,wBAAwB;CACtF,aAA8B,IAAI;;;;CAKlC,OAAO,0BAA0BC,OAAkD;AAC/E,gBACW,UAAU,YACjB,UAAU,QACT,MAAqC,iBAAiB,wBAAwB;CAEtF;;;;CAKD,SAASC,UAAiC;AACtC,OAAK,WAAW,IAAI,OAAO,SAAS,QAAQ,EAAE,SAAS;AACvD,SAAO;CACV;;;;CAKD,IAAIC,SAA4C;EAC5C,MAAM,WAAW,KAAK,WAAW,IAAI,OAAO,QAAQ,CAAC;AACrD,OAAK,SACD,OAAM,IAAI,OAAO,8CAA8C,OAAO,QAAQ,CAAC;AAEnF,SAAO;CACV;;;;CAKD,OAAmC;AAC/B,SAAO,CAAC,GAAG,KAAK,WAAW,QAAQ,AAAC;CACvC;AACJ;;;;AC1CD,SAAS,WAAWC,OAA+C;AAC/D,MAAK,MAAO,QAAO;CACnB,MAAM,SAAS,OAAO,MAAM;AAC5B,QAAO,OAAO,SAAS,OAAO,GAAG,SAAS;AAC7C;AAMM,SAAS,qBACZC,SACAC,MAKa;CACb,MAAM,cAAc,KAAK,UAAU,CAAE;AAErC,KAAI,KAAK,mBAAmB;EACxB,MAAM,SAAS,WAAW,KAAK,kBAAkB;EACjD,MAAM,UAAU,OAAO,QAAQ;EAC/B,MAAMC,SAAwB;GAC1B,KAAK,YAAY,OAAO,QAAQ;GAChC,MAAM,YAAY,QAAQ,QAAQ;GAClC,MAAM,YAAY,QAAQ,QAAQ;GAClC,UAAU,YAAY,YAAY,QAAQ;GAC1C,MAAM,YAAY,QAAQ,QAAQ;GAClC,UAAU,YAAY,YAAY,QAAQ;GAC1C,UAAU,YAAY,YAAY,QAAQ;GAC1C,gBAAgB,YAAY,kBAAkB,QAAQ;EACzD;AACD,MAAI,YAAY,QAAQ,OACpB,QAAO,WAAW,KAAK,cAAc,OAAO,YAAY;AAE5D,SAAO;CACV;AAED,KAAI,YAAY,QAAQ,SACpB,QAAO;EACH,KAAK,YAAY,OAAO,QAAQ,IAAI,sBAAsB,QAAQ,IAAI;EACtE,MAAM,YAAY,QAAQ,QAAQ,IAAI;EACtC,MAAM,YAAY,QAAQ,WAAW,QAAQ,IAAI,cAAc;EAC/D,UAAU,YAAY,YAAY,QAAQ,IAAI;EAC9C,MAAM,YAAY,QAAQ,QAAQ,IAAI;EACtC,UAAU,YAAY,YAAY,QAAQ,IAAI;EAC9C,gBAAgB,YAAY,kBAAkB;CACjD;AAGL,QAAO;EACH,UAAU,KAAK,cAAc,YAAY,YAAY,QAAQ,IAAI,yBAAyB;EAC1F,gBAAgB,YAAY,kBAAkB;CACjD;AACJ;;;;ICzCY,0BAAN,MAAM,wBAAmD;CAC5D,OAAgB,QAAQ;CACxB,eAA8D,wBAAwB;CACtF,UAA4B,QAAQ;CACpC,eAAiD;EAC7C,kBAAkB;EAClB,iBAAiB;EACjB,yBAAyB;EACzB,8BAA8B;EAC9B,eAAe;CAClB;;;;CAKD,OAAO,0BAA0BC,OAAkD;AAC/E,gBACW,UAAU,YACjB,UAAU,QACT,MAAqC,iBAAiB,wBAAwB;CAEtF;CAED,OAAe,gBAAgBC,gBAAiC;AAC5D,MAAI,eAAgB,QAAO;EAC3B,MAAM,SAAS,KAAK,QAAQ,CAAC,SAAS,GAAG,CAAC,MAAM,GAAG,EAAE;AACrD,UAAQ,aAAa,KAAK,KAAK,CAAC,GAAG,OAAO;CAC7C;;;;CAKD,MAAM,OAAOC,UAA0B,CAAE,GAA+B;EACpE,MAAM,SAAS,qBAAqB,QAAQ,UAAU;GAClD,QAAQ,QAAQ;GAChB,mBAAmB,QAAQ;EAC9B,EAAC;EAEF,MAAM,UAAU,IAAI;EACpB,MAAM,aAAa,wBAAwB,gBAAgB,QAAQ,OAAO;EAC1E,MAAM,YAAY,QAAQ,aAAa,UAAU;EACjD,IAAIC,SAA0B;EAE9B,MAAM,mBAAmB,YAA2B;GAChD,MAAM,WAAW;GACjB,MAAM,SAAS,qBAAmB,wBAAsB,YAAY,SAAS,EAAE,WAAW;AAC1F,SAAM,SAAS,OAAO,8BAA8B,OAAO,EAAE;AAC7D,SAAM,SAAS,OAAO,qBAAqB,OAAO,EAAE;EACvD;EAED,MAAM,iBAAiB,YAA2B;GAC9C,MAAM,WAAW;GACjB,MAAM,SAAS,qBAAmB,wBAAsB,YAAY,SAAS,EAAE,WAAW;AAC1F,SAAM,SAAS,OAAO,wBAAwB,OAAO,UAAU;AAC/D,SAAM,SAAS,OAAO,gBAAgB,OAAO,EAAE;AAC/C,SAAM,SAAS,OAAO,qBAAqB,OAAO,EAAE;EACvD;EAED,MAAMC,UAA8B;GAChC,SAAS,QAAQ;GACjB,cAAc,KAAK;GACnB;GACA,IAAI,WAAqB;AACrB,SAAK,OACD,OAAM,IAAI,MAAM;AAEpB,WAAO;GACV;GACD,MAAM,QAAuB;AACzB,aAAS,MAAM,QAAQ,QAAQ,OAAO;AACtC,UAAM,kBAAkB;GAC3B;GACD,MAAM,QAAuB;AACzB,SAAK,OACD,OAAM,IAAI,MAAM;AAEpB,QAAI,cAAc,UAAU,cAAc,cAAc,UAAU,aAAa;AAC3E,WAAM,gBAAgB;AACtB;IACH;IAED,MAAM,EAAE,MAAM,GAAG,MAAM,OAAO,OACzB,yGACD,CAAC,UAAW,EACf;AACD,SAAK,MAAM,OAAO,MAAM;KACpB,MAAM,SAAS,qBAAmB,wBAAsB,YAAY,SAAS,EAAE,WAAW;KAC1F,MAAM,QAAQ,qBACV,wBAAsB,OAAO,IAAI,WAAW,EAAE,QAAQ,EACtD,WACH;AACD,WAAM,OAAO,OAAO,iBAAiB,OAAO,GAAG,MAAM,2BAA2B;IACnF;AACD,UAAM,OAAO,OACR,qBAAqB,qBAAmB,wBAAsB,YAAY,SAAS,EAAE,WAAW,CAAC,EACrG;GACJ;GACD,MAAM,WAA0B;AAC5B,SAAK,OAAQ;AACb,QAAI;AACA,WAAM,OAAO,OACR,wBAAwB,qBAAmB,wBAAsB,YAAY,SAAS,EAAE,WAAW,CAAC,UACxG;IACJ,UAAS;AACN,WAAM,OAAO,OAAO;AACpB,cAAS;IACZ;GACJ;GACD,gBAAgBC,eAAwC;AACpD,SAAK,OACD,OAAM,IAAI,MAAM;AAEpB,WAAO,IAAI,kBAAgB,QAAQ,YAAgC;GACtE;EACJ;AAED,SAAO;CACV;AACJ;;;;ICrHY,wBAAN,MAAM,sBAAiD;CAC1D,OAAgB,QAAQ;CACxB,eAA4D,sBAAsB;CAClF,UAA4B,QAAQ;CACpC,eAAiD;EAC7C,kBAAkB;EAClB,iBAAiB;EACjB,yBAAyB;EACzB,8BAA8B;EAC9B,eAAe;CAClB;;;;CAKD,OAAO,wBAAwBC,OAAgD;AAC3E,gBACW,UAAU,YACjB,UAAU,QACT,MAAqC,iBAAiB,sBAAsB;CAEpF;CAED,aAAqB,cAAcC,QAAiC;EAChE,MAAM,EAAE,MAAM,GAAG,MAAM,OAAO,MAC1B,iFACH;AACD,OAAK,MAAM,OAAO,MAAM;GACpB,MAAM,QAAQ,qBAAmB,wBAAsB,OAAO,IAAI,KAAK,EAAE,QAAQ,EAAE,SAAS;AAC5F,SAAM,OAAO,OAAO,uBAAuB,MAAM,EAAE;EACtD;CACJ;;;;CAKD,MAAM,OAAOC,UAA0B,CAAE,GAA+B;EACpE,MAAM,SAAS,qBAAqB,QAAQ,QAAQ;GAChD,QAAQ,QAAQ;GAChB,mBAAmB,QAAQ;GAC3B,YAAY,QAAQ;EACvB,EAAC;EAEF,MAAM,UAAU,IAAI;EACpB,MAAM,YAAY,QAAQ,aAAa,UAAU;EACjD,IAAIC,SAA0B;EAE9B,MAAM,YAAY,YAA+B;AAC7C,YAAS,MAAM,QAAQ,QAAQ,OAAO;AACtC,UAAO;EACV;EAED,MAAMC,UAA8B;GAChC,SAAS,QAAQ;GACjB,cAAc,KAAK;GACnB;GACA,IAAI,WAAqB;AACrB,SAAK,OACD,OAAM,IAAI,MAAM;AAEpB,WAAO;GACV;GACD,MAAM,QAAuB;AACzB,UAAM,WAAW;GACpB;GACD,MAAM,QAAuB;AACzB,SAAK,OACD,OAAM,IAAI,MAAM;AAGpB,QAAI,cAAc,UAAU,cAAc,OAAO,YAAY,OAAO,aAAa,YAAY;AACzF,WAAM,OAAO,OAAO;AACpB,WAAM,GAAG,OAAO,UAAU,EAAE,OAAO,KAAM,EAAC;AAC1C,WAAM,WAAW;AACjB;IACH;AAED,UAAM,sBAAsB,cAAc,OAAO;GACpD;GACD,MAAM,WAA0B;AAC5B,QAAI,QAAQ;AACR,WAAM,OAAO,OAAO;AACpB,cAAS;IACZ;AACD,QAAI,OAAO,YAAY,OAAO,aAAa,WACvC,OAAM,GAAG,OAAO,UAAU,EAAE,OAAO,KAAM,EAAC;GAEjD;GACD,gBAAgBC,eAAwC;AACpD,SAAK,OACD,OAAM,IAAI,MAAM;AAEpB,WAAO,IAAI,gBAAgB,QAAQ,UAA8B;GACpE;EACJ;AAED,SAAO;CACV;AACJ;;;;IC7GY,cAAN,MAAM,YAAY;CACrB,OAAgB,QAAQ;CACxB,OAAe,kBAAkD;CACjE,eAAkD,YAAY;;;;CAK9D,OAAO,cAAcC,OAAsC;AACvD,gBACW,UAAU,YACjB,UAAU,QACT,MAAqC,iBAAiB,YAAY;CAE1E;;;;CAKD,OAAO,iBAAiBC,UAAiC;AACrD,OAAK,gBAAgB,CAAC,SAAS,SAAS;CAC3C;;;;CAKD,OAAO,cAAuC;AAC1C,SAAO,KAAK,gBAAgB;CAC/B;;;;CAKD,aAAa,WACTC,MACAC,UAC2B;EAC3B,MAAM,mBAAmB,YAAY,KAAK,gBAAgB;EAC1D,MAAM,WAAW,iBAAiB,IAAI,KAAK,QAAQ;AACnD,SAAO,SAAS,OAAO,KAAK,QAAQ;CACvC;;;;CAKD,aAAa,OAAOC,SAAuD;AACvE,SAAO,KAAK,WAAW;GAAE,SAAS,QAAQ;GAAQ;EAAS,EAAC;CAC/D;;;;CAKD,aAAa,SAASA,SAAuD;AACzE,SAAO,KAAK,WAAW;GAAE,SAAS,QAAQ;GAAU;EAAS,EAAC;CACjE;CAED,OAAe,iBAA0C;AACrD,MAAI,KAAK,gBAAiB,QAAO,KAAK;EAEtC,MAAM,WAAW,IAAI;AACrB,WAAS,SAAS,IAAI,wBAAwB;AAC9C,WAAS,SAAS,IAAI,0BAA0B;AAChD,OAAK,kBAAkB;AACvB,SAAO;CACV;AACJ;;;;ACnEM,eAAe,UAClBC,SACAC,OACAC,MACa;AACb,KAAI,KAAK,WAAW,EAChB;CAGJ,MAAM,UAAU,OAAO,KAAK,KAAK,MAAM,CAAE,EAAC;AAC1C,KAAI,QAAQ,WAAW,EACnB;CAGJ,MAAM,UAAU,QAAQ;CACxB,MAAM,YAAY,mBAAmB,sBAAsB,OAAO,QAAQ,EAAE,QAAQ;CACpF,MAAM,cAAc,QAAQ,IAAI,CAAC,WAC7B,mBAAmB,sBAAsB,QAAQ,UAAU,QAAQ,EAAE,QAAQ,CAChF;AAED,MAAK,MAAM,OAAO,MAAM;EACpB,MAAM,SAAS,QAAQ,IAAI,CAAC,WAAW;GACnC,MAAM,QAAQ,IAAI;AAClB,OAAI,QAAQ,YAAY,QAAQ,iBAAiB,UAAU,UACvD,QAAO,QAAQ,IAAI;AAEvB,UAAO;EACV,EAAC;EACF,MAAM,eACF,QAAQ,YAAY,QAAQ,WACtB,QAAQ,IAAI,CAAC,GAAG,WAAW,GAAG,QAAQ,EAAE,EAAE,CAAC,KAAK,KAAK,GACrD,QAAQ,IAAI,MAAM,IAAI,CAAC,KAAK,KAAK;AAE3C,QAAM,QAAQ,SAAS,OAClB,cAAc,UAAU,IAAI,YAAY,KAAK,KAAK,CAAC,YAAY,aAAa,IAC7E,OACH;CACJ;AACJ;;;;ACrCM,SAAS,sBAA8DC,OAGzD;CACjB,MAAMC,WAAkC;EACpC,MAAM,MAAM;EACZ,QAAQ,MAAM,QAAQ;EACtB,SAAS,MAAM,QAAQ;EACvB,KAAK,OAAO,aAAa;GACrB,MAAM,SAAS,MAAM,MAAM,QAAQ,SAAS,MAAc,SAAS,KAAK,SAAS,OAAO;AACxF,UAAO,OAAO;EACjB;CACJ;AAED,QAAO,IAAI,SAAiB;AAC/B;;;;ACnBM,eAAe,kBAAqBC,QAAwBC,UAA4B;CAC3F,MAAM,WAAW,MAAM;AAEvB,KAAI,KAAK,UAAU,SAAS,KAAK,KAAK,UAAU,SAAS,CACrD,OAAM,IAAI,OAAO,wBAAwB,KAAK,UAAU,SAAS,CAAC,QAAQ,KAAK,UAAU,SAAS,CAAC;AAE1G"}
@@ -1,9 +1 @@
1
- export interface DBClient {
2
- query<T = unknown>(sql: string, params?: unknown[]): Promise<{
3
- rows: T[];
4
- }>;
5
- begin(): Promise<void>;
6
- commit(): Promise<void>;
7
- rollback(): Promise<void>;
8
- close(): Promise<void>;
9
- }
1
+ export type { DBClient } from '@danceroutine/tango-orm';
@@ -1,12 +1,5 @@
1
- export type MockQuerySetResult<TModel extends Record<string, unknown>> = {
2
- filter: (input: unknown) => MockQuerySetResult<TModel>;
3
- orderBy: (...tokens: Array<keyof TModel | `-${string & keyof TModel}`>) => MockQuerySetResult<TModel>;
4
- limit: (n: number) => MockQuerySetResult<TModel>;
5
- offset: (n: number) => MockQuerySetResult<TModel>;
6
- fetch: () => Promise<{
7
- results: TModel[];
8
- nextCursor: string | null;
9
- }>;
10
- fetchOne: () => Promise<TModel | null>;
11
- count: () => Promise<number>;
12
- };
1
+ import type { QuerySet } from '@danceroutine/tango-orm';
2
+ /**
3
+ * Legacy alias for query-set test doubles.
4
+ */
5
+ export type MockQuerySetResult<TModel extends Record<string, unknown>> = QuerySet<TModel>;
@@ -0,0 +1,21 @@
1
+ import type { DBClient } from './DBClient';
2
+ /**
3
+ * Looser override type for `aDBClient` that accepts concrete-typed query mocks.
4
+ * Consumers should not need to cast their mocks to satisfy `DBClient['query']`'s generic.
5
+ */
6
+ type DBClientOverrides = {
7
+ query?: (sql: string, params?: readonly unknown[]) => Promise<{
8
+ rows: any[];
9
+ }>;
10
+ begin?: () => Promise<void>;
11
+ commit?: () => Promise<void>;
12
+ rollback?: () => Promise<void>;
13
+ close?: () => Promise<void>;
14
+ };
15
+ /**
16
+ * Create a lightweight `DBClient` test double with optional behavior overrides.
17
+ * The `query` override accepts any function returning `Promise<{ rows: any[] }>`,
18
+ * so concrete-typed Vitest mocks do not require a cast at the call site.
19
+ */
20
+ export declare function aDBClient(overrides?: DBClientOverrides): DBClient;
21
+ export {};
@@ -0,0 +1,17 @@
1
+ import type { ManagerLike, QuerySet } from '@danceroutine/tango-orm';
2
+ import type { TableMeta } from '@danceroutine/tango-orm/query';
3
+ export type ManagerOverrides<TModel extends Record<string, unknown>> = {
4
+ meta?: TableMeta;
5
+ querySet?: QuerySet<TModel>;
6
+ query?: ManagerLike<TModel>['query'];
7
+ findById?: ManagerLike<TModel>['findById'];
8
+ getOrThrow?: ManagerLike<TModel>['getOrThrow'];
9
+ create?: ManagerLike<TModel>['create'];
10
+ update?: ManagerLike<TModel>['update'];
11
+ delete?: ManagerLike<TModel>['delete'];
12
+ bulkCreate?: ManagerLike<TModel>['bulkCreate'];
13
+ };
14
+ /**
15
+ * Create a manager-shaped test double for resource and service tests.
16
+ */
17
+ export declare function aManager<TModel extends Record<string, unknown>>(overrides?: ManagerOverrides<TModel>): ManagerLike<TModel>;
@@ -0,0 +1,14 @@
1
+ import type { QueryExecutor } from '@danceroutine/tango-orm';
2
+ import type { Dialect, TableMeta } from '@danceroutine/tango-orm/query';
3
+ export type QueryExecutorOverrides<TModel extends Record<string, unknown>> = {
4
+ dialect?: Dialect;
5
+ meta?: TableMeta;
6
+ query?: (sql: string, params?: readonly unknown[]) => Promise<{
7
+ rows: unknown[];
8
+ }>;
9
+ run?: QueryExecutor<TModel>['run'];
10
+ };
11
+ /**
12
+ * Create a minimal `QueryExecutor` test double for `QuerySet` tests.
13
+ */
14
+ export declare function aQueryExecutor<TModel extends Record<string, unknown>>(overrides?: QueryExecutorOverrides<TModel>): QueryExecutor<TModel>;
@@ -0,0 +1,5 @@
1
+ import type { QueryResult } from '@danceroutine/tango-orm/query';
2
+ /**
3
+ * Create a query-result test value with optional overrides.
4
+ */
5
+ export declare function aQueryResult<TModel>(overrides?: Partial<QueryResult<TModel>>): QueryResult<TModel>;
@@ -0,0 +1,8 @@
1
+ import type { QuerySet } from '@danceroutine/tango-orm';
2
+ /**
3
+ * Create a chainable query-set test double with optional behavior overrides.
4
+ *
5
+ * All methods are wrapped in `vi.fn()` so they can be asserted on directly
6
+ * without an additional `vi.mocked()` call.
7
+ */
8
+ export declare function aQuerySet<TModel extends Record<string, unknown>>(overrides?: Partial<QuerySet<TModel>>): QuerySet<TModel>;
@@ -0,0 +1,22 @@
1
+ import { RequestContext, type BaseUser } from '@danceroutine/tango-resources/context';
2
+ type RequestContextFactory<TUser, TContext extends RequestContextLike<TUser>> = (request: Request, user: TUser | null) => TContext;
3
+ type RequestContextLike<TUser> = {
4
+ request: Request;
5
+ user: TUser | null;
6
+ params: Record<string, string>;
7
+ };
8
+ export type RequestContextFixtureOptions<TUser = BaseUser, TContext extends RequestContextLike<TUser> = RequestContext<TUser>> = {
9
+ method?: string;
10
+ url?: string;
11
+ body?: unknown;
12
+ user?: TUser | null;
13
+ params?: Record<string, string>;
14
+ headers?: HeadersInit;
15
+ contextFactory?: RequestContextFactory<TUser, TContext>;
16
+ };
17
+ /**
18
+ * Create a RequestContext fixture with optional method/url/body/user/params.
19
+ */
20
+ export declare function aRequestContext<TUser = BaseUser, TContext extends RequestContextLike<TUser> = RequestContext<TUser>>(method: string, url: string, body?: unknown): TContext;
21
+ export declare function aRequestContext<TUser = BaseUser, TContext extends RequestContextLike<TUser> = RequestContext<TUser>>(options?: RequestContextFixtureOptions<TUser, TContext>): TContext;
22
+ export {};
@@ -1,9 +1,14 @@
1
1
  /**
2
2
  * Domain boundary barrel: centralizes this subdomain's public contract.
3
3
  */
4
- export { aMockDBClient } from './aMockDBClient';
5
- export { aMockQuerySet } from './aMockQuerySet';
6
- export { aMockRepository } from './aMockRepository';
4
+ export { aDBClient } from './aDBClient';
5
+ export { aManager } from './aManager';
6
+ export { aQueryResult } from './aQueryResult';
7
+ export { aQuerySet } from './aQuerySet';
8
+ export { aRequestContext } from './aRequestContext';
9
+ export { aQueryExecutor } from './aQueryExecutor';
10
+ export type { QueryExecutorOverrides } from './aQueryExecutor';
11
+ export type { ManagerOverrides } from './aManager';
7
12
  export type { DBClient } from './DBClient';
8
13
  export type { MockQuerySetResult } from './MockQuerySetResult';
9
- export type { RepositoryLike } from './RepositoryLike';
14
+ export type { RequestContextFixtureOptions } from './aRequestContext';
@@ -1,6 +1,4 @@
1
- /**
2
- * Domain boundary barrel: centralizes this subdomain's public contract.
3
- */
4
- export { aMockDBClient } from './aMockDBClient';
5
- export { aMockQuerySet } from './aMockQuerySet';
6
- export { aMockRepository } from './aMockRepository';
1
+ import { aDBClient } from "../aDBClient-W6eXsK3X.js";
2
+ import { aManager, aQueryExecutor, aQueryResult, aQuerySet, aRequestContext } from "../mocks-BkwkXQQt.js";
3
+
4
+ export { aDBClient, aManager, aQueryExecutor, aQueryResult, aQuerySet, aRequestContext };
@@ -0,0 +1,136 @@
1
+ import { __export } from "./chunk-BkvOhyD0.js";
2
+ import { aDBClient } from "./aDBClient-W6eXsK3X.js";
3
+ import { vi, vi as vi$1, vi as vi$2 } from "vitest";
4
+ import { QuerySet as QuerySetClass } from "@danceroutine/tango-orm";
5
+ import { RequestContext } from "@danceroutine/tango-resources/context";
6
+
7
+ //#region src/mocks/aQueryResult.ts
8
+ function aQueryResult(overrides = {}) {
9
+ return {
10
+ results: [],
11
+ nextCursor: null,
12
+ ...overrides
13
+ };
14
+ }
15
+
16
+ //#endregion
17
+ //#region src/mocks/aQueryExecutor.ts
18
+ function aQueryExecutor(overrides = {}) {
19
+ const dialect = overrides.dialect ?? "postgres";
20
+ const meta = overrides.meta ?? {
21
+ table: "mock_table",
22
+ pk: "id",
23
+ columns: {}
24
+ };
25
+ const run = overrides.run ?? vi$2.fn(async () => []);
26
+ const client = aDBClient(overrides.query ? { query: overrides.query } : {});
27
+ return {
28
+ meta,
29
+ client,
30
+ dialect,
31
+ run
32
+ };
33
+ }
34
+
35
+ //#endregion
36
+ //#region src/mocks/aQuerySet.ts
37
+ function aQuerySet(overrides = {}) {
38
+ const queryset = new QuerySetClass(aQueryExecutor());
39
+ const filterImpl = overrides.filter ?? ((_input) => queryset);
40
+ const excludeImpl = overrides.exclude ?? ((_input) => queryset);
41
+ const orderByImpl = overrides.orderBy ?? ((..._tokens) => queryset);
42
+ const limitImpl = overrides.limit ?? ((_n) => queryset);
43
+ const offsetImpl = overrides.offset ?? ((_n) => queryset);
44
+ const selectImpl = overrides.select ?? ((_cols) => queryset);
45
+ const selectRelatedImpl = overrides.selectRelated ?? ((..._rels) => queryset);
46
+ const prefetchRelatedImpl = overrides.prefetchRelated ?? ((..._rels) => queryset);
47
+ const fetchImpl = overrides.fetch ?? (async (_shape) => aQueryResult());
48
+ const fetchOneImpl = overrides.fetchOne ?? (async (_shape) => null);
49
+ const countImpl = overrides.count ?? (async () => 0);
50
+ const existsImpl = overrides.exists ?? (async () => false);
51
+ queryset.filter = vi$1.fn((input) => filterImpl(input));
52
+ queryset.exclude = vi$1.fn((input) => excludeImpl(input));
53
+ queryset.orderBy = vi$1.fn((...tokens) => orderByImpl(...tokens));
54
+ queryset.limit = vi$1.fn((n) => limitImpl(n));
55
+ queryset.offset = vi$1.fn((n) => offsetImpl(n));
56
+ queryset.select = vi$1.fn((cols) => selectImpl(cols));
57
+ queryset.selectRelated = vi$1.fn((...rels) => selectRelatedImpl(...rels));
58
+ queryset.prefetchRelated = vi$1.fn((...rels) => prefetchRelatedImpl(...rels));
59
+ queryset.fetch = vi$1.fn(fetchImpl);
60
+ queryset.fetchOne = vi$1.fn(fetchOneImpl);
61
+ queryset.count = vi$1.fn(() => countImpl());
62
+ queryset.exists = vi$1.fn(() => existsImpl());
63
+ return queryset;
64
+ }
65
+
66
+ //#endregion
67
+ //#region src/mocks/aManager.ts
68
+ function aManager(overrides = {}) {
69
+ const meta = overrides.meta ?? {
70
+ table: "mock_table",
71
+ pk: "id",
72
+ columns: {}
73
+ };
74
+ const querySet = overrides.querySet ?? aQuerySet();
75
+ const queryImpl = overrides.query ?? (() => querySet);
76
+ const findByIdImpl = overrides.findById ?? (async () => null);
77
+ const getOrThrowImpl = overrides.getOrThrow ?? (async (id) => {
78
+ const record = await findByIdImpl(id);
79
+ if (!record) throw new Error(`No ${meta.table} record found for ${String(meta.pk)}=${String(id)}.`);
80
+ return record;
81
+ });
82
+ const createImpl = overrides.create ?? (async (input) => input);
83
+ const updateImpl = overrides.update ?? (async (_id, patch) => patch);
84
+ const deleteImpl = overrides.delete ?? (async (_id) => {});
85
+ const bulkCreateImpl = overrides.bulkCreate ?? (async (inputs) => inputs);
86
+ return {
87
+ meta,
88
+ query: vi.fn(() => queryImpl()),
89
+ findById: vi.fn((id) => findByIdImpl(id)),
90
+ getOrThrow: vi.fn((id) => getOrThrowImpl(id)),
91
+ create: vi.fn((input) => createImpl(input)),
92
+ update: vi.fn((id, patch) => updateImpl(id, patch)),
93
+ delete: vi.fn((id) => deleteImpl(id)),
94
+ bulkCreate: vi.fn((inputs) => bulkCreateImpl(inputs))
95
+ };
96
+ }
97
+
98
+ //#endregion
99
+ //#region src/mocks/aRequestContext.ts
100
+ function aRequestContext(optionsOrMethod = {}, urlArg, bodyArg) {
101
+ const resolvedOptions = typeof optionsOrMethod === "string" ? {
102
+ method: optionsOrMethod,
103
+ url: urlArg,
104
+ body: bodyArg
105
+ } : optionsOrMethod;
106
+ const { method = "GET", url = "https://example.test", body, user = null, params = {}, headers, contextFactory } = resolvedOptions;
107
+ const resolvedHeaders = body === undefined ? headers : {
108
+ "content-type": "application/json",
109
+ ...headers
110
+ };
111
+ const request = new Request(url, {
112
+ method,
113
+ headers: resolvedHeaders,
114
+ body: body === undefined ? undefined : JSON.stringify(body)
115
+ });
116
+ const createContext = contextFactory ?? ((req, currentUser) => RequestContext.create(req, currentUser));
117
+ const context = createContext(request, user);
118
+ context.params = params;
119
+ return context;
120
+ }
121
+
122
+ //#endregion
123
+ //#region src/mocks/index.ts
124
+ var mocks_exports = {};
125
+ __export(mocks_exports, {
126
+ aDBClient: () => aDBClient,
127
+ aManager: () => aManager,
128
+ aQueryExecutor: () => aQueryExecutor,
129
+ aQueryResult: () => aQueryResult,
130
+ aQuerySet: () => aQuerySet,
131
+ aRequestContext: () => aRequestContext
132
+ });
133
+
134
+ //#endregion
135
+ export { aManager, aQueryExecutor, aQueryResult, aQuerySet, aRequestContext, mocks_exports };
136
+ //# sourceMappingURL=mocks-BkwkXQQt.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"mocks-BkwkXQQt.js","names":["overrides: Partial<QueryResult<TModel>>","overrides: QueryExecutorOverrides<TModel>","meta: TableMeta","overrides: Partial<QuerySet<TModel>>","filterImpl: QuerySet<TModel>['filter']","excludeImpl: QuerySet<TModel>['exclude']","orderByImpl: QuerySet<TModel>['orderBy']","limitImpl: QuerySet<TModel>['limit']","offsetImpl: QuerySet<TModel>['offset']","selectImpl: QuerySet<TModel>['select']","selectRelatedImpl: QuerySet<TModel>['selectRelated']","prefetchRelatedImpl: QuerySet<TModel>['prefetchRelated']","fetchImpl: QuerySet<TModel>['fetch']","_shape?: ((r: TModel) => Out) | { parse: (r: TModel) => Out }","fetchOneImpl: QuerySet<TModel>['fetchOne']","countImpl: QuerySet<TModel>['count']","existsImpl: QuerySet<TModel>['exists']","input: Parameters<QuerySet<TModel>['filter']>[0]","input: Parameters<QuerySet<TModel>['exclude']>[0]","n: number","cols: Parameters<QuerySet<TModel>['select']>[0]","overrides: ManagerOverrides<TModel>","id: ModelId","input: Partial<TModel>","_id: ModelId","patch: Partial<TModel>","inputs: Partial<TModel>[]","optionsOrMethod: RequestContextFixtureOptions<TUser, TContext> | string","urlArg?: string","bodyArg?: unknown","resolvedOptions: RequestContextFixtureOptions<TUser, TContext>","resolvedHeaders: HeadersInit | undefined","req: Request","currentUser: TUser | null"],"sources":["../src/mocks/aQueryResult.ts","../src/mocks/aQueryExecutor.ts","../src/mocks/aQuerySet.ts","../src/mocks/aManager.ts","../src/mocks/aRequestContext.ts","../src/mocks/index.ts"],"sourcesContent":["import type { QueryResult } from '@danceroutine/tango-orm/query';\n\n/**\n * Create a query-result test value with optional overrides.\n */\nexport function aQueryResult<TModel>(overrides: Partial<QueryResult<TModel>> = {}): QueryResult<TModel> {\n return {\n results: [] as TModel[],\n nextCursor: null,\n ...overrides,\n };\n}\n","import { vi } from 'vitest';\nimport type { QueryExecutor } from '@danceroutine/tango-orm';\nimport type { Dialect, TableMeta } from '@danceroutine/tango-orm/query';\nimport { aDBClient } from './aDBClient';\n\nexport type QueryExecutorOverrides<TModel extends Record<string, unknown>> = {\n dialect?: Dialect;\n meta?: TableMeta;\n query?: (sql: string, params?: readonly unknown[]) => Promise<{ rows: unknown[] }>;\n run?: QueryExecutor<TModel>['run'];\n};\n\n/**\n * Create a minimal `QueryExecutor` test double for `QuerySet` tests.\n */\nexport function aQueryExecutor<TModel extends Record<string, unknown>>(\n overrides: QueryExecutorOverrides<TModel> = {}\n): QueryExecutor<TModel> {\n const dialect = overrides.dialect ?? 'postgres';\n const meta: TableMeta = overrides.meta ?? { table: 'mock_table', pk: 'id', columns: {} };\n const run = overrides.run ?? vi.fn(async () => [] as TModel[]);\n const client = aDBClient(overrides.query ? { query: overrides.query } : {});\n\n return { meta, client, dialect, run };\n}\n","import { vi } from 'vitest';\nimport type { QuerySet } from '@danceroutine/tango-orm';\nimport { QuerySet as QuerySetClass } from '@danceroutine/tango-orm';\nimport { aQueryResult } from './aQueryResult';\nimport { aQueryExecutor } from './aQueryExecutor';\n\n/**\n * Create a chainable query-set test double with optional behavior overrides.\n *\n * All methods are wrapped in `vi.fn()` so they can be asserted on directly\n * without an additional `vi.mocked()` call.\n */\nexport function aQuerySet<TModel extends Record<string, unknown>>(\n overrides: Partial<QuerySet<TModel>> = {}\n): QuerySet<TModel> {\n const queryset = new QuerySetClass<TModel>(aQueryExecutor<TModel>());\n const filterImpl: QuerySet<TModel>['filter'] = overrides.filter ?? ((_input) => queryset);\n const excludeImpl: QuerySet<TModel>['exclude'] = overrides.exclude ?? ((_input) => queryset);\n const orderByImpl: QuerySet<TModel>['orderBy'] = overrides.orderBy ?? ((..._tokens) => queryset);\n const limitImpl: QuerySet<TModel>['limit'] = overrides.limit ?? ((_n) => queryset);\n const offsetImpl: QuerySet<TModel>['offset'] = overrides.offset ?? ((_n) => queryset);\n const selectImpl: QuerySet<TModel>['select'] = overrides.select ?? ((_cols) => queryset);\n const selectRelatedImpl: QuerySet<TModel>['selectRelated'] = overrides.selectRelated ?? ((..._rels) => queryset);\n const prefetchRelatedImpl: QuerySet<TModel>['prefetchRelated'] =\n overrides.prefetchRelated ?? ((..._rels) => queryset);\n const fetchImpl: QuerySet<TModel>['fetch'] =\n overrides.fetch ??\n (async <Out = TModel>(_shape?: ((r: TModel) => Out) | { parse: (r: TModel) => Out }) => aQueryResult<Out>());\n const fetchOneImpl: QuerySet<TModel>['fetchOne'] =\n overrides.fetchOne ??\n (async <Out = TModel>(_shape?: ((r: TModel) => Out) | { parse: (r: TModel) => Out }) => null as Out | null);\n const countImpl: QuerySet<TModel>['count'] = overrides.count ?? (async () => 0);\n const existsImpl: QuerySet<TModel>['exists'] = overrides.exists ?? (async () => false);\n\n queryset.filter = vi.fn((input: Parameters<QuerySet<TModel>['filter']>[0]) =>\n filterImpl(input)\n ) as QuerySet<TModel>['filter'];\n queryset.exclude = vi.fn((input: Parameters<QuerySet<TModel>['exclude']>[0]) =>\n excludeImpl(input)\n ) as QuerySet<TModel>['exclude'];\n queryset.orderBy = vi.fn((...tokens: Parameters<QuerySet<TModel>['orderBy']>) =>\n orderByImpl(...tokens)\n ) as QuerySet<TModel>['orderBy'];\n queryset.limit = vi.fn((n: number) => limitImpl(n)) as QuerySet<TModel>['limit'];\n queryset.offset = vi.fn((n: number) => offsetImpl(n)) as QuerySet<TModel>['offset'];\n queryset.select = vi.fn((cols: Parameters<QuerySet<TModel>['select']>[0]) =>\n selectImpl(cols)\n ) as QuerySet<TModel>['select'];\n queryset.selectRelated = vi.fn((...rels: Parameters<QuerySet<TModel>['selectRelated']>) =>\n selectRelatedImpl(...rels)\n ) as QuerySet<TModel>['selectRelated'];\n queryset.prefetchRelated = vi.fn((...rels: Parameters<QuerySet<TModel>['prefetchRelated']>) =>\n prefetchRelatedImpl(...rels)\n ) as QuerySet<TModel>['prefetchRelated'];\n queryset.fetch = vi.fn(fetchImpl) as QuerySet<TModel>['fetch'];\n queryset.fetchOne = vi.fn(fetchOneImpl) as QuerySet<TModel>['fetchOne'];\n queryset.count = vi.fn(() => countImpl()) as QuerySet<TModel>['count'];\n queryset.exists = vi.fn(() => existsImpl()) as QuerySet<TModel>['exists'];\n\n return queryset;\n}\n","import { vi } from 'vitest';\nimport type { ManagerLike, QuerySet } from '@danceroutine/tango-orm';\nimport type { TableMeta } from '@danceroutine/tango-orm/query';\nimport { aQuerySet } from './aQuerySet';\n\nexport type ManagerOverrides<TModel extends Record<string, unknown>> = {\n meta?: TableMeta;\n querySet?: QuerySet<TModel>;\n query?: ManagerLike<TModel>['query'];\n findById?: ManagerLike<TModel>['findById'];\n getOrThrow?: ManagerLike<TModel>['getOrThrow'];\n create?: ManagerLike<TModel>['create'];\n update?: ManagerLike<TModel>['update'];\n delete?: ManagerLike<TModel>['delete'];\n bulkCreate?: ManagerLike<TModel>['bulkCreate'];\n};\n\n/**\n * Create a manager-shaped test double for resource and service tests.\n */\nexport function aManager<TModel extends Record<string, unknown>>(\n overrides: ManagerOverrides<TModel> = {}\n): ManagerLike<TModel> {\n const meta = overrides.meta ?? { table: 'mock_table', pk: 'id', columns: {} };\n const querySet = overrides.querySet ?? aQuerySet<TModel>();\n type ModelId = TModel[keyof TModel];\n\n const queryImpl = overrides.query ?? (() => querySet);\n const findByIdImpl = overrides.findById ?? (async () => null as TModel | null);\n const getOrThrowImpl =\n overrides.getOrThrow ??\n (async (id: ModelId) => {\n const record = await findByIdImpl(id);\n if (!record) {\n throw new Error(`No ${meta.table} record found for ${String(meta.pk)}=${String(id)}.`);\n }\n return record;\n });\n const createImpl = overrides.create ?? (async (input: Partial<TModel>) => input as TModel);\n const updateImpl = overrides.update ?? (async (_id: ModelId, patch: Partial<TModel>) => patch as TModel);\n const deleteImpl = overrides.delete ?? (async (_id: ModelId) => {});\n const bulkCreateImpl = overrides.bulkCreate ?? (async (inputs: Partial<TModel>[]) => inputs as TModel[]);\n\n return {\n meta,\n query: vi.fn(() => queryImpl()),\n findById: vi.fn((id: ModelId) => findByIdImpl(id)),\n getOrThrow: vi.fn((id: ModelId) => getOrThrowImpl(id)),\n create: vi.fn((input: Partial<TModel>) => createImpl(input)),\n update: vi.fn((id: ModelId, patch: Partial<TModel>) => updateImpl(id, patch)),\n delete: vi.fn((id: ModelId) => deleteImpl(id)),\n bulkCreate: vi.fn((inputs: Partial<TModel>[]) => bulkCreateImpl(inputs)),\n };\n}\n","// Import through the package subpath so fixtures stay aligned with the public nominal type.\nimport { RequestContext, type BaseUser } from '@danceroutine/tango-resources/context';\n\ntype RequestContextFactory<TUser, TContext extends RequestContextLike<TUser>> = (\n request: Request,\n user: TUser | null\n) => TContext;\n\ntype RequestContextLike<TUser> = {\n request: Request;\n user: TUser | null;\n params: Record<string, string>;\n};\n\nexport type RequestContextFixtureOptions<\n TUser = BaseUser,\n TContext extends RequestContextLike<TUser> = RequestContext<TUser>,\n> = {\n method?: string;\n url?: string;\n body?: unknown;\n user?: TUser | null;\n params?: Record<string, string>;\n headers?: HeadersInit;\n contextFactory?: RequestContextFactory<TUser, TContext>;\n};\n\n/**\n * Create a RequestContext fixture with optional method/url/body/user/params.\n */\nexport function aRequestContext<TUser = BaseUser, TContext extends RequestContextLike<TUser> = RequestContext<TUser>>(\n method: string,\n url: string,\n body?: unknown\n): TContext;\nexport function aRequestContext<TUser = BaseUser, TContext extends RequestContextLike<TUser> = RequestContext<TUser>>(\n options?: RequestContextFixtureOptions<TUser, TContext>\n): TContext;\nexport function aRequestContext<TUser = BaseUser, TContext extends RequestContextLike<TUser> = RequestContext<TUser>>(\n optionsOrMethod: RequestContextFixtureOptions<TUser, TContext> | string = {},\n urlArg?: string,\n bodyArg?: unknown\n): TContext {\n const resolvedOptions: RequestContextFixtureOptions<TUser, TContext> =\n typeof optionsOrMethod === 'string'\n ? {\n method: optionsOrMethod,\n url: urlArg,\n body: bodyArg,\n }\n : optionsOrMethod;\n const {\n method = 'GET',\n url = 'https://example.test',\n body,\n user = null,\n params = {},\n headers,\n contextFactory,\n } = resolvedOptions;\n\n const resolvedHeaders: HeadersInit | undefined =\n body === undefined ? headers : { 'content-type': 'application/json', ...headers };\n\n const request = new Request(url, {\n method,\n headers: resolvedHeaders,\n body: body === undefined ? undefined : JSON.stringify(body),\n });\n const createContext =\n contextFactory ?? ((req: Request, currentUser: TUser | null) => RequestContext.create<TUser>(req, currentUser));\n const context = createContext(request, user) as TContext;\n context.params = params;\n return context;\n}\n","/**\n * Domain boundary barrel: centralizes this subdomain's public contract.\n */\n\nexport { aDBClient } from './aDBClient';\nexport { aManager } from './aManager';\nexport { aQueryResult } from './aQueryResult';\nexport { aQuerySet } from './aQuerySet';\nexport { aRequestContext } from './aRequestContext';\nexport { aQueryExecutor } from './aQueryExecutor';\nexport type { QueryExecutorOverrides } from './aQueryExecutor';\nexport type { ManagerOverrides } from './aManager';\nexport type { DBClient } from './DBClient';\nexport type { MockQuerySetResult } from './MockQuerySetResult';\nexport type { RequestContextFixtureOptions } from './aRequestContext';\n"],"mappings":";;;;;;;AAKO,SAAS,aAAqBA,YAA0C,CAAE,GAAuB;AACpG,QAAO;EACH,SAAS,CAAE;EACX,YAAY;EACZ,GAAG;CACN;AACJ;;;;ACIM,SAAS,eACZC,YAA4C,CAAE,GACzB;CACrB,MAAM,UAAU,UAAU,WAAW;CACrC,MAAMC,OAAkB,UAAU,QAAQ;EAAE,OAAO;EAAc,IAAI;EAAM,SAAS,CAAE;CAAE;CACxF,MAAM,MAAM,UAAU,OAAO,KAAG,GAAG,YAAY,CAAE,EAAa;CAC9D,MAAM,SAAS,UAAU,UAAU,QAAQ,EAAE,OAAO,UAAU,MAAO,IAAG,CAAE,EAAC;AAE3E,QAAO;EAAE;EAAM;EAAQ;EAAS;CAAK;AACxC;;;;ACZM,SAAS,UACZC,YAAuC,CAAE,GACzB;CAChB,MAAM,WAAW,IAAI,cAAsB,gBAAwB;CACnE,MAAMC,aAAyC,UAAU,WAAW,CAAC,WAAW;CAChF,MAAMC,cAA2C,UAAU,YAAY,CAAC,WAAW;CACnF,MAAMC,cAA2C,UAAU,YAAY,CAAC,GAAG,YAAY;CACvF,MAAMC,YAAuC,UAAU,UAAU,CAAC,OAAO;CACzE,MAAMC,aAAyC,UAAU,WAAW,CAAC,OAAO;CAC5E,MAAMC,aAAyC,UAAU,WAAW,CAAC,UAAU;CAC/E,MAAMC,oBAAuD,UAAU,kBAAkB,CAAC,GAAG,UAAU;CACvG,MAAMC,sBACF,UAAU,oBAAoB,CAAC,GAAG,UAAU;CAChD,MAAMC,YACF,UAAU,UACT,OAAqBC,WAAkE,cAAmB;CAC/G,MAAMC,eACF,UAAU,aACT,OAAqBD,WAAkE;CAC5F,MAAME,YAAuC,UAAU,UAAU,YAAY;CAC7E,MAAMC,aAAyC,UAAU,WAAW,YAAY;AAEhF,UAAS,SAAS,KAAG,GAAG,CAACC,UACrB,WAAW,MAAM,CACpB;AACD,UAAS,UAAU,KAAG,GAAG,CAACC,UACtB,YAAY,MAAM,CACrB;AACD,UAAS,UAAU,KAAG,GAAG,CAAC,GAAG,WACzB,YAAY,GAAG,OAAO,CACzB;AACD,UAAS,QAAQ,KAAG,GAAG,CAACC,MAAc,UAAU,EAAE,CAAC;AACnD,UAAS,SAAS,KAAG,GAAG,CAACA,MAAc,WAAW,EAAE,CAAC;AACrD,UAAS,SAAS,KAAG,GAAG,CAACC,SACrB,WAAW,KAAK,CACnB;AACD,UAAS,gBAAgB,KAAG,GAAG,CAAC,GAAG,SAC/B,kBAAkB,GAAG,KAAK,CAC7B;AACD,UAAS,kBAAkB,KAAG,GAAG,CAAC,GAAG,SACjC,oBAAoB,GAAG,KAAK,CAC/B;AACD,UAAS,QAAQ,KAAG,GAAG,UAAU;AACjC,UAAS,WAAW,KAAG,GAAG,aAAa;AACvC,UAAS,QAAQ,KAAG,GAAG,MAAM,WAAW,CAAC;AACzC,UAAS,SAAS,KAAG,GAAG,MAAM,YAAY,CAAC;AAE3C,QAAO;AACV;;;;ACxCM,SAAS,SACZC,YAAsC,CAAE,GACrB;CACnB,MAAM,OAAO,UAAU,QAAQ;EAAE,OAAO;EAAc,IAAI;EAAM,SAAS,CAAE;CAAE;CAC7E,MAAM,WAAW,UAAU,YAAY,WAAmB;CAG1D,MAAM,YAAY,UAAU,UAAU,MAAM;CAC5C,MAAM,eAAe,UAAU,aAAa,YAAY;CACxD,MAAM,iBACF,UAAU,eACT,OAAOC,OAAgB;EACpB,MAAM,SAAS,MAAM,aAAa,GAAG;AACrC,OAAK,OACD,OAAM,IAAI,OAAO,KAAK,KAAK,MAAM,oBAAoB,OAAO,KAAK,GAAG,CAAC,GAAG,OAAO,GAAG,CAAC;AAEvF,SAAO;CACV;CACL,MAAM,aAAa,UAAU,WAAW,OAAOC,UAA2B;CAC1E,MAAM,aAAa,UAAU,WAAW,OAAOC,KAAcC,UAA2B;CACxF,MAAM,aAAa,UAAU,WAAW,OAAOD,QAAiB,CAAE;CAClE,MAAM,iBAAiB,UAAU,eAAe,OAAOE,WAA8B;AAErF,QAAO;EACH;EACA,OAAO,GAAG,GAAG,MAAM,WAAW,CAAC;EAC/B,UAAU,GAAG,GAAG,CAACJ,OAAgB,aAAa,GAAG,CAAC;EAClD,YAAY,GAAG,GAAG,CAACA,OAAgB,eAAe,GAAG,CAAC;EACtD,QAAQ,GAAG,GAAG,CAACC,UAA2B,WAAW,MAAM,CAAC;EAC5D,QAAQ,GAAG,GAAG,CAACD,IAAaG,UAA2B,WAAW,IAAI,MAAM,CAAC;EAC7E,QAAQ,GAAG,GAAG,CAACH,OAAgB,WAAW,GAAG,CAAC;EAC9C,YAAY,GAAG,GAAG,CAACI,WAA8B,eAAe,OAAO,CAAC;CAC3E;AACJ;;;;ACfM,SAAS,gBACZC,kBAA0E,CAAE,GAC5EC,QACAC,SACQ;CACR,MAAMC,yBACK,oBAAoB,WACrB;EACI,QAAQ;EACR,KAAK;EACL,MAAM;CACT,IACD;CACV,MAAM,EACF,SAAS,OACT,MAAM,wBACN,MACA,OAAO,MACP,SAAS,CAAE,GACX,SACA,gBACH,GAAG;CAEJ,MAAMC,kBACF,SAAS,YAAY,UAAU;EAAE,gBAAgB;EAAoB,GAAG;CAAS;CAErF,MAAM,UAAU,IAAI,QAAQ,KAAK;EAC7B;EACA,SAAS;EACT,MAAM,SAAS,YAAY,YAAY,KAAK,UAAU,KAAK;CAC9D;CACD,MAAM,gBACF,mBAAmB,CAACC,KAAcC,gBAA8B,eAAe,OAAc,KAAK,YAAY;CAClH,MAAM,UAAU,cAAc,SAAS,KAAK;AAC5C,SAAQ,SAAS;AACjB,QAAO;AACV"}
@@ -1,2 +1,3 @@
1
- import "../integration-CDdpboYz.js";
2
- import "../vitest-PxMJue7R.js";
1
+ import "../aDBClient-W6eXsK3X.js";
2
+ import "../integration-BrJw6NzG.js";
3
+ import "../vitest-37qN8D93.js";
@@ -19,10 +19,10 @@ export interface TangoVitestHelpers {
19
19
  }>;
20
20
  introspectSchema(harness?: IntegrationHarness): Promise<unknown>;
21
21
  seedTable<T extends Record<string, unknown>>(table: string, rows: T[], harness?: IntegrationHarness): Promise<void>;
22
- createRepositoryFixture<TModel extends Record<string, unknown>>(options: {
23
- meta: import('@danceroutine/tango-orm/query').RepoMeta;
22
+ createQuerySetFixture<TModel extends Record<string, unknown>>(options: {
23
+ meta: import('@danceroutine/tango-orm/query').TableMeta;
24
24
  harness?: IntegrationHarness;
25
- }): import('@danceroutine/tango-orm').Repository<TModel>;
25
+ }): import('@danceroutine/tango-orm').QuerySet<TModel>;
26
26
  expectQueryResult<T>(actual: Promise<T> | T, expected: T): Promise<void>;
27
27
  }
28
28
  declare module 'vitest' {
@@ -1,4 +1,4 @@
1
- import { TestHarness, applyAndVerifyMigrations, assertMigrationPlan, createRepositoryFixture, expectQueryResult, introspectSchema, seedTable } from "./integration-CDdpboYz.js";
1
+ import { TestHarness, applyAndVerifyMigrations, assertMigrationPlan, createQuerySetFixture, expectQueryResult, introspectSchema, seedTable } from "./integration-BrJw6NzG.js";
2
2
  import { expect, vi } from "vitest";
3
3
 
4
4
  //#region src/vitest/registerVitestTango.ts
@@ -45,8 +45,8 @@ const tangoHelpers = {
45
45
  async seedTable(table, rows, harness) {
46
46
  await seedTable(harness ?? tangoHelpers.getTestHarness(), table, rows);
47
47
  },
48
- createRepositoryFixture(options) {
49
- return createRepositoryFixture({
48
+ createQuerySetFixture(options) {
49
+ return createQuerySetFixture({
50
50
  harness: options.harness ?? tangoHelpers.getTestHarness(),
51
51
  meta: options.meta
52
52
  });
@@ -78,4 +78,4 @@ var vitest_exports = {};
78
78
 
79
79
  //#endregion
80
80
  export { vitest_exports };
81
- //# sourceMappingURL=vitest-PxMJue7R.js.map
81
+ //# sourceMappingURL=vitest-37qN8D93.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"vitest-37qN8D93.js","names":["value: unknown","activeHarness: IntegrationHarness | null","input: IntegrationHarness | (() => IntegrationHarness | Promise<IntegrationHarness>)","tangoHelpers: TangoVitestHelpers","harness?: IntegrationHarness","table: string","rows: T[]","options: {\n meta: import('@danceroutine/tango-orm/query').TableMeta;\n harness?: IntegrationHarness;\n }","actual: Promise<T> | T","expected: T","received: unknown","schema: Parseable"],"sources":["../src/vitest/registerVitestTango.ts","../src/vitest/index.ts"],"sourcesContent":["/**\n * Vitest custom matchers and helpers for Tango.\n *\n * Import this module in your Vitest setup file:\n *\n * ```typescript\n * import '@danceroutine/tango-testing/vitest';\n * ```\n */\nimport { expect, vi } from 'vitest';\nimport {\n TestHarness,\n applyAndVerifyMigrations as applyAndVerifyMigrationsFn,\n assertMigrationPlan as assertMigrationPlanFn,\n createQuerySetFixture as createQuerySetFixtureFn,\n expectQueryResult as expectQueryResultFn,\n introspectSchema as introspectSchemaFn,\n seedTable as seedTableFn,\n type ApplyAndVerifyMigrationsOptions,\n type AssertMigrationPlanOptions,\n type HarnessStrategyRegistry,\n type IntegrationHarness,\n} from '../integration';\n\ninterface Parseable {\n parse(data: unknown): unknown;\n}\n\nfunction isError(value: unknown): value is Error {\n return (\n typeof value === 'object' &&\n value !== null &&\n typeof (value as { name?: unknown }).name === 'string' &&\n typeof (value as { message?: unknown }).message === 'string'\n );\n}\n\nlet activeHarness: IntegrationHarness | null = null;\n\nasync function resolveHarness(\n input: IntegrationHarness | (() => IntegrationHarness | Promise<IntegrationHarness>)\n): Promise<IntegrationHarness> {\n if (typeof input === 'function') {\n return input();\n }\n return input;\n}\n\nexport interface TangoVitestHelpers {\n useHarness(\n input: IntegrationHarness | (() => IntegrationHarness | Promise<IntegrationHarness>)\n ): Promise<IntegrationHarness>;\n getTestHarness(): IntegrationHarness;\n getRegistry(): HarnessStrategyRegistry;\n assertMigrationPlan(options: AssertMigrationPlanOptions & { harness?: IntegrationHarness }): Promise<string>;\n applyAndVerifyMigrations(\n options: ApplyAndVerifyMigrationsOptions & { harness?: IntegrationHarness }\n ): Promise<{ statuses: { id: string; applied: boolean }[] }>;\n introspectSchema(harness?: IntegrationHarness): Promise<unknown>;\n seedTable<T extends Record<string, unknown>>(table: string, rows: T[], harness?: IntegrationHarness): Promise<void>;\n createQuerySetFixture<TModel extends Record<string, unknown>>(options: {\n meta: import('@danceroutine/tango-orm/query').TableMeta;\n harness?: IntegrationHarness;\n }): import('@danceroutine/tango-orm').QuerySet<TModel>;\n expectQueryResult<T>(actual: Promise<T> | T, expected: T): Promise<void>;\n}\n\nconst tangoHelpers: TangoVitestHelpers = {\n async useHarness(input): Promise<IntegrationHarness> {\n const harness = await resolveHarness(input);\n await harness.setup();\n activeHarness = harness;\n return harness;\n },\n getTestHarness(): IntegrationHarness {\n if (!activeHarness) {\n throw new Error('No active test harness. Call vi.tango.useHarness(...) in beforeAll first.');\n }\n return activeHarness;\n },\n getRegistry(): HarnessStrategyRegistry {\n return TestHarness.getRegistry();\n },\n async assertMigrationPlan(options): Promise<string> {\n const harness = options.harness ?? tangoHelpers.getTestHarness();\n return assertMigrationPlanFn(harness, {\n migrationsDir: options.migrationsDir,\n expectSqlContains: options.expectSqlContains,\n });\n },\n async applyAndVerifyMigrations(options): Promise<{ statuses: { id: string; applied: boolean }[] }> {\n const harness = options.harness ?? tangoHelpers.getTestHarness();\n return applyAndVerifyMigrationsFn(harness, {\n migrationsDir: options.migrationsDir,\n toId: options.toId,\n expectedAppliedIds: options.expectedAppliedIds,\n });\n },\n async introspectSchema(harness?: IntegrationHarness): Promise<unknown> {\n return introspectSchemaFn(harness ?? tangoHelpers.getTestHarness());\n },\n async seedTable<T extends Record<string, unknown>>(\n table: string,\n rows: T[],\n harness?: IntegrationHarness\n ): Promise<void> {\n await seedTableFn(harness ?? tangoHelpers.getTestHarness(), table, rows);\n },\n createQuerySetFixture<TModel extends Record<string, unknown>>(options: {\n meta: import('@danceroutine/tango-orm/query').TableMeta;\n harness?: IntegrationHarness;\n }): import('@danceroutine/tango-orm').QuerySet<TModel> {\n return createQuerySetFixtureFn<TModel>({\n harness: options.harness ?? tangoHelpers.getTestHarness(),\n meta: options.meta,\n });\n },\n async expectQueryResult<T>(actual: Promise<T> | T, expected: T): Promise<void> {\n await expectQueryResultFn(actual, expected);\n },\n};\n\nexpect.extend({\n toMatchSchema(received: unknown, schema: Parseable) {\n try {\n schema.parse(received);\n return {\n pass: true,\n message: () => 'expected data not to match schema',\n };\n } catch (error) {\n const detail = isError(error) ? error.message : String(error);\n return {\n pass: false,\n message: () => `expected data to match schema\\n\\n${detail}`,\n };\n }\n },\n});\n\n(vi as unknown as { tango?: TangoVitestHelpers }).tango = tangoHelpers;\n\ndeclare module 'vitest' {\n // oxlint-disable-next-line no-unused-vars\n interface Assertion<T> {\n toMatchSchema(schema: Parseable): void;\n }\n\n interface AsymmetricMatchersContaining {\n toMatchSchema(schema: Parseable): void;\n }\n\n interface VitestUtils {\n tango: TangoVitestHelpers;\n }\n}\n","/**\n * Domain boundary barrel: centralizes this subdomain's public contract.\n */\n\nexport * from './registerVitestTango';\n"],"mappings":";;;;AA4BA,SAAS,QAAQA,OAAgC;AAC7C,eACW,UAAU,YACjB,UAAU,eACF,MAA6B,SAAS,mBACtC,MAAgC,YAAY;AAE3D;AAED,IAAIC,gBAA2C;AAE/C,eAAe,eACXC,OAC2B;AAC3B,YAAW,UAAU,WACjB,QAAO,OAAO;AAElB,QAAO;AACV;AAqBD,MAAMC,eAAmC;CACrC,MAAM,WAAW,OAAoC;EACjD,MAAM,UAAU,MAAM,eAAe,MAAM;AAC3C,QAAM,QAAQ,OAAO;AACrB,kBAAgB;AAChB,SAAO;CACV;CACD,iBAAqC;AACjC,OAAK,cACD,OAAM,IAAI,MAAM;AAEpB,SAAO;CACV;CACD,cAAuC;AACnC,SAAO,YAAY,aAAa;CACnC;CACD,MAAM,oBAAoB,SAA0B;EAChD,MAAM,UAAU,QAAQ,WAAW,aAAa,gBAAgB;AAChE,SAAO,oBAAsB,SAAS;GAClC,eAAe,QAAQ;GACvB,mBAAmB,QAAQ;EAC9B,EAAC;CACL;CACD,MAAM,yBAAyB,SAAoE;EAC/F,MAAM,UAAU,QAAQ,WAAW,aAAa,gBAAgB;AAChE,SAAO,yBAA2B,SAAS;GACvC,eAAe,QAAQ;GACvB,MAAM,QAAQ;GACd,oBAAoB,QAAQ;EAC/B,EAAC;CACL;CACD,MAAM,iBAAiBC,SAAgD;AACnE,SAAO,iBAAmB,WAAW,aAAa,gBAAgB,CAAC;CACtE;CACD,MAAM,UACFC,OACAC,MACAF,SACa;AACb,QAAM,UAAY,WAAW,aAAa,gBAAgB,EAAE,OAAO,KAAK;CAC3E;CACD,sBAA8DG,SAGP;AACnD,SAAO,sBAAgC;GACnC,SAAS,QAAQ,WAAW,aAAa,gBAAgB;GACzD,MAAM,QAAQ;EACjB,EAAC;CACL;CACD,MAAM,kBAAqBC,QAAwBC,UAA4B;AAC3E,QAAM,kBAAoB,QAAQ,SAAS;CAC9C;AACJ;AAED,OAAO,OAAO,EACV,cAAcC,UAAmBC,QAAmB;AAChD,KAAI;AACA,SAAO,MAAM,SAAS;AACtB,SAAO;GACH,MAAM;GACN,SAAS,MAAM;EAClB;CACJ,SAAQ,OAAO;EACZ,MAAM,SAAS,QAAQ,MAAM,GAAG,MAAM,UAAU,OAAO,MAAM;AAC7D,SAAO;GACH,MAAM;GACN,SAAS,OAAO,mCAAmC,OAAO;EAC7D;CACJ;AACJ,EACJ,EAAC;AAEF,GAAkD,QAAQ"}