@contractspec/example.product-intent 1.57.0 → 1.58.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (42) hide show
  1. package/.turbo/turbo-build.log +21 -96
  2. package/.turbo/turbo-prebuild.log +1 -0
  3. package/CHANGELOG.md +15 -0
  4. package/dist/example.d.ts +2 -6
  5. package/dist/example.d.ts.map +1 -1
  6. package/dist/example.js +27 -37
  7. package/dist/index.d.ts +6 -4
  8. package/dist/index.d.ts.map +1 -0
  9. package/dist/index.js +342 -4
  10. package/dist/load-evidence.d.ts +13 -17
  11. package/dist/load-evidence.d.ts.map +1 -1
  12. package/dist/load-evidence.js +307 -68
  13. package/dist/load-evidence.test.d.ts +2 -0
  14. package/dist/load-evidence.test.d.ts.map +1 -0
  15. package/dist/node/example.js +28 -0
  16. package/dist/node/index.js +342 -0
  17. package/dist/node/load-evidence.js +313 -0
  18. package/dist/node/posthog-signals.js +248 -0
  19. package/dist/node/script.js +512 -0
  20. package/dist/node/sync-actions.js +491 -0
  21. package/dist/posthog-signals.d.ts +15 -19
  22. package/dist/posthog-signals.d.ts.map +1 -1
  23. package/dist/posthog-signals.js +222 -178
  24. package/dist/script.d.ts +2 -1
  25. package/dist/script.d.ts.map +1 -0
  26. package/dist/script.js +493 -152
  27. package/dist/sync-actions.d.ts +2 -1
  28. package/dist/sync-actions.d.ts.map +1 -0
  29. package/dist/sync-actions.js +466 -128
  30. package/package.json +57 -27
  31. package/tsdown.config.js +1 -2
  32. package/.turbo/turbo-build$colon$bundle.log +0 -99
  33. package/dist/example.js.map +0 -1
  34. package/dist/libs/analytics/dist/funnel/analyzer.js +0 -64
  35. package/dist/libs/analytics/dist/funnel/analyzer.js.map +0 -1
  36. package/dist/libs/analytics/dist/types.d.ts +0 -22
  37. package/dist/libs/analytics/dist/types.d.ts.map +0 -1
  38. package/dist/load-evidence.js.map +0 -1
  39. package/dist/posthog-signals.js.map +0 -1
  40. package/dist/script.js.map +0 -1
  41. package/dist/sync-actions.js.map +0 -1
  42. package/tsconfig.tsbuildinfo +0 -1
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@contractspec/example.product-intent",
3
- "version": "1.57.0",
3
+ "version": "1.58.0",
4
4
  "description": "Product intent example: evidence ingestion and prompt-ready outputs.",
5
5
  "keywords": [
6
6
  "contractspec",
@@ -12,48 +12,78 @@
12
12
  "type": "module",
13
13
  "types": "./dist/index.d.ts",
14
14
  "exports": {
15
- ".": "./dist/index.js",
16
- "./example": "./dist/example.js",
17
- "./load-evidence": "./dist/load-evidence.js",
18
- "./posthog-signals": "./dist/posthog-signals.js",
19
- "./script": "./dist/script.js",
20
- "./sync-actions": "./dist/sync-actions.js",
21
- "./*": "./*"
15
+ ".": "./src/index.ts",
16
+ "./example": "./src/example.ts",
17
+ "./load-evidence": "./src/load-evidence.ts",
18
+ "./posthog-signals": "./src/posthog-signals.ts",
19
+ "./script": "./src/script.ts",
20
+ "./sync-actions": "./src/sync-actions.ts"
22
21
  },
23
22
  "scripts": {
24
23
  "publish:pkg": "bun publish --tolerate-republish --ignore-scripts --verbose",
25
24
  "publish:pkg:canary": "bun publish:pkg --tag canary",
26
- "build": "bun build:types && bun build:bundle",
27
- "build:bundle": "tsdown",
28
- "build:types": "tsc --noEmit",
29
- "dev": "bun build:bundle --watch",
25
+ "build": "bun run prebuild && bun run build:bundle && bun run build:types",
26
+ "build:bundle": "contractspec-bun-build transpile",
27
+ "build:types": "contractspec-bun-build types",
28
+ "dev": "contractspec-bun-build dev",
30
29
  "clean": "rimraf dist .turbo",
31
30
  "lint": "bun lint:fix",
32
31
  "lint:fix": "eslint src --fix",
33
32
  "lint:check": "eslint src",
34
- "test": "bun test"
33
+ "test": "bun test",
34
+ "prebuild": "contractspec-bun-build prebuild",
35
+ "typecheck": "tsc --noEmit"
35
36
  },
36
37
  "dependencies": {
37
- "@contractspec/lib.contracts": "1.57.0",
38
- "@contractspec/lib.ai-agent": "1.57.0",
39
- "@contractspec/lib.product-intent-utils": "1.57.0",
40
- "@contractspec/integration.providers-impls": "1.57.0"
38
+ "@contractspec/lib.contracts": "1.58.0",
39
+ "@contractspec/lib.ai-agent": "1.58.0",
40
+ "@contractspec/lib.product-intent-utils": "1.58.0",
41
+ "@contractspec/integration.providers-impls": "1.58.0"
41
42
  },
42
43
  "devDependencies": {
43
- "@contractspec/tool.tsdown": "1.57.0",
44
- "@contractspec/tool.typescript": "1.57.0",
45
- "tsdown": "^0.20.3",
46
- "typescript": "^5.9.3"
44
+ "@contractspec/tool.typescript": "1.58.0",
45
+ "typescript": "^5.9.3",
46
+ "@contractspec/tool.bun": "1.57.0"
47
47
  },
48
48
  "publishConfig": {
49
49
  "access": "public",
50
50
  "exports": {
51
- ".": "./dist/index.js",
52
- "./example": "./dist/example.js",
53
- "./load-evidence": "./dist/load-evidence.js",
54
- "./script": "./dist/script.js",
55
- "./posthog-signals": "./dist/posthog-signals.js",
56
- "./*": "./*"
51
+ ".": {
52
+ "types": "./dist/index.d.ts",
53
+ "bun": "./dist/index.js",
54
+ "node": "./dist/node/index.mjs",
55
+ "default": "./dist/index.js"
56
+ },
57
+ "./example": {
58
+ "types": "./dist/example.d.ts",
59
+ "bun": "./dist/example.js",
60
+ "node": "./dist/node/example.mjs",
61
+ "default": "./dist/example.js"
62
+ },
63
+ "./load-evidence": {
64
+ "types": "./dist/load-evidence.d.ts",
65
+ "bun": "./dist/load-evidence.js",
66
+ "node": "./dist/node/load-evidence.mjs",
67
+ "default": "./dist/load-evidence.js"
68
+ },
69
+ "./posthog-signals": {
70
+ "types": "./dist/posthog-signals.d.ts",
71
+ "bun": "./dist/posthog-signals.js",
72
+ "node": "./dist/node/posthog-signals.mjs",
73
+ "default": "./dist/posthog-signals.js"
74
+ },
75
+ "./script": {
76
+ "types": "./dist/script.d.ts",
77
+ "bun": "./dist/script.js",
78
+ "node": "./dist/node/script.mjs",
79
+ "default": "./dist/script.js"
80
+ },
81
+ "./sync-actions": {
82
+ "types": "./dist/sync-actions.d.ts",
83
+ "bun": "./dist/sync-actions.js",
84
+ "node": "./dist/node/sync-actions.mjs",
85
+ "default": "./dist/sync-actions.js"
86
+ }
57
87
  },
58
88
  "registry": "https://registry.npmjs.org/"
59
89
  },
package/tsdown.config.js CHANGED
@@ -1,5 +1,4 @@
1
- import { defineConfig } from 'tsdown';
2
- import { moduleLibrary, withDevExports } from '@contractspec/tool.tsdown';
1
+ import { defineConfig, moduleLibrary, withDevExports } from '@contractspec/tool.bun';
3
2
 
4
3
  export default defineConfig(() => ({
5
4
  ...moduleLibrary,
@@ -1,99 +0,0 @@
1
- $ tsdown
2
- ℹ tsdown v0.20.3 powered by rolldown v1.0.0-rc.3
3
- ℹ config file: /home/runner/work/contractspec/contractspec/packages/examples/product-intent/tsdown.config.js
4
- ℹ entry: src/example.ts, src/index.ts, src/load-evidence.ts, src/posthog-signals.ts, src/script.ts, src/sync-actions.ts
5
- ℹ target: esnext
6
- ℹ tsconfig: tsconfig.json
7
- ℹ Build start
8
- ℹ Cleaning 28 files
9
- ℹ dist/posthog-signals.js  6.96 kB │ gzip: 2.20 kB
10
- ℹ dist/script.js  6.55 kB │ gzip: 2.29 kB
11
- ℹ dist/sync-actions.js  5.88 kB │ gzip: 2.07 kB
12
- ℹ dist/load-evidence.js  2.67 kB │ gzip: 1.12 kB
13
- ℹ dist/example.js  0.81 kB │ gzip: 0.44 kB
14
- ℹ dist/index.js  0.50 kB │ gzip: 0.21 kB
15
- ℹ dist/posthog-signals.js.map 14.02 kB │ gzip: 4.15 kB
16
- ℹ dist/script.js.map 11.64 kB │ gzip: 3.74 kB
17
- ℹ dist/sync-actions.js.map 10.53 kB │ gzip: 3.37 kB
18
- ℹ dist/load-evidence.js.map  4.79 kB │ gzip: 1.86 kB
19
- ℹ dist/libs/analytics/dist/funnel/analyzer.js.map  3.65 kB │ gzip: 1.39 kB
20
- ℹ dist/libs/analytics/dist/funnel/analyzer.js  2.00 kB │ gzip: 0.84 kB
21
- ℹ dist/example.js.map  1.22 kB │ gzip: 0.63 kB
22
- ℹ dist/libs/analytics/dist/types.d.ts.map  0.95 kB │ gzip: 0.48 kB
23
- ℹ dist/posthog-signals.d.ts.map  0.37 kB │ gzip: 0.23 kB
24
- ℹ dist/load-evidence.d.ts.map  0.37 kB │ gzip: 0.23 kB
25
- ℹ dist/example.d.ts.map  0.13 kB │ gzip: 0.13 kB
26
- ℹ dist/load-evidence.d.ts  1.10 kB │ gzip: 0.49 kB
27
- ℹ dist/posthog-signals.d.ts  1.00 kB │ gzip: 0.43 kB
28
- ℹ dist/index.d.ts  0.67 kB │ gzip: 0.24 kB
29
- ℹ dist/example.d.ts  0.25 kB │ gzip: 0.17 kB
30
- ℹ dist/script.d.ts  0.01 kB │ gzip: 0.03 kB
31
- ℹ dist/sync-actions.d.ts  0.01 kB │ gzip: 0.03 kB
32
- ℹ dist/libs/analytics/dist/types.d.ts  0.50 kB │ gzip: 0.28 kB
33
- ℹ 24 files, total: 76.58 kB
34
- src/load-evidence.ts (1:15) [UNRESOLVED_IMPORT] Warning: Could not resolve 'node:fs' in src/load-evidence.ts
35
- ╭─[ src/load-evidence.ts:1:16 ]
36
- │
37
- 1 │ import fs from 'node:fs';
38
-  │ ────┬────
39
-  │ ╰────── Module not found, treating it as an external dependency
40
-  │
41
-  │ Help: The "main" field here was ignored. Main fields must be configured explicitly when using the "neutral" platform.
42
- ───╯
43
-
44
- src/load-evidence.ts (2:17) [UNRESOLVED_IMPORT] Warning: Could not resolve 'node:path' in src/load-evidence.ts
45
- ╭─[ src/load-evidence.ts:2:18 ]
46
- │
47
- 2 │ import path from 'node:path';
48
-  │ ─────┬─────
49
-  │ ╰─────── Module not found, treating it as an external dependency
50
-  │
51
-  │ Help: The "main" field here was ignored. Main fields must be configured explicitly when using the "neutral" platform.
52
- ───╯
53
-
54
- src/load-evidence.ts (3:30) [UNRESOLVED_IMPORT] Warning: Could not resolve 'node:url' in src/load-evidence.ts
55
- ╭─[ src/load-evidence.ts:3:31 ]
56
- │
57
- 3 │ import { fileURLToPath } from 'node:url';
58
-  │ ─────┬────
59
-  │ ╰────── Module not found, treating it as an external dependency
60
-  │
61
-  │ Help: The "main" field here was ignored. Main fields must be configured explicitly when using the "neutral" platform.
62
- ───╯
63
-
64
- src/script.ts (1:15) [UNRESOLVED_IMPORT] Warning: Could not resolve 'node:fs' in src/script.ts
65
- ╭─[ src/script.ts:1:16 ]
66
- │
67
- 1 │ import fs from 'node:fs';
68
-  │ ────┬────
69
-  │ ╰────── Module not found, treating it as an external dependency
70
-  │
71
-  │ Help: The "main" field here was ignored. Main fields must be configured explicitly when using the "neutral" platform.
72
- ───╯
73
-
74
- src/script.ts (2:17) [UNRESOLVED_IMPORT] Warning: Could not resolve 'node:path' in src/script.ts
75
- ╭─[ src/script.ts:2:18 ]
76
- │
77
- 2 │ import path from 'node:path';
78
-  │ ─────┬─────
79
-  │ ╰─────── Module not found, treating it as an external dependency
80
-  │
81
-  │ Help: The "main" field here was ignored. Main fields must be configured explicitly when using the "neutral" platform.
82
- ───╯
83
-
84
- src/script.ts (3:30) [UNRESOLVED_IMPORT] Warning: Could not resolve 'node:url' in src/script.ts
85
- ╭─[ src/script.ts:3:31 ]
86
- │
87
- 3 │ import { fileURLToPath } from 'node:url';
88
-  │ ─────┬────
89
-  │ ╰────── Module not found, treating it as an external dependency
90
-  │
91
-  │ Help: The "main" field here was ignored. Main fields must be configured explicitly when using the "neutral" platform.
92
- ───╯
93
-
94
- [PLUGIN_TIMINGS] Warning: Your build spent significant time in plugins. Here is a breakdown:
95
- - tsdown:external (76%)
96
- - rolldown-plugin-dts:generate (24%)
97
- See https://rolldown.rs/options/checks#plugintimings for more details.
98
-
99
- ✔ Build complete in 21052ms
@@ -1 +0,0 @@
1
- {"version":3,"file":"example.js","names":[],"sources":["../src/example.ts"],"sourcesContent":["import { defineExample } from '@contractspec/lib.contracts';\n\nconst example = defineExample({\n meta: {\n key: 'product-intent',\n version: '1.0.0',\n title: 'Product Intent Discovery',\n description:\n 'Evidence ingestion and product-intent workflow for PM discovery.',\n kind: 'script',\n visibility: 'public',\n stability: 'experimental',\n owners: ['@platform.core'],\n tags: ['product-intent', 'discovery', 'pm', 'evidence', 'llm'],\n },\n entrypoints: {\n packageName: '@contractspec/example.product-intent',\n },\n surfaces: {\n templates: false,\n sandbox: { enabled: false, modes: [] },\n studio: { enabled: false, installable: false },\n mcp: { enabled: false },\n },\n});\n\nexport default example;\n"],"mappings":";;;AAEA,MAAM,UAAU,cAAc;CAC5B,MAAM;EACJ,KAAK;EACL,SAAS;EACT,OAAO;EACP,aACE;EACF,MAAM;EACN,YAAY;EACZ,WAAW;EACX,QAAQ,CAAC,iBAAiB;EAC1B,MAAM;GAAC;GAAkB;GAAa;GAAM;GAAY;GAAM;EAC/D;CACD,aAAa,EACX,aAAa,wCACd;CACD,UAAU;EACR,WAAW;EACX,SAAS;GAAE,SAAS;GAAO,OAAO,EAAE;GAAE;EACtC,QAAQ;GAAE,SAAS;GAAO,aAAa;GAAO;EAC9C,KAAK,EAAE,SAAS,OAAO;EACxB;CACF,CAAC"}
@@ -1,64 +0,0 @@
1
- //#region ../../libs/analytics/dist/funnel/analyzer.js
2
- var FunnelAnalyzer = class {
3
- analyze(events, definition) {
4
- const windowMs = (definition.windowHours ?? 72) * 60 * 60 * 1e3;
5
- const eventsByUser = groupByUser(events);
6
- const stepCounts = definition.steps.map(() => 0);
7
- for (const userEvents of eventsByUser.values()) this.evaluateUser(userEvents, definition.steps, windowMs).forEach((hit, stepIdx) => {
8
- if (hit) stepCounts[stepIdx] = (stepCounts[stepIdx] ?? 0) + 1;
9
- });
10
- const totalUsers = eventsByUser.size;
11
- return {
12
- definition,
13
- totalUsers,
14
- steps: definition.steps.map((step, index) => {
15
- const prevCount = index === 0 ? totalUsers : stepCounts[index - 1] || 1;
16
- const count = stepCounts[index] ?? 0;
17
- const conversionRate = prevCount === 0 ? 0 : Number((count / prevCount).toFixed(3));
18
- return {
19
- step,
20
- count,
21
- conversionRate,
22
- dropOffRate: Number((1 - conversionRate).toFixed(3))
23
- };
24
- })
25
- };
26
- }
27
- evaluateUser(events, steps, windowMs) {
28
- const sorted = [...events].sort((a, b) => new Date(a.timestamp).getTime() - new Date(b.timestamp).getTime());
29
- const completion = Array(steps.length).fill(false);
30
- let cursor = 0;
31
- let anchorTime;
32
- for (const event of sorted) {
33
- const step = steps[cursor];
34
- if (!step) break;
35
- if (event.name !== step.eventName) continue;
36
- if (step.match && !step.match(event)) continue;
37
- const eventTime = new Date(event.timestamp).getTime();
38
- if (cursor === 0) {
39
- anchorTime = eventTime;
40
- completion[cursor] = true;
41
- cursor += 1;
42
- continue;
43
- }
44
- if (anchorTime && eventTime - anchorTime <= windowMs) {
45
- completion[cursor] = true;
46
- cursor += 1;
47
- }
48
- }
49
- return completion;
50
- }
51
- };
52
- function groupByUser(events) {
53
- const map = /* @__PURE__ */ new Map();
54
- for (const event of events) {
55
- const list = map.get(event.userId) ?? [];
56
- list.push(event);
57
- map.set(event.userId, list);
58
- }
59
- return map;
60
- }
61
-
62
- //#endregion
63
- export { FunnelAnalyzer };
64
- //# sourceMappingURL=analyzer.js.map
@@ -1 +0,0 @@
1
- {"version":3,"file":"analyzer.js","names":[],"sources":["../../../../../../../libs/analytics/dist/funnel/analyzer.js"],"sourcesContent":["//#region src/funnel/analyzer.ts\nvar FunnelAnalyzer = class {\n\tanalyze(events, definition) {\n\t\tconst windowMs = (definition.windowHours ?? 72) * 60 * 60 * 1e3;\n\t\tconst eventsByUser = groupByUser(events);\n\t\tconst stepCounts = definition.steps.map(() => 0);\n\t\tfor (const userEvents of eventsByUser.values()) this.evaluateUser(userEvents, definition.steps, windowMs).forEach((hit, stepIdx) => {\n\t\t\tif (hit) stepCounts[stepIdx] = (stepCounts[stepIdx] ?? 0) + 1;\n\t\t});\n\t\tconst totalUsers = eventsByUser.size;\n\t\treturn {\n\t\t\tdefinition,\n\t\t\ttotalUsers,\n\t\t\tsteps: definition.steps.map((step, index) => {\n\t\t\t\tconst prevCount = index === 0 ? totalUsers : stepCounts[index - 1] || 1;\n\t\t\t\tconst count = stepCounts[index] ?? 0;\n\t\t\t\tconst conversionRate = prevCount === 0 ? 0 : Number((count / prevCount).toFixed(3));\n\t\t\t\treturn {\n\t\t\t\t\tstep,\n\t\t\t\t\tcount,\n\t\t\t\t\tconversionRate,\n\t\t\t\t\tdropOffRate: Number((1 - conversionRate).toFixed(3))\n\t\t\t\t};\n\t\t\t})\n\t\t};\n\t}\n\tevaluateUser(events, steps, windowMs) {\n\t\tconst sorted = [...events].sort((a, b) => new Date(a.timestamp).getTime() - new Date(b.timestamp).getTime());\n\t\tconst completion = Array(steps.length).fill(false);\n\t\tlet cursor = 0;\n\t\tlet anchorTime;\n\t\tfor (const event of sorted) {\n\t\t\tconst step = steps[cursor];\n\t\t\tif (!step) break;\n\t\t\tif (event.name !== step.eventName) continue;\n\t\t\tif (step.match && !step.match(event)) continue;\n\t\t\tconst eventTime = new Date(event.timestamp).getTime();\n\t\t\tif (cursor === 0) {\n\t\t\t\tanchorTime = eventTime;\n\t\t\t\tcompletion[cursor] = true;\n\t\t\t\tcursor += 1;\n\t\t\t\tcontinue;\n\t\t\t}\n\t\t\tif (anchorTime && eventTime - anchorTime <= windowMs) {\n\t\t\t\tcompletion[cursor] = true;\n\t\t\t\tcursor += 1;\n\t\t\t}\n\t\t}\n\t\treturn completion;\n\t}\n};\nfunction groupByUser(events) {\n\tconst map = /* @__PURE__ */ new Map();\n\tfor (const event of events) {\n\t\tconst list = map.get(event.userId) ?? [];\n\t\tlist.push(event);\n\t\tmap.set(event.userId, list);\n\t}\n\treturn map;\n}\n\n//#endregion\nexport { FunnelAnalyzer };\n//# sourceMappingURL=analyzer.js.map"],"mappings":";AACA,IAAI,iBAAiB,MAAM;CAC1B,QAAQ,QAAQ,YAAY;EAC3B,MAAM,YAAY,WAAW,eAAe,MAAM,KAAK,KAAK;EAC5D,MAAM,eAAe,YAAY,OAAO;EACxC,MAAM,aAAa,WAAW,MAAM,UAAU,EAAE;AAChD,OAAK,MAAM,cAAc,aAAa,QAAQ,CAAE,MAAK,aAAa,YAAY,WAAW,OAAO,SAAS,CAAC,SAAS,KAAK,YAAY;AACnI,OAAI,IAAK,YAAW,YAAY,WAAW,YAAY,KAAK;IAC3D;EACF,MAAM,aAAa,aAAa;AAChC,SAAO;GACN;GACA;GACA,OAAO,WAAW,MAAM,KAAK,MAAM,UAAU;IAC5C,MAAM,YAAY,UAAU,IAAI,aAAa,WAAW,QAAQ,MAAM;IACtE,MAAM,QAAQ,WAAW,UAAU;IACnC,MAAM,iBAAiB,cAAc,IAAI,IAAI,QAAQ,QAAQ,WAAW,QAAQ,EAAE,CAAC;AACnF,WAAO;KACN;KACA;KACA;KACA,aAAa,QAAQ,IAAI,gBAAgB,QAAQ,EAAE,CAAC;KACpD;KACA;GACF;;CAEF,aAAa,QAAQ,OAAO,UAAU;EACrC,MAAM,SAAS,CAAC,GAAG,OAAO,CAAC,MAAM,GAAG,MAAM,IAAI,KAAK,EAAE,UAAU,CAAC,SAAS,GAAG,IAAI,KAAK,EAAE,UAAU,CAAC,SAAS,CAAC;EAC5G,MAAM,aAAa,MAAM,MAAM,OAAO,CAAC,KAAK,MAAM;EAClD,IAAI,SAAS;EACb,IAAI;AACJ,OAAK,MAAM,SAAS,QAAQ;GAC3B,MAAM,OAAO,MAAM;AACnB,OAAI,CAAC,KAAM;AACX,OAAI,MAAM,SAAS,KAAK,UAAW;AACnC,OAAI,KAAK,SAAS,CAAC,KAAK,MAAM,MAAM,CAAE;GACtC,MAAM,YAAY,IAAI,KAAK,MAAM,UAAU,CAAC,SAAS;AACrD,OAAI,WAAW,GAAG;AACjB,iBAAa;AACb,eAAW,UAAU;AACrB,cAAU;AACV;;AAED,OAAI,cAAc,YAAY,cAAc,UAAU;AACrD,eAAW,UAAU;AACrB,cAAU;;;AAGZ,SAAO;;;AAGT,SAAS,YAAY,QAAQ;CAC5B,MAAM,sBAAsB,IAAI,KAAK;AACrC,MAAK,MAAM,SAAS,QAAQ;EAC3B,MAAM,OAAO,IAAI,IAAI,MAAM,OAAO,IAAI,EAAE;AACxC,OAAK,KAAK,MAAM;AAChB,MAAI,IAAI,MAAM,QAAQ,KAAK;;AAE5B,QAAO"}
@@ -1,22 +0,0 @@
1
- //#region ../../libs/analytics/dist/types.d.ts
2
- //#region src/types.d.ts
3
- interface AnalyticsEvent {
4
- name: string;
5
- userId: string;
6
- tenantId?: string;
7
- timestamp: string | Date;
8
- properties?: Record<string, unknown>;
9
- }
10
- interface FunnelStep {
11
- id: string;
12
- eventName: string;
13
- match?: (event: AnalyticsEvent) => boolean;
14
- }
15
- interface FunnelDefinition {
16
- name: string;
17
- steps: FunnelStep[];
18
- windowHours?: number;
19
- }
20
- //#endregion
21
- export { FunnelDefinition };
22
- //# sourceMappingURL=types.d.ts.map
@@ -1 +0,0 @@
1
- {"version":3,"file":"types.d.ts","names":["AnalyticsEvent","Date","Record","name","userId","tenantId","timestamp","properties","FunnelStep","id","eventName","match","event","FunnelDefinition","steps","windowHours","FunnelStepResult","step","count","conversionRate","dropOffRate","FunnelAnalysis","definition","totalUsers","CohortEvent","amount","CohortDefinition","bucket","periods","startDate","CohortStats","cohortKey","users","retention","ltv","CohortAnalysis","cohorts","ChurnSignal","score","drivers","GrowthMetric","current","previous","target","GrowthHypothesis","statement","metric","confidence","impact"],"sources":["../../../../../../libs/analytics/dist/types.d.ts"],"mappings":";;UACUA,cAAAA;EACRG,IAAAA;EACAC,MAAAA;EACAC,QAAAA;EACAC,SAAAA,WAAoBL,IAAAA;EACpBM,UAAAA,GAAaL,MAAAA;AAAAA;AAAAA,UAELM,UAAAA;EACRC,EAAAA;EACAC,SAAAA;EACAC,KAAAA,IAASC,KAAAA,EAAOZ,cAAAA;AAAAA;AAAAA,UAERa,gBAAAA;EACRV,IAAAA;EACAW,KAAAA,EAAON,UAAAA;EACPO,WAAAA;AAAAA"}
@@ -1 +0,0 @@
1
- {"version":3,"file":"load-evidence.js","names":[],"sources":["../src/load-evidence.ts"],"sourcesContent":["import fs from 'node:fs';\nimport path from 'node:path';\nimport { fileURLToPath } from 'node:url';\nimport type { EvidenceChunk } from '@contractspec/lib.contracts/product-intent/types';\nimport type { PosthogEvidenceOptions } from './posthog-signals';\nimport { loadPosthogEvidenceChunks } from './posthog-signals';\n\nconst MODULE_DIR = path.dirname(fileURLToPath(import.meta.url));\n\nexport const DEFAULT_EVIDENCE_ROOT = path.join(MODULE_DIR, '../evidence');\nexport const DEFAULT_TRANSCRIPT_DIRS = ['interviews', 'tickets', 'public'];\nexport const DEFAULT_CHUNK_SIZE = 800;\n\nexport interface EvidenceLoadOptions {\n evidenceRoot?: string;\n transcriptDirs?: string[];\n chunkSize?: number;\n}\n\nexport interface EvidenceLoadWithSignalsOptions extends EvidenceLoadOptions {\n posthog?: PosthogEvidenceOptions;\n}\n\n/**\n * Remove YAML front matter from a file. Synthetic interview and ticket\n * files include a YAML header delimited by triple dashes.\n */\nfunction stripYamlFrontMatter(contents: string): string {\n const start = contents.indexOf('---');\n if (start === -1) return contents;\n const end = contents.indexOf('---', start + 3);\n if (end === -1) return contents;\n return contents.slice(end + 3).trimStart();\n}\n\n/**\n * Split a transcript into fixed-size chunks.\n */\nfunction chunkTranscript(\n fileId: string,\n text: string,\n chunkSize: number\n): EvidenceChunk[] {\n const chunks: EvidenceChunk[] = [];\n const clean = text.trim();\n for (let offset = 0, idx = 0; offset < clean.length; idx += 1) {\n const slice = clean.slice(offset, offset + chunkSize);\n chunks.push({\n chunkId: `${fileId}#c_${String(idx).padStart(2, '0')}`,\n text: slice,\n meta: { source: fileId },\n });\n offset += chunkSize;\n }\n return chunks;\n}\n\n/**\n * Load all transcript files under the given directories and return\n * EvidenceChunk objects ready for prompt formatting.\n */\nexport function loadEvidenceChunks(\n options: EvidenceLoadOptions = {}\n): EvidenceChunk[] {\n const evidenceRoot = options.evidenceRoot ?? DEFAULT_EVIDENCE_ROOT;\n const transcriptDirs = options.transcriptDirs ?? DEFAULT_TRANSCRIPT_DIRS;\n const chunkSize = options.chunkSize ?? DEFAULT_CHUNK_SIZE;\n\n const chunks: EvidenceChunk[] = [];\n for (const dir of transcriptDirs) {\n const fullDir = path.join(evidenceRoot, dir);\n if (!fs.existsSync(fullDir)) continue;\n for (const fileName of fs.readdirSync(fullDir)) {\n const ext = path.extname(fileName).toLowerCase();\n if (ext !== '.md') continue;\n const filePath = path.join(fullDir, fileName);\n const raw = fs.readFileSync(filePath, 'utf8');\n const withoutFrontMatter = stripYamlFrontMatter(raw);\n const baseId = path.parse(fileName).name;\n const fileChunks = chunkTranscript(baseId, withoutFrontMatter, chunkSize);\n chunks.push(...fileChunks);\n }\n }\n return chunks;\n}\n\nexport async function loadEvidenceChunksWithSignals(\n options: EvidenceLoadWithSignalsOptions = {}\n): Promise<EvidenceChunk[]> {\n const baseChunks = loadEvidenceChunks(options);\n if (!options.posthog) return baseChunks;\n const posthogChunks = await loadPosthogEvidenceChunks(options.posthog);\n return [...baseChunks, ...posthogChunks];\n}\n"],"mappings":";;;;;;AAOA,MAAM,aAAa,KAAK,QAAQ,cAAc,OAAO,KAAK,IAAI,CAAC;AAE/D,MAAa,wBAAwB,KAAK,KAAK,YAAY,cAAc;AACzE,MAAa,0BAA0B;CAAC;CAAc;CAAW;CAAS;AAC1E,MAAa,qBAAqB;;;;;AAgBlC,SAAS,qBAAqB,UAA0B;CACtD,MAAM,QAAQ,SAAS,QAAQ,MAAM;AACrC,KAAI,UAAU,GAAI,QAAO;CACzB,MAAM,MAAM,SAAS,QAAQ,OAAO,QAAQ,EAAE;AAC9C,KAAI,QAAQ,GAAI,QAAO;AACvB,QAAO,SAAS,MAAM,MAAM,EAAE,CAAC,WAAW;;;;;AAM5C,SAAS,gBACP,QACA,MACA,WACiB;CACjB,MAAM,SAA0B,EAAE;CAClC,MAAM,QAAQ,KAAK,MAAM;AACzB,MAAK,IAAI,SAAS,GAAG,MAAM,GAAG,SAAS,MAAM,QAAQ,OAAO,GAAG;EAC7D,MAAM,QAAQ,MAAM,MAAM,QAAQ,SAAS,UAAU;AACrD,SAAO,KAAK;GACV,SAAS,GAAG,OAAO,KAAK,OAAO,IAAI,CAAC,SAAS,GAAG,IAAI;GACpD,MAAM;GACN,MAAM,EAAE,QAAQ,QAAQ;GACzB,CAAC;AACF,YAAU;;AAEZ,QAAO;;;;;;AAOT,SAAgB,mBACd,UAA+B,EAAE,EAChB;CACjB,MAAM,eAAe,QAAQ,gBAAgB;CAC7C,MAAM,iBAAiB,QAAQ,kBAAkB;CACjD,MAAM,YAAY,QAAQ,aAAa;CAEvC,MAAM,SAA0B,EAAE;AAClC,MAAK,MAAM,OAAO,gBAAgB;EAChC,MAAM,UAAU,KAAK,KAAK,cAAc,IAAI;AAC5C,MAAI,CAAC,GAAG,WAAW,QAAQ,CAAE;AAC7B,OAAK,MAAM,YAAY,GAAG,YAAY,QAAQ,EAAE;AAE9C,OADY,KAAK,QAAQ,SAAS,CAAC,aAAa,KACpC,MAAO;GACnB,MAAM,WAAW,KAAK,KAAK,SAAS,SAAS;GAE7C,MAAM,qBAAqB,qBADf,GAAG,aAAa,UAAU,OAAO,CACO;GACpD,MAAM,SAAS,KAAK,MAAM,SAAS,CAAC;GACpC,MAAM,aAAa,gBAAgB,QAAQ,oBAAoB,UAAU;AACzE,UAAO,KAAK,GAAG,WAAW;;;AAG9B,QAAO;;AAGT,eAAsB,8BACpB,UAA0C,EAAE,EAClB;CAC1B,MAAM,aAAa,mBAAmB,QAAQ;AAC9C,KAAI,CAAC,QAAQ,QAAS,QAAO;CAC7B,MAAM,gBAAgB,MAAM,0BAA0B,QAAQ,QAAQ;AACtE,QAAO,CAAC,GAAG,YAAY,GAAG,cAAc"}
@@ -1 +0,0 @@
1
- {"version":3,"file":"posthog-signals.js","names":[],"sources":["../src/posthog-signals.ts"],"sourcesContent":["import type {\n AnalyticsReader,\n AnalyticsQueryResult,\n DateRangeInput,\n} from '@contractspec/lib.contracts/integrations/providers/analytics';\nimport type { EvidenceChunk } from '@contractspec/lib.contracts/product-intent/types';\nimport { FunnelAnalyzer } from '@contractspec/lib.analytics/funnel';\nimport type {\n AnalyticsEvent,\n FunnelDefinition,\n} from '@contractspec/lib.analytics';\nimport { PosthogAnalyticsProvider } from '@contractspec/integration.providers-impls/impls/posthog';\n\nexport interface PosthogEvidenceOptions {\n reader: AnalyticsReader;\n dateRange?: DateRangeInput;\n eventNames?: string[];\n limit?: number;\n funnel?: FunnelDefinition;\n includeFeatureFlags?: boolean;\n}\n\nexport interface PosthogEvidenceEnvOptions {\n defaultLookbackDays?: number;\n defaultLimit?: number;\n}\n\nexport async function loadPosthogEvidenceChunks(\n options: PosthogEvidenceOptions\n): Promise<EvidenceChunk[]> {\n const chunks: EvidenceChunk[] = [];\n const range = resolveRange(options.dateRange);\n\n const eventSummary = await loadEventSummary(options, range);\n if (eventSummary) {\n chunks.push(eventSummary);\n }\n\n const funnelEvidence = await loadFunnelEvidence(options, range);\n if (funnelEvidence) {\n chunks.push(funnelEvidence);\n }\n\n const featureFlags = await loadFeatureFlagEvidence(options);\n if (featureFlags) {\n chunks.push(featureFlags);\n }\n\n return chunks;\n}\n\nasync function loadEventSummary(\n options: PosthogEvidenceOptions,\n range: { from: Date; to: Date }\n): Promise<EvidenceChunk | null> {\n if (!options.reader.queryHogQL) return null;\n const eventFilter = buildEventFilter(options.eventNames);\n const limit = options.limit ?? 10;\n const result = await options.reader.queryHogQL({\n query: [\n 'select',\n ' event as eventName,',\n ' count() as total',\n 'from events',\n 'where timestamp >= {dateFrom} and timestamp < {dateTo}',\n eventFilter.clause ? `and ${eventFilter.clause}` : '',\n 'group by eventName',\n 'order by total desc',\n `limit ${limit}`,\n ]\n .filter(Boolean)\n .join('\\n'),\n values: {\n dateFrom: range.from.toISOString(),\n dateTo: range.to.toISOString(),\n ...eventFilter.values,\n },\n });\n const rows = mapRows(result);\n if (rows.length === 0) return null;\n const lines = rows.map((row) => {\n const name = asString(row.eventName) ?? 'unknown';\n const total = asNumber(row.total);\n return `- ${name}: ${total}`;\n });\n return {\n chunkId: `posthog:event_summary:${range.from.toISOString()}`,\n text: [\n `PostHog event summary (${range.from.toISOString()} → ${range.to.toISOString()}):`,\n ...lines,\n ].join('\\n'),\n meta: {\n source: 'posthog',\n kind: 'event_summary',\n dateFrom: range.from.toISOString(),\n dateTo: range.to.toISOString(),\n },\n };\n}\n\nasync function loadFunnelEvidence(\n options: PosthogEvidenceOptions,\n range: { from: Date; to: Date }\n): Promise<EvidenceChunk | null> {\n if (!options.funnel) return null;\n if (!options.reader.getEvents) return null;\n const events: AnalyticsEvent[] = [];\n const eventNames = options.funnel.steps.map((step) => step.eventName);\n for (const eventName of eventNames) {\n const response = await options.reader.getEvents({\n event: eventName,\n dateRange: {\n from: range.from,\n to: range.to,\n },\n limit: options.limit ?? 500,\n });\n response.results.forEach((event) => {\n events.push({\n name: event.event,\n userId: event.distinctId,\n tenantId:\n typeof event.properties?.tenantId === 'string'\n ? event.properties.tenantId\n : undefined,\n timestamp: event.timestamp,\n properties: event.properties,\n });\n });\n }\n if (events.length === 0) return null;\n const analyzer = new FunnelAnalyzer();\n const analysis = analyzer.analyze(events, options.funnel);\n const lines = analysis.steps.map((step) => {\n return `- ${step.step.eventName}: ${step.count} (conversion ${step.conversionRate}, drop-off ${step.dropOffRate})`;\n });\n return {\n chunkId: `posthog:funnel:${options.funnel.name}`,\n text: [`PostHog funnel analysis — ${options.funnel.name}:`, ...lines].join(\n '\\n'\n ),\n meta: {\n source: 'posthog',\n kind: 'funnel',\n funnelName: options.funnel.name,\n dateFrom: range.from.toISOString(),\n dateTo: range.to.toISOString(),\n },\n };\n}\n\nasync function loadFeatureFlagEvidence(\n options: PosthogEvidenceOptions\n): Promise<EvidenceChunk | null> {\n if (!options.includeFeatureFlags) return null;\n if (!options.reader.getFeatureFlags) return null;\n const response = await options.reader.getFeatureFlags({ limit: 10 });\n if (!response.results.length) return null;\n const lines = response.results.map((flag) => {\n const key = flag.key ?? 'unknown';\n const active = flag.active ? 'active' : 'inactive';\n return `- ${key}: ${active}`;\n });\n return {\n chunkId: 'posthog:feature_flags',\n text: ['PostHog feature flags:', ...lines].join('\\n'),\n meta: {\n source: 'posthog',\n kind: 'feature_flags',\n },\n };\n}\n\nfunction resolveRange(dateRange: DateRangeInput | undefined): {\n from: Date;\n to: Date;\n} {\n const now = new Date();\n const from =\n dateRange?.from instanceof Date\n ? dateRange.from\n : dateRange?.from\n ? new Date(dateRange.from)\n : new Date(now.getTime() - 30 * 24 * 60 * 60 * 1000);\n const to =\n dateRange?.to instanceof Date\n ? dateRange.to\n : dateRange?.to\n ? new Date(dateRange.to)\n : now;\n return { from, to };\n}\n\nfunction buildEventFilter(eventNames?: string[]): {\n clause?: string;\n values?: Record<string, unknown>;\n} {\n if (!eventNames || eventNames.length === 0) return {};\n if (eventNames.length === 1) {\n return {\n clause: 'event = {event0}',\n values: { event0: eventNames[0] },\n };\n }\n const values: Record<string, unknown> = {};\n const clauses = eventNames.map((eventName, index) => {\n values[`event${index}`] = eventName;\n return `event = {event${index}}`;\n });\n return {\n clause: `(${clauses.join(' or ')})`,\n values,\n };\n}\n\nfunction mapRows(result: AnalyticsQueryResult): Record<string, unknown>[] {\n if (!Array.isArray(result.results) || !Array.isArray(result.columns)) {\n return [];\n }\n const columns = result.columns;\n return result.results.flatMap((row) => {\n if (!Array.isArray(row)) return [];\n const record: Record<string, unknown> = {};\n columns.forEach((column, index) => {\n record[column] = row[index];\n });\n return [record];\n });\n}\n\nfunction asString(value: unknown): string | null {\n if (typeof value === 'string' && value.trim()) return value;\n if (typeof value === 'number') return String(value);\n return null;\n}\n\nfunction asNumber(value: unknown): number {\n if (typeof value === 'number' && Number.isFinite(value)) return value;\n if (typeof value === 'string' && value.trim()) {\n const parsed = Number(value);\n if (Number.isFinite(parsed)) return parsed;\n }\n return 0;\n}\n\nexport function resolvePosthogEvidenceOptionsFromEnv(\n options: PosthogEvidenceEnvOptions = {}\n): PosthogEvidenceOptions | null {\n const projectId = process.env.POSTHOG_PROJECT_ID;\n const personalApiKey = process.env.POSTHOG_PERSONAL_API_KEY;\n if (!projectId || !personalApiKey) return null;\n\n const lookbackDays = resolveNumberEnv(\n 'POSTHOG_EVIDENCE_LOOKBACK_DAYS',\n options.defaultLookbackDays ?? 30\n );\n const limit = resolveNumberEnv(\n 'POSTHOG_EVIDENCE_LIMIT',\n options.defaultLimit ?? 10\n );\n\n const now = new Date();\n const from = new Date(now.getTime() - lookbackDays * 24 * 60 * 60 * 1000);\n const eventNames = resolveCsvEnv('POSTHOG_EVIDENCE_EVENTS');\n const funnelSteps = resolveCsvEnv('POSTHOG_EVIDENCE_FUNNEL_STEPS');\n const funnel =\n funnelSteps && funnelSteps.length\n ? {\n name: 'posthog-evidence-funnel',\n steps: funnelSteps.map((eventName, index) => ({\n id: `step_${index + 1}`,\n eventName,\n })),\n }\n : undefined;\n\n const reader = new PosthogAnalyticsProvider({\n host: process.env.POSTHOG_HOST,\n projectId,\n personalApiKey,\n });\n\n return {\n reader,\n dateRange: { from, to: now },\n eventNames,\n limit,\n funnel,\n includeFeatureFlags: resolveBooleanEnv(\n 'POSTHOG_EVIDENCE_FEATURE_FLAGS',\n true\n ),\n };\n}\n\nfunction resolveCsvEnv(key: string): string[] | undefined {\n const value = process.env[key];\n if (!value) return undefined;\n return value\n .split(',')\n .map((item) => item.trim())\n .filter(Boolean);\n}\n\nfunction resolveNumberEnv(key: string, fallback: number): number {\n const value = process.env[key];\n if (!value) return fallback;\n const parsed = Number(value);\n return Number.isFinite(parsed) ? parsed : fallback;\n}\n\nfunction resolveBooleanEnv(key: string, fallback: boolean): boolean {\n const value = process.env[key];\n if (value === undefined) return fallback;\n return value.toLowerCase() === 'true';\n}\n"],"mappings":";;;;AA2BA,eAAsB,0BACpB,SAC0B;CAC1B,MAAM,SAA0B,EAAE;CAClC,MAAM,QAAQ,aAAa,QAAQ,UAAU;CAE7C,MAAM,eAAe,MAAM,iBAAiB,SAAS,MAAM;AAC3D,KAAI,aACF,QAAO,KAAK,aAAa;CAG3B,MAAM,iBAAiB,MAAM,mBAAmB,SAAS,MAAM;AAC/D,KAAI,eACF,QAAO,KAAK,eAAe;CAG7B,MAAM,eAAe,MAAM,wBAAwB,QAAQ;AAC3D,KAAI,aACF,QAAO,KAAK,aAAa;AAG3B,QAAO;;AAGT,eAAe,iBACb,SACA,OAC+B;AAC/B,KAAI,CAAC,QAAQ,OAAO,WAAY,QAAO;CACvC,MAAM,cAAc,iBAAiB,QAAQ,WAAW;CACxD,MAAM,QAAQ,QAAQ,SAAS;CAqB/B,MAAM,OAAO,QApBE,MAAM,QAAQ,OAAO,WAAW;EAC7C,OAAO;GACL;GACA;GACA;GACA;GACA;GACA,YAAY,SAAS,OAAO,YAAY,WAAW;GACnD;GACA;GACA,SAAS;GACV,CACE,OAAO,QAAQ,CACf,KAAK,KAAK;EACb,QAAQ;GACN,UAAU,MAAM,KAAK,aAAa;GAClC,QAAQ,MAAM,GAAG,aAAa;GAC9B,GAAG,YAAY;GAChB;EACF,CAAC,CAC0B;AAC5B,KAAI,KAAK,WAAW,EAAG,QAAO;CAC9B,MAAM,QAAQ,KAAK,KAAK,QAAQ;AAG9B,SAAO,KAFM,SAAS,IAAI,UAAU,IAAI,UAEvB,IADH,SAAS,IAAI,MAAM;GAEjC;AACF,QAAO;EACL,SAAS,yBAAyB,MAAM,KAAK,aAAa;EAC1D,MAAM,CACJ,0BAA0B,MAAM,KAAK,aAAa,CAAC,KAAK,MAAM,GAAG,aAAa,CAAC,KAC/E,GAAG,MACJ,CAAC,KAAK,KAAK;EACZ,MAAM;GACJ,QAAQ;GACR,MAAM;GACN,UAAU,MAAM,KAAK,aAAa;GAClC,QAAQ,MAAM,GAAG,aAAa;GAC/B;EACF;;AAGH,eAAe,mBACb,SACA,OAC+B;AAC/B,KAAI,CAAC,QAAQ,OAAQ,QAAO;AAC5B,KAAI,CAAC,QAAQ,OAAO,UAAW,QAAO;CACtC,MAAM,SAA2B,EAAE;CACnC,MAAM,aAAa,QAAQ,OAAO,MAAM,KAAK,SAAS,KAAK,UAAU;AACrE,MAAK,MAAM,aAAa,WAStB,EARiB,MAAM,QAAQ,OAAO,UAAU;EAC9C,OAAO;EACP,WAAW;GACT,MAAM,MAAM;GACZ,IAAI,MAAM;GACX;EACD,OAAO,QAAQ,SAAS;EACzB,CAAC,EACO,QAAQ,SAAS,UAAU;AAClC,SAAO,KAAK;GACV,MAAM,MAAM;GACZ,QAAQ,MAAM;GACd,UACE,OAAO,MAAM,YAAY,aAAa,WAClC,MAAM,WAAW,WACjB;GACN,WAAW,MAAM;GACjB,YAAY,MAAM;GACnB,CAAC;GACF;AAEJ,KAAI,OAAO,WAAW,EAAG,QAAO;CAGhC,MAAM,QAFW,IAAI,gBAAgB,CACX,QAAQ,QAAQ,QAAQ,OAAO,CAClC,MAAM,KAAK,SAAS;AACzC,SAAO,KAAK,KAAK,KAAK,UAAU,IAAI,KAAK,MAAM,eAAe,KAAK,eAAe,aAAa,KAAK,YAAY;GAChH;AACF,QAAO;EACL,SAAS,kBAAkB,QAAQ,OAAO;EAC1C,MAAM,CAAC,6BAA6B,QAAQ,OAAO,KAAK,IAAI,GAAG,MAAM,CAAC,KACpE,KACD;EACD,MAAM;GACJ,QAAQ;GACR,MAAM;GACN,YAAY,QAAQ,OAAO;GAC3B,UAAU,MAAM,KAAK,aAAa;GAClC,QAAQ,MAAM,GAAG,aAAa;GAC/B;EACF;;AAGH,eAAe,wBACb,SAC+B;AAC/B,KAAI,CAAC,QAAQ,oBAAqB,QAAO;AACzC,KAAI,CAAC,QAAQ,OAAO,gBAAiB,QAAO;CAC5C,MAAM,WAAW,MAAM,QAAQ,OAAO,gBAAgB,EAAE,OAAO,IAAI,CAAC;AACpE,KAAI,CAAC,SAAS,QAAQ,OAAQ,QAAO;AAMrC,QAAO;EACL,SAAS;EACT,MAAM,CAAC,0BAA0B,GAPrB,SAAS,QAAQ,KAAK,SAAS;AAG3C,UAAO,KAFK,KAAK,OAAO,UAER,IADD,KAAK,SAAS,WAAW;IAExC,CAG0C,CAAC,KAAK,KAAK;EACrD,MAAM;GACJ,QAAQ;GACR,MAAM;GACP;EACF;;AAGH,SAAS,aAAa,WAGpB;CACA,MAAM,sBAAM,IAAI,MAAM;AAatB,QAAO;EAAE,MAXP,WAAW,gBAAgB,OACvB,UAAU,OACV,WAAW,OACT,IAAI,KAAK,UAAU,KAAK,mBACxB,IAAI,KAAK,IAAI,SAAS,GAAG,MAAU,KAAK,KAAK,IAAK;EAO3C,IALb,WAAW,cAAc,OACrB,UAAU,KACV,WAAW,KACT,IAAI,KAAK,UAAU,GAAG,GACtB;EACW;;AAGrB,SAAS,iBAAiB,YAGxB;AACA,KAAI,CAAC,cAAc,WAAW,WAAW,EAAG,QAAO,EAAE;AACrD,KAAI,WAAW,WAAW,EACxB,QAAO;EACL,QAAQ;EACR,QAAQ,EAAE,QAAQ,WAAW,IAAI;EAClC;CAEH,MAAM,SAAkC,EAAE;AAK1C,QAAO;EACL,QAAQ,IALM,WAAW,KAAK,WAAW,UAAU;AACnD,UAAO,QAAQ,WAAW;AAC1B,UAAO,iBAAiB,MAAM;IAC9B,CAEoB,KAAK,OAAO,CAAC;EACjC;EACD;;AAGH,SAAS,QAAQ,QAAyD;AACxE,KAAI,CAAC,MAAM,QAAQ,OAAO,QAAQ,IAAI,CAAC,MAAM,QAAQ,OAAO,QAAQ,CAClE,QAAO,EAAE;CAEX,MAAM,UAAU,OAAO;AACvB,QAAO,OAAO,QAAQ,SAAS,QAAQ;AACrC,MAAI,CAAC,MAAM,QAAQ,IAAI,CAAE,QAAO,EAAE;EAClC,MAAM,SAAkC,EAAE;AAC1C,UAAQ,SAAS,QAAQ,UAAU;AACjC,UAAO,UAAU,IAAI;IACrB;AACF,SAAO,CAAC,OAAO;GACf;;AAGJ,SAAS,SAAS,OAA+B;AAC/C,KAAI,OAAO,UAAU,YAAY,MAAM,MAAM,CAAE,QAAO;AACtD,KAAI,OAAO,UAAU,SAAU,QAAO,OAAO,MAAM;AACnD,QAAO;;AAGT,SAAS,SAAS,OAAwB;AACxC,KAAI,OAAO,UAAU,YAAY,OAAO,SAAS,MAAM,CAAE,QAAO;AAChE,KAAI,OAAO,UAAU,YAAY,MAAM,MAAM,EAAE;EAC7C,MAAM,SAAS,OAAO,MAAM;AAC5B,MAAI,OAAO,SAAS,OAAO,CAAE,QAAO;;AAEtC,QAAO;;AAGT,SAAgB,qCACd,UAAqC,EAAE,EACR;CAC/B,MAAM,YAAY,QAAQ,IAAI;CAC9B,MAAM,iBAAiB,QAAQ,IAAI;AACnC,KAAI,CAAC,aAAa,CAAC,eAAgB,QAAO;CAE1C,MAAM,eAAe,iBACnB,kCACA,QAAQ,uBAAuB,GAChC;CACD,MAAM,QAAQ,iBACZ,0BACA,QAAQ,gBAAgB,GACzB;CAED,MAAM,sBAAM,IAAI,MAAM;CACtB,MAAM,uBAAO,IAAI,KAAK,IAAI,SAAS,GAAG,eAAe,KAAK,KAAK,KAAK,IAAK;CACzE,MAAM,aAAa,cAAc,0BAA0B;CAC3D,MAAM,cAAc,cAAc,gCAAgC;CAClE,MAAM,SACJ,eAAe,YAAY,SACvB;EACE,MAAM;EACN,OAAO,YAAY,KAAK,WAAW,WAAW;GAC5C,IAAI,QAAQ,QAAQ;GACpB;GACD,EAAE;EACJ,GACD;AAQN,QAAO;EACL,QAPa,IAAI,yBAAyB;GAC1C,MAAM,QAAQ,IAAI;GAClB;GACA;GACD,CAAC;EAIA,WAAW;GAAE;GAAM,IAAI;GAAK;EAC5B;EACA;EACA;EACA,qBAAqB,kBACnB,kCACA,KACD;EACF;;AAGH,SAAS,cAAc,KAAmC;CACxD,MAAM,QAAQ,QAAQ,IAAI;AAC1B,KAAI,CAAC,MAAO,QAAO;AACnB,QAAO,MACJ,MAAM,IAAI,CACV,KAAK,SAAS,KAAK,MAAM,CAAC,CAC1B,OAAO,QAAQ;;AAGpB,SAAS,iBAAiB,KAAa,UAA0B;CAC/D,MAAM,QAAQ,QAAQ,IAAI;AAC1B,KAAI,CAAC,MAAO,QAAO;CACnB,MAAM,SAAS,OAAO,MAAM;AAC5B,QAAO,OAAO,SAAS,OAAO,GAAG,SAAS;;AAG5C,SAAS,kBAAkB,KAAa,UAA4B;CAClE,MAAM,QAAQ,QAAQ,IAAI;AAC1B,KAAI,UAAU,OAAW,QAAO;AAChC,QAAO,MAAM,aAAa,KAAK"}
@@ -1 +0,0 @@
1
- {"version":3,"file":"script.js","names":[],"sources":["../src/script.ts"],"sourcesContent":["import fs from 'node:fs';\nimport path from 'node:path';\nimport { fileURLToPath } from 'node:url';\nimport { createAgentJsonRunner } from '@contractspec/lib.ai-agent';\nimport {\n extractEvidence,\n generateTickets,\n groupProblems,\n impactEngine,\n type RepoScanFile,\n suggestPatch,\n} from '@contractspec/lib.product-intent-utils';\nimport { loadEvidenceChunksWithSignals } from './load-evidence';\nimport { resolvePosthogEvidenceOptionsFromEnv } from './posthog-signals';\n\nconst QUESTION =\n 'Which activation and onboarding friction should we prioritize next?';\nconst DEFAULT_PROVIDER = 'openai';\nconst DEFAULT_MODEL = 'gpt-5.2';\nconst DEFAULT_TEMPERATURE = 0;\nconst DEFAULT_MAX_ATTEMPTS = 2;\n\nconst MODULE_DIR = path.dirname(fileURLToPath(import.meta.url));\nconst REPO_ROOT = path.resolve(MODULE_DIR, '../../../..');\nconst REPO_SCAN_FILES = [\n 'packages/examples/product-intent/src/load-evidence.ts',\n 'packages/examples/product-intent/src/script.ts',\n 'packages/libs/contracts/src/product-intent/contract-patch-intent.ts',\n 'packages/libs/contracts/src/product-intent/spec.ts',\n 'packages/libs/product-intent-utils/src/impact-engine.ts',\n];\n\nfunction collectRepoFiles(root: string, files: string[]): RepoScanFile[] {\n const collected: RepoScanFile[] = [];\n for (const relativePath of files) {\n const fullPath = path.join(root, relativePath);\n if (!fs.existsSync(fullPath)) continue;\n const content = fs.readFileSync(fullPath, 'utf8');\n collected.push({ path: relativePath, content });\n }\n return collected;\n}\n\ninterface TicketPipelineLogger {\n log: (entry: {\n stage: string;\n phase: string;\n attempt: number;\n prompt: string;\n response?: string;\n error?: string;\n timestamp: string;\n }) => void | Promise<void>;\n}\n\ntype ProviderName = 'openai' | 'anthropic' | 'mistral' | 'gemini' | 'ollama';\n\nfunction resolveProviderName(): ProviderName {\n const raw =\n process.env.CONTRACTSPEC_AI_PROVIDER ??\n process.env.AI_PROVIDER ??\n DEFAULT_PROVIDER;\n const normalized = raw.toLowerCase();\n const allowed: ProviderName[] = [\n 'openai',\n 'anthropic',\n 'mistral',\n 'gemini',\n 'ollama',\n ];\n if (!allowed.includes(normalized as ProviderName)) {\n throw new Error(\n `Unsupported AI provider '${raw}'. Allowed: ${allowed.join(', ')}`\n );\n }\n return normalized as ProviderName;\n}\n\nfunction resolveApiKey(provider: ProviderName): string | undefined {\n switch (provider) {\n case 'openai':\n return process.env.OPENAI_API_KEY;\n case 'anthropic':\n return process.env.ANTHROPIC_API_KEY;\n case 'mistral':\n return process.env.MISTRAL_API_KEY;\n case 'gemini':\n return process.env.GOOGLE_API_KEY ?? process.env.GEMINI_API_KEY;\n case 'ollama':\n return undefined;\n }\n}\n\nfunction resolveTemperature(): number {\n const raw =\n process.env.CONTRACTSPEC_AI_TEMPERATURE ?? process.env.AI_TEMPERATURE;\n if (!raw) return DEFAULT_TEMPERATURE;\n const value = Number.parseFloat(raw);\n return Number.isNaN(value) ? DEFAULT_TEMPERATURE : value;\n}\n\nfunction resolveMaxAttempts(): number {\n const raw =\n process.env.CONTRACTSPEC_AI_MAX_ATTEMPTS ?? process.env.AI_MAX_ATTEMPTS;\n if (!raw) return DEFAULT_MAX_ATTEMPTS;\n const value = Number.parseInt(raw, 10);\n return Number.isNaN(value) ? DEFAULT_MAX_ATTEMPTS : Math.max(1, value);\n}\n\nfunction writeArtifact(dir: string, name: string, contents: string): string {\n const filePath = path.join(dir, name);\n fs.writeFileSync(filePath, contents, 'utf8');\n return filePath;\n}\n\nfunction createPipelineLogger(\n logDir: string,\n runId: string\n): TicketPipelineLogger {\n const tracePath = path.join(logDir, 'trace.jsonl');\n\n return {\n log(entry) {\n const baseName = `${entry.stage}-attempt-${entry.attempt}-${entry.phase}`;\n const payload: Record<string, unknown> = {\n runId,\n stage: entry.stage,\n phase: entry.phase,\n attempt: entry.attempt,\n timestamp: entry.timestamp,\n };\n\n if (entry.prompt) {\n payload.promptPath = path.relative(\n REPO_ROOT,\n writeArtifact(logDir, `${baseName}.prompt.txt`, entry.prompt)\n );\n }\n\n if (entry.response) {\n payload.responsePath = path.relative(\n REPO_ROOT,\n writeArtifact(logDir, `${baseName}.response.txt`, entry.response)\n );\n }\n\n if (entry.error) {\n payload.errorPath = path.relative(\n REPO_ROOT,\n writeArtifact(logDir, `${baseName}.error.txt`, entry.error)\n );\n }\n\n fs.appendFileSync(tracePath, `${JSON.stringify(payload)}\\n`, 'utf8');\n },\n };\n}\n\nasync function main() {\n const provider = resolveProviderName();\n const temperature = resolveTemperature();\n const maxAttempts = resolveMaxAttempts();\n const apiKey = resolveApiKey(provider);\n const proxyUrl = process.env.CONTRACTSPEC_AI_PROXY_URL;\n const organizationId = process.env.CONTRACTSPEC_ORG_ID;\n const baseUrl = process.env.OLLAMA_BASE_URL;\n const model =\n process.env.CONTRACTSPEC_AI_MODEL ??\n process.env.AI_MODEL ??\n (provider === 'mistral' ? DEFAULT_MODEL : undefined);\n\n if (provider !== 'ollama' && !apiKey && !proxyUrl && !organizationId) {\n throw new Error(\n `Missing API credentials for ${provider}. Set the provider API key or CONTRACTSPEC_AI_PROXY_URL.`\n );\n }\n\n const runId = new Date().toISOString().replace(/[:.]/g, '-');\n const logDir = path.join(MODULE_DIR, '../logs', `run-${runId}`);\n fs.mkdirSync(logDir, { recursive: true });\n const logger = createPipelineLogger(logDir, runId);\n\n const modelRunner = await createAgentJsonRunner({\n provider: {\n provider,\n model,\n apiKey,\n baseUrl,\n proxyUrl,\n organizationId,\n },\n temperature,\n system:\n 'You are a product discovery analyst. Respond with strict JSON only and use exact quotes for citations.',\n });\n\n console.log(`AI provider: ${provider}`);\n console.log(`Model: ${model ?? '(provider default)'}`);\n console.log(`Temperature: ${temperature}`);\n console.log(`Max attempts: ${maxAttempts}`);\n console.log(`Trace log: ${path.relative(REPO_ROOT, logDir)}/trace.jsonl`);\n\n const posthogEvidence = resolvePosthogEvidenceOptionsFromEnv();\n const evidenceChunks = await loadEvidenceChunksWithSignals({\n posthog: posthogEvidence ?? undefined,\n });\n console.log(`Loaded ${evidenceChunks.length} evidence chunks`);\n\n const findings = await extractEvidence(evidenceChunks, QUESTION, {\n maxFindings: 12,\n modelRunner,\n logger,\n maxAttempts,\n });\n console.log('\\nEvidence findings:\\n');\n console.log(JSON.stringify(findings, null, 2));\n\n const problems = await groupProblems(findings, QUESTION, {\n modelRunner,\n logger,\n maxAttempts,\n });\n console.log('\\nProblems:\\n');\n console.log(JSON.stringify(problems, null, 2));\n\n const tickets = await generateTickets(problems, findings, QUESTION, {\n modelRunner,\n logger,\n maxAttempts,\n });\n console.log('\\nTickets:\\n');\n console.log(JSON.stringify(tickets, null, 2));\n\n if (!tickets[0]) {\n console.log('\\nNo tickets generated.');\n return;\n }\n\n const patchIntent = await suggestPatch(tickets[0], {\n modelRunner,\n logger,\n maxAttempts,\n });\n console.log('\\nPatch intent:\\n');\n console.log(JSON.stringify(patchIntent, null, 2));\n\n const repoFiles = collectRepoFiles(REPO_ROOT, REPO_SCAN_FILES);\n const impact = impactEngine(patchIntent, {\n repoFiles,\n maxHitsPerChange: 3,\n });\n console.log('\\nImpact report (deterministic):\\n');\n console.log(JSON.stringify(impact, null, 2));\n}\n\nmain().catch((error) => {\n console.error(error);\n process.exitCode = 1;\n});\n"],"mappings":";;;;;;;;;AAeA,MAAM,WACJ;AACF,MAAM,mBAAmB;AACzB,MAAM,gBAAgB;AACtB,MAAM,sBAAsB;AAC5B,MAAM,uBAAuB;AAE7B,MAAM,aAAa,KAAK,QAAQ,cAAc,OAAO,KAAK,IAAI,CAAC;AAC/D,MAAM,YAAY,KAAK,QAAQ,YAAY,cAAc;AACzD,MAAM,kBAAkB;CACtB;CACA;CACA;CACA;CACA;CACD;AAED,SAAS,iBAAiB,MAAc,OAAiC;CACvE,MAAM,YAA4B,EAAE;AACpC,MAAK,MAAM,gBAAgB,OAAO;EAChC,MAAM,WAAW,KAAK,KAAK,MAAM,aAAa;AAC9C,MAAI,CAAC,GAAG,WAAW,SAAS,CAAE;EAC9B,MAAM,UAAU,GAAG,aAAa,UAAU,OAAO;AACjD,YAAU,KAAK;GAAE,MAAM;GAAc;GAAS,CAAC;;AAEjD,QAAO;;AAiBT,SAAS,sBAAoC;CAC3C,MAAM,MACJ,QAAQ,IAAI,4BACZ,QAAQ,IAAI,eACZ;CACF,MAAM,aAAa,IAAI,aAAa;CACpC,MAAM,UAA0B;EAC9B;EACA;EACA;EACA;EACA;EACD;AACD,KAAI,CAAC,QAAQ,SAAS,WAA2B,CAC/C,OAAM,IAAI,MACR,4BAA4B,IAAI,cAAc,QAAQ,KAAK,KAAK,GACjE;AAEH,QAAO;;AAGT,SAAS,cAAc,UAA4C;AACjE,SAAQ,UAAR;EACE,KAAK,SACH,QAAO,QAAQ,IAAI;EACrB,KAAK,YACH,QAAO,QAAQ,IAAI;EACrB,KAAK,UACH,QAAO,QAAQ,IAAI;EACrB,KAAK,SACH,QAAO,QAAQ,IAAI,kBAAkB,QAAQ,IAAI;EACnD,KAAK,SACH;;;AAIN,SAAS,qBAA6B;CACpC,MAAM,MACJ,QAAQ,IAAI,+BAA+B,QAAQ,IAAI;AACzD,KAAI,CAAC,IAAK,QAAO;CACjB,MAAM,QAAQ,OAAO,WAAW,IAAI;AACpC,QAAO,OAAO,MAAM,MAAM,GAAG,sBAAsB;;AAGrD,SAAS,qBAA6B;CACpC,MAAM,MACJ,QAAQ,IAAI,gCAAgC,QAAQ,IAAI;AAC1D,KAAI,CAAC,IAAK,QAAO;CACjB,MAAM,QAAQ,OAAO,SAAS,KAAK,GAAG;AACtC,QAAO,OAAO,MAAM,MAAM,GAAG,uBAAuB,KAAK,IAAI,GAAG,MAAM;;AAGxE,SAAS,cAAc,KAAa,MAAc,UAA0B;CAC1E,MAAM,WAAW,KAAK,KAAK,KAAK,KAAK;AACrC,IAAG,cAAc,UAAU,UAAU,OAAO;AAC5C,QAAO;;AAGT,SAAS,qBACP,QACA,OACsB;CACtB,MAAM,YAAY,KAAK,KAAK,QAAQ,cAAc;AAElD,QAAO,EACL,IAAI,OAAO;EACT,MAAM,WAAW,GAAG,MAAM,MAAM,WAAW,MAAM,QAAQ,GAAG,MAAM;EAClE,MAAM,UAAmC;GACvC;GACA,OAAO,MAAM;GACb,OAAO,MAAM;GACb,SAAS,MAAM;GACf,WAAW,MAAM;GAClB;AAED,MAAI,MAAM,OACR,SAAQ,aAAa,KAAK,SACxB,WACA,cAAc,QAAQ,GAAG,SAAS,cAAc,MAAM,OAAO,CAC9D;AAGH,MAAI,MAAM,SACR,SAAQ,eAAe,KAAK,SAC1B,WACA,cAAc,QAAQ,GAAG,SAAS,gBAAgB,MAAM,SAAS,CAClE;AAGH,MAAI,MAAM,MACR,SAAQ,YAAY,KAAK,SACvB,WACA,cAAc,QAAQ,GAAG,SAAS,aAAa,MAAM,MAAM,CAC5D;AAGH,KAAG,eAAe,WAAW,GAAG,KAAK,UAAU,QAAQ,CAAC,KAAK,OAAO;IAEvE;;AAGH,eAAe,OAAO;CACpB,MAAM,WAAW,qBAAqB;CACtC,MAAM,cAAc,oBAAoB;CACxC,MAAM,cAAc,oBAAoB;CACxC,MAAM,SAAS,cAAc,SAAS;CACtC,MAAM,WAAW,QAAQ,IAAI;CAC7B,MAAM,iBAAiB,QAAQ,IAAI;CACnC,MAAM,UAAU,QAAQ,IAAI;CAC5B,MAAM,QACJ,QAAQ,IAAI,yBACZ,QAAQ,IAAI,aACX,aAAa,YAAY,gBAAgB;AAE5C,KAAI,aAAa,YAAY,CAAC,UAAU,CAAC,YAAY,CAAC,eACpD,OAAM,IAAI,MACR,+BAA+B,SAAS,0DACzC;CAGH,MAAM,yBAAQ,IAAI,MAAM,EAAC,aAAa,CAAC,QAAQ,SAAS,IAAI;CAC5D,MAAM,SAAS,KAAK,KAAK,YAAY,WAAW,OAAO,QAAQ;AAC/D,IAAG,UAAU,QAAQ,EAAE,WAAW,MAAM,CAAC;CACzC,MAAM,SAAS,qBAAqB,QAAQ,MAAM;CAElD,MAAM,cAAc,MAAM,sBAAsB;EAC9C,UAAU;GACR;GACA;GACA;GACA;GACA;GACA;GACD;EACD;EACA,QACE;EACH,CAAC;AAEF,SAAQ,IAAI,gBAAgB,WAAW;AACvC,SAAQ,IAAI,UAAU,SAAS,uBAAuB;AACtD,SAAQ,IAAI,gBAAgB,cAAc;AAC1C,SAAQ,IAAI,iBAAiB,cAAc;AAC3C,SAAQ,IAAI,cAAc,KAAK,SAAS,WAAW,OAAO,CAAC,cAAc;CAGzE,MAAM,iBAAiB,MAAM,8BAA8B,EACzD,SAFsB,sCAAsC,IAEhC,QAC7B,CAAC;AACF,SAAQ,IAAI,UAAU,eAAe,OAAO,kBAAkB;CAE9D,MAAM,WAAW,MAAM,gBAAgB,gBAAgB,UAAU;EAC/D,aAAa;EACb;EACA;EACA;EACD,CAAC;AACF,SAAQ,IAAI,yBAAyB;AACrC,SAAQ,IAAI,KAAK,UAAU,UAAU,MAAM,EAAE,CAAC;CAE9C,MAAM,WAAW,MAAM,cAAc,UAAU,UAAU;EACvD;EACA;EACA;EACD,CAAC;AACF,SAAQ,IAAI,gBAAgB;AAC5B,SAAQ,IAAI,KAAK,UAAU,UAAU,MAAM,EAAE,CAAC;CAE9C,MAAM,UAAU,MAAM,gBAAgB,UAAU,UAAU,UAAU;EAClE;EACA;EACA;EACD,CAAC;AACF,SAAQ,IAAI,eAAe;AAC3B,SAAQ,IAAI,KAAK,UAAU,SAAS,MAAM,EAAE,CAAC;AAE7C,KAAI,CAAC,QAAQ,IAAI;AACf,UAAQ,IAAI,0BAA0B;AACtC;;CAGF,MAAM,cAAc,MAAM,aAAa,QAAQ,IAAI;EACjD;EACA;EACA;EACD,CAAC;AACF,SAAQ,IAAI,oBAAoB;AAChC,SAAQ,IAAI,KAAK,UAAU,aAAa,MAAM,EAAE,CAAC;CAGjD,MAAM,SAAS,aAAa,aAAa;EACvC,WAFgB,iBAAiB,WAAW,gBAAgB;EAG5D,kBAAkB;EACnB,CAAC;AACF,SAAQ,IAAI,qCAAqC;AACjD,SAAQ,IAAI,KAAK,UAAU,QAAQ,MAAM,EAAE,CAAC;;AAG9C,MAAM,CAAC,OAAO,UAAU;AACtB,SAAQ,MAAM,MAAM;AACpB,SAAQ,WAAW;EACnB"}
@@ -1 +0,0 @@
1
- {"version":3,"file":"sync-actions.js","names":[],"sources":["../src/sync-actions.ts"],"sourcesContent":["import { createAgentJsonRunner } from '@contractspec/lib.ai-agent';\nimport {\n buildProjectManagementSyncPayload,\n extractEvidence,\n generateTickets,\n groupProblems,\n impactEngine,\n suggestPatch,\n type TicketPipelineLogger,\n} from '@contractspec/lib.product-intent-utils';\nimport type { TicketPipelineModelRunner } from '@contractspec/lib.product-intent-utils';\nimport { loadEvidenceChunksWithSignals } from './load-evidence';\nimport { resolvePosthogEvidenceOptionsFromEnv } from './posthog-signals';\nimport { LinearProjectManagementProvider } from '@contractspec/integration.providers-impls/impls/linear';\nimport { JiraProjectManagementProvider } from '@contractspec/integration.providers-impls/impls/jira';\nimport { NotionProjectManagementProvider } from '@contractspec/integration.providers-impls/impls/notion';\nimport type { ProjectManagementProvider } from '@contractspec/lib.contracts/integrations/providers/project-management';\n\nconst QUESTION =\n 'Which activation and onboarding friction should we prioritize next?';\n\ntype ProviderName = 'linear' | 'jira' | 'notion';\ntype AiProviderName = 'openai' | 'anthropic' | 'mistral' | 'gemini' | 'ollama';\n\nfunction resolveProvider(): ProviderName {\n const raw = (process.env.CONTRACTSPEC_PM_PROVIDER ?? '').toLowerCase();\n if (raw === 'linear' || raw === 'jira' || raw === 'notion') return raw;\n throw new Error(\n 'Set CONTRACTSPEC_PM_PROVIDER to one of: linear, jira, notion'\n );\n}\n\nasync function resolveModelRunner(): Promise<TicketPipelineModelRunner> {\n const provider = resolveAiProviderName();\n const apiKey = resolveApiKey(provider);\n const proxyUrl = process.env.CONTRACTSPEC_AI_PROXY_URL;\n const organizationId = process.env.CONTRACTSPEC_ORG_ID;\n const baseUrl = process.env.OLLAMA_BASE_URL;\n const model =\n process.env.CONTRACTSPEC_AI_MODEL ?? process.env.AI_MODEL ?? undefined;\n\n if (provider !== 'ollama' && !apiKey && !proxyUrl && !organizationId) {\n throw new Error(\n `Missing API credentials for ${provider}. Set provider API key or CONTRACTSPEC_AI_PROXY_URL.`\n );\n }\n\n return createAgentJsonRunner({\n provider: {\n provider,\n model,\n apiKey,\n baseUrl,\n proxyUrl,\n organizationId,\n },\n temperature: 0,\n system:\n 'You are a product discovery analyst. Respond with strict JSON only and use exact quotes for citations.',\n });\n}\n\nfunction resolveAiProviderName(): AiProviderName {\n const raw =\n process.env.CONTRACTSPEC_AI_PROVIDER ?? process.env.AI_PROVIDER ?? 'openai';\n const normalized = raw.toLowerCase();\n const allowed: AiProviderName[] = [\n 'openai',\n 'anthropic',\n 'mistral',\n 'gemini',\n 'ollama',\n ];\n if (!allowed.includes(normalized as AiProviderName)) {\n throw new Error(\n `Unsupported AI provider: ${raw}. Use one of: ${allowed.join(', ')}`\n );\n }\n return normalized as AiProviderName;\n}\n\nfunction resolveApiKey(provider: AiProviderName): string | undefined {\n switch (provider.toLowerCase()) {\n case 'openai':\n return process.env.OPENAI_API_KEY;\n case 'anthropic':\n return process.env.ANTHROPIC_API_KEY;\n case 'mistral':\n return process.env.MISTRAL_API_KEY;\n case 'gemini':\n return process.env.GOOGLE_API_KEY ?? process.env.GEMINI_API_KEY;\n case 'ollama':\n return undefined;\n default:\n return undefined;\n }\n}\n\nfunction createLogger(): TicketPipelineLogger {\n return {\n log: () => undefined,\n };\n}\n\nasync function run() {\n const provider = resolveProvider();\n const modelRunner = await resolveModelRunner();\n\n const evidenceChunks = await loadEvidenceChunksWithSignals({\n posthog: resolvePosthogEvidenceOptionsFromEnv() ?? undefined,\n });\n const findings = await extractEvidence(evidenceChunks, QUESTION, {\n modelRunner,\n logger: createLogger(),\n maxAttempts: 2,\n });\n const problems = await groupProblems(findings, QUESTION, {\n modelRunner,\n logger: createLogger(),\n maxAttempts: 2,\n });\n const tickets = await generateTickets(problems, findings, QUESTION, {\n modelRunner,\n logger: createLogger(),\n maxAttempts: 2,\n });\n const patchIntent = tickets[0]\n ? await suggestPatch(tickets[0], {\n modelRunner,\n logger: createLogger(),\n maxAttempts: 2,\n })\n : undefined;\n const impact = patchIntent ? impactEngine(patchIntent) : undefined;\n\n const payload = buildProjectManagementSyncPayload({\n question: QUESTION,\n tickets,\n patchIntent,\n impact,\n options: {\n includeSummary: true,\n baseTags: ['product-intent'],\n defaultPriority: 'medium',\n },\n });\n\n if (process.env.CONTRACTSPEC_PM_DRY_RUN === 'true') {\n console.log(JSON.stringify(payload, null, 2));\n return;\n }\n\n const created = await syncToProvider(provider, payload);\n console.log(JSON.stringify(created, null, 2));\n}\n\nasync function syncToProvider(\n provider: ProviderName,\n payload: ReturnType<typeof buildProjectManagementSyncPayload>\n) {\n if (provider === 'linear') {\n const client = new LinearProjectManagementProvider({\n apiKey: requireEnv('LINEAR_API_KEY'),\n teamId: requireEnv('LINEAR_TEAM_ID'),\n projectId: process.env.LINEAR_PROJECT_ID,\n stateId: process.env.LINEAR_STATE_ID,\n assigneeId: process.env.LINEAR_ASSIGNEE_ID,\n labelIds: splitList(process.env.LINEAR_LABEL_IDS),\n });\n return executeSync(client, payload);\n }\n\n if (provider === 'jira') {\n const client = new JiraProjectManagementProvider({\n siteUrl: requireEnv('JIRA_SITE_URL'),\n email: requireEnv('JIRA_EMAIL'),\n apiToken: requireEnv('JIRA_API_TOKEN'),\n projectKey: process.env.JIRA_PROJECT_KEY,\n issueType: process.env.JIRA_ISSUE_TYPE,\n defaultLabels: splitList(process.env.JIRA_DEFAULT_LABELS),\n });\n return executeSync(client, payload);\n }\n\n const client = new NotionProjectManagementProvider({\n apiKey: requireEnv('NOTION_API_KEY'),\n databaseId: process.env.NOTION_DATABASE_ID,\n summaryParentPageId: process.env.NOTION_SUMMARY_PARENT_PAGE_ID,\n titleProperty: process.env.NOTION_TITLE_PROPERTY,\n statusProperty: process.env.NOTION_STATUS_PROPERTY,\n priorityProperty: process.env.NOTION_PRIORITY_PROPERTY,\n tagsProperty: process.env.NOTION_TAGS_PROPERTY,\n dueDateProperty: process.env.NOTION_DUE_DATE_PROPERTY,\n descriptionProperty: process.env.NOTION_DESCRIPTION_PROPERTY,\n });\n return executeSync(client, payload);\n}\n\nasync function executeSync(\n client: ProjectManagementProvider,\n payload: ReturnType<typeof buildProjectManagementSyncPayload>\n) {\n const summary = payload.summary\n ? await client.createWorkItem(payload.summary)\n : undefined;\n const items = await client.createWorkItems(payload.items);\n return { summary, items };\n}\n\nfunction requireEnv(key: string): string {\n const value = process.env[key];\n if (!value) {\n throw new Error(`Missing required env var: ${key}`);\n }\n return value;\n}\n\nfunction splitList(value?: string): string[] | undefined {\n if (!value) return undefined;\n const items = value\n .split(',')\n .map((item) => item.trim())\n .filter(Boolean);\n return items.length > 0 ? items : undefined;\n}\n\nrun().catch((error) => {\n console.error(error);\n process.exitCode = 1;\n});\n"],"mappings":";;;;;;;;;AAkBA,MAAM,WACJ;AAKF,SAAS,kBAAgC;CACvC,MAAM,OAAO,QAAQ,IAAI,4BAA4B,IAAI,aAAa;AACtE,KAAI,QAAQ,YAAY,QAAQ,UAAU,QAAQ,SAAU,QAAO;AACnE,OAAM,IAAI,MACR,+DACD;;AAGH,eAAe,qBAAyD;CACtE,MAAM,WAAW,uBAAuB;CACxC,MAAM,SAAS,cAAc,SAAS;CACtC,MAAM,WAAW,QAAQ,IAAI;CAC7B,MAAM,iBAAiB,QAAQ,IAAI;CACnC,MAAM,UAAU,QAAQ,IAAI;CAC5B,MAAM,QACJ,QAAQ,IAAI,yBAAyB,QAAQ,IAAI,YAAY;AAE/D,KAAI,aAAa,YAAY,CAAC,UAAU,CAAC,YAAY,CAAC,eACpD,OAAM,IAAI,MACR,+BAA+B,SAAS,sDACzC;AAGH,QAAO,sBAAsB;EAC3B,UAAU;GACR;GACA;GACA;GACA;GACA;GACA;GACD;EACD,aAAa;EACb,QACE;EACH,CAAC;;AAGJ,SAAS,wBAAwC;CAC/C,MAAM,MACJ,QAAQ,IAAI,4BAA4B,QAAQ,IAAI,eAAe;CACrE,MAAM,aAAa,IAAI,aAAa;CACpC,MAAM,UAA4B;EAChC;EACA;EACA;EACA;EACA;EACD;AACD,KAAI,CAAC,QAAQ,SAAS,WAA6B,CACjD,OAAM,IAAI,MACR,4BAA4B,IAAI,gBAAgB,QAAQ,KAAK,KAAK,GACnE;AAEH,QAAO;;AAGT,SAAS,cAAc,UAA8C;AACnE,SAAQ,SAAS,aAAa,EAA9B;EACE,KAAK,SACH,QAAO,QAAQ,IAAI;EACrB,KAAK,YACH,QAAO,QAAQ,IAAI;EACrB,KAAK,UACH,QAAO,QAAQ,IAAI;EACrB,KAAK,SACH,QAAO,QAAQ,IAAI,kBAAkB,QAAQ,IAAI;EACnD,KAAK,SACH;EACF,QACE;;;AAIN,SAAS,eAAqC;AAC5C,QAAO,EACL,WAAW,QACZ;;AAGH,eAAe,MAAM;CACnB,MAAM,WAAW,iBAAiB;CAClC,MAAM,cAAc,MAAM,oBAAoB;CAK9C,MAAM,WAAW,MAAM,gBAHA,MAAM,8BAA8B,EACzD,SAAS,sCAAsC,IAAI,QACpD,CAAC,EACqD,UAAU;EAC/D;EACA,QAAQ,cAAc;EACtB,aAAa;EACd,CAAC;CAMF,MAAM,UAAU,MAAM,gBALL,MAAM,cAAc,UAAU,UAAU;EACvD;EACA,QAAQ,cAAc;EACtB,aAAa;EACd,CAAC,EAC8C,UAAU,UAAU;EAClE;EACA,QAAQ,cAAc;EACtB,aAAa;EACd,CAAC;CACF,MAAM,cAAc,QAAQ,KACxB,MAAM,aAAa,QAAQ,IAAI;EAC7B;EACA,QAAQ,cAAc;EACtB,aAAa;EACd,CAAC,GACF;CAGJ,MAAM,UAAU,kCAAkC;EAChD,UAAU;EACV;EACA;EACA,QANa,cAAc,aAAa,YAAY,GAAG;EAOvD,SAAS;GACP,gBAAgB;GAChB,UAAU,CAAC,iBAAiB;GAC5B,iBAAiB;GAClB;EACF,CAAC;AAEF,KAAI,QAAQ,IAAI,4BAA4B,QAAQ;AAClD,UAAQ,IAAI,KAAK,UAAU,SAAS,MAAM,EAAE,CAAC;AAC7C;;CAGF,MAAM,UAAU,MAAM,eAAe,UAAU,QAAQ;AACvD,SAAQ,IAAI,KAAK,UAAU,SAAS,MAAM,EAAE,CAAC;;AAG/C,eAAe,eACb,UACA,SACA;AACA,KAAI,aAAa,SASf,QAAO,YARQ,IAAI,gCAAgC;EACjD,QAAQ,WAAW,iBAAiB;EACpC,QAAQ,WAAW,iBAAiB;EACpC,WAAW,QAAQ,IAAI;EACvB,SAAS,QAAQ,IAAI;EACrB,YAAY,QAAQ,IAAI;EACxB,UAAU,UAAU,QAAQ,IAAI,iBAAiB;EAClD,CAAC,EACyB,QAAQ;AAGrC,KAAI,aAAa,OASf,QAAO,YARQ,IAAI,8BAA8B;EAC/C,SAAS,WAAW,gBAAgB;EACpC,OAAO,WAAW,aAAa;EAC/B,UAAU,WAAW,iBAAiB;EACtC,YAAY,QAAQ,IAAI;EACxB,WAAW,QAAQ,IAAI;EACvB,eAAe,UAAU,QAAQ,IAAI,oBAAoB;EAC1D,CAAC,EACyB,QAAQ;AAcrC,QAAO,YAXQ,IAAI,gCAAgC;EACjD,QAAQ,WAAW,iBAAiB;EACpC,YAAY,QAAQ,IAAI;EACxB,qBAAqB,QAAQ,IAAI;EACjC,eAAe,QAAQ,IAAI;EAC3B,gBAAgB,QAAQ,IAAI;EAC5B,kBAAkB,QAAQ,IAAI;EAC9B,cAAc,QAAQ,IAAI;EAC1B,iBAAiB,QAAQ,IAAI;EAC7B,qBAAqB,QAAQ,IAAI;EAClC,CAAC,EACyB,QAAQ;;AAGrC,eAAe,YACb,QACA,SACA;AAKA,QAAO;EAAE,SAJO,QAAQ,UACpB,MAAM,OAAO,eAAe,QAAQ,QAAQ,GAC5C;EAEc,OADJ,MAAM,OAAO,gBAAgB,QAAQ,MAAM;EAChC;;AAG3B,SAAS,WAAW,KAAqB;CACvC,MAAM,QAAQ,QAAQ,IAAI;AAC1B,KAAI,CAAC,MACH,OAAM,IAAI,MAAM,6BAA6B,MAAM;AAErD,QAAO;;AAGT,SAAS,UAAU,OAAsC;AACvD,KAAI,CAAC,MAAO,QAAO;CACnB,MAAM,QAAQ,MACX,MAAM,IAAI,CACV,KAAK,SAAS,KAAK,MAAM,CAAC,CAC1B,OAAO,QAAQ;AAClB,QAAO,MAAM,SAAS,IAAI,QAAQ;;AAGpC,KAAK,CAAC,OAAO,UAAU;AACrB,SAAQ,MAAM,MAAM;AACpB,SAAQ,WAAW;EACnB"}