@contractspec/example.product-intent 1.57.0 → 1.58.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (42) hide show
  1. package/.turbo/turbo-build.log +21 -96
  2. package/.turbo/turbo-prebuild.log +1 -0
  3. package/CHANGELOG.md +15 -0
  4. package/dist/example.d.ts +2 -6
  5. package/dist/example.d.ts.map +1 -1
  6. package/dist/example.js +27 -37
  7. package/dist/index.d.ts +6 -4
  8. package/dist/index.d.ts.map +1 -0
  9. package/dist/index.js +342 -4
  10. package/dist/load-evidence.d.ts +13 -17
  11. package/dist/load-evidence.d.ts.map +1 -1
  12. package/dist/load-evidence.js +307 -68
  13. package/dist/load-evidence.test.d.ts +2 -0
  14. package/dist/load-evidence.test.d.ts.map +1 -0
  15. package/dist/node/example.js +28 -0
  16. package/dist/node/index.js +342 -0
  17. package/dist/node/load-evidence.js +313 -0
  18. package/dist/node/posthog-signals.js +248 -0
  19. package/dist/node/script.js +512 -0
  20. package/dist/node/sync-actions.js +491 -0
  21. package/dist/posthog-signals.d.ts +15 -19
  22. package/dist/posthog-signals.d.ts.map +1 -1
  23. package/dist/posthog-signals.js +222 -178
  24. package/dist/script.d.ts +2 -1
  25. package/dist/script.d.ts.map +1 -0
  26. package/dist/script.js +493 -152
  27. package/dist/sync-actions.d.ts +2 -1
  28. package/dist/sync-actions.d.ts.map +1 -0
  29. package/dist/sync-actions.js +466 -128
  30. package/package.json +57 -27
  31. package/tsdown.config.js +1 -2
  32. package/.turbo/turbo-build$colon$bundle.log +0 -99
  33. package/dist/example.js.map +0 -1
  34. package/dist/libs/analytics/dist/funnel/analyzer.js +0 -64
  35. package/dist/libs/analytics/dist/funnel/analyzer.js.map +0 -1
  36. package/dist/libs/analytics/dist/types.d.ts +0 -22
  37. package/dist/libs/analytics/dist/types.d.ts.map +0 -1
  38. package/dist/load-evidence.js.map +0 -1
  39. package/dist/posthog-signals.js.map +0 -1
  40. package/dist/script.js.map +0 -1
  41. package/dist/sync-actions.js.map +0 -1
  42. package/tsconfig.tsbuildinfo +0 -1
@@ -1,154 +1,492 @@
1
- import { resolvePosthogEvidenceOptionsFromEnv } from "./posthog-signals.js";
2
- import { loadEvidenceChunksWithSignals } from "./load-evidence.js";
1
+ // @bun
2
+ // src/posthog-signals.ts
3
+ import { FunnelAnalyzer } from "@contractspec/lib.analytics/funnel";
4
+ import { PosthogAnalyticsProvider } from "@contractspec/integration.providers-impls/impls/posthog";
5
+ async function loadPosthogEvidenceChunks(options) {
6
+ const chunks = [];
7
+ const range = resolveRange(options.dateRange);
8
+ const eventSummary = await loadEventSummary(options, range);
9
+ if (eventSummary) {
10
+ chunks.push(eventSummary);
11
+ }
12
+ const funnelEvidence = await loadFunnelEvidence(options, range);
13
+ if (funnelEvidence) {
14
+ chunks.push(funnelEvidence);
15
+ }
16
+ const featureFlags = await loadFeatureFlagEvidence(options);
17
+ if (featureFlags) {
18
+ chunks.push(featureFlags);
19
+ }
20
+ return chunks;
21
+ }
22
+ async function loadEventSummary(options, range) {
23
+ if (!options.reader.queryHogQL)
24
+ return null;
25
+ const eventFilter = buildEventFilter(options.eventNames);
26
+ const limit = options.limit ?? 10;
27
+ const result = await options.reader.queryHogQL({
28
+ query: [
29
+ "select",
30
+ " event as eventName,",
31
+ " count() as total",
32
+ "from events",
33
+ "where timestamp >= {dateFrom} and timestamp < {dateTo}",
34
+ eventFilter.clause ? `and ${eventFilter.clause}` : "",
35
+ "group by eventName",
36
+ "order by total desc",
37
+ `limit ${limit}`
38
+ ].filter(Boolean).join(`
39
+ `),
40
+ values: {
41
+ dateFrom: range.from.toISOString(),
42
+ dateTo: range.to.toISOString(),
43
+ ...eventFilter.values
44
+ }
45
+ });
46
+ const rows = mapRows(result);
47
+ if (rows.length === 0)
48
+ return null;
49
+ const lines = rows.map((row) => {
50
+ const name = asString(row.eventName) ?? "unknown";
51
+ const total = asNumber(row.total);
52
+ return `- ${name}: ${total}`;
53
+ });
54
+ return {
55
+ chunkId: `posthog:event_summary:${range.from.toISOString()}`,
56
+ text: [
57
+ `PostHog event summary (${range.from.toISOString()} \u2192 ${range.to.toISOString()}):`,
58
+ ...lines
59
+ ].join(`
60
+ `),
61
+ meta: {
62
+ source: "posthog",
63
+ kind: "event_summary",
64
+ dateFrom: range.from.toISOString(),
65
+ dateTo: range.to.toISOString()
66
+ }
67
+ };
68
+ }
69
+ async function loadFunnelEvidence(options, range) {
70
+ if (!options.funnel)
71
+ return null;
72
+ if (!options.reader.getEvents)
73
+ return null;
74
+ const events = [];
75
+ const eventNames = options.funnel.steps.map((step) => step.eventName);
76
+ for (const eventName of eventNames) {
77
+ const response = await options.reader.getEvents({
78
+ event: eventName,
79
+ dateRange: {
80
+ from: range.from,
81
+ to: range.to
82
+ },
83
+ limit: options.limit ?? 500
84
+ });
85
+ response.results.forEach((event) => {
86
+ events.push({
87
+ name: event.event,
88
+ userId: event.distinctId,
89
+ tenantId: typeof event.properties?.tenantId === "string" ? event.properties.tenantId : undefined,
90
+ timestamp: event.timestamp,
91
+ properties: event.properties
92
+ });
93
+ });
94
+ }
95
+ if (events.length === 0)
96
+ return null;
97
+ const analyzer = new FunnelAnalyzer;
98
+ const analysis = analyzer.analyze(events, options.funnel);
99
+ const lines = analysis.steps.map((step) => {
100
+ return `- ${step.step.eventName}: ${step.count} (conversion ${step.conversionRate}, drop-off ${step.dropOffRate})`;
101
+ });
102
+ return {
103
+ chunkId: `posthog:funnel:${options.funnel.name}`,
104
+ text: [`PostHog funnel analysis \u2014 ${options.funnel.name}:`, ...lines].join(`
105
+ `),
106
+ meta: {
107
+ source: "posthog",
108
+ kind: "funnel",
109
+ funnelName: options.funnel.name,
110
+ dateFrom: range.from.toISOString(),
111
+ dateTo: range.to.toISOString()
112
+ }
113
+ };
114
+ }
115
+ async function loadFeatureFlagEvidence(options) {
116
+ if (!options.includeFeatureFlags)
117
+ return null;
118
+ if (!options.reader.getFeatureFlags)
119
+ return null;
120
+ const response = await options.reader.getFeatureFlags({ limit: 10 });
121
+ if (!response.results.length)
122
+ return null;
123
+ const lines = response.results.map((flag) => {
124
+ const key = flag.key ?? "unknown";
125
+ const active = flag.active ? "active" : "inactive";
126
+ return `- ${key}: ${active}`;
127
+ });
128
+ return {
129
+ chunkId: "posthog:feature_flags",
130
+ text: ["PostHog feature flags:", ...lines].join(`
131
+ `),
132
+ meta: {
133
+ source: "posthog",
134
+ kind: "feature_flags"
135
+ }
136
+ };
137
+ }
138
+ function resolveRange(dateRange) {
139
+ const now = new Date;
140
+ const from = dateRange?.from instanceof Date ? dateRange.from : dateRange?.from ? new Date(dateRange.from) : new Date(now.getTime() - 30 * 24 * 60 * 60 * 1000);
141
+ const to = dateRange?.to instanceof Date ? dateRange.to : dateRange?.to ? new Date(dateRange.to) : now;
142
+ return { from, to };
143
+ }
144
+ function buildEventFilter(eventNames) {
145
+ if (!eventNames || eventNames.length === 0)
146
+ return {};
147
+ if (eventNames.length === 1) {
148
+ return {
149
+ clause: "event = {event0}",
150
+ values: { event0: eventNames[0] }
151
+ };
152
+ }
153
+ const values = {};
154
+ const clauses = eventNames.map((eventName, index) => {
155
+ values[`event${index}`] = eventName;
156
+ return `event = {event${index}}`;
157
+ });
158
+ return {
159
+ clause: `(${clauses.join(" or ")})`,
160
+ values
161
+ };
162
+ }
163
+ function mapRows(result) {
164
+ if (!Array.isArray(result.results) || !Array.isArray(result.columns)) {
165
+ return [];
166
+ }
167
+ const columns = result.columns;
168
+ return result.results.flatMap((row) => {
169
+ if (!Array.isArray(row))
170
+ return [];
171
+ const record = {};
172
+ columns.forEach((column, index) => {
173
+ record[column] = row[index];
174
+ });
175
+ return [record];
176
+ });
177
+ }
178
+ function asString(value) {
179
+ if (typeof value === "string" && value.trim())
180
+ return value;
181
+ if (typeof value === "number")
182
+ return String(value);
183
+ return null;
184
+ }
185
+ function asNumber(value) {
186
+ if (typeof value === "number" && Number.isFinite(value))
187
+ return value;
188
+ if (typeof value === "string" && value.trim()) {
189
+ const parsed = Number(value);
190
+ if (Number.isFinite(parsed))
191
+ return parsed;
192
+ }
193
+ return 0;
194
+ }
195
+ function resolvePosthogEvidenceOptionsFromEnv(options = {}) {
196
+ const projectId = process.env.POSTHOG_PROJECT_ID;
197
+ const personalApiKey = process.env.POSTHOG_PERSONAL_API_KEY;
198
+ if (!projectId || !personalApiKey)
199
+ return null;
200
+ const lookbackDays = resolveNumberEnv("POSTHOG_EVIDENCE_LOOKBACK_DAYS", options.defaultLookbackDays ?? 30);
201
+ const limit = resolveNumberEnv("POSTHOG_EVIDENCE_LIMIT", options.defaultLimit ?? 10);
202
+ const now = new Date;
203
+ const from = new Date(now.getTime() - lookbackDays * 24 * 60 * 60 * 1000);
204
+ const eventNames = resolveCsvEnv("POSTHOG_EVIDENCE_EVENTS");
205
+ const funnelSteps = resolveCsvEnv("POSTHOG_EVIDENCE_FUNNEL_STEPS");
206
+ const funnel = funnelSteps && funnelSteps.length ? {
207
+ name: "posthog-evidence-funnel",
208
+ steps: funnelSteps.map((eventName, index) => ({
209
+ id: `step_${index + 1}`,
210
+ eventName
211
+ }))
212
+ } : undefined;
213
+ const reader = new PosthogAnalyticsProvider({
214
+ host: process.env.POSTHOG_HOST,
215
+ projectId,
216
+ personalApiKey
217
+ });
218
+ return {
219
+ reader,
220
+ dateRange: { from, to: now },
221
+ eventNames,
222
+ limit,
223
+ funnel,
224
+ includeFeatureFlags: resolveBooleanEnv("POSTHOG_EVIDENCE_FEATURE_FLAGS", true)
225
+ };
226
+ }
227
+ function resolveCsvEnv(key) {
228
+ const value = process.env[key];
229
+ if (!value)
230
+ return;
231
+ return value.split(",").map((item) => item.trim()).filter(Boolean);
232
+ }
233
+ function resolveNumberEnv(key, fallback) {
234
+ const value = process.env[key];
235
+ if (!value)
236
+ return fallback;
237
+ const parsed = Number(value);
238
+ return Number.isFinite(parsed) ? parsed : fallback;
239
+ }
240
+ function resolveBooleanEnv(key, fallback) {
241
+ const value = process.env[key];
242
+ if (value === undefined)
243
+ return fallback;
244
+ return value.toLowerCase() === "true";
245
+ }
246
+
247
+ // src/load-evidence.ts
248
+ import fs from "fs";
249
+ import path from "path";
250
+ import { fileURLToPath } from "url";
251
+ var MODULE_DIR = path.dirname(fileURLToPath(import.meta.url));
252
+ var DEFAULT_EVIDENCE_ROOT = path.join(MODULE_DIR, "../evidence");
253
+ var DEFAULT_TRANSCRIPT_DIRS = ["interviews", "tickets", "public"];
254
+ var DEFAULT_CHUNK_SIZE = 800;
255
+ function stripYamlFrontMatter(contents) {
256
+ const start = contents.indexOf("---");
257
+ if (start === -1)
258
+ return contents;
259
+ const end = contents.indexOf("---", start + 3);
260
+ if (end === -1)
261
+ return contents;
262
+ return contents.slice(end + 3).trimStart();
263
+ }
264
+ function chunkTranscript(fileId, text, chunkSize) {
265
+ const chunks = [];
266
+ const clean = text.trim();
267
+ for (let offset = 0, idx = 0;offset < clean.length; idx += 1) {
268
+ const slice = clean.slice(offset, offset + chunkSize);
269
+ chunks.push({
270
+ chunkId: `${fileId}#c_${String(idx).padStart(2, "0")}`,
271
+ text: slice,
272
+ meta: { source: fileId }
273
+ });
274
+ offset += chunkSize;
275
+ }
276
+ return chunks;
277
+ }
278
+ function loadEvidenceChunks(options = {}) {
279
+ const evidenceRoot = options.evidenceRoot ?? DEFAULT_EVIDENCE_ROOT;
280
+ const transcriptDirs = options.transcriptDirs ?? DEFAULT_TRANSCRIPT_DIRS;
281
+ const chunkSize = options.chunkSize ?? DEFAULT_CHUNK_SIZE;
282
+ const chunks = [];
283
+ for (const dir of transcriptDirs) {
284
+ const fullDir = path.join(evidenceRoot, dir);
285
+ if (!fs.existsSync(fullDir))
286
+ continue;
287
+ for (const fileName of fs.readdirSync(fullDir)) {
288
+ const ext = path.extname(fileName).toLowerCase();
289
+ if (ext !== ".md")
290
+ continue;
291
+ const filePath = path.join(fullDir, fileName);
292
+ const raw = fs.readFileSync(filePath, "utf8");
293
+ const withoutFrontMatter = stripYamlFrontMatter(raw);
294
+ const baseId = path.parse(fileName).name;
295
+ const fileChunks = chunkTranscript(baseId, withoutFrontMatter, chunkSize);
296
+ chunks.push(...fileChunks);
297
+ }
298
+ }
299
+ return chunks;
300
+ }
301
+ async function loadEvidenceChunksWithSignals(options = {}) {
302
+ const baseChunks = loadEvidenceChunks(options);
303
+ if (!options.posthog)
304
+ return baseChunks;
305
+ const posthogChunks = await loadPosthogEvidenceChunks(options.posthog);
306
+ return [...baseChunks, ...posthogChunks];
307
+ }
308
+
309
+ // src/sync-actions.ts
3
310
  import { createAgentJsonRunner } from "@contractspec/lib.ai-agent";
4
- import { buildProjectManagementSyncPayload, extractEvidence, generateTickets, groupProblems, impactEngine, suggestPatch } from "@contractspec/lib.product-intent-utils";
311
+ import {
312
+ buildProjectManagementSyncPayload,
313
+ extractEvidence,
314
+ generateTickets,
315
+ groupProblems,
316
+ impactEngine,
317
+ suggestPatch
318
+ } from "@contractspec/lib.product-intent-utils";
5
319
  import { LinearProjectManagementProvider } from "@contractspec/integration.providers-impls/impls/linear";
6
320
  import { JiraProjectManagementProvider } from "@contractspec/integration.providers-impls/impls/jira";
7
321
  import { NotionProjectManagementProvider } from "@contractspec/integration.providers-impls/impls/notion";
8
-
9
- //#region src/sync-actions.ts
10
- const QUESTION = "Which activation and onboarding friction should we prioritize next?";
322
+ var QUESTION = "Which activation and onboarding friction should we prioritize next?";
11
323
  function resolveProvider() {
12
- const raw = (process.env.CONTRACTSPEC_PM_PROVIDER ?? "").toLowerCase();
13
- if (raw === "linear" || raw === "jira" || raw === "notion") return raw;
14
- throw new Error("Set CONTRACTSPEC_PM_PROVIDER to one of: linear, jira, notion");
324
+ const raw = (process.env.CONTRACTSPEC_PM_PROVIDER ?? "").toLowerCase();
325
+ if (raw === "linear" || raw === "jira" || raw === "notion")
326
+ return raw;
327
+ throw new Error("Set CONTRACTSPEC_PM_PROVIDER to one of: linear, jira, notion");
15
328
  }
16
329
  async function resolveModelRunner() {
17
- const provider = resolveAiProviderName();
18
- const apiKey = resolveApiKey(provider);
19
- const proxyUrl = process.env.CONTRACTSPEC_AI_PROXY_URL;
20
- const organizationId = process.env.CONTRACTSPEC_ORG_ID;
21
- const baseUrl = process.env.OLLAMA_BASE_URL;
22
- const model = process.env.CONTRACTSPEC_AI_MODEL ?? process.env.AI_MODEL ?? void 0;
23
- if (provider !== "ollama" && !apiKey && !proxyUrl && !organizationId) throw new Error(`Missing API credentials for ${provider}. Set provider API key or CONTRACTSPEC_AI_PROXY_URL.`);
24
- return createAgentJsonRunner({
25
- provider: {
26
- provider,
27
- model,
28
- apiKey,
29
- baseUrl,
30
- proxyUrl,
31
- organizationId
32
- },
33
- temperature: 0,
34
- system: "You are a product discovery analyst. Respond with strict JSON only and use exact quotes for citations."
35
- });
330
+ const provider = resolveAiProviderName();
331
+ const apiKey = resolveApiKey(provider);
332
+ const proxyUrl = process.env.CONTRACTSPEC_AI_PROXY_URL;
333
+ const organizationId = process.env.CONTRACTSPEC_ORG_ID;
334
+ const baseUrl = process.env.OLLAMA_BASE_URL;
335
+ const model = process.env.CONTRACTSPEC_AI_MODEL ?? process.env.AI_MODEL ?? undefined;
336
+ if (provider !== "ollama" && !apiKey && !proxyUrl && !organizationId) {
337
+ throw new Error(`Missing API credentials for ${provider}. Set provider API key or CONTRACTSPEC_AI_PROXY_URL.`);
338
+ }
339
+ return createAgentJsonRunner({
340
+ provider: {
341
+ provider,
342
+ model,
343
+ apiKey,
344
+ baseUrl,
345
+ proxyUrl,
346
+ organizationId
347
+ },
348
+ temperature: 0,
349
+ system: "You are a product discovery analyst. Respond with strict JSON only and use exact quotes for citations."
350
+ });
36
351
  }
37
352
  function resolveAiProviderName() {
38
- const raw = process.env.CONTRACTSPEC_AI_PROVIDER ?? process.env.AI_PROVIDER ?? "openai";
39
- const normalized = raw.toLowerCase();
40
- const allowed = [
41
- "openai",
42
- "anthropic",
43
- "mistral",
44
- "gemini",
45
- "ollama"
46
- ];
47
- if (!allowed.includes(normalized)) throw new Error(`Unsupported AI provider: ${raw}. Use one of: ${allowed.join(", ")}`);
48
- return normalized;
353
+ const raw = process.env.CONTRACTSPEC_AI_PROVIDER ?? process.env.AI_PROVIDER ?? "openai";
354
+ const normalized = raw.toLowerCase();
355
+ const allowed = [
356
+ "openai",
357
+ "anthropic",
358
+ "mistral",
359
+ "gemini",
360
+ "ollama"
361
+ ];
362
+ if (!allowed.includes(normalized)) {
363
+ throw new Error(`Unsupported AI provider: ${raw}. Use one of: ${allowed.join(", ")}`);
364
+ }
365
+ return normalized;
49
366
  }
50
367
  function resolveApiKey(provider) {
51
- switch (provider.toLowerCase()) {
52
- case "openai": return process.env.OPENAI_API_KEY;
53
- case "anthropic": return process.env.ANTHROPIC_API_KEY;
54
- case "mistral": return process.env.MISTRAL_API_KEY;
55
- case "gemini": return process.env.GOOGLE_API_KEY ?? process.env.GEMINI_API_KEY;
56
- case "ollama": return;
57
- default: return;
58
- }
368
+ switch (provider.toLowerCase()) {
369
+ case "openai":
370
+ return process.env.OPENAI_API_KEY;
371
+ case "anthropic":
372
+ return process.env.ANTHROPIC_API_KEY;
373
+ case "mistral":
374
+ return process.env.MISTRAL_API_KEY;
375
+ case "gemini":
376
+ return process.env.GOOGLE_API_KEY ?? process.env.GEMINI_API_KEY;
377
+ case "ollama":
378
+ return;
379
+ default:
380
+ return;
381
+ }
59
382
  }
60
383
  function createLogger() {
61
- return { log: () => void 0 };
384
+ return {
385
+ log: () => {
386
+ return;
387
+ }
388
+ };
62
389
  }
63
390
  async function run() {
64
- const provider = resolveProvider();
65
- const modelRunner = await resolveModelRunner();
66
- const findings = await extractEvidence(await loadEvidenceChunksWithSignals({ posthog: resolvePosthogEvidenceOptionsFromEnv() ?? void 0 }), QUESTION, {
67
- modelRunner,
68
- logger: createLogger(),
69
- maxAttempts: 2
70
- });
71
- const tickets = await generateTickets(await groupProblems(findings, QUESTION, {
72
- modelRunner,
73
- logger: createLogger(),
74
- maxAttempts: 2
75
- }), findings, QUESTION, {
76
- modelRunner,
77
- logger: createLogger(),
78
- maxAttempts: 2
79
- });
80
- const patchIntent = tickets[0] ? await suggestPatch(tickets[0], {
81
- modelRunner,
82
- logger: createLogger(),
83
- maxAttempts: 2
84
- }) : void 0;
85
- const payload = buildProjectManagementSyncPayload({
86
- question: QUESTION,
87
- tickets,
88
- patchIntent,
89
- impact: patchIntent ? impactEngine(patchIntent) : void 0,
90
- options: {
91
- includeSummary: true,
92
- baseTags: ["product-intent"],
93
- defaultPriority: "medium"
94
- }
95
- });
96
- if (process.env.CONTRACTSPEC_PM_DRY_RUN === "true") {
97
- console.log(JSON.stringify(payload, null, 2));
98
- return;
99
- }
100
- const created = await syncToProvider(provider, payload);
101
- console.log(JSON.stringify(created, null, 2));
391
+ const provider = resolveProvider();
392
+ const modelRunner = await resolveModelRunner();
393
+ const evidenceChunks = await loadEvidenceChunksWithSignals({
394
+ posthog: resolvePosthogEvidenceOptionsFromEnv() ?? undefined
395
+ });
396
+ const findings = await extractEvidence(evidenceChunks, QUESTION, {
397
+ modelRunner,
398
+ logger: createLogger(),
399
+ maxAttempts: 2
400
+ });
401
+ const problems = await groupProblems(findings, QUESTION, {
402
+ modelRunner,
403
+ logger: createLogger(),
404
+ maxAttempts: 2
405
+ });
406
+ const tickets = await generateTickets(problems, findings, QUESTION, {
407
+ modelRunner,
408
+ logger: createLogger(),
409
+ maxAttempts: 2
410
+ });
411
+ const patchIntent = tickets[0] ? await suggestPatch(tickets[0], {
412
+ modelRunner,
413
+ logger: createLogger(),
414
+ maxAttempts: 2
415
+ }) : undefined;
416
+ const impact = patchIntent ? impactEngine(patchIntent) : undefined;
417
+ const payload = buildProjectManagementSyncPayload({
418
+ question: QUESTION,
419
+ tickets,
420
+ patchIntent,
421
+ impact,
422
+ options: {
423
+ includeSummary: true,
424
+ baseTags: ["product-intent"],
425
+ defaultPriority: "medium"
426
+ }
427
+ });
428
+ if (process.env.CONTRACTSPEC_PM_DRY_RUN === "true") {
429
+ console.log(JSON.stringify(payload, null, 2));
430
+ return;
431
+ }
432
+ const created = await syncToProvider(provider, payload);
433
+ console.log(JSON.stringify(created, null, 2));
102
434
  }
103
435
  async function syncToProvider(provider, payload) {
104
- if (provider === "linear") return executeSync(new LinearProjectManagementProvider({
105
- apiKey: requireEnv("LINEAR_API_KEY"),
106
- teamId: requireEnv("LINEAR_TEAM_ID"),
107
- projectId: process.env.LINEAR_PROJECT_ID,
108
- stateId: process.env.LINEAR_STATE_ID,
109
- assigneeId: process.env.LINEAR_ASSIGNEE_ID,
110
- labelIds: splitList(process.env.LINEAR_LABEL_IDS)
111
- }), payload);
112
- if (provider === "jira") return executeSync(new JiraProjectManagementProvider({
113
- siteUrl: requireEnv("JIRA_SITE_URL"),
114
- email: requireEnv("JIRA_EMAIL"),
115
- apiToken: requireEnv("JIRA_API_TOKEN"),
116
- projectKey: process.env.JIRA_PROJECT_KEY,
117
- issueType: process.env.JIRA_ISSUE_TYPE,
118
- defaultLabels: splitList(process.env.JIRA_DEFAULT_LABELS)
119
- }), payload);
120
- return executeSync(new NotionProjectManagementProvider({
121
- apiKey: requireEnv("NOTION_API_KEY"),
122
- databaseId: process.env.NOTION_DATABASE_ID,
123
- summaryParentPageId: process.env.NOTION_SUMMARY_PARENT_PAGE_ID,
124
- titleProperty: process.env.NOTION_TITLE_PROPERTY,
125
- statusProperty: process.env.NOTION_STATUS_PROPERTY,
126
- priorityProperty: process.env.NOTION_PRIORITY_PROPERTY,
127
- tagsProperty: process.env.NOTION_TAGS_PROPERTY,
128
- dueDateProperty: process.env.NOTION_DUE_DATE_PROPERTY,
129
- descriptionProperty: process.env.NOTION_DESCRIPTION_PROPERTY
130
- }), payload);
436
+ if (provider === "linear") {
437
+ const client2 = new LinearProjectManagementProvider({
438
+ apiKey: requireEnv("LINEAR_API_KEY"),
439
+ teamId: requireEnv("LINEAR_TEAM_ID"),
440
+ projectId: process.env.LINEAR_PROJECT_ID,
441
+ stateId: process.env.LINEAR_STATE_ID,
442
+ assigneeId: process.env.LINEAR_ASSIGNEE_ID,
443
+ labelIds: splitList(process.env.LINEAR_LABEL_IDS)
444
+ });
445
+ return executeSync(client2, payload);
446
+ }
447
+ if (provider === "jira") {
448
+ const client2 = new JiraProjectManagementProvider({
449
+ siteUrl: requireEnv("JIRA_SITE_URL"),
450
+ email: requireEnv("JIRA_EMAIL"),
451
+ apiToken: requireEnv("JIRA_API_TOKEN"),
452
+ projectKey: process.env.JIRA_PROJECT_KEY,
453
+ issueType: process.env.JIRA_ISSUE_TYPE,
454
+ defaultLabels: splitList(process.env.JIRA_DEFAULT_LABELS)
455
+ });
456
+ return executeSync(client2, payload);
457
+ }
458
+ const client = new NotionProjectManagementProvider({
459
+ apiKey: requireEnv("NOTION_API_KEY"),
460
+ databaseId: process.env.NOTION_DATABASE_ID,
461
+ summaryParentPageId: process.env.NOTION_SUMMARY_PARENT_PAGE_ID,
462
+ titleProperty: process.env.NOTION_TITLE_PROPERTY,
463
+ statusProperty: process.env.NOTION_STATUS_PROPERTY,
464
+ priorityProperty: process.env.NOTION_PRIORITY_PROPERTY,
465
+ tagsProperty: process.env.NOTION_TAGS_PROPERTY,
466
+ dueDateProperty: process.env.NOTION_DUE_DATE_PROPERTY,
467
+ descriptionProperty: process.env.NOTION_DESCRIPTION_PROPERTY
468
+ });
469
+ return executeSync(client, payload);
131
470
  }
132
471
  async function executeSync(client, payload) {
133
- return {
134
- summary: payload.summary ? await client.createWorkItem(payload.summary) : void 0,
135
- items: await client.createWorkItems(payload.items)
136
- };
472
+ const summary = payload.summary ? await client.createWorkItem(payload.summary) : undefined;
473
+ const items = await client.createWorkItems(payload.items);
474
+ return { summary, items };
137
475
  }
138
476
  function requireEnv(key) {
139
- const value = process.env[key];
140
- if (!value) throw new Error(`Missing required env var: ${key}`);
141
- return value;
477
+ const value = process.env[key];
478
+ if (!value) {
479
+ throw new Error(`Missing required env var: ${key}`);
480
+ }
481
+ return value;
142
482
  }
143
483
  function splitList(value) {
144
- if (!value) return void 0;
145
- const items = value.split(",").map((item) => item.trim()).filter(Boolean);
146
- return items.length > 0 ? items : void 0;
484
+ if (!value)
485
+ return;
486
+ const items = value.split(",").map((item) => item.trim()).filter(Boolean);
487
+ return items.length > 0 ? items : undefined;
147
488
  }
148
489
  run().catch((error) => {
149
- console.error(error);
150
- process.exitCode = 1;
490
+ console.error(error);
491
+ process.exitCode = 1;
151
492
  });
152
-
153
- //#endregion
154
- //# sourceMappingURL=sync-actions.js.map