@zibby/cli 0.1.90 → 0.2.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (37) hide show
  1. package/README.md +94 -848
  2. package/dist/bin/zibby.js +47 -2
  3. package/dist/commands/chat.js +1 -1
  4. package/dist/commands/studio.js +1 -1
  5. package/dist/commands/workflow.js +19 -19
  6. package/dist/commands/workflows/agent-helpers.js +18 -0
  7. package/dist/commands/workflows/deploy.js +47 -24
  8. package/dist/commands/workflows/env-helpers.js +2 -0
  9. package/dist/commands/workflows/env.js +39 -0
  10. package/dist/commands/workflows/generate.js +48 -31
  11. package/dist/commands/workflows/input-helpers.js +7 -0
  12. package/dist/commands/workflows/list.js +2 -2
  13. package/dist/commands/workflows/logs.js +22 -22
  14. package/dist/commands/workflows/run-local.js +19 -0
  15. package/dist/commands/workflows/run.js +5 -5
  16. package/dist/commands/workflows/start.js +11 -11
  17. package/dist/commands/workflows/trigger.js +9 -9
  18. package/dist/package.json +4 -4
  19. package/dist/templates/zibby-workflow-claude/agents-md-block.md +113 -0
  20. package/dist/templates/zibby-workflow-claude/claude/agents/zibby-test-author.md +72 -0
  21. package/dist/templates/zibby-workflow-claude/claude/agents/zibby-workflow-builder.md +81 -0
  22. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-add-node.md +75 -0
  23. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-debug.md +67 -0
  24. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-delete.md +37 -0
  25. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-deploy.md +77 -0
  26. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-list.md +30 -0
  27. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-static-ip.md +68 -0
  28. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-tail.md +53 -0
  29. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-test-debug.md +59 -0
  30. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-test-generate.md +39 -0
  31. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-test-run.md +48 -0
  32. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-test-write.md +46 -0
  33. package/dist/templates/zibby-workflow-claude/claude/commands/zibby-trigger.md +52 -0
  34. package/dist/templates/zibby-workflow-claude/claude/settings.json +10 -0
  35. package/dist/templates/zibby-workflow-claude/cursor/rules/zibby-workflows.mdc +56 -0
  36. package/dist/templates/zibby-workflow-claude/manifest.json +43 -0
  37. package/package.json +4 -4
@@ -0,0 +1,39 @@
1
+ import a from"chalk";import x from"ora";import{existsSync as ao,readFileSync as uo}from"fs";import{join as po}from"path";import A from"chalk";import{confirm as fo}from"@inquirer/prompts";import c from"chalk";import T from"ora";import{spawn as ro}from"child_process";var m={local:{name:"Local Development",apiUrl:"http://localhost:3001",accountApiUrl:"http://localhost:3001",frontendUrl:"http://localhost:3000",description:"Local backend running on port 3001"},prod:{name:"Production",apiUrl:process.env.ZIBBY_PROD_API_URL||"https://api-prod.zibby.app",accountApiUrl:process.env.ZIBBY_PROD_ACCOUNT_API_URL||"https://account-api-prod.zibby.app",frontendUrl:process.env.ZIBBY_PROD_FRONTEND_URL||"https://studio.zibby.app",description:"Production environment"}};function v(){let o;if(process.env.ZIBBY_API_URL)o=process.env.ZIBBY_API_URL;else{let e=process.env.ZIBBY_ENV||"prod";m[e]?o=m[e].apiUrl:o=m.prod.apiUrl}try{let e=new URL(o);return e.protocol!=="http:"&&e.protocol!=="https:"?(console.error(`\u26A0\uFE0F Invalid API URL protocol: ${e.protocol} (only http/https allowed)`),m.prod.apiUrl):o}catch{return console.error(`\u26A0\uFE0F Invalid API URL: ${o}`),m.prod.apiUrl}}import{existsSync as L,mkdirSync as Q,readFileSync as oo,writeFileSync as eo}from"fs";import{homedir as z}from"os";import{join as U}from"path";function O(){return process.env.ZIBBY_CONFIG_DIR||U(z(),".zibby")}function R(){return U(O(),"config.json")}var no=U(z(),".zibby"),ko=U(no,"config.json");function to(){let o=O();L(o)||Q(o,{recursive:!0})}function g(){try{let o=R();if(L(o)){let e=oo(o,"utf-8");return JSON.parse(e)}}catch{}return{}}function w(o){to(),eo(R(),JSON.stringify(o,null,2))}function Y(){return g().sessionToken||null}function D(o){let e=g();e.sessionToken=o,w(e)}function F(){return g().user||null}function K(o){let e=g();e.user=o,w(e)}function Z(o){let e=g();e.proxyUrl=o,w(e)}function V(o){let e=g();e.mem0ProxyUrl=o,w(e)}function J(o){let e=g();e.projects=o,w(e)}import{existsSync as Io,mkdirSync as bo,readFileSync as To,writeFileSync as So,unlinkSync as Ao}from"fs";import{resolve as Eo}from"path";import{homedir as _o}from"os";function so(o){let e=process.platform;try{let n,t;return e==="darwin"?(n="open",t=[o]):e==="win32"?(n="cmd",t=["/c","start","",o]):(n="xdg-open",t=[o]),ro(n,t,{detached:!0,stdio:"ignore"}).unref(),!0}catch{return!1}}function io(){let o=Y(),e=F();return o&&e?{loggedIn:!0,user:e,token:o}:{loggedIn:!1}}async function M(){try{console.log(c.cyan(`
2
+ \u{1F510} Initiating login...
3
+ `));let o=io();if(o.loggedIn){console.log(c.green("\u2705 Already logged in!")),console.log(c.gray(`User: ${o.user.email}`)),console.log(c.gray(`Name: ${o.user.name}
4
+ `));let{createInterface:e}=await import("readline"),n=e({input:process.stdin,output:process.stdout});return new Promise((t,r)=>{let s=()=>{n.close(),process.stdin.isTTY&&process.stdin.setRawMode(!1)},i=()=>{console.log(c.yellow(`
5
+
6
+ \u26A0\uFE0F Login cancelled
7
+ `)),s(),process.exit(0)};process.on("SIGINT",i),n.question(c.yellow("Continue with this session? (Y/n): "),async l=>{process.removeListener("SIGINT",i),s();try{if(l.toLowerCase()==="n"||l.toLowerCase()==="no"){console.log(c.gray(`Starting new login...
8
+ `));let f=await H();t(f)}else console.log(c.green(`Using existing session.
9
+ `)),t({success:!0,...o})}catch(f){r(f)}})})}return await H()}catch(o){return console.error(c.red(`
10
+ \u274C Login failed:`,o.message)),{success:!1,error:o.message}}}async function lo(o){let e=v();try{let n=await fetch(`${e}/projects`,{headers:{Authorization:`Bearer ${o}`}});if(n.ok){let r=((await n.json()).projects||[]).map(s=>({name:s.name,projectId:s.projectId,apiToken:s.apiToken}));return J(r),r}}catch(n){console.log(c.gray(`\u26A0\uFE0F Could not fetch projects: ${n.message}`))}return[]}async function H(){let o=v(),e=T("Requesting login code...").start(),n=await fetch(`${o}/cli/login/initiate`,{method:"POST",headers:{"Content-Type":"application/json"}});if(!n.ok){e.fail("Failed to request login code");let y=await n.json();throw new Error(y.error||"Failed to initiate login")}let{deviceCode:t,userCode:r,verificationUrl:s,expiresIn:i,interval:l}=await n.json();e.succeed("Login code generated"),console.log(""),console.log(c.cyan("\u2554\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2557")),console.log(c.cyan("\u2551")+c.white.bold(" Complete login in your browser ")+c.cyan("\u2551")),console.log(c.cyan("\u255A\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u255D")),console.log(""),console.log(c.white("Opening browser to login page...")),console.log(c.gray(`Code expires in ${Math.floor(i/60)} minutes`)),console.log(""),await so(s)||(console.log(c.yellow("\u26A0\uFE0F Could not open browser automatically.")),console.log(c.white("Please open this URL manually: ")+c.blue(s)),console.log(""));let u=T("Waiting for authorization...").start(),d=(l||3)*1e3,X=Math.floor(i/(l||3)),_=0,N=!1,B=()=>{N=!0,u.stop(),console.log(c.yellow(`
11
+
12
+ \u26A0\uFE0F Login cancelled
13
+ `)),process.exit(0)};process.on("SIGINT",B);try{for(;_<X&&!N;){await co(d),_++;let y=await fetch(`${o}/cli/login/poll`,{method:"POST",headers:{"Content-Type":"application/json"},body:JSON.stringify({deviceCode:t})});if(y.status===202)continue;if(!y.ok){u.fail("Authorization failed");let b=await y.json();throw new Error(b.error||"Authorization failed")}let p=await y.json();if(p.status==="authorized"){u.succeed(c.white("Authorization successful!")),D(p.token),K(p.user),p.proxyUrl&&Z(p.proxyUrl),p.mem0ProxyUrl&&V(p.mem0ProxyUrl),console.log(""),console.log(c.gray(`User: ${p.user.email}`));let b=T("Fetching projects...").start(),C=await lo(p.token);return b.succeed(`Fetched ${C.length} project${C.length!==1?"s":""}`),console.log(c.gray(`Session saved to: ~/.zibby/config.json
14
+ `)),{success:!0,loggedIn:!0,user:p.user,token:p.token}}if(p.status==="denied")throw u.fail("Authorization denied"),new Error("User denied authorization")}throw u.fail("Login timeout"),new Error("Login timed out - please try again")}finally{process.removeListener("SIGINT",B)}}function co(o){return new Promise(e=>setTimeout(e,o))}function q(){try{let o=process.env.HOME||process.env.USERPROFILE;if(!o)return null;let e=po(o,".zibby","config.json");return ao(e)&&JSON.parse(uo(e,"utf-8")).sessionToken||null}catch{return null}}function S(){console.log(`
15
+ Not authenticated.`),console.log(` Run ${A.cyan("zibby login")} or set ${A.cyan("ZIBBY_API_KEY")} in your environment.
16
+ `)}async function G(o={}){let e=o.apiKey||process.env.ZIBBY_API_KEY||null,n=q();if(n||e)return{sessionToken:n,apiKey:e};if(!process.stdin.isTTY){if(o.optional)return{sessionToken:null,apiKey:null};S(),process.exit(1)}console.log(A.yellow(`
17
+ Not logged in.`));let t;try{t=await fo({message:"Open browser to log in now?",default:!0})}catch{t=!1}if(!t){if(o.optional)return{sessionToken:null,apiKey:null};S(),process.exit(1)}if(await M(),n=q(),!n){if(o.optional)return{sessionToken:null,apiKey:null};S(),process.exit(1)}return{sessionToken:n,apiKey:null}}import{readFileSync as go,existsSync as yo}from"fs";import ho from"dotenv";var j=/^[A-Z_][A-Z0-9_]*$/;function k(o){return typeof o=="string"&&j.test(o)}function W(o){if(typeof o!="string"||o.length===0)throw new Error("Expected KEY=value, got empty argument");let e=o.indexOf("=");if(e<=0)throw new Error(`Expected KEY=value, got "${o}" \u2014 missing '=' or empty key`);let n=o.slice(0,e),t=o.slice(e+1);if(!k(n))throw new Error(`Invalid env var name "${n}" \u2014 must match ${j} (uppercase letters, digits, underscores; can't start with a digit)`);return{key:n,value:t}}function E(o){if(!Array.isArray(o)||o.length===0)return{};let e={};for(let n of o){if(!yo(n))throw new Error(`--env file not found: ${n}`);let t=ho.parse(go(n,"utf-8"));for(let[r,s]of Object.entries(t)){if(!k(r))throw new Error(`Invalid env var name "${r}" in ${n} \u2014 must match ${j} (uppercase letters, digits, underscores; can't start with a digit)`);e[r]=s}}return e}function h(o,e,n){let t=String(o).replace(/\/+$/,"");return n?`${t}/workflows/${e}/env/${encodeURIComponent(n)}`:`${t}/workflows/${e}/env`}function P(o){return!Array.isArray(o)||o.length===0?"No env vars set on this workflow.":[...o].sort().map(n=>` ${n}`).join(`
18
+ `)}async function $(o){let{sessionToken:e,apiKey:n}=await G({apiKey:o.apiKey}),t=e||n;return t||(console.log(a.red("\n Not authenticated \u2014 run `zibby login` or pass --api-key.\n")),process.exit(1)),{authToken:t,apiUrl:v()}}async function I(o,e,n,t,r){let s={method:o,headers:{Authorization:`Bearer ${n}`,...t?{"Content-Type":"application/json"}:{}},...t?{body:JSON.stringify(t)}:{}},i;try{i=await fetch(e,s)}catch(f){r&&r.fail(`Network error: ${f.message}`),process.exit(1)}let l;try{l=await i.json()}catch{l={}}return i.ok||(r?r.fail(l.error||l.message||`HTTP ${i.status}`):console.log(a.red(`
19
+ ${l.error||l.message||`HTTP ${i.status}`}
20
+ `)),process.exit(1)),l}async function te(o,e={}){o||(console.log(a.red(`
21
+ Workflow UUID is required`)),console.log(a.gray(` Usage: zibby workflow env list <uuid>
22
+ `)),process.exit(1));let{authToken:n,apiUrl:t}=await $(e),r=x(`Fetching env keys for ${o}...`).start(),s=await I("GET",h(t,o),n,null,r),i=s.keys||[];r.succeed(`${i.length} env var${i.length===1?"":"s"} on workflow ${o}`),console.log(""),console.log(P(i)),s.envUpdatedAt&&console.log(a.gray(`
23
+ Last updated: ${s.envUpdatedAt}`)),console.log("")}async function re(o,e,n={}){(!o||!e)&&(console.log(a.red(`
24
+ Both UUID and KEY=value are required`)),console.log(a.gray(` Usage: zibby workflow env set <uuid> KEY=value
25
+ `)),process.exit(1));let t,r;try{({key:t,value:r}=W(e))}catch(d){console.log(a.red(`
26
+ ${d.message}
27
+ `)),process.exit(1)}let{authToken:s,apiUrl:i}=await $(n),l=x(`Setting ${t} on workflow ${o}...`).start(),u=((await I("PATCH",h(i,o,t),s,{value:r},l)).keys||[]).length;l.succeed(`Set ${a.cyan(t)} (workflow now has ${u} env var${u===1?"":"s"})`)}async function se(o,e,n={}){(!o||!e)&&(console.log(a.red(`
28
+ Both UUID and KEY are required`)),console.log(a.gray(` Usage: zibby workflow env unset <uuid> KEY
29
+ `)),process.exit(1)),k(e)||(console.log(a.red(`
30
+ Invalid env var name "${e}"
31
+ `)),process.exit(1));let{authToken:t,apiUrl:r}=await $(n),s=x(`Unsetting ${e} on workflow ${o}...`).start(),i=await I("DELETE",h(r,o,e),t,null,s),l=(i.keys||[]).length;i.removed?s.succeed(`Unset ${a.cyan(e)} (workflow now has ${l} env var${l===1?"":"s"})`):s.warn(`Key ${a.cyan(e)} was not set on workflow ${o} \u2014 nothing to do`)}async function ie(o,e={}){o||(console.log(a.red(`
32
+ Workflow UUID is required`)),console.log(a.gray(` Usage: zibby workflow env push <uuid> --file .env [--file .env.prod]
33
+ `)),process.exit(1));let n=Array.isArray(e.file)?e.file:e.file?[e.file]:[];n.length===0&&(console.log(a.red(`
34
+ At least one --file is required`)),console.log(a.gray(` Usage: zibby workflow env push <uuid> --file .env
35
+ `)),process.exit(1));let t;try{t=E(n)}catch(f){console.log(a.red(`
36
+ ${f.message}
37
+ `)),process.exit(1)}let r=Object.keys(t);r.length===0&&(console.log(a.yellow(`
38
+ No env vars found in ${n.join(", ")} \u2014 nothing to push.
39
+ `)),process.exit(0));let{authToken:s,apiUrl:i}=await $(e),l=x(`Pushing ${r.length} env var${r.length===1?"":"s"} to ${o}...`).start();await I("PUT",h(i,o),s,{env:t},l),l.succeed(`Pushed ${r.length} env var${r.length===1?"":"s"} from ${n.join(", ")}`),console.log(""),console.log(P(r)),console.log("")}async function le({uuid:o,files:e,authToken:n,apiUrl:t,spinner:r}){let s=E(e),i=Object.keys(s);if(i.length===0)return{count:0,files:e};let l=h(t,o),f={method:"PUT",headers:{Authorization:`Bearer ${n}`,"Content-Type":"application/json"},body:JSON.stringify({env:s})},u=await fetch(l,f);if(!u.ok){let d=await u.json().catch(()=>({}));throw new Error(`Env push failed: ${d.error||d.message||`HTTP ${u.status}`}`)}return r&&(r.text=`Synced ${i.length} env var${i.length===1?"":"s"} to workflow`),{count:i.length,files:e}}export{te as listEnvCommand,ie as pushEnvCommand,re as setEnvCommand,le as syncEnvFromFiles,se as unsetEnvCommand};
@@ -1,4 +1,4 @@
1
- import{mkdir as v,writeFile as p}from"fs/promises";import{existsSync as k}from"fs";import{spawn as M}from"child_process";import{join as c}from"path";import t from"chalk";import I from"ora";import{input as j}from"@inquirer/prompts";import{existsSync as P}from"fs";import{join as E}from"path";import{pathToFileURL as $}from"url";async function b(o){let n=E(o,".zibby.config.mjs");if(!P(n))throw new Error(".zibby.config.mjs not found");try{let e=await import($(n).href);return e.default||e}catch(e){throw new Error(`Failed to load .zibby.config.mjs: ${e.message}`,{cause:e})}}import{mkdir as le,writeFile as ce,readFile as de}from"fs/promises";import{existsSync as me,readdirSync as ge}from"fs";import{join as fe,resolve as ye,dirname as _}from"path";import{homedir as we}from"os";import xe from"inquirer";import ke from"chalk";import Ae from"ora";import{spawn as Pe,execSync as Ee}from"child_process";import{fileURLToPath as S}from"url";import{createRequire as z}from"module";import{existsSync as U,readFileSync as q,writeFileSync as J,mkdirSync as X}from"fs";import{join as ee}from"path";import{homedir as oe}from"os";var K=S(import.meta.url),ze=_(K),Ke=z(import.meta.url);function x(o,n={},e={}){let r=["dolt","mem0"].includes(String(e.memoryBackend||"").toLowerCase())?String(e.memoryBackend).toLowerCase():"dolt",l={claude:`
1
+ import{mkdir as T,writeFile as k}from"fs/promises";import{existsSync as K}from"fs";import{spawn as ae}from"child_process";import{join as w}from"path";import a from"chalk";import M from"ora";import{input as le}from"@inquirer/prompts";import{existsSync as Y}from"fs";import{join as N}from"path";import{pathToFileURL as L}from"url";async function E(n){let e=N(n,".zibby.config.mjs");if(!Y(e))throw new Error(".zibby.config.mjs not found");try{let t=await import(L(e).href);return t.default||t}catch(t){throw new Error(`Failed to load .zibby.config.mjs: ${t.message}`,{cause:t})}}import{mkdir as Me,writeFile as Re,readFile as Oe}from"fs/promises";import{existsSync as Ne,readdirSync as Le}from"fs";import{join as Ge,resolve as Ve,dirname as D}from"path";import{homedir as He}from"os";import Fe from"inquirer";import qe from"chalk";import Xe from"ora";import{spawn as et,execSync as tt}from"child_process";import{fileURLToPath as G}from"url";import{createRequire as V}from"module";import{existsSync as Ae,readFileSync as Ce,writeFileSync as Ie,mkdirSync as $e}from"fs";import{join as Ee}from"path";import{homedir as _e}from"os";var Z=G(import.meta.url),rt=D(Z),it=V(import.meta.url);function S(n,e={},t={}){let o=["dolt","mem0"].includes(String(t.memoryBackend||"").toLowerCase())?String(t.memoryBackend).toLowerCase():"dolt",s={claude:`
2
2
  claude: {
3
3
  model: 'auto', // Options: 'auto', 'sonnet-4.6', 'opus-4.6', 'sonnet-4.5', 'opus-4.5'
4
4
  maxTokens: 4096,
@@ -11,13 +11,13 @@ import{mkdir as v,writeFile as p}from"fs/promises";import{existsSync as k}from"f
11
11
  },`,gemini:`
12
12
  gemini: {
13
13
  model: 'gemini-2.5-pro', // Options: 'auto', 'gemini-2.5-pro', 'gemini-2.5-flash'
14
- },`},m=o.agent,s=Object.entries(l).filter(([i])=>i!==m).map(([,i])=>i.split(`
15
- `).map(d=>d.trim()?` // ${d.trimStart()}`:d).join(`
14
+ },`},i=n.agent,l=Object.entries(s).filter(([r])=>r!==i).map(([,r])=>r.split(`
15
+ `).map(c=>c.trim()?` // ${c.trimStart()}`:c).join(`
16
16
  `)).join(`
17
17
  `);return`export default {
18
18
  // AI agent settings
19
- agent: {${l[m]}
20
- ${s}
19
+ agent: {${s[i]}
20
+ ${l}
21
21
  strictMode: false,
22
22
  },
23
23
 
@@ -25,12 +25,12 @@ ${s}
25
25
  // and workflow config. Runtime strategies attach MCP per run (no global Gemini settings mutation).
26
26
  browser: {
27
27
  mcp: 'playwright',
28
- headless: ${o.browserMode==="headless"},
28
+ headless: ${n.browserMode==="headless"},
29
29
  },
30
30
 
31
31
  // Chat memory backend adapter (dolt | mem0)
32
32
  memory: {
33
- backend: '${r}',
33
+ backend: '${o}',
34
34
  },
35
35
 
36
36
  // Advanced: Override models per node (optional)
@@ -44,7 +44,7 @@ ${s}
44
44
  specs: 'test-specs', // Where your .txt test specs are
45
45
  generated: 'tests', // Where generated .spec.js files go
46
46
  output: '.zibby/output', // Where workflow execution results are saved (default: .zibby/output)
47
- workflows: '${o.workflowsPath||".zibby/workflows"}', // Where custom workflows are stored
47
+ workflows: '${n.workflowsPath||".zibby/workflows"}', // Where custom workflows are stored
48
48
  // sessionPrefix: 'run', // Optional: prefix for session folders (e.g., run_1772788458045)
49
49
  },
50
50
 
@@ -79,10 +79,27 @@ ${s}
79
79
  },
80
80
 
81
81
  // Cloud sync - auto-upload test results & videos (requires ZIBBY_API_KEY in .env)
82
- cloudSync: ${o.cloudSync||!1}
82
+ cloudSync: ${n.cloudSync||!1}
83
83
  };
84
- `}function B(o){return new Promise(n=>{let e=M("npm",["install","--no-audit","--no-fund"],{cwd:o,stdio:"pipe",shell:!1}),r="";e.stderr.on("data",l=>{r+=l.toString()}),e.on("error",()=>n({ok:!1,stderr:"npm not found on PATH"})),e.on("close",l=>n({ok:l===0,stderr:r}))})}var T=/^[a-z][a-z0-9-]{0,62}[a-z0-9]$/,A=["stellar","quantum","cosmic","nova","nebula","solar","lunar","atomic","plasma","fusion","pulse","flux","spark","blaze","ember","radiant","luminous","electric","magnetic","kinetic","neon","cyber","pixel","matrix","vector","synth","neural","prism","zenith","phoenix","catalyst","nexus","echo","wave","crystal","jade","ruby","emerald","onyx","amber","silver","turbo","lightning","thunder","storm","arcane","mystic","ethereal","celestial","swift","crimson","iron","cobalt"],C=["flow","runner","pipeline","stream","circuit","engine","beacon","forge","relay","shuttle","conduit","gateway","sentinel","scout","pilot","voyager","ranger","dispatch","signal","pulse","agent","daemon","spark","orbit","vector","nexus","matrix","grid","mesh","bridge","link","node","craft","bolt","ray","arc","wave","hook","probe","shard"];function Y(){let o=A[Math.floor(Math.random()*A.length)],n=C[Math.floor(Math.random()*C.length)];return`${o}-${n}`}function R(o){return`${o.split("-").map(n=>n.charAt(0).toUpperCase()+n.slice(1)).join("")}Workflow`}function O(o,n){return`/**
85
- * ${o}
84
+ `}import{readFileSync as x,existsSync as m,mkdirSync as $,writeFileSync as v,statSync as H,chmodSync as W}from"fs";import{join as g,dirname as A}from"path";import{fileURLToPath as F}from"url";import{select as U}from"@inquirer/prompts";import h from"chalk";var q=A(F(import.meta.url)),P=g(q,"..","..","..","templates","zibby-workflow-claude"),J="<!-- BEGIN zibby-workflows",_="<!-- END zibby-workflows -->",I={claude:"Claude Code",cursor:"Cursor",codex:"Codex (AGENTS.md)"};function X(n){let e=process.argv[1];if(!e||!m(e))return null;let t=g(n,".zibby","bin"),o=g(t,"zibby");$(t,{recursive:!0});let s=`#!/bin/sh
85
+ # Auto-generated by 'zibby workflow generate'. Re-runs of generate update this path.
86
+ # Lets agents (Claude / Cursor / Codex) and CI scripts call zibby reliably even
87
+ # when the CLI isn't on PATH. Delete this file or .zibby/bin/ to disable.
88
+ exec node ${JSON.stringify(e)} "$@"
89
+ `;v(o,s);try{H(o)}catch{}try{W(o,493)}catch{}return o}function Q(n){let e=[];return m(g(n,".claude"))&&e.push("claude"),(m(g(n,".cursor"))||m(g(n,".cursorrules")))&&e.push("cursor"),m(g(n,"AGENTS.md"))&&e.push("codex"),e}function ee(){let n=g(P,"manifest.json");return JSON.parse(x(n,"utf8"))}function z(n){let e=n.match(/<!--\s*(?:BEGIN zibby-workflows\s+)?zibby-template-version:\s*(\d+)\s*-->/);return e?parseInt(e[1],10):null}function te(n,e,t){if(!m(n))return{write:!0,reason:"new"};if(t)return{write:!0,reason:"force"};let o=x(n,"utf8"),s=z(o),i=z(e);return s==null?{write:!1,reason:"user-edited (no version marker)"}:i==null?{write:!1,reason:"shipped file has no version (skipping for safety)"}:s>i?{write:!1,reason:`current version ${s} > shipped ${i}`}:{write:!0,reason:s===i?"same version, refresh":`upgrade ${s} \u2192 ${i}`}}function ne(n,e){let t=`${e.trim()}
90
+ `;if(!m(n))return v(n,t),"created";let o=x(n,"utf8"),s=o.indexOf(J),i=o.indexOf(_);if(s!==-1&&i!==-1&&i>s){let r=o.slice(0,s),c=o.slice(i+_.length),p=r+t.trim()+c;return p===o?"unchanged":(v(n,p),"updated")}let l=o.endsWith(`
91
+ `)?`
92
+ `:`
93
+
94
+ `;return v(n,o+l+t),"appended"}function oe(n,e,t,o){let s=t.agents[n];if(!s)return[];let i=[],l=r=>g(P,r.split("/").map(c=>c.startsWith(".")?c.slice(1):c).join("/"));for(let r of s.files||[]){let c=l(r);if(!m(c))continue;let p=x(c,"utf8"),f=g(e,r),d=te(f,p,o);d.write?($(A(f),{recursive:!0}),v(f,p),i.push({path:r,action:"write",detail:d.reason})):i.push({path:r,action:"skip",detail:d.reason})}for(let r of s.writeOnceFiles||[]){let c=l(r);if(!m(c))continue;let p=g(e,r);if(m(p)){i.push({path:r,action:"skip",detail:"write-once: already exists, leaving alone"});continue}$(A(p),{recursive:!0}),v(p,x(c,"utf8")),i.push({path:r,action:"write",detail:"write-once: created"})}if(s.rootBlock){let r=g(P,s.rootBlock.source),c=x(r,"utf8"),p=g(e,s.rootBlock.target),f=ne(p,c);i.push({path:s.rootBlock.target,action:f})}return i}function se(n){let e=n;for(let t=0;t<6;t++){if(m(g(e,".git"))||m(g(e,".zibby.config.mjs"))||m(g(e,"package.json")))return e;let o=A(e);if(o===e)break;e=o}return n}async function re(n){let e=g(n,".zibby.config.mjs");if(!m(e))return null;try{return((await import(`file://${e}?t=${Date.now()}`)).default||{}).agentHelpers||null}catch(t){return console.log(h.gray(` (couldn't parse .zibby.config.mjs agentHelpers section: ${t.message})`)),null}}function j(n,e){let t=g(n,".zibby.config.mjs");if(!m(t)){console.log(h.yellow(" \u26A0\uFE0F .zibby.config.mjs not found; agentHelpers state not persisted"));return}let o=x(t,"utf8"),s=` agentHelpers: {
95
+ templateVersion: ${e.templateVersion},
96
+ agents: ${JSON.stringify(e.agents)},
97
+ updatedAt: '${new Date().toISOString()}',
98
+ },
99
+ `,i=/\s*agentHelpers\s*:\s*\{[^}]*\}\s*,?\n?/m,l;if(i.test(o))l=o.replace(i,`
100
+ ${s}`);else{let r=o.lastIndexOf("}");if(r===-1)return;let c=o.slice(0,r),p=c.replace(/\s+$/,"").slice(-1);l=c+(p&&p!==","&&p!=="{"?`,
101
+ `:"")+s+o.slice(r)}v(t,l)}async function ie(n){let e=n[0]||"claude",t=[{name:`${I.claude}${n.includes("claude")?" (detected \u2014 recommended)":""}`,value:"claude"},{name:`${I.codex}${n.includes("codex")?" (detected)":""}`,value:"codex"},{name:`${I.cursor}${n.includes("cursor")?" (detected)":""}`,value:"cursor"},{name:"All of the above",value:"all"},{name:"Skip \u2014 I'll set up later",value:"none"}];if(e!=="claude"){let s=t.findIndex(i=>i.value===e);if(s>0){let[i]=t.splice(s,1);t.unshift(i)}}let o=await U({message:"Set up agent helpers (slash commands, sub-agents, context)?",choices:t,default:t[0].value});return o==="all"?["claude","cursor","codex"]:[o]}async function B(n={}){let{forcedAgents:e,force:t=!1,silent:o=!1}=n,s=n.projectRoot||se(process.cwd()),i=ee(),l=await re(s),r,c=n&&n.forcePrompt===!0;if(e&&e.length>0)r=e;else if(!c&&l&&Array.isArray(l.agents)&&l.agents.length>0)r=l.agents;else{if(o)return;if(!process.stdin.isTTY)console.log(h.gray(" (non-TTY; defaulting agent helpers to Claude Code \u2014 pass --agent to override)")),r=["claude"];else{let d=Q(s);try{r=await ie(d)}catch{r=["none"]}}}if(r.length===1&&r[0]==="none"){j(s,{templateVersion:i.templateVersion,agents:["none"]}),o||console.log(h.gray(" Skipped agent helpers. Run `zibby agents add` later to set up."));return}let p=X(s),f=[];for(let d of r){if(!i.agents[d])continue;oe(d,s,i,t).forEach(b=>f.push({...b,agent:d}))}if(j(s,{templateVersion:i.templateVersion,agents:r}),!o){let d=f.filter(u=>["write","created","updated","appended"].includes(u.action)),y=f.filter(u=>u.action==="skip"),b=f.filter(u=>u.action==="unchanged");if(d.length>0){console.log(h.green(` \u2713 Wrote ${d.length} agent helper file(s):`));for(let u of d)console.log(h.gray(` ${u.path} (${u.detail||u.action})`))}if(y.length>0){console.log(h.yellow(` \u26A0\uFE0F Skipped ${y.length} (use --force-claude to overwrite):`));for(let u of y)console.log(h.gray(` ${u.path} \u2014 ${u.detail}`))}d.length===0&&y.length===0&&b.length>0&&console.log(h.gray(` Agent helpers up to date (template v${i.templateVersion}).`)),console.log(h.gray(` agentHelpers: { agents: [${r.map(u=>`"${u}"`).join(", ")}] } saved to .zibby.config.mjs`)),p&&console.log(h.gray(` CLI shim: ${p} (agents fall back to ./.zibby/bin/zibby if zibby isn't on PATH)`)),r.includes("claude")&&console.log(h.gray(" \u2192 Run `claude` and try /zibby-add-node"))}}function ce(n){return new Promise(e=>{let t=ae("npm",["install","--no-audit","--no-fund"],{cwd:n,stdio:"pipe",shell:!1}),o="";t.stderr.on("data",s=>{o+=s.toString()}),t.on("error",()=>e({ok:!1,stderr:"npm not found on PATH"})),t.on("close",s=>e({ok:s===0,stderr:o}))})}var de=/^[a-z][a-z0-9-]{0,62}[a-z0-9]$/,R=["stellar","quantum","cosmic","nova","nebula","solar","lunar","atomic","plasma","fusion","pulse","flux","spark","blaze","ember","radiant","luminous","electric","magnetic","kinetic","neon","cyber","pixel","matrix","vector","synth","neural","prism","zenith","phoenix","catalyst","nexus","echo","wave","crystal","jade","ruby","emerald","onyx","amber","silver","turbo","lightning","thunder","storm","arcane","mystic","ethereal","celestial","swift","crimson","iron","cobalt"],O=["flow","runner","pipeline","stream","circuit","engine","beacon","forge","relay","shuttle","conduit","gateway","sentinel","scout","pilot","voyager","ranger","dispatch","signal","pulse","agent","daemon","spark","orbit","vector","nexus","matrix","grid","mesh","bridge","link","node","craft","bolt","ray","arc","wave","hook","probe","shard"];function ue(){let n=R[Math.floor(Math.random()*R.length)],e=O[Math.floor(Math.random()*O.length)];return`${n}-${e}`}function pe(n){return`${n.split("-").map(e=>e.charAt(0).toUpperCase()+e.slice(1)).join("")}Workflow`}function ge(n,e){return`/**
102
+ * ${n}
86
103
  *
87
104
  * buildGraph() \u2014 define nodes, edges, conditional routing
88
105
  * onComplete(result) \u2014 post-processing after the graph finishes
@@ -91,7 +108,7 @@ ${s}
91
108
  import { WorkflowAgent, WorkflowGraph } from '@zibby/core';
92
109
  import { exampleNode } from './nodes/index.mjs';
93
110
 
94
- export class ${o} extends WorkflowAgent {
111
+ export class ${n} extends WorkflowAgent {
95
112
  buildGraph() {
96
113
  const graph = new WorkflowGraph();
97
114
 
@@ -104,11 +121,11 @@ export class ${o} extends WorkflowAgent {
104
121
  }
105
122
 
106
123
  async onComplete(result) {
107
- console.log(\`[${n}] workflow complete \u2014 success: \${result.success !== false}\`);
124
+ console.log(\`[${e}] workflow complete \u2014 success: \${result.success !== false}\`);
108
125
  }
109
126
  }
110
- `}function N(){return`export { exampleNode } from './example.mjs';
111
- `}function L(){return`import { z } from '@zibby/core';
127
+ `}function me(){return`export { exampleNode } from './example.mjs';
128
+ `}function fe(){return`import { z } from '@zibby/core';
112
129
 
113
130
  const ExampleOutputSchema = z.object({
114
131
  summary: z.string().describe('A short summary of the result'),
@@ -127,19 +144,19 @@ Analyze the input and return a summary with a status.\`,
127
144
 
128
145
  outputSchema: ExampleOutputSchema,
129
146
  };
130
- `}function D(o,n){return`${JSON.stringify({name:o,description:`${n} workflow`,entryClass:n,triggers:{api:!0}},null,2)}
131
- `}function Z(){return`${JSON.stringify({type:"module",dependencies:{"@zibby/core":"^0.1.48"}},null,2)}
132
- `}async function Ze(o,n={}){let e;o?e=o.toLowerCase():(e=Y(),console.log(t.gray(`
133
- No name provided \u2014 generated: ${t.white(e)}`))),T.test(e)||(console.log(t.red(`
134
- Invalid workflow name: "${o}"`)),console.log(t.gray(" Must be lowercase, start with a letter, use only a-z, 0-9, hyphens")),console.log(t.gray(" Length: 2\u201364 characters")),console.log(t.gray(` Example: ticket-triage, pr-review, deploy-checker
135
- `)),process.exit(1));let r=process.cwd(),l=c(r,".zibby.config.mjs"),m=k(l),s=".zibby/workflows";if(m)try{let a=await b(r);a?.paths?.workflows&&(s=a.paths.workflows)}catch(a){console.log(t.yellow(` \u26A0\uFE0F Could not load .zibby.config.mjs: ${a.message}`)),console.log(t.gray(` Using default path: ${s}`))}else if(process.stdin.isTTY){console.log(t.gray(`
136
- Common picks: ${t.white(".zibby/workflows")} (hidden, default) \xB7 ${t.white("workflows")} (visible at project root)
137
- `));let g=(await j({message:`Where to save workflows? ${t.gray("[Enter for .zibby/workflows]")}`})).trim();s=g===""?".zibby/workflows":g}else console.log(t.gray(` Using default workflows path: ${s} (run interactively to customize)`));let i=c(r,s,e),d=c(i,"nodes");k(i)&&(console.log(t.red(`
138
- Workflow already exists: ${s}/${e}/`)),console.log(t.gray(` Choose a different name or delete the existing folder.
139
- `)),process.exit(1));let y=R(e),h=I(` Scaffolding workflow "${e}"...`).start();try{await v(c(r,s),{recursive:!0}),await v(d,{recursive:!0}),await Promise.all([p(c(i,"graph.mjs"),O(y,e)),p(c(d,"index.mjs"),N()),p(c(d,"example.mjs"),L()),p(c(i,"workflow.json"),D(e,y)),p(c(i,"package.json"),Z())]);let a=!1;if(!m){let f=x({agent:"claude",browserMode:"headless",workflowsPath:s},{},{memoryBackend:"dolt"});await p(l,f),a=!0}h.succeed(` Scaffolded ${t.bold(e)}`),console.log(t.green(`
140
- Created:`)),console.log(t.white(` ${s}/${e}/`)),console.log(t.gray(` graph.mjs ${y} (entry)`)),console.log(t.gray(" nodes/index.mjs barrel export")),console.log(t.gray(" nodes/example.mjs starter node (prompt + schema)")),console.log(t.gray(" workflow.json manifest")),console.log(t.gray(" package.json dependencies (@zibby/core)")),a&&console.log(t.white(" .zibby.config.mjs project config (agent: claude \u2014 edit to switch)"));let g=!1;if(n.skipInstall)console.log(t.gray(` Skipped npm install (--skip-install). Run manually: cd ${s}/${e} && npm install`));else{let f=I({text:` Installing dependencies in ${s}/${e}/...`,prefixText:""}).start(),w=await B(i);w.ok?(f.succeed(` Installed dependencies in ${s}/${e}/`),g=!0):(f.warn(` Could not install dependencies (${w.stderr.trim().split(`
141
- `)[0]||"unknown error"})`),console.log(t.gray(` Run manually: cd ${s}/${e} && npm install`)))}console.log(t.white(`
142
- Next steps:`));let u=1;!g&&n.skipInstall!==!0&&console.log(t.cyan(` ${u++}. Install deps: cd ${s}/${e} && npm install`)),console.log(t.cyan(` ${u++}. Edit nodes in ${s}/${e}/nodes/`)),console.log(t.cyan(` ${u++}. Wire them in graph.mjs`)),console.log(t.cyan(` ${u++}. Test locally: zibby workflow start ${e}`)),console.log(t.cyan(` ${u++}. Deploy to cloud: zibby workflow deploy ${e}
143
- `))}catch(a){h.fail(" Scaffold failed"),console.log(t.red(`
144
- ${a.message}
145
- `)),process.exit(1)}}export{Ze as generateWorkflowCommand};
147
+ `}function ye(n,e){return`${JSON.stringify({name:n,description:`${e} workflow`,entryClass:e,triggers:{api:!0}},null,2)}
148
+ `}function he(){return`${JSON.stringify({type:"module",dependencies:{"@zibby/core":"^0.3.0"}},null,2)}
149
+ `}async function Ct(n,e={}){let t;n?t=n.toLowerCase():(t=ue(),console.log(a.gray(`
150
+ No name provided \u2014 generated: ${a.white(t)}`))),de.test(t)||(console.log(a.red(`
151
+ Invalid workflow name: "${n}"`)),console.log(a.gray(" Must be lowercase, start with a letter, use only a-z, 0-9, hyphens")),console.log(a.gray(" Length: 2\u201364 characters")),console.log(a.gray(` Example: ticket-triage, pr-review, deploy-checker
152
+ `)),process.exit(1));let o=process.cwd(),s=w(o,".zibby.config.mjs"),i=K(s),l=".zibby/workflows";if(i)try{let d=await E(o);d?.paths?.workflows&&(l=d.paths.workflows)}catch(d){console.log(a.yellow(` \u26A0\uFE0F Could not load .zibby.config.mjs: ${d.message}`)),console.log(a.gray(` Using default path: ${l}`))}else if(process.stdin.isTTY){console.log(a.gray(`
153
+ Common picks: ${a.white(".zibby/workflows")} (hidden, default) \xB7 ${a.white("workflows")} (visible at project root)
154
+ `));let y=(await le({message:`Where to save workflows? ${a.gray("[Enter for .zibby/workflows]")}`})).trim();l=y===""?".zibby/workflows":y}else console.log(a.gray(` Using default workflows path: ${l} (run interactively to customize)`));let r=w(o,l,t),c=w(r,"nodes");K(r)&&(console.log(a.red(`
155
+ Workflow already exists: ${l}/${t}/`)),console.log(a.gray(` Choose a different name or delete the existing folder.
156
+ `)),process.exit(1));let p=pe(t),f=M(` Scaffolding workflow "${t}"...`).start();try{await T(w(o,l),{recursive:!0}),await T(c,{recursive:!0}),await Promise.all([k(w(r,"graph.mjs"),ge(p,t)),k(w(c,"index.mjs"),me()),k(w(c,"example.mjs"),fe()),k(w(r,"workflow.json"),ye(t,p)),k(w(r,"package.json"),he())]);let d=!1;if(!i){let u=S({agent:"claude",browserMode:"headless",workflowsPath:l},{},{memoryBackend:"dolt"});await k(s,u),d=!0}f.succeed(` Scaffolded ${a.bold(t)}`),console.log(a.green(`
157
+ Created:`)),console.log(a.white(` ${l}/${t}/`)),console.log(a.gray(` graph.mjs ${p} (entry)`)),console.log(a.gray(" nodes/index.mjs barrel export")),console.log(a.gray(" nodes/example.mjs starter node (prompt + schema)")),console.log(a.gray(" workflow.json manifest")),console.log(a.gray(" package.json dependencies (@zibby/core)")),d&&console.log(a.white(" .zibby.config.mjs project config (agent: claude \u2014 edit to switch)"));let y=!1;if(e.skipInstall)console.log(a.gray(` Skipped npm install (--skip-install). Run manually: cd ${l}/${t} && npm install`));else{let u=M({text:` Installing dependencies in ${l}/${t}/...`,prefixText:""}).start(),C=await ce(r);C.ok?(u.succeed(` Installed dependencies in ${l}/${t}/`),y=!0):(u.warn(` Could not install dependencies (${C.stderr.trim().split(`
158
+ `)[0]||"unknown error"})`),console.log(a.gray(` Run manually: cd ${l}/${t} && npm install`)))}try{let u=e.agent;await B({projectRoot:o,forcedAgents:u?u==="all"?["claude","cursor","codex"]:u==="none"?["none"]:[u]:void 0,force:e.forceClaude===!0||e.forceAgents===!0})}catch(u){console.log(a.yellow(` \u26A0\uFE0F Agent helpers setup failed: ${u.message}`)),console.log(a.gray(" Run `zibby agents add` later to retry."))}console.log(a.white(`
159
+ Next steps:`));let b=1;!y&&e.skipInstall!==!0&&console.log(a.cyan(` ${b++}. Install deps: cd ${l}/${t} && npm install`)),console.log(a.cyan(` ${b++}. Edit nodes in ${l}/${t}/nodes/`)),console.log(a.cyan(` ${b++}. Wire them in graph.mjs`)),console.log(a.cyan(` ${b++}. Run locally: zibby workflow run ${t}`)),console.log(a.cyan(` ${b++}. Deploy to cloud: zibby workflow deploy ${t}
160
+ `))}catch(d){f.fail(" Scaffold failed"),console.log(a.red(`
161
+ ${d.message}
162
+ `)),process.exit(1)}}export{Ct as generateWorkflowCommand};
@@ -0,0 +1,7 @@
1
+ import{existsSync as c,readFileSync as f}from"fs";import{resolve as a}from"path";function p(e){return e==="true"?!0:e==="false"?!1:e==="null"?null:e!==""&&!isNaN(Number(e))?Number(e):e}function m(e){let r={};for(let t of e||[]){let l=t.indexOf("=");if(l===-1){console.warn(` Warning: ignored param "${t}" \u2014 expected key=value format`);continue}let s=t.slice(0,l).trim(),u=p(t.slice(l+1)),n=s.split("."),o=r;for(let i=0;i<n.length-1;i++)(typeof o[n[i]]!="object"||o[n[i]]===null)&&(o[n[i]]={}),o=o[n[i]];o[n[n.length-1]]=u}return r}function g(e){let r=a(e);c(r)||(console.log(`
2
+ Error: --input-file not found: ${e}
3
+ `),process.exit(1));try{return JSON.parse(f(r,"utf-8"))}catch(t){console.log(`
4
+ Error: --input-file is not valid JSON: ${t.message}
5
+ `),process.exit(1)}}function N(e){let r={};if(e.inputFile&&(r={...g(e.inputFile)}),e.input)try{r={...r,...JSON.parse(e.input)}}catch(t){console.log(`
6
+ Error: --input is not valid JSON`),console.log(` ${t.message}
7
+ `),process.exit(1)}return e.param?.length&&(r={...r,...m(e.param)}),r}export{p as coerceValue,m as parseParams,N as resolveInput};
@@ -22,7 +22,7 @@ import{readdir as wo,stat as ho}from"fs/promises";import{existsSync as P}from"fs
22
22
  Local Workflows (${t})
23
23
  `)),console.log(r.gray(" ".padEnd(60,"-"))),console.log(r.white(" Name".padEnd(35))+r.white("Files".padEnd(25))),console.log(r.gray(" ".padEnd(60,"-")));for(let c of i){let a=[];c.hasGraph&&a.push("graph"),c.hasManifest&&a.push("manifest"),console.log(` ${r.cyan(c.name.padEnd(33))}${r.gray(a.join(", "))}`)}return console.log(r.gray(" ".padEnd(60,"-"))),console.log(r.gray(`
24
24
  Total: ${i.length} workflow${i.length===1?"":"s"}
25
- `)),console.log(r.white(" Commands:")),console.log(r.cyan(" zibby workflow start <name> ")+r.gray("Test locally")),console.log(r.cyan(" zibby workflow deploy <name> ")+r.gray("Deploy to cloud")),console.log(r.cyan(" zibby workflow trigger <uuid> ")+r.gray("Run workflow (returns job ID)")),console.log(r.cyan(" zibby workflow logs <uuid> ")+r.gray("Tail execution logs")),console.log(""),i}catch(u){if(o.quiet)return[];console.log(r.red(`
25
+ `)),console.log(r.white(" Commands:")),console.log(r.cyan(" zibby workflow run <name> ")+r.gray("Run locally")),console.log(r.cyan(" zibby workflow deploy <name> ")+r.gray("Deploy to cloud")),console.log(r.cyan(" zibby workflow trigger <uuid> ")+r.gray("Run workflow (returns job ID)")),console.log(r.cyan(" zibby workflow logs <uuid> ")+r.gray("Tail execution logs")),console.log(""),i}catch(u){if(o.quiet)return[];console.log(r.red(`
26
26
  Error reading workflows: ${u.message}
27
27
  `)),process.exit(1)}}async function pe(o={}){let e=await bo({...o,quiet:!0}),{sessionToken:t,apiKey:l}=await q({apiKey:o.apiKey,optional:!0}),u=[],i=t||l;if(i)try{let n=I(),p=await fetch(`${n}/projects`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`}});if(p.ok){let h=(await p.json()).projects||[];for(let v of h){let B=await fetch(`${n}/projects/${v.projectId}/workflows`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`}});if(B.ok){let H=await B.json(),X=["analysis","implementation","run_test"];for(let k of H)X.includes(k.workflowType)||u.push({uuid:k.uuid,name:k.workflowType,projectId:v.projectId,projectName:v.name||v.projectId,version:k.version||0,nodes:k.graphJson?JSON.parse(k.graphJson).nodes?.length:0,updatedAt:k.updatedAt||null,source:"remote"})}}}}catch{}let c=[],a=new Map(e.map(n=>[n.name,n]));for(let n of u)c.push({uuid:n.uuid,name:n.name,project:n.projectName,version:n.version});for(let[n,p]of a.entries())u.some(h=>h.name===n)||c.push({uuid:null,name:p.name,project:"-",version:"-"});if(c.length===0){console.log(`
28
28
  No workflows found
@@ -31,5 +31,5 @@ import{readdir as wo,stat as ho}from"fs/promises";import{existsSync as P}from"fs
31
31
  Workflows
32
32
  `),console.log(` ${T}`),console.log(` ${r.bold(m)}`),console.log(` ${x}`);for(let n of y){let p=g.map(d=>{let h=String(n[d]).padEnd(w[d]);return` ${d==="uuid"&&n.uuid!=="-"?r.cyan(h):d==="name"?r.white(h):h} `});console.log(` \u2502${p.join("\u2502")}\u2502`)}console.log(` ${$}`),console.log(`
33
33
  Total: ${c.length} workflow${c.length===1?"":"s"}
34
- `),console.log(" Commands:"),console.log(" zibby workflow new <name> Scaffold a new workflow"),console.log(" zibby workflow start <name> Run locally"),console.log(" zibby workflow deploy <name> Ship to cloud (generates UUID on first deploy)"),console.log(" zibby workflow trigger <uuid> Run a deployed workflow remotely"),console.log(" zibby workflow logs <uuid> Tail logs from a remote run"),console.log(" zibby workflow download <uuid> Pull remote back to local (then edit + redeploy)"),console.log(" zibby workflow delete <uuid> Delete a deployed workflow"),i||console.log(`
34
+ `),console.log(" Commands:"),console.log(" zibby workflow new <name> Scaffold a new workflow"),console.log(" zibby workflow run <name> Run locally"),console.log(" zibby workflow deploy <name> Ship to cloud (generates UUID on first deploy)"),console.log(" zibby workflow trigger <uuid> Run a deployed workflow remotely"),console.log(" zibby workflow logs <uuid> Tail logs from a remote run"),console.log(" zibby workflow download <uuid> Pull remote back to local (then edit + redeploy)"),console.log(" zibby workflow delete <uuid> Delete a deployed workflow"),i||console.log(`
35
35
  Set ZIBBY_API_KEY to see remote workflows`),console.log("")}export{pe as listAllWorkflowsCommand,bo as listLocalWorkflowsCommand};
@@ -1,42 +1,42 @@
1
- import e from"chalk";import{readFileSync as v,existsSync as N}from"fs";import{homedir as R}from"os";import{join as C}from"path";var x="https://logs.workflows.zibby.app",j="https://logs-stream.zibby.app/",b=null;async function _(i){return b||(process.env.ZIBBY_SSE_ENDPOINT?(b=process.env.ZIBBY_SSE_ENDPOINT,b):(b=j,b))}function A(i){let r=C(R(),".zibby","config.json");N(r)||(console.log(e.red(`
1
+ import e from"chalk";import{readFileSync as v,existsSync as N,writeSync as R}from"fs";import{homedir as C}from"os";import{join as _}from"path";var x="https://logs.workflows.zibby.app",j="https://logs-stream.zibby.app/",b=null;async function A(i){return b||(process.env.ZIBBY_SSE_ENDPOINT?(b=process.env.ZIBBY_SSE_ENDPOINT,b):(b=j,b))}function L(i){let r=_(C(),".zibby","config.json");N(r)||(console.log(e.red(`
2
2
  Not authenticated`)),console.log(e.gray(` Run: zibby login
3
- `)),process.exit(1));let n;try{n=JSON.parse(v(r,"utf-8"))}catch{console.log(e.red(`
3
+ `)),process.exit(1));let s;try{s=JSON.parse(v(r,"utf-8"))}catch{console.log(e.red(`
4
4
  Config file corrupt`)),console.log(e.gray(` Run: zibby login
5
- `)),process.exit(1)}let o=n.sessionToken;o||(console.log(e.red(`
5
+ `)),process.exit(1)}let o=s.sessionToken;o||(console.log(e.red(`
6
6
  Not authenticated`)),console.log(e.gray(` Run: zibby login
7
- `)),process.exit(1));let g=i.project;return{token:o,projectId:g}}function k(i){return new Date(i).toISOString().replace("T"," ").replace("Z","")}async function E(i,r){let n=await fetch(i,{headers:{Authorization:`Bearer ${r}`}});if(!n.ok){let o=await n.text();throw new Error(`API ${n.status}: ${o}`)}return n.json()}async function O(i,r,n,o){return i||(console.log(e.red(`
7
+ `)),process.exit(1));let f=i.project;return{token:o,projectId:f}}function k(i){return new Date(i).toISOString().replace("T"," ").replace("Z","")}async function E(i,r){let s=await fetch(i,{headers:{Authorization:`Bearer ${r}`}});if(!s.ok){let o=await s.text();throw new Error(`API ${s.status}: ${o}`)}return s.json()}async function O(i,r,s,o){return i||(console.log(e.red(`
8
8
  Workflow UUID is required`)),console.log(e.gray(" Usage: zibby workflow logs <workflow-uuid>")),console.log(e.gray(` zibby workflow logs <workflow-uuid> -t
9
9
  `)),process.exit(1)),i}function F(i,r){let o=((i||"")+r).split(`
10
- `),g=o.pop()||"",a=[],u=null,s=null;for(let d=0;d<o.length;d++){let l=o[d];if(l.trim()){if(l.startsWith("id:")){u=l.slice(3).trim();continue}if(l.startsWith("event:")){let f=l.slice(6).trim();if(f==="log")continue;if(f==="status"){let c=o[d+1];if(c&&c.startsWith("data:"))try{let y=JSON.parse(c.slice(5).trim());y.status==="new_execution"?a.push({type:"newExecution",executionId:y.executionId,taskId:y.taskId}):y.status==="waiting"&&a.push({type:"waiting"})}catch{}continue}if(f==="complete"){a.push({type:"complete"});continue}if(f==="error"){let c=o[d+1];if(c&&c.startsWith("data:"))try{if(JSON.parse(c.slice(5).trim()).error==="No executions found for workflow"){s={type:"notFound"};break}}catch{}s={type:"failed"};break}continue}if(l.startsWith("data:")){let f=l.slice(5).trim();if(!f)continue;try{let c=JSON.parse(f);c.timestamp&&c.message&&a.push({type:"log",timestamp:c.timestamp,message:c.message,taskId:c.taskId})}catch{}}}}return{actions:a,remainder:g,lastEventId:u,returnSignal:s}}async function J({token:i,executionId:r,sseEndpoint:n,stopped:o}){let g=null;try{let a=new URL(n);a.searchParams.set("jobId",r),g&&a.searchParams.set("lastEventId",g);let u=await fetch(a.toString(),{headers:{Authorization:`Bearer ${i}`,Accept:"text/event-stream"}});if(!u.ok)throw new Error(`SSE connection failed: ${u.status} ${u.statusText}`);let s=u.body.getReader(),d=new TextDecoder,l="",f=!1;for(;!o.value;){let{done:c,value:y}=await s.read();if(c)break;let w=F(l,d.decode(y,{stream:!0}));l=w.remainder,w.lastEventId&&(g=w.lastEventId);for(let t of w.actions)switch(t.type){case"newExecution":{let m=`${t.executionId.slice(0,8)}...${t.executionId.slice(-4)}`,h=t.taskId?t.taskId.slice(-8):"pending";console.log(e.cyan(`
10
+ `),f=o.pop()||"",a=[],u=null,n=null;for(let d=0;d<o.length;d++){let l=o[d];if(l.trim()){if(l.startsWith("id:")){u=l.slice(3).trim();continue}if(l.startsWith("event:")){let g=l.slice(6).trim();if(g==="log")continue;if(g==="status"){let c=o[d+1];if(c&&c.startsWith("data:"))try{let y=JSON.parse(c.slice(5).trim());y.status==="new_execution"?a.push({type:"newExecution",executionId:y.executionId,taskId:y.taskId}):y.status==="waiting"&&a.push({type:"waiting"})}catch{}continue}if(g==="complete"){a.push({type:"complete"});continue}if(g==="error"){let c=o[d+1];if(c&&c.startsWith("data:"))try{if(JSON.parse(c.slice(5).trim()).error==="No executions found for workflow"){n={type:"notFound"};break}}catch{}n={type:"failed"};break}continue}if(l.startsWith("data:")){let g=l.slice(5).trim();if(!g)continue;try{let c=JSON.parse(g);c.timestamp&&c.message&&a.push({type:"log",timestamp:c.timestamp,message:c.message,taskId:c.taskId})}catch{}}}}return{actions:a,remainder:f,lastEventId:u,returnSignal:n}}async function J({token:i,executionId:r,sseEndpoint:s,stopped:o}){let f=null;try{let a=new URL(s);a.searchParams.set("jobId",r),f&&a.searchParams.set("lastEventId",f);let u=await fetch(a.toString(),{headers:{Authorization:`Bearer ${i}`,Accept:"text/event-stream"}});if(!u.ok)throw new Error(`SSE connection failed: ${u.status} ${u.statusText}`);let n=u.body.getReader(),d=new TextDecoder,l="",g=!1;for(;!o.value;){let{done:c,value:y}=await n.read();if(c)break;let w=F(l,d.decode(y,{stream:!0}));l=w.remainder,w.lastEventId&&(f=w.lastEventId);for(let t of w.actions)switch(t.type){case"newExecution":{let m=`${t.executionId.slice(0,8)}...${t.executionId.slice(-4)}`,h=t.taskId?t.taskId.slice(-8):"pending";console.log(e.cyan(`
11
11
  \u250C\u2500 Execution: ${m} (task: ${h})`)),console.log(e.cyan(` \u2514\u2500 Streaming logs...
12
12
  `));break}case"waiting":console.log(e.gray(`
13
- Waiting for next execution...`));break;case"complete":f=!0;break;case"log":{let m=e.gray(k(t.timestamp)),h=t.taskId?e.gray(`(${t.taskId.slice(-8)}) `):"";console.log(`${m} ${h}${t.message.replace(/\n$/,"")}`);break}}if(w.returnSignal)return w.returnSignal.type==="notFound"?{notFound:!0}:{failed:!0}}return{completed:f}}catch(a){if(a.name==="AbortError")return{aborted:!0};throw a}}function M(i,{baseMs:r=500,capMs:n=3e4,rand:o=Math.random}={}){let g=Math.min(n,r*Math.pow(2,Math.max(0,i)));return Math.floor(o()*g)}async function z({attemptStream:i,stopped:r,follow:n,logger:o,sleep:g=d=>new Promise(l=>setTimeout(l,d)),exit:a=d=>{throw new Error(`exit:${d}`)},backoff:u=M,notFoundPollMs:s=5e3}){let d=0,l=!1;for(;!r.value;){let f;try{f=await i(),d=0}catch(c){if(c.name==="AbortError"||r.value)return{reason:"aborted"};if(l||(o.error(` SSE Error: ${c.message}`),n&&o.gray(" Reconnecting..."),l=!0),!n)return a("error")??{reason:"error"};let y=u(d);d++,await g(y);continue}if(f.aborted||r.value)return{reason:"aborted"};if(f.notFound){if(n){l||(o.yellow(" No executions found yet. Waiting for workflow to be triggered..."),o.gray(" Press Ctrl+C to stop."),l=!0),await g(s);continue}return o.yellow(`
13
+ Waiting for next execution...`));break;case"complete":g=!0;break;case"log":{let m=e.gray(k(t.timestamp)),h=t.taskId?e.gray(`(${t.taskId.slice(-8)}) `):"";console.log(`${m} ${h}${t.message.replace(/\n$/,"")}`);break}}if(w.returnSignal)return w.returnSignal.type==="notFound"?{notFound:!0}:{failed:!0}}return{completed:g}}catch(a){if(a.name==="AbortError")return{aborted:!0};throw a}}function M(i,{baseMs:r=500,capMs:s=3e4,rand:o=Math.random}={}){let f=Math.min(s,r*Math.pow(2,Math.max(0,i)));return Math.floor(o()*f)}async function z({attemptStream:i,stopped:r,follow:s,logger:o,sleep:f=d=>new Promise(l=>setTimeout(l,d)),exit:a=d=>{throw new Error(`exit:${d}`)},backoff:u=M,notFoundPollMs:n=5e3}){let d=0,l=!1;for(;!r.value;){let g;try{g=await i(),d=0}catch(c){if(c.name==="AbortError"||r.value)return{reason:"aborted"};if(l||(o.error(` SSE Error: ${c.message}`),s&&o.gray(" Reconnecting..."),l=!0),!s)return a("error")??{reason:"error"};let y=u(d);d++,await f(y);continue}if(g.aborted||r.value)return{reason:"aborted"};if(g.notFound){if(s){l||(o.yellow(" No executions found yet. Waiting for workflow to be triggered..."),o.gray(" Press Ctrl+C to stop."),l=!0),await f(n);continue}return o.yellow(`
14
14
  No executions found for this workflow. Trigger the workflow first.
15
15
  `),a("notFound")??{reason:"notFound"}}if(l&&(o.gray(` Reconnected.
16
- `),l=!1),f.failed)return o.red(`
17
- Execution failed.`),n?{reason:"failed"}:a("failed")??{reason:"failed"};if(f.completed)return a("completed")??{reason:"completed"};if(!n)return{reason:"disconnected"}}return{reason:"stopped"}}async function L({token:i,jobId:r,follow:n,projectId:o}){console.log(e.gray(` Streaming logs for workflow ${e.cyan(r)}...`)),console.log(n?e.gray(` Press Ctrl+C to stop.
18
- `):"");let g=await _(i);if(!g)return console.log(e.yellow(` SSE endpoint not configured, using CloudWatch polling...
19
- `)),T({token:i,projectId:null,jobId:r,follow:n,limit:1e5});let a={value:!1},u=()=>{a.value=!0,console.log(e.gray(`
16
+ `),l=!1),g.failed)return o.red(`
17
+ Execution failed.`),s?{reason:"failed"}:a("failed")??{reason:"failed"};if(g.completed)return a("completed")??{reason:"completed"};if(!s)return{reason:"disconnected"}}return{reason:"stopped"}}async function W({token:i,jobId:r,follow:s,projectId:o}){console.log(e.gray(` Streaming logs for workflow ${e.cyan(r)}...`)),console.log(s?e.gray(` Press Ctrl+C to stop.
18
+ `):"");let f=await A(i);if(!f)return console.log(e.yellow(` SSE endpoint not configured, using CloudWatch polling...
19
+ `)),T({token:i,projectId:null,jobId:r,follow:s,limit:1e5});let a={value:!1},u=()=>{a.value=!0;try{R(2,`
20
20
  Stopped streaming.
21
- `)),process.exit(0)};process.on("SIGINT",u),process.on("SIGTERM",u),await z({attemptStream:()=>J({token:i,executionId:r,sseEndpoint:g,stopped:a}),stopped:a,follow:n,logger:{gray:s=>console.log(e.gray(s)),red:s=>console.log(e.red(s)),yellow:s=>console.log(e.yellow(s)),error:s=>console.error(e.red(s))},exit:s=>{s==="completed"&&process.exit(0),(s==="error"||s==="notFound"||s==="failed")&&process.exit(1)}})}async function T({token:i,projectId:r,jobId:n,follow:o,limit:g}){let a=r?`${x}/logs/${r}/${n}`:`${x}/job/${n}`,u=null,s=0,d=new Set,l=!1,f=0,c=5,y=()=>{l=!0,console.log(e.gray(`
21
+ `)}catch{}process.exit(0)};process.prependListener("SIGINT",u),process.prependListener("SIGTERM",u),await z({attemptStream:()=>J({token:i,executionId:r,sseEndpoint:f,stopped:a}),stopped:a,follow:s,logger:{gray:n=>console.log(e.gray(n)),red:n=>console.log(e.red(n)),yellow:n=>console.log(e.yellow(n)),error:n=>console.error(e.red(n))},exit:n=>{n==="completed"&&process.exit(0),(n==="error"||n==="notFound"||n==="failed")&&process.exit(1)}})}async function T({token:i,projectId:r,jobId:s,follow:o,limit:f}){let a=r?`${x}/logs/${r}/${s}`:`${x}/job/${s}`,u=null,n=0,d=new Set,l=!1,g=0,c=5,y=()=>{l=!0,console.log(e.gray(`
22
22
  Stopped tailing.
23
- `)),process.exit(0)};for(process.on("SIGINT",y),process.on("SIGTERM",y),console.log(e.gray(` Fetching logs for workflow ${e.cyan(n)}...`)),console.log(o?e.gray(` Press Ctrl+C to stop.
24
- `):"");!l;)try{let w=new URLSearchParams({limit:String(g)});u&&w.set("nextToken",u);let t=await E(`${a}?${w}`,i);f=0,t.message&&t.lines?.length===0&&s===0&&console.log(e.gray(` ${t.message}`)),t.status==="starting"&&t.lines?.length===0&&s===0&&console.log(e.gray(" Container starting..."));for(let p of t.lines||[]){let $=`${p.timestamp}:${p.message}`;if(d.has($))continue;d.add($);let I=e.gray(k(p.timestamp)),P=t.taskId?e.gray(`(${t.taskId.slice(-8)}) `):"";console.log(`${I} ${P}${p.message.replace(/\n$/,"")}`)}if(s=t.lines?.length>0?0:s+1,u=t.nextForwardToken||null,t.status==="completed"||t.status==="failed"){let p=t.status==="completed"?e.green:e.red;console.log(p(`
23
+ `)),process.exit(0)};for(process.on("SIGINT",y),process.on("SIGTERM",y),console.log(e.gray(` Fetching logs for workflow ${e.cyan(s)}...`)),console.log(o?e.gray(` Press Ctrl+C to stop.
24
+ `):"");!l;)try{let w=new URLSearchParams({limit:String(f)});u&&w.set("nextToken",u);let t=await E(`${a}?${w}`,i);g=0,t.message&&t.lines?.length===0&&n===0&&console.log(e.gray(` ${t.message}`)),t.status==="starting"&&t.lines?.length===0&&n===0&&console.log(e.gray(" Container starting..."));for(let p of t.lines||[]){let $=`${p.timestamp}:${p.message}`;if(d.has($))continue;d.add($);let I=e.gray(k(p.timestamp)),P=t.taskId?e.gray(`(${t.taskId.slice(-8)}) `):"";console.log(`${I} ${P}${p.message.replace(/\n$/,"")}`)}if(n=t.lines?.length>0?0:n+1,u=t.nextForwardToken||null,t.status==="completed"||t.status==="failed"){let p=t.status==="completed"?e.green:e.red;console.log(p(`
25
25
  Job ${t.status}.`)),process.exit(t.status==="completed"?0:1)}if(!o){t.status&&console.log(e.gray(`
26
- Status: ${t.status}`));break}let S=t.lines?.length>0?500:s>5?5e3:2e3;await new Promise(p=>setTimeout(p,S))}catch(w){if(w.name==="AbortError")break;w.message.match(/API (400|401|403|404):/)&&(console.error(e.red(`
26
+ Status: ${t.status}`));break}let S=t.lines?.length>0?500:n>5?5e3:2e3;await new Promise(p=>setTimeout(p,S))}catch(w){if(w.name==="AbortError")break;w.message.match(/API (400|401|403|404):/)&&(console.error(e.red(`
27
27
  ${w.message}
28
- `)),process.exit(1)),f++,console.error(e.red(` Error: ${w.message}`)),f>=c&&(console.error(e.red(`
28
+ `)),process.exit(1)),g++,console.error(e.red(` Error: ${w.message}`)),g>=c&&(console.error(e.red(`
29
29
  Too many consecutive errors (${c}). Stopping.
30
- `)),process.exit(1)),o||process.exit(1),await new Promise(m=>setTimeout(m,3e3))}}async function U({token:i,projectId:r,workflow:n,follow:o,limit:g}){let a=`${x}/all/${r}`,u=null,s=0,d=new Set,l=null,f=!1,c=0,y=5,w=()=>{f=!0,console.log(e.gray(`
30
+ `)),process.exit(1)),o||process.exit(1),await new Promise(m=>setTimeout(m,3e3))}}async function U({token:i,projectId:r,workflow:s,follow:o,limit:f}){let a=`${x}/all/${r}`,u=null,n=0,d=new Set,l=null,g=!1,c=0,y=5,w=()=>{g=!0,console.log(e.gray(`
31
31
  Stopped tailing.
32
32
  `)),process.exit(0)};for(process.on("SIGINT",w),process.on("SIGTERM",w),console.log(e.gray(`
33
- Tailing all runs for ${e.cyan(n)}...`)),console.log(o?e.gray(` Press Ctrl+C to stop.
34
- `):"");!f;)try{let t=new URLSearchParams({workflow:n,limit:String(g)});u&&t.set("nextToken",u);let m=await E(`${a}?${t}`,i);c=0,m.message&&m.lines?.length===0&&s===0&&console.log(e.gray(` ${m.message}`));for(let p of m.lines||[]){let $=`${p.timestamp}:${p.jobId}:${p.message}`;if(d.has($))continue;d.add($),p.jobId!==l&&(l!==null&&console.log(""),console.log(e.dim(` \u2500\u2500 ${p.jobId} \u2500\u2500`)),l=p.jobId);let I=e.gray(k(p.timestamp));console.log(`${I} ${p.message.replace(/\n$/,"")}`)}if(s=m.lines?.length>0?0:s+1,u=m.nextToken||null,!o){u&&console.log(e.gray(`
35
- ... more logs available. Run again or use --follow to stream.`)),m.jobCount&&console.log(e.gray(` ${m.jobCount} job(s) found.`));break}if(!m.hasRunning&&!u&&s>2){console.log(e.gray(`
36
- No running jobs. All caught up.`));break}let S=m.lines?.length>0?500:s>5?5e3:2e3;await new Promise(p=>setTimeout(p,S))}catch(t){if(t.name==="AbortError")break;t.message.match(/API (400|401|403|404):/)&&(console.error(e.red(`
33
+ Tailing all runs for ${e.cyan(s)}...`)),console.log(o?e.gray(` Press Ctrl+C to stop.
34
+ `):"");!g;)try{let t=new URLSearchParams({workflow:s,limit:String(f)});u&&t.set("nextToken",u);let m=await E(`${a}?${t}`,i);c=0,m.message&&m.lines?.length===0&&n===0&&console.log(e.gray(` ${m.message}`));for(let p of m.lines||[]){let $=`${p.timestamp}:${p.jobId}:${p.message}`;if(d.has($))continue;d.add($),p.jobId!==l&&(l!==null&&console.log(""),console.log(e.dim(` \u2500\u2500 ${p.jobId} \u2500\u2500`)),l=p.jobId);let I=e.gray(k(p.timestamp));console.log(`${I} ${p.message.replace(/\n$/,"")}`)}if(n=m.lines?.length>0?0:n+1,u=m.nextToken||null,!o){u&&console.log(e.gray(`
35
+ ... more logs available. Run again or use --follow to stream.`)),m.jobCount&&console.log(e.gray(` ${m.jobCount} job(s) found.`));break}if(!m.hasRunning&&!u&&n>2){console.log(e.gray(`
36
+ No running jobs. All caught up.`));break}let S=m.lines?.length>0?500:n>5?5e3:2e3;await new Promise(p=>setTimeout(p,S))}catch(t){if(t.name==="AbortError")break;t.message.match(/API (400|401|403|404):/)&&(console.error(e.red(`
37
37
  ${t.message}
38
38
  `)),process.exit(1)),c++,console.error(e.red(` Error: ${t.message}`)),c>=y&&(console.error(e.red(`
39
39
  Too many consecutive errors (${y}). Stopping.
40
- `)),process.exit(1)),o||process.exit(1),await new Promise(h=>setTimeout(h,3e3))}}async function H(i,r){let{token:n,projectId:o}=A(r),g=r.follow===!0,a=r.lines?parseInt(r.lines,10):1e5;if(r.all){let s=r.workflow;return s||(console.log(e.red(`
40
+ `)),process.exit(1)),o||process.exit(1),await new Promise(h=>setTimeout(h,3e3))}}async function V(i,r){let{token:s,projectId:o}=L(r),f=r.follow===!0,a=r.lines?parseInt(r.lines,10):1e5;if(r.all){let n=r.workflow;return n||(console.log(e.red(`
41
41
  --workflow is required with --all`)),console.log(e.gray(` Example: zibby workflow logs --workflow ticket-triage --all --project <id>
42
- `)),process.exit(1)),U({token:n,projectId:o,workflow:s,follow:g,limit:a})}let u=await O(i,r,n,o);return g?L({token:n,jobId:u,follow:g,projectId:o}):T({token:n,projectId:o,jobId:u,follow:!1,limit:a})}export{H as logsCommand,F as parseSseChunk,z as runReconnectLoop,M as sseBackoffMs};
42
+ `)),process.exit(1)),U({token:s,projectId:o,workflow:n,follow:f,limit:a})}let u=await O(i,r,s,o);return f?W({token:s,jobId:u,follow:f,projectId:o}):T({token:s,projectId:o,jobId:u,follow:!1,limit:a})}export{V as logsCommand,F as parseSseChunk,z as runReconnectLoop,M as sseBackoffMs};
@@ -0,0 +1,19 @@
1
+ import{existsSync as d}from"fs";import{readFile as U}from"fs/promises";import{join as g}from"path";import{pathToFileURL as W}from"url";import s from"chalk";import D from"ora";import{existsSync as F}from"fs";import{join as N}from"path";import{pathToFileURL as S}from"url";async function k(o){let r=N(o,".zibby.config.mjs");if(!F(r))throw new Error(".zibby.config.mjs not found");try{let e=await import(S(r).href);return e.default||e}catch(e){throw new Error(`Failed to load .zibby.config.mjs: ${e.message}`,{cause:e})}}import{existsSync as B,readFileSync as I}from"fs";import{resolve as P}from"path";function z(o){return o==="true"?!0:o==="false"?!1:o==="null"?null:o!==""&&!isNaN(Number(o))?Number(o):o}function L(o){let r={};for(let e of o||[]){let t=e.indexOf("=");if(t===-1){console.warn(` Warning: ignored param "${e}" \u2014 expected key=value format`);continue}let l=e.slice(0,t).trim(),a=z(e.slice(t+1)),n=l.split("."),c=r;for(let i=0;i<n.length-1;i++)(typeof c[n[i]]!="object"||c[n[i]]===null)&&(c[n[i]]={}),c=c[n[i]];c[n[n.length-1]]=a}return r}function O(o){let r=P(o);B(r)||(console.log(`
2
+ Error: --input-file not found: ${o}
3
+ `),process.exit(1));try{return JSON.parse(I(r,"utf-8"))}catch(e){console.log(`
4
+ Error: --input-file is not valid JSON: ${e.message}
5
+ `),process.exit(1)}}function x(o){let r={};if(o.inputFile&&(r={...O(o.inputFile)}),o.input)try{r={...r,...JSON.parse(o.input)}}catch(e){console.log(`
6
+ Error: --input is not valid JSON`),console.log(` ${e.message}
7
+ `),process.exit(1)}return o.param?.length&&(r={...r,...L(o.param)}),r}function j({workflowType:o,jobId:r,projectId:e,agentType:t,model:l}){let a="\u2500".repeat(60),n=`${t||"default"} (model: ${l||"auto"})`;return["",a,` Workflow: ${o}`,` Job: ${r||"local"}`,` Project: ${e||"none"}`,` Agent: ${n}`,a].join(`
8
+ `)}async function G(o){try{let r=await k(o);return{userConfig:r,workflowsBasePath:r?.paths?.workflows||".zibby/workflows"}}catch{return{userConfig:null,workflowsBasePath:".zibby/workflows"}}}async function J(o,r,e){let t=g(o,"graph.mjs");if(!d(t))throw new Error(`graph.mjs not found in ${e}/${r}/`);let l=await T(o,r),a=await import(W(t).href),n=l.entryClass,c=n&&a[n]||a.default||Object.values(a).find(i=>typeof i=="function"&&i.prototype?.buildGraph);if(!c)throw new Error("No WorkflowAgent class found in graph.mjs. Export a class with buildGraph() method.");return{AgentClass:c,manifest:l}}async function T(o,r){let e=g(o,"workflow.json");if(!d(e))return{name:r,triggers:{api:!0}};let t=await U(e,"utf-8");return JSON.parse(t)}async function so(o,r={}){o||(console.log(s.red(`
9
+ Workflow name is required`)),console.log(s.gray(" Usage: zibby workflow run <workflow-name>")),console.log(s.gray(` Example: zibby workflow run my-pipeline -p ticket=BUG-123
10
+ `)),process.exit(1));let e=o.toLowerCase(),t=process.cwd(),{userConfig:l,workflowsBasePath:a}=await G(t),n=g(t,a,e);d(n)||(console.log(s.red(`
11
+ Workflow not found: ${a}/${e}/`)),console.log(s.gray(" Create one first:")),console.log(s.cyan(` zibby workflow new ${e}
12
+ `)),process.exit(1));let c=x(r),i=D(` Loading workflow "${e}"...`).start(),m,w;try{({AgentClass:m,manifest:w}=await J(n,e,a)),i.succeed(` Loaded ${s.bold(w.entryClass||m.name)} (${e})`)}catch(u){i.fail(" Failed to load workflow"),console.log(s.red(`
13
+ ${u.message}
14
+ `)),process.exit(1)}let y=`local-${Date.now()}`,C=j({workflowType:e,jobId:y,projectId:l?.projectId,agentType:l?.agent?.provider||process.env.AGENT_TYPE,model:l?.agent?.model||process.env.MODEL});console.log(C);let h=Date.now(),E={input:c,cwd:t,runId:y,config:l||{}},f,p;try{p=new m({workflow:e}),f=await p.buildGraph().run(p,E)}catch(u){let v=((Date.now()-h)/1e3).toFixed(1);console.log(s.red(`
15
+ \u2716 ${e} failed after ${v}s`)),console.log(s.red(` ${u.message}
16
+ `)),process.env.ZIBBY_DEBUG&&console.error(u.stack),process.exit(1)}let $=((Date.now()-h)/1e3).toFixed(1),b=f?.success!==!1;b?console.log(s.green(`
17
+ \u2714 ${e} completed in ${$}s`)):(console.log(s.red(`
18
+ \u2716 ${e} failed after ${$}s`)),f?.error&&console.log(s.red(` ${f.error}
19
+ `))),p.onComplete&&await p.onComplete(f),b||process.exit(1)}export{so as runLocalWorkflowCommand};
@@ -1,6 +1,6 @@
1
1
  #!/usr/bin/env node
2
- import{mkdirSync as W,writeFileSync as Y,existsSync as w}from"fs";import{join as g,dirname as Z,resolve as B}from"path";import{pathToFileURL as A}from"url";import{execSync as ee,spawn as D}from"node:child_process";import{SQSClient as H,SendMessageCommand as M}from"@aws-sdk/client-sqs";var P=null;function q(){return P||(P=new H({region:process.env.AWS_REGION||"ap-southeast-2"})),P}async function F(i,{status:e,error:o}){let{EXECUTION_ID:n,SQS_AUTH_TOKEN:c,PROGRESS_API_URL:r,PROGRESS_QUEUE_URL:s,PROJECT_API_TOKEN:a}=i;if(!n)return;let f={executionId:n,...c&&{sqsAuthToken:c},status:e,...o&&{error:o},timestamp:new Date().toISOString()},l=r?"HTTP":s?"SQS":"NONE",u=JSON.stringify(f).length;console.log(`Sending final status: ${e} via ${l} (${(u/1024).toFixed(1)}KB)`);try{if(r)await X(r,n,f,a);else if(s){let d=["completed","failed","insufficient_context","blocked"].includes(e)?"execution_completed":"progress_update";await V(s,n,f,d)}else{console.warn("No transport configured for final status \u2014 neither PROGRESS_API_URL nor PROGRESS_QUEUE_URL set");return}console.log(`Final status ${e} sent via ${l}`)}catch(d){console.error(`Failed to send final status (${e}) via ${l}:`),console.error(` Payload: ${(u/1024).toFixed(1)}KB`),console.error(` Error: ${d.message}`),d.name&&console.error(` Error type: ${d.name}`),d.code&&console.error(` Error code: ${d.code}`)}}async function X(i,e,o,n){let c=`${i}/${e}/progress`,r={"Content-Type":"application/json"};n&&(r.Authorization=`Bearer ${n}`);let s=await fetch(c,{method:"POST",headers:r,body:JSON.stringify(o)});if(!s.ok){let a=await s.text();throw new Error(`HTTP ${s.status}: ${a}`)}}async function V(i,e,o,n="progress_update"){let c=JSON.stringify(o),r=(c.length/1024).toFixed(1);c.length>256*1024&&console.error(`\u274C SQS message too large: ${r}KB (limit 256KB) for ${e} [${n}]`),await q().send(new M({QueueUrl:i,MessageBody:c,MessageGroupId:e,MessageAttributes:{executionId:{DataType:"String",StringValue:e},messageType:{DataType:"String",StringValue:n}}}))}function L({workflowType:i,jobId:e,projectId:o,agentType:n,model:c}){let r="\u2500".repeat(60),s=`${n||"default"} (model: ${c||"auto"})`;return["",r,` Workflow: ${i}`,` Job: ${e||"local"}`,` Project: ${o||"none"}`,` Agent: ${s}`,r].join(`
3
- `)}import"@zibby/core";var T=process.env.WORKSPACE||"/workspace";async function oe(i,e){W(e,{recursive:!0});let o=Date.now();console.log("[setup] Fetching bundle...");let n=setInterval(()=>{let r=((Date.now()-o)/1e3).toFixed(1);console.log(`[setup] still fetching (${r}s elapsed)`)},3e3);try{await new Promise((r,s)=>{let a=D("curl",["-fsSL",i],{stdio:["ignore","pipe","inherit"]}),f=D("tar",["-xzf","-","-C",e],{stdio:["pipe","inherit","inherit"]});a.stdout.pipe(f.stdin);let l,u,d=()=>{if(l!==void 0&&u!==void 0){if(l!==0)return s(new Error(`curl exited ${l}`));if(u!==0)return s(new Error(`tar exited ${u}`));r()}};a.on("close",S=>{l=S,d()}),f.on("close",S=>{u=S,d()}),a.on("error",s),f.on("error",s)})}finally{clearInterval(n)}let c=((Date.now()-o)/1e3).toFixed(1);return console.log(`[setup] Bundle extracted (${c}s)`),e}async function K(){let i=process.env.WORKFLOW_SOURCES_URL;if(!i)throw new Error("WORKFLOW_SOURCES_URL env var is required");let e=await fetch(i);if(!e.ok)throw new Error(`Failed to fetch sources: ${e.status} ${e.statusText}`);let o=await e.json();if(!o.sources||typeof o.sources!="object")throw new Error('Invalid sources payload \u2014 missing "sources" map');return o}function te(i,e){let o=B(e),n=0;for(let[c,r]of Object.entries(i)){let s=B(e,c);if(!s.startsWith(`${o}/`)&&s!==o){console.error(` \u26D4 Skipping unsafe path: ${c}`);continue}W(Z(s),{recursive:!0}),Y(s,r,"utf-8"),n++}return n}async function ne(i,e){let o=g(i,"graph.mjs");if(!w(o))throw new Error(`graph.mjs not found at ${o}`);let n=await import(A(o).href),c=e?.entryClass,r=c&&n[c]||n.default||Object.values(n).find(s=>typeof s=="function"&&s.prototype?.buildGraph);if(!r)throw new Error("No WorkflowAgent class found in graph.mjs");return r}async function ge(){if(!process.env.NODE_PATH){process.env.NODE_PATH="/opt/zibby/packages";let t=await import("module");t.default._initPaths&&t.default._initPaths()}let{WORKFLOW_JOB_ID:i,WORKFLOW_TYPE:e,PROJECT_ID:o,AGENT_TYPE:n,MODEL:c}=process.env;e||(console.error("Missing WORKFLOW_TYPE env var"),process.exit(1));let r=L({workflowType:e,jobId:i,projectId:o,agentType:n,model:c});console.log(r);let s=process.env.WORKFLOW_BUNDLE_URL,a,f={},l,u;if(s){l=e,a=g(T,".zibby","workflows",l);try{await oe(s,a);try{let t=await K();f=t.input||{},u=t.version}catch(t){console.warn(`[setup] Could not fetch input payload: ${t.message}`)}}catch(t){console.warn(`[setup] Bundle extract failed (${t.message}); falling back to source install`),a=null}}if(!a){let t=await K(),{sources:p,input:y,workflowType:O,version:E}=t;f=y||{},l=O||e,u=E,console.log(`[setup] Workflow v${u||"?"} (${Object.keys(p).length} files)`),a=g(T,".zibby","workflows",l);let b=te(p,a);console.log(`[setup] Wrote ${b} files`),console.log("[setup] Installing dependencies...");try{ee("npm install --silent --no-audit --no-fund",{cwd:a,stdio:"inherit"}),console.log("[setup] Dependencies installed")}catch($){console.warn(`[setup] npm install failed: ${$.message}`)}}let d={},S=g(a,"workflow.json");if(w(S)){let{readFileSync:t}=await import("fs");d=JSON.parse(t(S,"utf-8"))}let I=await ne(a,d);console.log(`[setup] Loaded ${I.name}`);let m=[],x=g(a,"node_modules","@zibby","agent-workflow"),C=g(a,"node_modules","@zibby","core","node_modules","@zibby","agent-workflow");w(x)&&m.push({kind:"hoisted",path:x}),w(C)&&m.push({kind:"nested",path:C});let R=process.env.ZIBBY_RUN_DIAG==="1";if(R){let{readdirSync:t}=await import("fs");console.log(` [diag] @zibby/agent-workflow copies in bundle: ${m.length}`);for(let p of m)console.log(` [diag] ${p.kind}: ${p.path}`);try{let p=g(a,"node_modules","@zibby");w(p)&&console.log(` [diag] node_modules/@zibby/ contents: [${t(p).join(", ")}]`)}catch{}}let k=g(a,"node_modules","@zibby","core","dist","index.js");if(w(k)&&m.length>0)try{let t=await import(A(k).href),p=[t.AssistantStrategy,t.CursorAgentStrategy,t.ClaudeAgentStrategy,t.CodexAgentStrategy,t.GeminiAgentStrategy].filter(Boolean);for(let y of m){let O=g(y.path,"dist","index.js");if(!w(O))continue;let E=await import(A(O).href),b=R?E.listStrategies():null;for(let $ of p)try{E.registerStrategy(new $)}catch(J){console.warn(` register ${$.name} into ${y.kind} failed: ${J.message}`)}R&&console.log(` [diag] ${y.kind} registry: before=[${b.join(",")||"empty"}] after=[${E.listStrategies().join(",")||"empty"}]`)}console.log("[setup] Registered 5 agent strategies (assistant, cursor, claude, codex, gemini)")}catch(t){console.warn(`[setup] Failed to bridge strategies: ${t.message}`)}else console.warn("[setup] No @zibby/core or @zibby/agent-workflow in bundle \u2014 agent strategies may be unavailable");let G=Date.now(),_=new I({workflow:l||e}),z=_.buildGraph(),Q={input:f||{},cwd:T,runId:i||`run-${Date.now()}`};console.log("");let h;try{h=await z.run(_,Q)}catch(t){console.error(`
4
- Workflow execution failed: ${t.message}`),console.error(t.stack),await U("failed",t.message),process.exit(1)}let v=((Date.now()-G)/1e3).toFixed(1),j=h?.success!==!1,N=l||e;j?(console.log(`
5
- [done] ${N} completed in ${v}s`),await U("completed")):(console.error(`
6
- [done] ${N} failed after ${v}s`),await U("failed",h?.error||"Workflow execution failed"),process.exit(1)),_.onComplete&&await _.onComplete(h)}async function U(i,e=null){let o={EXECUTION_ID:process.env.WORKFLOW_JOB_ID,PROGRESS_API_URL:process.env.PROGRESS_API_URL,PROGRESS_QUEUE_URL:process.env.PROGRESS_QUEUE_URL,PROJECT_API_TOKEN:process.env.PROJECT_API_TOKEN,SQS_AUTH_TOKEN:process.env.SQS_AUTH_TOKEN};if(o.EXECUTION_ID)try{await F(o,{status:i,...e&&{error:e}})}catch(n){console.error(`\u26A0\uFE0F Failed to report status: ${n.message}`)}}export{ge as runWorkflowCommand};
2
+ import{mkdirSync as j,writeFileSync as oe,existsSync as w,readFileSync as K}from"fs";import{join as g,dirname as te,resolve as G}from"path";import{pathToFileURL as x}from"url";import{execSync as ne,spawn as W}from"node:child_process";import{SQSClient as Y,SendMessageCommand as X}from"@aws-sdk/client-sqs";var P=null;function V(){return P||(P=new Y({region:process.env.AWS_REGION||"ap-southeast-2"})),P}async function L(s,{status:e,error:o}){let{EXECUTION_ID:t,SQS_AUTH_TOKEN:c,PROGRESS_API_URL:i,PROGRESS_QUEUE_URL:r,PROJECT_API_TOKEN:a}=s;if(!t)return;let f={executionId:t,...c&&{sqsAuthToken:c},status:e,...o&&{error:o},timestamp:new Date().toISOString()},l=i?"HTTP":r?"SQS":"NONE",u=JSON.stringify(f).length;console.log(`Sending final status: ${e} via ${l} (${(u/1024).toFixed(1)}KB)`);try{if(i)await Z(i,t,f,a);else if(r){let d=["completed","failed","insufficient_context","blocked"].includes(e)?"execution_completed":"progress_update";await ee(r,t,f,d)}else{console.warn("No transport configured for final status \u2014 neither PROGRESS_API_URL nor PROGRESS_QUEUE_URL set");return}console.log(`Final status ${e} sent via ${l}`)}catch(d){console.error(`Failed to send final status (${e}) via ${l}:`),console.error(` Payload: ${(u/1024).toFixed(1)}KB`),console.error(` Error: ${d.message}`),d.name&&console.error(` Error type: ${d.name}`),d.code&&console.error(` Error code: ${d.code}`)}}async function Z(s,e,o,t){let c=`${s}/${e}/progress`,i={"Content-Type":"application/json"};t&&(i.Authorization=`Bearer ${t}`);let r=await fetch(c,{method:"POST",headers:i,body:JSON.stringify(o)});if(!r.ok){let a=await r.text();throw new Error(`HTTP ${r.status}: ${a}`)}}async function ee(s,e,o,t="progress_update"){let c=JSON.stringify(o),i=(c.length/1024).toFixed(1);c.length>256*1024&&console.error(`\u274C SQS message too large: ${i}KB (limit 256KB) for ${e} [${t}]`),await V().send(new X({QueueUrl:s,MessageBody:c,MessageGroupId:e,MessageAttributes:{executionId:{DataType:"String",StringValue:e},messageType:{DataType:"String",StringValue:t}}}))}function D({workflowType:s,jobId:e,projectId:o,agentType:t,model:c}){let i="\u2500".repeat(60),r=`${t||"default"} (model: ${c||"auto"})`;return["",i,` Workflow: ${s}`,` Job: ${e||"local"}`,` Project: ${o||"none"}`,` Agent: ${r}`,i].join(`
3
+ `)}import"@zibby/core";var T=process.env.WORKSPACE||"/workspace";async function se(s,e){j(e,{recursive:!0});let o=Date.now();console.log("[setup] Fetching bundle...");let t=setInterval(()=>{let i=((Date.now()-o)/1e3).toFixed(1);console.log(`[setup] still fetching (${i}s elapsed)`)},3e3);try{await new Promise((i,r)=>{let a=W("curl",["-fsSL",s],{stdio:["ignore","pipe","inherit"]}),f=W("tar",["-xzf","-","-C",e],{stdio:["pipe","inherit","inherit"]});a.stdout.pipe(f.stdin);let l,u,d=()=>{if(l!==void 0&&u!==void 0){if(l!==0)return r(new Error(`curl exited ${l}`));if(u!==0)return r(new Error(`tar exited ${u}`));i()}};a.on("close",S=>{l=S,d()}),f.on("close",S=>{u=S,d()}),a.on("error",r),f.on("error",r)})}finally{clearInterval(t)}let c=((Date.now()-o)/1e3).toFixed(1);return console.log(`[setup] Bundle extracted (${c}s)`),e}async function z(){let s=process.env.WORKFLOW_SOURCES_URL;if(!s)throw new Error("WORKFLOW_SOURCES_URL env var is required");let e=await fetch(s);if(!e.ok)throw new Error(`Failed to fetch sources: ${e.status} ${e.statusText}`);let o=await e.json();if(!o.sources||typeof o.sources!="object")throw new Error('Invalid sources payload \u2014 missing "sources" map');return o}function re(s,e){let o=G(e),t=0;for(let[c,i]of Object.entries(s)){let r=G(e,c);if(!r.startsWith(`${o}/`)&&r!==o){console.error(` \u26D4 Skipping unsafe path: ${c}`);continue}j(te(r),{recursive:!0}),oe(r,i,"utf-8"),t++}return t}async function ie(){let s=process.env.ZIBBY_EGRESS_PROXY_URL,e=process.env.ZIBBY_EGRESS_TOKEN;if(!(!s||!e))try{let o=await import("undici"),t=new o.ProxyAgent({uri:s,token:`Bearer ${e}`});o.setGlobalDispatcher(t),console.log(`[setup] Egress proxy active \u2192 ${s}`)}catch(o){console.warn(`[setup] Failed to install egress proxy dispatcher: ${o.message}`)}}async function ae(s,e){let o=g(s,"graph.mjs");if(!w(o))throw new Error(`graph.mjs not found at ${o}`);let t=await import(x(o).href),c=e?.entryClass,i=c&&t[c]||t.default||Object.values(t).find(r=>typeof r=="function"&&r.prototype?.buildGraph);if(!i)throw new Error("No WorkflowAgent class found in graph.mjs");return i}async function me(){if(!process.env.NODE_PATH){process.env.NODE_PATH="/opt/zibby/packages";let n=await import("module");n.default._initPaths&&n.default._initPaths()}await ie();let{WORKFLOW_JOB_ID:s,WORKFLOW_TYPE:e,PROJECT_ID:o,AGENT_TYPE:t,MODEL:c}=process.env;e||(console.error("Missing WORKFLOW_TYPE env var"),process.exit(1));let i=D({workflowType:e,jobId:s,projectId:o,agentType:t,model:c});console.log(i);let r=process.env.WORKFLOW_BUNDLE_URL,a,f={},l,u;if(r){l=e,a=g(T,".zibby","workflows",l);try{await se(r,a);try{let n=await z();f=n.input||{},u=n.version}catch(n){console.warn(`[setup] Could not fetch input payload: ${n.message}`)}}catch(n){console.warn(`[setup] Bundle extract failed (${n.message}); falling back to source install`),a=null}}if(!a){let n=await z(),{sources:p,input:m,workflowType:O,version:E}=n;f=m||{},l=O||e,u=E,console.log(`[setup] Workflow v${u||"?"} (${Object.keys(p).length} files)`),a=g(T,".zibby","workflows",l);let b=re(p,a);console.log(`[setup] Wrote ${b} files`),console.log("[setup] Installing dependencies...");try{ne("npm install --silent --no-audit --no-fund",{cwd:a,stdio:"inherit"}),console.log("[setup] Dependencies installed")}catch($){console.warn(`[setup] npm install failed: ${$.message}`)}}let d={},S=g(a,"workflow.json");w(S)&&(d=JSON.parse(K(S,"utf-8")));let A={},I=g(a,"zibby.config.json");if(w(I))try{A=JSON.parse(K(I,"utf-8")),console.log("[setup] Loaded user config from zibby.config.json")}catch(n){console.warn(`[setup] Failed to parse zibby.config.json: ${n.message} \u2014 falling back to defaults`)}let k=await ae(a,d);console.log(`[setup] Loaded ${k.name}`);let y=[],C=g(a,"node_modules","@zibby","agent-workflow"),v=g(a,"node_modules","@zibby","core","node_modules","@zibby","agent-workflow");w(C)&&y.push({kind:"hoisted",path:C}),w(v)&&y.push({kind:"nested",path:v});let R=process.env.ZIBBY_RUN_DIAG==="1";if(R){let{readdirSync:n}=await import("fs");console.log(` [diag] @zibby/agent-workflow copies in bundle: ${y.length}`);for(let p of y)console.log(` [diag] ${p.kind}: ${p.path}`);try{let p=g(a,"node_modules","@zibby");w(p)&&console.log(` [diag] node_modules/@zibby/ contents: [${n(p).join(", ")}]`)}catch{}}let N=g(a,"node_modules","@zibby","core","dist","index.js");if(w(N)&&y.length>0)try{let n=await import(x(N).href),p=[n.AssistantStrategy,n.CursorAgentStrategy,n.ClaudeAgentStrategy,n.CodexAgentStrategy,n.GeminiAgentStrategy].filter(Boolean);for(let m of y){let O=g(m.path,"dist","index.js");if(!w(O))continue;let E=await import(x(O).href),b=R?E.listStrategies():null;for(let $ of p)try{E.registerStrategy(new $)}catch(q){console.warn(` register ${$.name} into ${m.kind} failed: ${q.message}`)}R&&console.log(` [diag] ${m.kind} registry: before=[${b.join(",")||"empty"}] after=[${E.listStrategies().join(",")||"empty"}]`)}console.log("[setup] Registered 5 agent strategies (assistant, cursor, claude, codex, gemini)")}catch(n){console.warn(`[setup] Failed to bridge strategies: ${n.message}`)}else console.warn("[setup] No @zibby/core or @zibby/agent-workflow in bundle \u2014 agent strategies may be unavailable");let Q=Date.now(),_=new k({workflow:l||e}),J=_.buildGraph(),H={input:f||{},cwd:T,runId:s||`run-${Date.now()}`,config:A};console.log("");let h;try{h=await J.run(_,H)}catch(n){console.error(`
4
+ Workflow execution failed: ${n.message}`),console.error(n.stack),await U("failed",n.message),process.exit(1)}let F=((Date.now()-Q)/1e3).toFixed(1),M=h?.success!==!1,B=l||e;M?(console.log(`
5
+ [done] ${B} completed in ${F}s`),await U("completed")):(console.error(`
6
+ [done] ${B} failed after ${F}s`),await U("failed",h?.error||"Workflow execution failed"),process.exit(1)),_.onComplete&&await _.onComplete(h)}async function U(s,e=null){let o={EXECUTION_ID:process.env.WORKFLOW_JOB_ID,PROGRESS_API_URL:process.env.PROGRESS_API_URL,PROGRESS_QUEUE_URL:process.env.PROGRESS_QUEUE_URL,PROJECT_API_TOKEN:process.env.PROJECT_API_TOKEN,SQS_AUTH_TOKEN:process.env.SQS_AUTH_TOKEN};if(o.EXECUTION_ID)try{await L(o,{status:s,...e&&{error:e}})}catch(t){console.error(`\u26A0\uFE0F Failed to report status: ${t.message}`)}}export{me as runWorkflowCommand};
@@ -1,16 +1,16 @@
1
- import{existsSync as h}from"fs";import{readFile as W}from"fs/promises";import{join as k}from"path";import{pathToFileURL as F}from"url";import o from"chalk";import R from"ora";import{existsSync as T}from"fs";import{join as z}from"path";import{pathToFileURL as S}from"url";async function j(n){let e=z(n,".zibby.config.mjs");if(!T(e))throw new Error(".zibby.config.mjs not found");try{let t=await import(S(e).href);return t.default||t}catch(t){throw new Error(`Failed to load .zibby.config.mjs: ${t.message}`,{cause:t})}}var L=3848;async function O(n){try{return(await j(n))?.paths?.workflows||".zibby/workflows"}catch{return".zibby/workflows"}}async function U(n,e,t){let s=k(n,"graph.mjs");if(!h(s))throw new Error(`graph.mjs not found in ${t}/${e}/`);let a=await v(n,e),l=await import(F(s).href),i=a.entryClass,r=i&&l[i]||l.default||Object.values(l).find(f=>typeof f=="function"&&f.prototype?.buildGraph);if(!r)throw new Error("No WorkflowAgent class found in graph.mjs. Export a class with buildGraph() method.");return{AgentClass:r,manifest:a}}async function v(n,e){let t=k(n,"workflow.json");if(!h(t))return{name:e,triggers:{api:!0}};let s=await W(t,"utf-8");return JSON.parse(s)}async function X(n,e){n||(console.log(o.red(`
1
+ import{existsSync as h}from"fs";import{readFile as R}from"fs/promises";import{join as k}from"path";import{pathToFileURL as W}from"url";import o from"chalk";import L from"ora";import{existsSync as z}from"fs";import{join as S}from"path";import{pathToFileURL as F}from"url";async function x(e){let n=S(e,".zibby.config.mjs");if(!z(n))throw new Error(".zibby.config.mjs not found");try{let t=await import(F(n).href);return t.default||t}catch(t){throw new Error(`Failed to load .zibby.config.mjs: ${t.message}`,{cause:t})}}var O=3848;async function U(e){try{let n=await x(e);return{userConfig:n,workflowsBasePath:n?.paths?.workflows||".zibby/workflows"}}catch{return{userConfig:null,workflowsBasePath:".zibby/workflows"}}}async function A(e,n,t){let s=k(e,"graph.mjs");if(!h(s))throw new Error(`graph.mjs not found in ${t}/${n}/`);let w=await D(e,n),a=await import(W(s).href),c=w.entryClass,i=c&&a[c]||a.default||Object.values(a).find(r=>typeof r=="function"&&r.prototype?.buildGraph);if(!i)throw new Error("No WorkflowAgent class found in graph.mjs. Export a class with buildGraph() method.");return{AgentClass:i,manifest:w}}async function D(e,n){let t=k(e,"workflow.json");if(!h(t))return{name:n,triggers:{api:!0}};let s=await R(t,"utf-8");return JSON.parse(s)}async function Z(e,n){e||(console.log(o.red(`
2
2
  Workflow name is required`)),console.log(o.gray(" Usage: zibby workflow start <workflow-name>")),console.log(o.gray(` Example: zibby workflow start ticket-triage
3
- `)),process.exit(1));let t=n.toLowerCase(),s=process.cwd(),a=await O(s),l=k(s,a,t);h(l)||(console.log(o.red(`
3
+ `)),process.exit(1));let t=e.toLowerCase(),s=process.cwd(),{userConfig:w,workflowsBasePath:a}=await U(s),c=k(s,a,t);h(c)||(console.log(o.red(`
4
4
  Workflow not found: ${a}/${t}/`)),console.log(o.gray(" Create one first:")),console.log(o.cyan(` zibby workflow new ${t}
5
- `)),process.exit(1));let i=R(` Loading workflow "${t}"...`).start(),r,f;try{({AgentClass:r,manifest:f}=await U(l,t,a)),i.succeed(` Loaded ${o.bold(f.entryClass||r.name)} (${t})`)}catch(w){i.fail(" Failed to load workflow"),console.log(o.red(`
6
- ${w.message}
7
- `)),process.exit(1)}let g=parseInt(e.port,10)||L,u;try{u=(await import("express")).default}catch{console.log(o.red(`
5
+ `)),process.exit(1));let i=L(` Loading workflow "${t}"...`).start(),r,b;try{({AgentClass:r,manifest:b}=await A(c,t,a)),i.succeed(` Loaded ${o.bold(b.entryClass||r.name)} (${t})`)}catch(g){i.fail(" Failed to load workflow"),console.log(o.red(`
6
+ ${g.message}
7
+ `)),process.exit(1)}let f=parseInt(n.port,10)||O,u;try{u=(await import("express")).default}catch{console.log(o.red(`
8
8
  express is required for local workflow server`)),console.log(o.gray(` npm install express
9
- `)),process.exit(1)}let p=u();p.use(u.json({limit:"1mb"})),p.get("/health",(w,y)=>{y.json({status:"ok",workflow:t,class:r.name})}),p.post("/trigger",async(w,y)=>{let c=`local-${Date.now()}`,b=w.body.input||w.body||{};console.log(o.cyan(`
10
- \u25B6 Run ${c} triggered`)),console.log(o.gray(` input: ${JSON.stringify(b).slice(0,200)}`)),y.status(202).json({runId:c,status:"running",workflow:t});try{let m=Date.now(),d=new r({workflow:t}),x=d.buildGraph(),E={input:b,cwd:process.cwd(),runId:c},$=await x.run(d,E),C=((Date.now()-m)/1e3).toFixed(1),P=$?.success!==!1;console.log(P?o.green(` \u2714 Run ${c} succeeded (${C}s)`):o.red(` \u2716 Run ${c} failed (${C}s)`)),d.onComplete&&await d.onComplete($)}catch(m){console.log(o.red(` \u2716 Run ${c} error: ${m.message}`))}}),p.listen(g,()=>{console.log(o.bold.cyan(`
9
+ `)),process.exit(1)}let p=u();p.use(u.json({limit:"1mb"})),p.get("/health",(g,y)=>{y.json({status:"ok",workflow:t,class:r.name})}),p.post("/trigger",async(g,y)=>{let l=`local-${Date.now()}`,$=g.body.input||g.body||{};console.log(o.cyan(`
10
+ \u25B6 Run ${l} triggered`)),console.log(o.gray(` input: ${JSON.stringify($).slice(0,200)}`)),y.status(202).json({runId:l,status:"running",workflow:t});try{let m=Date.now(),d=new r({workflow:t}),E=d.buildGraph(),P={input:$,cwd:process.cwd(),runId:l,config:w||{}},C=await E.run(d,P),j=((Date.now()-m)/1e3).toFixed(1),T=C?.success!==!1;console.log(T?o.green(` \u2714 Run ${l} succeeded (${j}s)`):o.red(` \u2716 Run ${l} failed (${j}s)`)),d.onComplete&&await d.onComplete(C)}catch(m){console.log(o.red(` \u2716 Run ${l} error: ${m.message}`))}}),p.listen(f,()=>{console.log(o.bold.cyan(`
11
11
  Zibby Workflow Server \u2014 ${t}
12
- `)),console.log(o.gray(" ".padEnd(56,"-"))),console.log(o.white(` Workflow: ${o.cyan(t)}`)),console.log(o.white(` Class: ${o.cyan(r.name)}`)),console.log(o.white(` Port: ${o.cyan(g)}`)),console.log(o.gray(" ".padEnd(56,"-"))),console.log(o.white(`
13
- Endpoints:`)),console.log(o.gray(` GET http://localhost:${g}/health`)),console.log(o.cyan(` POST http://localhost:${g}/trigger`)),console.log(o.white(`
14
- Test with:`)),console.log(o.gray(` curl -X POST http://localhost:${g}/trigger \\`)),console.log(o.gray(' -H "Content-Type: application/json" \\')),console.log(o.gray(` -d '{"input": {"key": "value"}}'
12
+ `)),console.log(o.gray(" ".padEnd(56,"-"))),console.log(o.white(` Workflow: ${o.cyan(t)}`)),console.log(o.white(` Class: ${o.cyan(r.name)}`)),console.log(o.white(` Port: ${o.cyan(f)}`)),console.log(o.gray(" ".padEnd(56,"-"))),console.log(o.white(`
13
+ Endpoints:`)),console.log(o.gray(` GET http://localhost:${f}/health`)),console.log(o.cyan(` POST http://localhost:${f}/trigger`)),console.log(o.white(`
14
+ Test with:`)),console.log(o.gray(` curl -X POST http://localhost:${f}/trigger \\`)),console.log(o.gray(' -H "Content-Type: application/json" \\')),console.log(o.gray(` -d '{"input": {"key": "value"}}'
15
15
  `)),console.log(o.gray(` Ctrl+C to stop
16
- `))})}export{X as startWorkflowCommand};
16
+ `))})}export{Z as startWorkflowCommand};
@@ -1,22 +1,22 @@
1
- import h from"ora";import{select as I}from"@inquirer/prompts";import{readFileSync as $,existsSync as U}from"fs";import{homedir as b}from"os";import{join as j,resolve as _}from"path";var g={local:{name:"Local Development",apiUrl:"http://localhost:3001",accountApiUrl:"http://localhost:3001",frontendUrl:"http://localhost:3000",description:"Local backend running on port 3001"},prod:{name:"Production",apiUrl:process.env.ZIBBY_PROD_API_URL||"https://api-prod.zibby.app",accountApiUrl:process.env.ZIBBY_PROD_ACCOUNT_API_URL||"https://account-api-prod.zibby.app",frontendUrl:process.env.ZIBBY_PROD_FRONTEND_URL||"https://studio.zibby.app",description:"Production environment"}};function u(){let o;if(process.env.ZIBBY_API_URL)o=process.env.ZIBBY_API_URL;else{let e=process.env.ZIBBY_ENV||"prod";g[e]?o=g[e].apiUrl:o=g.prod.apiUrl}try{let e=new URL(o);return e.protocol!=="http:"&&e.protocol!=="https:"?(console.error(`\u26A0\uFE0F Invalid API URL protocol: ${e.protocol} (only http/https allowed)`),g.prod.apiUrl):o}catch{return console.error(`\u26A0\uFE0F Invalid API URL: ${o}`),g.prod.apiUrl}}var v=/^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$/i;function m(o){return o?v.test(o)?{ok:!0}:{ok:!1,error:`'${o}' is not a UUID. Cloud workflows are identified by UUID only. Run \`zibby workflow list\` to find yours, or run \`zibby workflow trigger\` with no argument for interactive selection.`}:{ok:!0}}function B(o){return o==="true"?!0:o==="false"?!1:o==="null"?null:o!==""&&!isNaN(Number(o))?Number(o):o}function E(o){let e={};for(let r of o||[]){let i=r.indexOf("=");if(i===-1){console.warn(` Warning: ignored param "${r}" \u2014 expected key=value format`);continue}let t=r.slice(0,i).trim(),n=B(r.slice(i+1)),l=t.split("."),c=e;for(let s=0;s<l.length-1;s++)(typeof c[l[s]]!="object"||c[l[s]]===null)&&(c[l[s]]={}),c=c[l[s]];c[l[l.length-1]]=n}return e}function N(o){let e=_(o);U(e)||(console.log(`
1
+ import h from"ora";import{select as v}from"@inquirer/prompts";import{readFileSync as x,existsSync as N}from"fs";import{homedir as R}from"os";import{join as T}from"path";var u={local:{name:"Local Development",apiUrl:"http://localhost:3001",accountApiUrl:"http://localhost:3001",frontendUrl:"http://localhost:3000",description:"Local backend running on port 3001"},prod:{name:"Production",apiUrl:process.env.ZIBBY_PROD_API_URL||"https://api-prod.zibby.app",accountApiUrl:process.env.ZIBBY_PROD_ACCOUNT_API_URL||"https://account-api-prod.zibby.app",frontendUrl:process.env.ZIBBY_PROD_FRONTEND_URL||"https://studio.zibby.app",description:"Production environment"}};function g(){let o;if(process.env.ZIBBY_API_URL)o=process.env.ZIBBY_API_URL;else{let e=process.env.ZIBBY_ENV||"prod";u[e]?o=u[e].apiUrl:o=u.prod.apiUrl}try{let e=new URL(o);return e.protocol!=="http:"&&e.protocol!=="https:"?(console.error(`\u26A0\uFE0F Invalid API URL protocol: ${e.protocol} (only http/https allowed)`),u.prod.apiUrl):o}catch{return console.error(`\u26A0\uFE0F Invalid API URL: ${o}`),u.prod.apiUrl}}var b=/^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$/i;function y(o){return o?b.test(o)?{ok:!0}:{ok:!1,error:`'${o}' is not a UUID. Cloud workflows are identified by UUID only. Run \`zibby workflow list\` to find yours, or run \`zibby workflow trigger\` with no argument for interactive selection.`}:{ok:!0}}import{existsSync as j,readFileSync as _}from"fs";import{resolve as B}from"path";function I(o){return o==="true"?!0:o==="false"?!1:o==="null"?null:o!==""&&!isNaN(Number(o))?Number(o):o}function $(o){let e={};for(let r of o||[]){let i=r.indexOf("=");if(i===-1){console.warn(` Warning: ignored param "${r}" \u2014 expected key=value format`);continue}let t=r.slice(0,i).trim(),n=I(r.slice(i+1)),l=t.split("."),c=e;for(let s=0;s<l.length-1;s++)(typeof c[l[s]]!="object"||c[l[s]]===null)&&(c[l[s]]={}),c=c[l[s]];c[l[l.length-1]]=n}return e}function E(o){let e=B(o);j(e)||(console.log(`
2
2
  Error: --input-file not found: ${o}
3
- `),process.exit(1));try{return JSON.parse($(e,"utf-8"))}catch(r){console.log(`
3
+ `),process.exit(1));try{return JSON.parse(_(e,"utf-8"))}catch(r){console.log(`
4
4
  Error: --input-file is not valid JSON: ${r.message}
5
- `),process.exit(1)}}function x(o){let e={};if(o.inputFile&&(e={...N(o.inputFile)}),o.input)try{e={...e,...JSON.parse(o.input)}}catch(r){console.log(`
5
+ `),process.exit(1)}}function U(o){let e={};if(o.inputFile&&(e={...E(o.inputFile)}),o.input)try{e={...e,...JSON.parse(o.input)}}catch(r){console.log(`
6
6
  Error: --input is not valid JSON`),console.log(` ${r.message}
7
- `),process.exit(1)}return o.param?.length&&(e={...e,...E(o.param)}),e}function R(){let o=j(b(),".zibby","config.json");if(U(o))try{let r=JSON.parse($(o,"utf-8"));if(r.sessionToken)return r.sessionToken}catch{}let e=process.env.ZIBBY_API_KEY;if(e)return e;console.log(`
7
+ `),process.exit(1)}return o.param?.length&&(e={...e,...$(o.param)}),e}function A(){let o=T(R(),".zibby","config.json");if(N(o))try{let r=JSON.parse(x(o,"utf-8"));if(r.sessionToken)return r.sessionToken}catch{}let e=process.env.ZIBBY_API_KEY;if(e)return e;console.log(`
8
8
  Not authenticated`),console.log(" Run: zibby login"),console.log(` OR set ZIBBY_API_KEY env var (for CI/CD)
9
- `),process.exit(1)}async function T(o){let e=u(),r=h("Fetching projects...").start();try{let i=await fetch(`${e}/projects`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${o}`}});i.ok||(r.fail("Failed to fetch projects"),process.exit(1));let t=await i.json();Array.isArray(t)||(t.projects?t=t.projects:t.data&&(t=t.data)),(!t||t.length===0)&&(r.fail("No projects found"),process.exit(1)),r.succeed(`Found ${t.length} project${t.length===1?"":"s"}`),console.log("");let n=t.map(l=>({name:`${l.name||"Unnamed"} (${l.projectId||l.id})`,value:l.projectId||l.id}));return await I({message:"Select a project:",choices:n})}catch(i){r.fail(`Error: ${i.message}`),process.exit(1)}}async function A(o,e){let r=u(),i=h("Fetching deployed workflows...").start();try{let t=["analysis","implementation","run_test"],n=[];for(let c of t){let s=await fetch(`${r}/projects/${o}/workflows/${c}`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${e}`}});if(s.ok){let a=await s.json();a.graph&&n.push({name:c,version:a.version||0,isDefault:a.isDefault!==!1})}}n.length===0&&(i.fail("No deployed workflows found for this project"),process.exit(1)),i.succeed(`Found ${n.length} deployed workflow${n.length===1?"":"s"}`),console.log("");let l=n.map(c=>({name:`${c.name} (v${c.version})${c.isDefault?" [default]":""}`,value:c.name}));return await I({message:"Select a workflow to trigger:",choices:l})}catch(t){i.fail(`Error: ${t.message}`),process.exit(1)}}async function J(o,e={}){let r=m(o);r.ok||(console.log(`
9
+ `),process.exit(1)}async function D(o){let e=g(),r=h("Fetching projects...").start();try{let i=await fetch(`${e}/projects`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${o}`}});i.ok||(r.fail("Failed to fetch projects"),process.exit(1));let t=await i.json();Array.isArray(t)||(t.projects?t=t.projects:t.data&&(t=t.data)),(!t||t.length===0)&&(r.fail("No projects found"),process.exit(1)),r.succeed(`Found ${t.length} project${t.length===1?"":"s"}`),console.log("");let n=t.map(l=>({name:`${l.name||"Unnamed"} (${l.projectId||l.id})`,value:l.projectId||l.id}));return await v({message:"Select a project:",choices:n})}catch(i){r.fail(`Error: ${i.message}`),process.exit(1)}}async function P(o,e){let r=g(),i=h("Fetching deployed workflows...").start();try{let t=["analysis","implementation","run_test"],n=[];for(let c of t){let s=await fetch(`${r}/projects/${o}/workflows/${c}`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${e}`}});if(s.ok){let a=await s.json();a.graph&&n.push({name:c,version:a.version||0,isDefault:a.isDefault!==!1})}}n.length===0&&(i.fail("No deployed workflows found for this project"),process.exit(1)),i.succeed(`Found ${n.length} deployed workflow${n.length===1?"":"s"}`),console.log("");let l=n.map(c=>({name:`${c.name} (v${c.version})${c.isDefault?" [default]":""}`,value:c.name}));return await v({message:"Select a workflow to trigger:",choices:l})}catch(t){i.fail(`Error: ${t.message}`),process.exit(1)}}async function M(o,e={}){let r=y(o);r.ok||(console.log(`
10
10
  Error: ${r.error}
11
- `),process.exit(1));let i=R(),t=e.project||process.env.ZIBBY_PROJECT_ID,n;if(o){let s=u();try{let a=await fetch(`${s}/projects`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`}});if(a.ok){let d=(await a.json()).projects||[];for(let p of d){let y=await fetch(`${s}/projects/${p.projectId}/workflows`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`}});if(y.ok){let w=(await y.json()).find(k=>k.uuid===o);if(w){t=p.projectId,n=w.workflowType||w.name,console.log(`
11
+ `),process.exit(1));let i=A(),t=e.project||process.env.ZIBBY_PROJECT_ID,n;if(o){let s=g();try{let a=await fetch(`${s}/projects`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`}});if(a.ok){let d=(await a.json()).projects||[];for(let p of d){let m=await fetch(`${s}/projects/${p.projectId}/workflows`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`}});if(m.ok){let w=(await m.json()).find(k=>k.uuid===o);if(w){t=p.projectId,n=w.workflowType||w.name,console.log(`
12
12
  \u2713 Found workflow "${n}" (UUID: ${o})
13
13
  `);break}}}(!n||n===o)&&(console.log(`
14
14
  Error: Workflow with UUID "${o}" not found`),console.log(` Check: zibby workflow list
15
15
  `),process.exit(1))}}catch(a){console.log(`
16
16
  Error looking up workflow UUID: ${a.message}
17
- `),process.exit(1)}}t||(console.log(""),t=await T(i)),n||(console.log(""),n=await A(t,i));let l=x(e);if(console.log(`
17
+ `),process.exit(1)}}t||(console.log(""),t=await D(i)),n||(console.log(""),n=await P(t,i));let l=U(e);if(console.log(`
18
18
  Triggering Workflow
19
- `),console.log(" ".padEnd(60,"-")),console.log(` Workflow: ${n}`),console.log(` Project: ${t}`),Object.keys(l).length>0){let s=JSON.stringify(l);console.log(` Input: ${s.length>60?`${s.substring(0,57)}...`:s}`)}e.idempotencyKey&&console.log(` Idempotency: ${e.idempotencyKey}`),console.log(" ".padEnd(60,"-")),console.log("");let c=h("Triggering workflow execution...").start();try{let s=u(),a={input:l};e.idempotencyKey&&(a.idempotencyKey=e.idempotencyKey);let f=await fetch(`${s}/projects/${t}/workflows/${n}/trigger`,{method:"POST",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`},body:JSON.stringify(a)});if(!f.ok){let p=await f.json().catch(()=>({}));f.status===429&&(c.fail("Quota exceeded"),console.log(`
19
+ `),console.log(" ".padEnd(60,"-")),console.log(` Workflow: ${n}`),console.log(` Project: ${t}`),Object.keys(l).length>0){let s=JSON.stringify(l);console.log(` Input: ${s.length>60?`${s.substring(0,57)}...`:s}`)}e.idempotencyKey&&console.log(` Idempotency: ${e.idempotencyKey}`),console.log(" ".padEnd(60,"-")),console.log("");let c=h("Triggering workflow execution...").start();try{let s=g(),a={input:l};e.idempotencyKey&&(a.idempotencyKey=e.idempotencyKey);let f=await fetch(`${s}/projects/${t}/workflows/${n}/trigger`,{method:"POST",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`},body:JSON.stringify(a)});if(!f.ok){let p=await f.json().catch(()=>({}));f.status===429&&(c.fail("Quota exceeded"),console.log(`
20
20
  Your workflow execution quota has been exceeded`),p.quotaInfo&&(console.log(` Used: ${p.quotaInfo.used}/${p.quotaInfo.limit} executions`),console.log(` Plan: ${p.quotaInfo.planId}`),p.quotaInfo.periodEnd&&console.log(` Resets: ${new Date(p.quotaInfo.periodEnd).toLocaleDateString()}`)),console.log(""),process.exit(1)),c.fail("Trigger failed"),console.log(` Error: ${p.message||f.statusText}
21
21
  `),process.exit(1)}let d=await f.json();c.succeed("Workflow triggered successfully"),console.log(""),console.log(" Job Details:"),console.log(` Job ID: ${d.jobId}`),console.log(` Status: ${d.status}`),console.log(` Version: ${d.version}`),console.log(` Triggered: ${new Date(d.triggeredAt).toLocaleString()}`),console.log(""),console.log(" Monitor execution:"),o?(console.log(` zibby workflow logs ${o}`),console.log(` zibby workflow logs ${o} -t`)):(console.log(` zibby workflow logs --workflow ${n} --project ${t}`),console.log(` zibby workflow logs --workflow ${n} --project ${t} -t`)),console.log("")}catch(s){c.fail("Trigger failed"),console.log(` Error: ${s.message}
22
- `),process.exit(1)}}export{B as coerceValue,E as parseParams,x as resolveInput,J as triggerWorkflowCommand};
22
+ `),process.exit(1)}}export{I as coerceValue,$ as parseParams,U as resolveInput,M as triggerWorkflowCommand};