@zibby/cli 0.1.90 → 0.1.95
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +94 -848
- package/dist/bin/zibby.js +47 -2
- package/dist/commands/chat.js +1 -1
- package/dist/commands/workflow.js +19 -19
- package/dist/commands/workflows/agent-helpers.js +18 -0
- package/dist/commands/workflows/deploy.js +47 -24
- package/dist/commands/workflows/env-helpers.js +2 -0
- package/dist/commands/workflows/env.js +39 -0
- package/dist/commands/workflows/generate.js +48 -31
- package/dist/commands/workflows/input-helpers.js +7 -0
- package/dist/commands/workflows/list.js +2 -2
- package/dist/commands/workflows/logs.js +22 -22
- package/dist/commands/workflows/run-local.js +19 -0
- package/dist/commands/workflows/run.js +5 -5
- package/dist/commands/workflows/start.js +11 -11
- package/dist/commands/workflows/trigger.js +9 -9
- package/dist/package.json +4 -4
- package/dist/templates/zibby-workflow-claude/agents-md-block.md +113 -0
- package/dist/templates/zibby-workflow-claude/claude/agents/zibby-test-author.md +72 -0
- package/dist/templates/zibby-workflow-claude/claude/agents/zibby-workflow-builder.md +81 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-add-node.md +75 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-debug.md +67 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-delete.md +37 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-deploy.md +77 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-list.md +30 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-static-ip.md +68 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-tail.md +53 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-test-debug.md +59 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-test-generate.md +39 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-test-run.md +48 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-test-write.md +46 -0
- package/dist/templates/zibby-workflow-claude/claude/commands/zibby-trigger.md +52 -0
- package/dist/templates/zibby-workflow-claude/claude/settings.json +10 -0
- package/dist/templates/zibby-workflow-claude/cursor/rules/zibby-workflows.mdc +56 -0
- package/dist/templates/zibby-workflow-claude/manifest.json +43 -0
- package/package.json +4 -4
|
@@ -0,0 +1,39 @@
|
|
|
1
|
+
import a from"chalk";import x from"ora";import{existsSync as ao,readFileSync as uo}from"fs";import{join as po}from"path";import A from"chalk";import{confirm as fo}from"@inquirer/prompts";import c from"chalk";import T from"ora";import{spawn as ro}from"child_process";var m={local:{name:"Local Development",apiUrl:"http://localhost:3001",accountApiUrl:"http://localhost:3001",frontendUrl:"http://localhost:3000",description:"Local backend running on port 3001"},prod:{name:"Production",apiUrl:process.env.ZIBBY_PROD_API_URL||"https://api-prod.zibby.app",accountApiUrl:process.env.ZIBBY_PROD_ACCOUNT_API_URL||"https://account-api-prod.zibby.app",frontendUrl:process.env.ZIBBY_PROD_FRONTEND_URL||"https://studio.zibby.app",description:"Production environment"}};function v(){let o;if(process.env.ZIBBY_API_URL)o=process.env.ZIBBY_API_URL;else{let e=process.env.ZIBBY_ENV||"prod";m[e]?o=m[e].apiUrl:o=m.prod.apiUrl}try{let e=new URL(o);return e.protocol!=="http:"&&e.protocol!=="https:"?(console.error(`\u26A0\uFE0F Invalid API URL protocol: ${e.protocol} (only http/https allowed)`),m.prod.apiUrl):o}catch{return console.error(`\u26A0\uFE0F Invalid API URL: ${o}`),m.prod.apiUrl}}import{existsSync as L,mkdirSync as Q,readFileSync as oo,writeFileSync as eo}from"fs";import{homedir as z}from"os";import{join as U}from"path";function O(){return process.env.ZIBBY_CONFIG_DIR||U(z(),".zibby")}function R(){return U(O(),"config.json")}var no=U(z(),".zibby"),ko=U(no,"config.json");function to(){let o=O();L(o)||Q(o,{recursive:!0})}function g(){try{let o=R();if(L(o)){let e=oo(o,"utf-8");return JSON.parse(e)}}catch{}return{}}function w(o){to(),eo(R(),JSON.stringify(o,null,2))}function Y(){return g().sessionToken||null}function D(o){let e=g();e.sessionToken=o,w(e)}function F(){return g().user||null}function K(o){let e=g();e.user=o,w(e)}function Z(o){let e=g();e.proxyUrl=o,w(e)}function V(o){let e=g();e.mem0ProxyUrl=o,w(e)}function J(o){let e=g();e.projects=o,w(e)}import{existsSync as Io,mkdirSync as bo,readFileSync as To,writeFileSync as So,unlinkSync as Ao}from"fs";import{resolve as Eo}from"path";import{homedir as _o}from"os";function so(o){let e=process.platform;try{let n,t;return e==="darwin"?(n="open",t=[o]):e==="win32"?(n="cmd",t=["/c","start","",o]):(n="xdg-open",t=[o]),ro(n,t,{detached:!0,stdio:"ignore"}).unref(),!0}catch{return!1}}function io(){let o=Y(),e=F();return o&&e?{loggedIn:!0,user:e,token:o}:{loggedIn:!1}}async function M(){try{console.log(c.cyan(`
|
|
2
|
+
\u{1F510} Initiating login...
|
|
3
|
+
`));let o=io();if(o.loggedIn){console.log(c.green("\u2705 Already logged in!")),console.log(c.gray(`User: ${o.user.email}`)),console.log(c.gray(`Name: ${o.user.name}
|
|
4
|
+
`));let{createInterface:e}=await import("readline"),n=e({input:process.stdin,output:process.stdout});return new Promise((t,r)=>{let s=()=>{n.close(),process.stdin.isTTY&&process.stdin.setRawMode(!1)},i=()=>{console.log(c.yellow(`
|
|
5
|
+
|
|
6
|
+
\u26A0\uFE0F Login cancelled
|
|
7
|
+
`)),s(),process.exit(0)};process.on("SIGINT",i),n.question(c.yellow("Continue with this session? (Y/n): "),async l=>{process.removeListener("SIGINT",i),s();try{if(l.toLowerCase()==="n"||l.toLowerCase()==="no"){console.log(c.gray(`Starting new login...
|
|
8
|
+
`));let f=await H();t(f)}else console.log(c.green(`Using existing session.
|
|
9
|
+
`)),t({success:!0,...o})}catch(f){r(f)}})})}return await H()}catch(o){return console.error(c.red(`
|
|
10
|
+
\u274C Login failed:`,o.message)),{success:!1,error:o.message}}}async function lo(o){let e=v();try{let n=await fetch(`${e}/projects`,{headers:{Authorization:`Bearer ${o}`}});if(n.ok){let r=((await n.json()).projects||[]).map(s=>({name:s.name,projectId:s.projectId,apiToken:s.apiToken}));return J(r),r}}catch(n){console.log(c.gray(`\u26A0\uFE0F Could not fetch projects: ${n.message}`))}return[]}async function H(){let o=v(),e=T("Requesting login code...").start(),n=await fetch(`${o}/cli/login/initiate`,{method:"POST",headers:{"Content-Type":"application/json"}});if(!n.ok){e.fail("Failed to request login code");let y=await n.json();throw new Error(y.error||"Failed to initiate login")}let{deviceCode:t,userCode:r,verificationUrl:s,expiresIn:i,interval:l}=await n.json();e.succeed("Login code generated"),console.log(""),console.log(c.cyan("\u2554\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2557")),console.log(c.cyan("\u2551")+c.white.bold(" Complete login in your browser ")+c.cyan("\u2551")),console.log(c.cyan("\u255A\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u2550\u255D")),console.log(""),console.log(c.white("Opening browser to login page...")),console.log(c.gray(`Code expires in ${Math.floor(i/60)} minutes`)),console.log(""),await so(s)||(console.log(c.yellow("\u26A0\uFE0F Could not open browser automatically.")),console.log(c.white("Please open this URL manually: ")+c.blue(s)),console.log(""));let u=T("Waiting for authorization...").start(),d=(l||3)*1e3,X=Math.floor(i/(l||3)),_=0,N=!1,B=()=>{N=!0,u.stop(),console.log(c.yellow(`
|
|
11
|
+
|
|
12
|
+
\u26A0\uFE0F Login cancelled
|
|
13
|
+
`)),process.exit(0)};process.on("SIGINT",B);try{for(;_<X&&!N;){await co(d),_++;let y=await fetch(`${o}/cli/login/poll`,{method:"POST",headers:{"Content-Type":"application/json"},body:JSON.stringify({deviceCode:t})});if(y.status===202)continue;if(!y.ok){u.fail("Authorization failed");let b=await y.json();throw new Error(b.error||"Authorization failed")}let p=await y.json();if(p.status==="authorized"){u.succeed(c.white("Authorization successful!")),D(p.token),K(p.user),p.proxyUrl&&Z(p.proxyUrl),p.mem0ProxyUrl&&V(p.mem0ProxyUrl),console.log(""),console.log(c.gray(`User: ${p.user.email}`));let b=T("Fetching projects...").start(),C=await lo(p.token);return b.succeed(`Fetched ${C.length} project${C.length!==1?"s":""}`),console.log(c.gray(`Session saved to: ~/.zibby/config.json
|
|
14
|
+
`)),{success:!0,loggedIn:!0,user:p.user,token:p.token}}if(p.status==="denied")throw u.fail("Authorization denied"),new Error("User denied authorization")}throw u.fail("Login timeout"),new Error("Login timed out - please try again")}finally{process.removeListener("SIGINT",B)}}function co(o){return new Promise(e=>setTimeout(e,o))}function q(){try{let o=process.env.HOME||process.env.USERPROFILE;if(!o)return null;let e=po(o,".zibby","config.json");return ao(e)&&JSON.parse(uo(e,"utf-8")).sessionToken||null}catch{return null}}function S(){console.log(`
|
|
15
|
+
Not authenticated.`),console.log(` Run ${A.cyan("zibby login")} or set ${A.cyan("ZIBBY_API_KEY")} in your environment.
|
|
16
|
+
`)}async function G(o={}){let e=o.apiKey||process.env.ZIBBY_API_KEY||null,n=q();if(n||e)return{sessionToken:n,apiKey:e};if(!process.stdin.isTTY){if(o.optional)return{sessionToken:null,apiKey:null};S(),process.exit(1)}console.log(A.yellow(`
|
|
17
|
+
Not logged in.`));let t;try{t=await fo({message:"Open browser to log in now?",default:!0})}catch{t=!1}if(!t){if(o.optional)return{sessionToken:null,apiKey:null};S(),process.exit(1)}if(await M(),n=q(),!n){if(o.optional)return{sessionToken:null,apiKey:null};S(),process.exit(1)}return{sessionToken:n,apiKey:null}}import{readFileSync as go,existsSync as yo}from"fs";import ho from"dotenv";var j=/^[A-Z_][A-Z0-9_]*$/;function k(o){return typeof o=="string"&&j.test(o)}function W(o){if(typeof o!="string"||o.length===0)throw new Error("Expected KEY=value, got empty argument");let e=o.indexOf("=");if(e<=0)throw new Error(`Expected KEY=value, got "${o}" \u2014 missing '=' or empty key`);let n=o.slice(0,e),t=o.slice(e+1);if(!k(n))throw new Error(`Invalid env var name "${n}" \u2014 must match ${j} (uppercase letters, digits, underscores; can't start with a digit)`);return{key:n,value:t}}function E(o){if(!Array.isArray(o)||o.length===0)return{};let e={};for(let n of o){if(!yo(n))throw new Error(`--env file not found: ${n}`);let t=ho.parse(go(n,"utf-8"));for(let[r,s]of Object.entries(t)){if(!k(r))throw new Error(`Invalid env var name "${r}" in ${n} \u2014 must match ${j} (uppercase letters, digits, underscores; can't start with a digit)`);e[r]=s}}return e}function h(o,e,n){let t=String(o).replace(/\/+$/,"");return n?`${t}/workflows/${e}/env/${encodeURIComponent(n)}`:`${t}/workflows/${e}/env`}function P(o){return!Array.isArray(o)||o.length===0?"No env vars set on this workflow.":[...o].sort().map(n=>` ${n}`).join(`
|
|
18
|
+
`)}async function $(o){let{sessionToken:e,apiKey:n}=await G({apiKey:o.apiKey}),t=e||n;return t||(console.log(a.red("\n Not authenticated \u2014 run `zibby login` or pass --api-key.\n")),process.exit(1)),{authToken:t,apiUrl:v()}}async function I(o,e,n,t,r){let s={method:o,headers:{Authorization:`Bearer ${n}`,...t?{"Content-Type":"application/json"}:{}},...t?{body:JSON.stringify(t)}:{}},i;try{i=await fetch(e,s)}catch(f){r&&r.fail(`Network error: ${f.message}`),process.exit(1)}let l;try{l=await i.json()}catch{l={}}return i.ok||(r?r.fail(l.error||l.message||`HTTP ${i.status}`):console.log(a.red(`
|
|
19
|
+
${l.error||l.message||`HTTP ${i.status}`}
|
|
20
|
+
`)),process.exit(1)),l}async function te(o,e={}){o||(console.log(a.red(`
|
|
21
|
+
Workflow UUID is required`)),console.log(a.gray(` Usage: zibby workflow env list <uuid>
|
|
22
|
+
`)),process.exit(1));let{authToken:n,apiUrl:t}=await $(e),r=x(`Fetching env keys for ${o}...`).start(),s=await I("GET",h(t,o),n,null,r),i=s.keys||[];r.succeed(`${i.length} env var${i.length===1?"":"s"} on workflow ${o}`),console.log(""),console.log(P(i)),s.envUpdatedAt&&console.log(a.gray(`
|
|
23
|
+
Last updated: ${s.envUpdatedAt}`)),console.log("")}async function re(o,e,n={}){(!o||!e)&&(console.log(a.red(`
|
|
24
|
+
Both UUID and KEY=value are required`)),console.log(a.gray(` Usage: zibby workflow env set <uuid> KEY=value
|
|
25
|
+
`)),process.exit(1));let t,r;try{({key:t,value:r}=W(e))}catch(d){console.log(a.red(`
|
|
26
|
+
${d.message}
|
|
27
|
+
`)),process.exit(1)}let{authToken:s,apiUrl:i}=await $(n),l=x(`Setting ${t} on workflow ${o}...`).start(),u=((await I("PATCH",h(i,o,t),s,{value:r},l)).keys||[]).length;l.succeed(`Set ${a.cyan(t)} (workflow now has ${u} env var${u===1?"":"s"})`)}async function se(o,e,n={}){(!o||!e)&&(console.log(a.red(`
|
|
28
|
+
Both UUID and KEY are required`)),console.log(a.gray(` Usage: zibby workflow env unset <uuid> KEY
|
|
29
|
+
`)),process.exit(1)),k(e)||(console.log(a.red(`
|
|
30
|
+
Invalid env var name "${e}"
|
|
31
|
+
`)),process.exit(1));let{authToken:t,apiUrl:r}=await $(n),s=x(`Unsetting ${e} on workflow ${o}...`).start(),i=await I("DELETE",h(r,o,e),t,null,s),l=(i.keys||[]).length;i.removed?s.succeed(`Unset ${a.cyan(e)} (workflow now has ${l} env var${l===1?"":"s"})`):s.warn(`Key ${a.cyan(e)} was not set on workflow ${o} \u2014 nothing to do`)}async function ie(o,e={}){o||(console.log(a.red(`
|
|
32
|
+
Workflow UUID is required`)),console.log(a.gray(` Usage: zibby workflow env push <uuid> --file .env [--file .env.prod]
|
|
33
|
+
`)),process.exit(1));let n=Array.isArray(e.file)?e.file:e.file?[e.file]:[];n.length===0&&(console.log(a.red(`
|
|
34
|
+
At least one --file is required`)),console.log(a.gray(` Usage: zibby workflow env push <uuid> --file .env
|
|
35
|
+
`)),process.exit(1));let t;try{t=E(n)}catch(f){console.log(a.red(`
|
|
36
|
+
${f.message}
|
|
37
|
+
`)),process.exit(1)}let r=Object.keys(t);r.length===0&&(console.log(a.yellow(`
|
|
38
|
+
No env vars found in ${n.join(", ")} \u2014 nothing to push.
|
|
39
|
+
`)),process.exit(0));let{authToken:s,apiUrl:i}=await $(e),l=x(`Pushing ${r.length} env var${r.length===1?"":"s"} to ${o}...`).start();await I("PUT",h(i,o),s,{env:t},l),l.succeed(`Pushed ${r.length} env var${r.length===1?"":"s"} from ${n.join(", ")}`),console.log(""),console.log(P(r)),console.log("")}async function le({uuid:o,files:e,authToken:n,apiUrl:t,spinner:r}){let s=E(e),i=Object.keys(s);if(i.length===0)return{count:0,files:e};let l=h(t,o),f={method:"PUT",headers:{Authorization:`Bearer ${n}`,"Content-Type":"application/json"},body:JSON.stringify({env:s})},u=await fetch(l,f);if(!u.ok){let d=await u.json().catch(()=>({}));throw new Error(`Env push failed: ${d.error||d.message||`HTTP ${u.status}`}`)}return r&&(r.text=`Synced ${i.length} env var${i.length===1?"":"s"} to workflow`),{count:i.length,files:e}}export{te as listEnvCommand,ie as pushEnvCommand,re as setEnvCommand,le as syncEnvFromFiles,se as unsetEnvCommand};
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import{mkdir as
|
|
1
|
+
import{mkdir as T,writeFile as k}from"fs/promises";import{existsSync as K}from"fs";import{spawn as ae}from"child_process";import{join as w}from"path";import a from"chalk";import M from"ora";import{input as le}from"@inquirer/prompts";import{existsSync as Y}from"fs";import{join as N}from"path";import{pathToFileURL as L}from"url";async function E(n){let e=N(n,".zibby.config.mjs");if(!Y(e))throw new Error(".zibby.config.mjs not found");try{let t=await import(L(e).href);return t.default||t}catch(t){throw new Error(`Failed to load .zibby.config.mjs: ${t.message}`,{cause:t})}}import{mkdir as Me,writeFile as Re,readFile as Oe}from"fs/promises";import{existsSync as Ne,readdirSync as Le}from"fs";import{join as Ge,resolve as Ve,dirname as D}from"path";import{homedir as He}from"os";import Fe from"inquirer";import qe from"chalk";import Xe from"ora";import{spawn as et,execSync as tt}from"child_process";import{fileURLToPath as G}from"url";import{createRequire as V}from"module";import{existsSync as Ae,readFileSync as Ce,writeFileSync as Ie,mkdirSync as $e}from"fs";import{join as Ee}from"path";import{homedir as _e}from"os";var Z=G(import.meta.url),rt=D(Z),it=V(import.meta.url);function S(n,e={},t={}){let o=["dolt","mem0"].includes(String(t.memoryBackend||"").toLowerCase())?String(t.memoryBackend).toLowerCase():"dolt",s={claude:`
|
|
2
2
|
claude: {
|
|
3
3
|
model: 'auto', // Options: 'auto', 'sonnet-4.6', 'opus-4.6', 'sonnet-4.5', 'opus-4.5'
|
|
4
4
|
maxTokens: 4096,
|
|
@@ -11,13 +11,13 @@ import{mkdir as v,writeFile as p}from"fs/promises";import{existsSync as k}from"f
|
|
|
11
11
|
},`,gemini:`
|
|
12
12
|
gemini: {
|
|
13
13
|
model: 'gemini-2.5-pro', // Options: 'auto', 'gemini-2.5-pro', 'gemini-2.5-flash'
|
|
14
|
-
},`},
|
|
15
|
-
`).map(
|
|
14
|
+
},`},i=n.agent,l=Object.entries(s).filter(([r])=>r!==i).map(([,r])=>r.split(`
|
|
15
|
+
`).map(c=>c.trim()?` // ${c.trimStart()}`:c).join(`
|
|
16
16
|
`)).join(`
|
|
17
17
|
`);return`export default {
|
|
18
18
|
// AI agent settings
|
|
19
|
-
agent: {${
|
|
20
|
-
${
|
|
19
|
+
agent: {${s[i]}
|
|
20
|
+
${l}
|
|
21
21
|
strictMode: false,
|
|
22
22
|
},
|
|
23
23
|
|
|
@@ -25,12 +25,12 @@ ${s}
|
|
|
25
25
|
// and workflow config. Runtime strategies attach MCP per run (no global Gemini settings mutation).
|
|
26
26
|
browser: {
|
|
27
27
|
mcp: 'playwright',
|
|
28
|
-
headless: ${
|
|
28
|
+
headless: ${n.browserMode==="headless"},
|
|
29
29
|
},
|
|
30
30
|
|
|
31
31
|
// Chat memory backend adapter (dolt | mem0)
|
|
32
32
|
memory: {
|
|
33
|
-
backend: '${
|
|
33
|
+
backend: '${o}',
|
|
34
34
|
},
|
|
35
35
|
|
|
36
36
|
// Advanced: Override models per node (optional)
|
|
@@ -44,7 +44,7 @@ ${s}
|
|
|
44
44
|
specs: 'test-specs', // Where your .txt test specs are
|
|
45
45
|
generated: 'tests', // Where generated .spec.js files go
|
|
46
46
|
output: '.zibby/output', // Where workflow execution results are saved (default: .zibby/output)
|
|
47
|
-
workflows: '${
|
|
47
|
+
workflows: '${n.workflowsPath||".zibby/workflows"}', // Where custom workflows are stored
|
|
48
48
|
// sessionPrefix: 'run', // Optional: prefix for session folders (e.g., run_1772788458045)
|
|
49
49
|
},
|
|
50
50
|
|
|
@@ -79,10 +79,27 @@ ${s}
|
|
|
79
79
|
},
|
|
80
80
|
|
|
81
81
|
// Cloud sync - auto-upload test results & videos (requires ZIBBY_API_KEY in .env)
|
|
82
|
-
cloudSync: ${
|
|
82
|
+
cloudSync: ${n.cloudSync||!1}
|
|
83
83
|
};
|
|
84
|
-
`}
|
|
85
|
-
|
|
84
|
+
`}import{readFileSync as x,existsSync as m,mkdirSync as $,writeFileSync as v,statSync as H,chmodSync as W}from"fs";import{join as g,dirname as A}from"path";import{fileURLToPath as F}from"url";import{select as U}from"@inquirer/prompts";import h from"chalk";var q=A(F(import.meta.url)),P=g(q,"..","..","..","templates","zibby-workflow-claude"),J="<!-- BEGIN zibby-workflows",_="<!-- END zibby-workflows -->",I={claude:"Claude Code",cursor:"Cursor",codex:"Codex (AGENTS.md)"};function X(n){let e=process.argv[1];if(!e||!m(e))return null;let t=g(n,".zibby","bin"),o=g(t,"zibby");$(t,{recursive:!0});let s=`#!/bin/sh
|
|
85
|
+
# Auto-generated by 'zibby workflow generate'. Re-runs of generate update this path.
|
|
86
|
+
# Lets agents (Claude / Cursor / Codex) and CI scripts call zibby reliably even
|
|
87
|
+
# when the CLI isn't on PATH. Delete this file or .zibby/bin/ to disable.
|
|
88
|
+
exec node ${JSON.stringify(e)} "$@"
|
|
89
|
+
`;v(o,s);try{H(o)}catch{}try{W(o,493)}catch{}return o}function Q(n){let e=[];return m(g(n,".claude"))&&e.push("claude"),(m(g(n,".cursor"))||m(g(n,".cursorrules")))&&e.push("cursor"),m(g(n,"AGENTS.md"))&&e.push("codex"),e}function ee(){let n=g(P,"manifest.json");return JSON.parse(x(n,"utf8"))}function z(n){let e=n.match(/<!--\s*(?:BEGIN zibby-workflows\s+)?zibby-template-version:\s*(\d+)\s*-->/);return e?parseInt(e[1],10):null}function te(n,e,t){if(!m(n))return{write:!0,reason:"new"};if(t)return{write:!0,reason:"force"};let o=x(n,"utf8"),s=z(o),i=z(e);return s==null?{write:!1,reason:"user-edited (no version marker)"}:i==null?{write:!1,reason:"shipped file has no version (skipping for safety)"}:s>i?{write:!1,reason:`current version ${s} > shipped ${i}`}:{write:!0,reason:s===i?"same version, refresh":`upgrade ${s} \u2192 ${i}`}}function ne(n,e){let t=`${e.trim()}
|
|
90
|
+
`;if(!m(n))return v(n,t),"created";let o=x(n,"utf8"),s=o.indexOf(J),i=o.indexOf(_);if(s!==-1&&i!==-1&&i>s){let r=o.slice(0,s),c=o.slice(i+_.length),p=r+t.trim()+c;return p===o?"unchanged":(v(n,p),"updated")}let l=o.endsWith(`
|
|
91
|
+
`)?`
|
|
92
|
+
`:`
|
|
93
|
+
|
|
94
|
+
`;return v(n,o+l+t),"appended"}function oe(n,e,t,o){let s=t.agents[n];if(!s)return[];let i=[],l=r=>g(P,r.split("/").map(c=>c.startsWith(".")?c.slice(1):c).join("/"));for(let r of s.files||[]){let c=l(r);if(!m(c))continue;let p=x(c,"utf8"),f=g(e,r),d=te(f,p,o);d.write?($(A(f),{recursive:!0}),v(f,p),i.push({path:r,action:"write",detail:d.reason})):i.push({path:r,action:"skip",detail:d.reason})}for(let r of s.writeOnceFiles||[]){let c=l(r);if(!m(c))continue;let p=g(e,r);if(m(p)){i.push({path:r,action:"skip",detail:"write-once: already exists, leaving alone"});continue}$(A(p),{recursive:!0}),v(p,x(c,"utf8")),i.push({path:r,action:"write",detail:"write-once: created"})}if(s.rootBlock){let r=g(P,s.rootBlock.source),c=x(r,"utf8"),p=g(e,s.rootBlock.target),f=ne(p,c);i.push({path:s.rootBlock.target,action:f})}return i}function se(n){let e=n;for(let t=0;t<6;t++){if(m(g(e,".git"))||m(g(e,".zibby.config.mjs"))||m(g(e,"package.json")))return e;let o=A(e);if(o===e)break;e=o}return n}async function re(n){let e=g(n,".zibby.config.mjs");if(!m(e))return null;try{return((await import(`file://${e}?t=${Date.now()}`)).default||{}).agentHelpers||null}catch(t){return console.log(h.gray(` (couldn't parse .zibby.config.mjs agentHelpers section: ${t.message})`)),null}}function j(n,e){let t=g(n,".zibby.config.mjs");if(!m(t)){console.log(h.yellow(" \u26A0\uFE0F .zibby.config.mjs not found; agentHelpers state not persisted"));return}let o=x(t,"utf8"),s=` agentHelpers: {
|
|
95
|
+
templateVersion: ${e.templateVersion},
|
|
96
|
+
agents: ${JSON.stringify(e.agents)},
|
|
97
|
+
updatedAt: '${new Date().toISOString()}',
|
|
98
|
+
},
|
|
99
|
+
`,i=/\s*agentHelpers\s*:\s*\{[^}]*\}\s*,?\n?/m,l;if(i.test(o))l=o.replace(i,`
|
|
100
|
+
${s}`);else{let r=o.lastIndexOf("}");if(r===-1)return;let c=o.slice(0,r),p=c.replace(/\s+$/,"").slice(-1);l=c+(p&&p!==","&&p!=="{"?`,
|
|
101
|
+
`:"")+s+o.slice(r)}v(t,l)}async function ie(n){let e=n[0]||"claude",t=[{name:`${I.claude}${n.includes("claude")?" (detected \u2014 recommended)":""}`,value:"claude"},{name:`${I.codex}${n.includes("codex")?" (detected)":""}`,value:"codex"},{name:`${I.cursor}${n.includes("cursor")?" (detected)":""}`,value:"cursor"},{name:"All of the above",value:"all"},{name:"Skip \u2014 I'll set up later",value:"none"}];if(e!=="claude"){let s=t.findIndex(i=>i.value===e);if(s>0){let[i]=t.splice(s,1);t.unshift(i)}}let o=await U({message:"Set up agent helpers (slash commands, sub-agents, context)?",choices:t,default:t[0].value});return o==="all"?["claude","cursor","codex"]:[o]}async function B(n={}){let{forcedAgents:e,force:t=!1,silent:o=!1}=n,s=n.projectRoot||se(process.cwd()),i=ee(),l=await re(s),r,c=n&&n.forcePrompt===!0;if(e&&e.length>0)r=e;else if(!c&&l&&Array.isArray(l.agents)&&l.agents.length>0)r=l.agents;else{if(o)return;if(!process.stdin.isTTY)console.log(h.gray(" (non-TTY; defaulting agent helpers to Claude Code \u2014 pass --agent to override)")),r=["claude"];else{let d=Q(s);try{r=await ie(d)}catch{r=["none"]}}}if(r.length===1&&r[0]==="none"){j(s,{templateVersion:i.templateVersion,agents:["none"]}),o||console.log(h.gray(" Skipped agent helpers. Run `zibby agents add` later to set up."));return}let p=X(s),f=[];for(let d of r){if(!i.agents[d])continue;oe(d,s,i,t).forEach(b=>f.push({...b,agent:d}))}if(j(s,{templateVersion:i.templateVersion,agents:r}),!o){let d=f.filter(u=>["write","created","updated","appended"].includes(u.action)),y=f.filter(u=>u.action==="skip"),b=f.filter(u=>u.action==="unchanged");if(d.length>0){console.log(h.green(` \u2713 Wrote ${d.length} agent helper file(s):`));for(let u of d)console.log(h.gray(` ${u.path} (${u.detail||u.action})`))}if(y.length>0){console.log(h.yellow(` \u26A0\uFE0F Skipped ${y.length} (use --force-claude to overwrite):`));for(let u of y)console.log(h.gray(` ${u.path} \u2014 ${u.detail}`))}d.length===0&&y.length===0&&b.length>0&&console.log(h.gray(` Agent helpers up to date (template v${i.templateVersion}).`)),console.log(h.gray(` agentHelpers: { agents: [${r.map(u=>`"${u}"`).join(", ")}] } saved to .zibby.config.mjs`)),p&&console.log(h.gray(` CLI shim: ${p} (agents fall back to ./.zibby/bin/zibby if zibby isn't on PATH)`)),r.includes("claude")&&console.log(h.gray(" \u2192 Run `claude` and try /zibby-add-node"))}}function ce(n){return new Promise(e=>{let t=ae("npm",["install","--no-audit","--no-fund"],{cwd:n,stdio:"pipe",shell:!1}),o="";t.stderr.on("data",s=>{o+=s.toString()}),t.on("error",()=>e({ok:!1,stderr:"npm not found on PATH"})),t.on("close",s=>e({ok:s===0,stderr:o}))})}var de=/^[a-z][a-z0-9-]{0,62}[a-z0-9]$/,R=["stellar","quantum","cosmic","nova","nebula","solar","lunar","atomic","plasma","fusion","pulse","flux","spark","blaze","ember","radiant","luminous","electric","magnetic","kinetic","neon","cyber","pixel","matrix","vector","synth","neural","prism","zenith","phoenix","catalyst","nexus","echo","wave","crystal","jade","ruby","emerald","onyx","amber","silver","turbo","lightning","thunder","storm","arcane","mystic","ethereal","celestial","swift","crimson","iron","cobalt"],O=["flow","runner","pipeline","stream","circuit","engine","beacon","forge","relay","shuttle","conduit","gateway","sentinel","scout","pilot","voyager","ranger","dispatch","signal","pulse","agent","daemon","spark","orbit","vector","nexus","matrix","grid","mesh","bridge","link","node","craft","bolt","ray","arc","wave","hook","probe","shard"];function ue(){let n=R[Math.floor(Math.random()*R.length)],e=O[Math.floor(Math.random()*O.length)];return`${n}-${e}`}function pe(n){return`${n.split("-").map(e=>e.charAt(0).toUpperCase()+e.slice(1)).join("")}Workflow`}function ge(n,e){return`/**
|
|
102
|
+
* ${n}
|
|
86
103
|
*
|
|
87
104
|
* buildGraph() \u2014 define nodes, edges, conditional routing
|
|
88
105
|
* onComplete(result) \u2014 post-processing after the graph finishes
|
|
@@ -91,7 +108,7 @@ ${s}
|
|
|
91
108
|
import { WorkflowAgent, WorkflowGraph } from '@zibby/core';
|
|
92
109
|
import { exampleNode } from './nodes/index.mjs';
|
|
93
110
|
|
|
94
|
-
export class ${
|
|
111
|
+
export class ${n} extends WorkflowAgent {
|
|
95
112
|
buildGraph() {
|
|
96
113
|
const graph = new WorkflowGraph();
|
|
97
114
|
|
|
@@ -104,11 +121,11 @@ export class ${o} extends WorkflowAgent {
|
|
|
104
121
|
}
|
|
105
122
|
|
|
106
123
|
async onComplete(result) {
|
|
107
|
-
console.log(\`[${
|
|
124
|
+
console.log(\`[${e}] workflow complete \u2014 success: \${result.success !== false}\`);
|
|
108
125
|
}
|
|
109
126
|
}
|
|
110
|
-
`}function
|
|
111
|
-
`}function
|
|
127
|
+
`}function me(){return`export { exampleNode } from './example.mjs';
|
|
128
|
+
`}function fe(){return`import { z } from '@zibby/core';
|
|
112
129
|
|
|
113
130
|
const ExampleOutputSchema = z.object({
|
|
114
131
|
summary: z.string().describe('A short summary of the result'),
|
|
@@ -127,19 +144,19 @@ Analyze the input and return a summary with a status.\`,
|
|
|
127
144
|
|
|
128
145
|
outputSchema: ExampleOutputSchema,
|
|
129
146
|
};
|
|
130
|
-
`}function
|
|
131
|
-
`}function
|
|
132
|
-
`}async function
|
|
133
|
-
No name provided \u2014 generated: ${
|
|
134
|
-
Invalid workflow name: "${
|
|
135
|
-
`)),process.exit(1));let
|
|
136
|
-
Common picks: ${
|
|
137
|
-
`));let
|
|
138
|
-
Workflow already exists: ${
|
|
139
|
-
`)),process.exit(1));let
|
|
140
|
-
Created:`)),console.log(
|
|
141
|
-
`)[0]||"unknown error"})`),console.log(
|
|
142
|
-
Next steps:`));let
|
|
143
|
-
`))}catch(
|
|
144
|
-
${
|
|
145
|
-
`)),process.exit(1)}}export{
|
|
147
|
+
`}function ye(n,e){return`${JSON.stringify({name:n,description:`${e} workflow`,entryClass:e,triggers:{api:!0}},null,2)}
|
|
148
|
+
`}function he(){return`${JSON.stringify({type:"module",dependencies:{"@zibby/core":"^0.1.48"}},null,2)}
|
|
149
|
+
`}async function Ct(n,e={}){let t;n?t=n.toLowerCase():(t=ue(),console.log(a.gray(`
|
|
150
|
+
No name provided \u2014 generated: ${a.white(t)}`))),de.test(t)||(console.log(a.red(`
|
|
151
|
+
Invalid workflow name: "${n}"`)),console.log(a.gray(" Must be lowercase, start with a letter, use only a-z, 0-9, hyphens")),console.log(a.gray(" Length: 2\u201364 characters")),console.log(a.gray(` Example: ticket-triage, pr-review, deploy-checker
|
|
152
|
+
`)),process.exit(1));let o=process.cwd(),s=w(o,".zibby.config.mjs"),i=K(s),l=".zibby/workflows";if(i)try{let d=await E(o);d?.paths?.workflows&&(l=d.paths.workflows)}catch(d){console.log(a.yellow(` \u26A0\uFE0F Could not load .zibby.config.mjs: ${d.message}`)),console.log(a.gray(` Using default path: ${l}`))}else if(process.stdin.isTTY){console.log(a.gray(`
|
|
153
|
+
Common picks: ${a.white(".zibby/workflows")} (hidden, default) \xB7 ${a.white("workflows")} (visible at project root)
|
|
154
|
+
`));let y=(await le({message:`Where to save workflows? ${a.gray("[Enter for .zibby/workflows]")}`})).trim();l=y===""?".zibby/workflows":y}else console.log(a.gray(` Using default workflows path: ${l} (run interactively to customize)`));let r=w(o,l,t),c=w(r,"nodes");K(r)&&(console.log(a.red(`
|
|
155
|
+
Workflow already exists: ${l}/${t}/`)),console.log(a.gray(` Choose a different name or delete the existing folder.
|
|
156
|
+
`)),process.exit(1));let p=pe(t),f=M(` Scaffolding workflow "${t}"...`).start();try{await T(w(o,l),{recursive:!0}),await T(c,{recursive:!0}),await Promise.all([k(w(r,"graph.mjs"),ge(p,t)),k(w(c,"index.mjs"),me()),k(w(c,"example.mjs"),fe()),k(w(r,"workflow.json"),ye(t,p)),k(w(r,"package.json"),he())]);let d=!1;if(!i){let u=S({agent:"claude",browserMode:"headless",workflowsPath:l},{},{memoryBackend:"dolt"});await k(s,u),d=!0}f.succeed(` Scaffolded ${a.bold(t)}`),console.log(a.green(`
|
|
157
|
+
Created:`)),console.log(a.white(` ${l}/${t}/`)),console.log(a.gray(` graph.mjs ${p} (entry)`)),console.log(a.gray(" nodes/index.mjs barrel export")),console.log(a.gray(" nodes/example.mjs starter node (prompt + schema)")),console.log(a.gray(" workflow.json manifest")),console.log(a.gray(" package.json dependencies (@zibby/core)")),d&&console.log(a.white(" .zibby.config.mjs project config (agent: claude \u2014 edit to switch)"));let y=!1;if(e.skipInstall)console.log(a.gray(` Skipped npm install (--skip-install). Run manually: cd ${l}/${t} && npm install`));else{let u=M({text:` Installing dependencies in ${l}/${t}/...`,prefixText:""}).start(),C=await ce(r);C.ok?(u.succeed(` Installed dependencies in ${l}/${t}/`),y=!0):(u.warn(` Could not install dependencies (${C.stderr.trim().split(`
|
|
158
|
+
`)[0]||"unknown error"})`),console.log(a.gray(` Run manually: cd ${l}/${t} && npm install`)))}try{let u=e.agent;await B({projectRoot:o,forcedAgents:u?u==="all"?["claude","cursor","codex"]:u==="none"?["none"]:[u]:void 0,force:e.forceClaude===!0||e.forceAgents===!0})}catch(u){console.log(a.yellow(` \u26A0\uFE0F Agent helpers setup failed: ${u.message}`)),console.log(a.gray(" Run `zibby agents add` later to retry."))}console.log(a.white(`
|
|
159
|
+
Next steps:`));let b=1;!y&&e.skipInstall!==!0&&console.log(a.cyan(` ${b++}. Install deps: cd ${l}/${t} && npm install`)),console.log(a.cyan(` ${b++}. Edit nodes in ${l}/${t}/nodes/`)),console.log(a.cyan(` ${b++}. Wire them in graph.mjs`)),console.log(a.cyan(` ${b++}. Run locally: zibby workflow run ${t}`)),console.log(a.cyan(` ${b++}. Deploy to cloud: zibby workflow deploy ${t}
|
|
160
|
+
`))}catch(d){f.fail(" Scaffold failed"),console.log(a.red(`
|
|
161
|
+
${d.message}
|
|
162
|
+
`)),process.exit(1)}}export{Ct as generateWorkflowCommand};
|
|
@@ -0,0 +1,7 @@
|
|
|
1
|
+
import{existsSync as c,readFileSync as f}from"fs";import{resolve as a}from"path";function p(e){return e==="true"?!0:e==="false"?!1:e==="null"?null:e!==""&&!isNaN(Number(e))?Number(e):e}function m(e){let r={};for(let t of e||[]){let l=t.indexOf("=");if(l===-1){console.warn(` Warning: ignored param "${t}" \u2014 expected key=value format`);continue}let s=t.slice(0,l).trim(),u=p(t.slice(l+1)),n=s.split("."),o=r;for(let i=0;i<n.length-1;i++)(typeof o[n[i]]!="object"||o[n[i]]===null)&&(o[n[i]]={}),o=o[n[i]];o[n[n.length-1]]=u}return r}function g(e){let r=a(e);c(r)||(console.log(`
|
|
2
|
+
Error: --input-file not found: ${e}
|
|
3
|
+
`),process.exit(1));try{return JSON.parse(f(r,"utf-8"))}catch(t){console.log(`
|
|
4
|
+
Error: --input-file is not valid JSON: ${t.message}
|
|
5
|
+
`),process.exit(1)}}function N(e){let r={};if(e.inputFile&&(r={...g(e.inputFile)}),e.input)try{r={...r,...JSON.parse(e.input)}}catch(t){console.log(`
|
|
6
|
+
Error: --input is not valid JSON`),console.log(` ${t.message}
|
|
7
|
+
`),process.exit(1)}return e.param?.length&&(r={...r,...m(e.param)}),r}export{p as coerceValue,m as parseParams,N as resolveInput};
|
|
@@ -22,7 +22,7 @@ import{readdir as wo,stat as ho}from"fs/promises";import{existsSync as P}from"fs
|
|
|
22
22
|
Local Workflows (${t})
|
|
23
23
|
`)),console.log(r.gray(" ".padEnd(60,"-"))),console.log(r.white(" Name".padEnd(35))+r.white("Files".padEnd(25))),console.log(r.gray(" ".padEnd(60,"-")));for(let c of i){let a=[];c.hasGraph&&a.push("graph"),c.hasManifest&&a.push("manifest"),console.log(` ${r.cyan(c.name.padEnd(33))}${r.gray(a.join(", "))}`)}return console.log(r.gray(" ".padEnd(60,"-"))),console.log(r.gray(`
|
|
24
24
|
Total: ${i.length} workflow${i.length===1?"":"s"}
|
|
25
|
-
`)),console.log(r.white(" Commands:")),console.log(r.cyan(" zibby workflow
|
|
25
|
+
`)),console.log(r.white(" Commands:")),console.log(r.cyan(" zibby workflow run <name> ")+r.gray("Run locally")),console.log(r.cyan(" zibby workflow deploy <name> ")+r.gray("Deploy to cloud")),console.log(r.cyan(" zibby workflow trigger <uuid> ")+r.gray("Run workflow (returns job ID)")),console.log(r.cyan(" zibby workflow logs <uuid> ")+r.gray("Tail execution logs")),console.log(""),i}catch(u){if(o.quiet)return[];console.log(r.red(`
|
|
26
26
|
Error reading workflows: ${u.message}
|
|
27
27
|
`)),process.exit(1)}}async function pe(o={}){let e=await bo({...o,quiet:!0}),{sessionToken:t,apiKey:l}=await q({apiKey:o.apiKey,optional:!0}),u=[],i=t||l;if(i)try{let n=I(),p=await fetch(`${n}/projects`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`}});if(p.ok){let h=(await p.json()).projects||[];for(let v of h){let B=await fetch(`${n}/projects/${v.projectId}/workflows`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`}});if(B.ok){let H=await B.json(),X=["analysis","implementation","run_test"];for(let k of H)X.includes(k.workflowType)||u.push({uuid:k.uuid,name:k.workflowType,projectId:v.projectId,projectName:v.name||v.projectId,version:k.version||0,nodes:k.graphJson?JSON.parse(k.graphJson).nodes?.length:0,updatedAt:k.updatedAt||null,source:"remote"})}}}}catch{}let c=[],a=new Map(e.map(n=>[n.name,n]));for(let n of u)c.push({uuid:n.uuid,name:n.name,project:n.projectName,version:n.version});for(let[n,p]of a.entries())u.some(h=>h.name===n)||c.push({uuid:null,name:p.name,project:"-",version:"-"});if(c.length===0){console.log(`
|
|
28
28
|
No workflows found
|
|
@@ -31,5 +31,5 @@ import{readdir as wo,stat as ho}from"fs/promises";import{existsSync as P}from"fs
|
|
|
31
31
|
Workflows
|
|
32
32
|
`),console.log(` ${T}`),console.log(` ${r.bold(m)}`),console.log(` ${x}`);for(let n of y){let p=g.map(d=>{let h=String(n[d]).padEnd(w[d]);return` ${d==="uuid"&&n.uuid!=="-"?r.cyan(h):d==="name"?r.white(h):h} `});console.log(` \u2502${p.join("\u2502")}\u2502`)}console.log(` ${$}`),console.log(`
|
|
33
33
|
Total: ${c.length} workflow${c.length===1?"":"s"}
|
|
34
|
-
`),console.log(" Commands:"),console.log(" zibby workflow new <name> Scaffold a new workflow"),console.log(" zibby workflow
|
|
34
|
+
`),console.log(" Commands:"),console.log(" zibby workflow new <name> Scaffold a new workflow"),console.log(" zibby workflow run <name> Run locally"),console.log(" zibby workflow deploy <name> Ship to cloud (generates UUID on first deploy)"),console.log(" zibby workflow trigger <uuid> Run a deployed workflow remotely"),console.log(" zibby workflow logs <uuid> Tail logs from a remote run"),console.log(" zibby workflow download <uuid> Pull remote back to local (then edit + redeploy)"),console.log(" zibby workflow delete <uuid> Delete a deployed workflow"),i||console.log(`
|
|
35
35
|
Set ZIBBY_API_KEY to see remote workflows`),console.log("")}export{pe as listAllWorkflowsCommand,bo as listLocalWorkflowsCommand};
|
|
@@ -1,42 +1,42 @@
|
|
|
1
|
-
import e from"chalk";import{readFileSync as v,existsSync as N}from"fs";import{homedir as
|
|
1
|
+
import e from"chalk";import{readFileSync as v,existsSync as N,writeSync as R}from"fs";import{homedir as C}from"os";import{join as _}from"path";var x="https://logs.workflows.zibby.app",j="https://logs-stream.zibby.app/",b=null;async function A(i){return b||(process.env.ZIBBY_SSE_ENDPOINT?(b=process.env.ZIBBY_SSE_ENDPOINT,b):(b=j,b))}function L(i){let r=_(C(),".zibby","config.json");N(r)||(console.log(e.red(`
|
|
2
2
|
Not authenticated`)),console.log(e.gray(` Run: zibby login
|
|
3
|
-
`)),process.exit(1));let
|
|
3
|
+
`)),process.exit(1));let s;try{s=JSON.parse(v(r,"utf-8"))}catch{console.log(e.red(`
|
|
4
4
|
Config file corrupt`)),console.log(e.gray(` Run: zibby login
|
|
5
|
-
`)),process.exit(1)}let o=
|
|
5
|
+
`)),process.exit(1)}let o=s.sessionToken;o||(console.log(e.red(`
|
|
6
6
|
Not authenticated`)),console.log(e.gray(` Run: zibby login
|
|
7
|
-
`)),process.exit(1));let
|
|
7
|
+
`)),process.exit(1));let f=i.project;return{token:o,projectId:f}}function k(i){return new Date(i).toISOString().replace("T"," ").replace("Z","")}async function E(i,r){let s=await fetch(i,{headers:{Authorization:`Bearer ${r}`}});if(!s.ok){let o=await s.text();throw new Error(`API ${s.status}: ${o}`)}return s.json()}async function O(i,r,s,o){return i||(console.log(e.red(`
|
|
8
8
|
Workflow UUID is required`)),console.log(e.gray(" Usage: zibby workflow logs <workflow-uuid>")),console.log(e.gray(` zibby workflow logs <workflow-uuid> -t
|
|
9
9
|
`)),process.exit(1)),i}function F(i,r){let o=((i||"")+r).split(`
|
|
10
|
-
`),
|
|
10
|
+
`),f=o.pop()||"",a=[],u=null,n=null;for(let d=0;d<o.length;d++){let l=o[d];if(l.trim()){if(l.startsWith("id:")){u=l.slice(3).trim();continue}if(l.startsWith("event:")){let g=l.slice(6).trim();if(g==="log")continue;if(g==="status"){let c=o[d+1];if(c&&c.startsWith("data:"))try{let y=JSON.parse(c.slice(5).trim());y.status==="new_execution"?a.push({type:"newExecution",executionId:y.executionId,taskId:y.taskId}):y.status==="waiting"&&a.push({type:"waiting"})}catch{}continue}if(g==="complete"){a.push({type:"complete"});continue}if(g==="error"){let c=o[d+1];if(c&&c.startsWith("data:"))try{if(JSON.parse(c.slice(5).trim()).error==="No executions found for workflow"){n={type:"notFound"};break}}catch{}n={type:"failed"};break}continue}if(l.startsWith("data:")){let g=l.slice(5).trim();if(!g)continue;try{let c=JSON.parse(g);c.timestamp&&c.message&&a.push({type:"log",timestamp:c.timestamp,message:c.message,taskId:c.taskId})}catch{}}}}return{actions:a,remainder:f,lastEventId:u,returnSignal:n}}async function J({token:i,executionId:r,sseEndpoint:s,stopped:o}){let f=null;try{let a=new URL(s);a.searchParams.set("jobId",r),f&&a.searchParams.set("lastEventId",f);let u=await fetch(a.toString(),{headers:{Authorization:`Bearer ${i}`,Accept:"text/event-stream"}});if(!u.ok)throw new Error(`SSE connection failed: ${u.status} ${u.statusText}`);let n=u.body.getReader(),d=new TextDecoder,l="",g=!1;for(;!o.value;){let{done:c,value:y}=await n.read();if(c)break;let w=F(l,d.decode(y,{stream:!0}));l=w.remainder,w.lastEventId&&(f=w.lastEventId);for(let t of w.actions)switch(t.type){case"newExecution":{let m=`${t.executionId.slice(0,8)}...${t.executionId.slice(-4)}`,h=t.taskId?t.taskId.slice(-8):"pending";console.log(e.cyan(`
|
|
11
11
|
\u250C\u2500 Execution: ${m} (task: ${h})`)),console.log(e.cyan(` \u2514\u2500 Streaming logs...
|
|
12
12
|
`));break}case"waiting":console.log(e.gray(`
|
|
13
|
-
Waiting for next execution...`));break;case"complete":
|
|
13
|
+
Waiting for next execution...`));break;case"complete":g=!0;break;case"log":{let m=e.gray(k(t.timestamp)),h=t.taskId?e.gray(`(${t.taskId.slice(-8)}) `):"";console.log(`${m} ${h}${t.message.replace(/\n$/,"")}`);break}}if(w.returnSignal)return w.returnSignal.type==="notFound"?{notFound:!0}:{failed:!0}}return{completed:g}}catch(a){if(a.name==="AbortError")return{aborted:!0};throw a}}function M(i,{baseMs:r=500,capMs:s=3e4,rand:o=Math.random}={}){let f=Math.min(s,r*Math.pow(2,Math.max(0,i)));return Math.floor(o()*f)}async function z({attemptStream:i,stopped:r,follow:s,logger:o,sleep:f=d=>new Promise(l=>setTimeout(l,d)),exit:a=d=>{throw new Error(`exit:${d}`)},backoff:u=M,notFoundPollMs:n=5e3}){let d=0,l=!1;for(;!r.value;){let g;try{g=await i(),d=0}catch(c){if(c.name==="AbortError"||r.value)return{reason:"aborted"};if(l||(o.error(` SSE Error: ${c.message}`),s&&o.gray(" Reconnecting..."),l=!0),!s)return a("error")??{reason:"error"};let y=u(d);d++,await f(y);continue}if(g.aborted||r.value)return{reason:"aborted"};if(g.notFound){if(s){l||(o.yellow(" No executions found yet. Waiting for workflow to be triggered..."),o.gray(" Press Ctrl+C to stop."),l=!0),await f(n);continue}return o.yellow(`
|
|
14
14
|
No executions found for this workflow. Trigger the workflow first.
|
|
15
15
|
`),a("notFound")??{reason:"notFound"}}if(l&&(o.gray(` Reconnected.
|
|
16
|
-
`),l=!1),
|
|
17
|
-
Execution failed.`),
|
|
18
|
-
`):"");let
|
|
19
|
-
`)),T({token:i,projectId:null,jobId:r,follow:
|
|
16
|
+
`),l=!1),g.failed)return o.red(`
|
|
17
|
+
Execution failed.`),s?{reason:"failed"}:a("failed")??{reason:"failed"};if(g.completed)return a("completed")??{reason:"completed"};if(!s)return{reason:"disconnected"}}return{reason:"stopped"}}async function W({token:i,jobId:r,follow:s,projectId:o}){console.log(e.gray(` Streaming logs for workflow ${e.cyan(r)}...`)),console.log(s?e.gray(` Press Ctrl+C to stop.
|
|
18
|
+
`):"");let f=await A(i);if(!f)return console.log(e.yellow(` SSE endpoint not configured, using CloudWatch polling...
|
|
19
|
+
`)),T({token:i,projectId:null,jobId:r,follow:s,limit:1e5});let a={value:!1},u=()=>{a.value=!0;try{R(2,`
|
|
20
20
|
Stopped streaming.
|
|
21
|
-
`)
|
|
21
|
+
`)}catch{}process.exit(0)};process.prependListener("SIGINT",u),process.prependListener("SIGTERM",u),await z({attemptStream:()=>J({token:i,executionId:r,sseEndpoint:f,stopped:a}),stopped:a,follow:s,logger:{gray:n=>console.log(e.gray(n)),red:n=>console.log(e.red(n)),yellow:n=>console.log(e.yellow(n)),error:n=>console.error(e.red(n))},exit:n=>{n==="completed"&&process.exit(0),(n==="error"||n==="notFound"||n==="failed")&&process.exit(1)}})}async function T({token:i,projectId:r,jobId:s,follow:o,limit:f}){let a=r?`${x}/logs/${r}/${s}`:`${x}/job/${s}`,u=null,n=0,d=new Set,l=!1,g=0,c=5,y=()=>{l=!0,console.log(e.gray(`
|
|
22
22
|
Stopped tailing.
|
|
23
|
-
`)),process.exit(0)};for(process.on("SIGINT",y),process.on("SIGTERM",y),console.log(e.gray(` Fetching logs for workflow ${e.cyan(
|
|
24
|
-
`):"");!l;)try{let w=new URLSearchParams({limit:String(
|
|
23
|
+
`)),process.exit(0)};for(process.on("SIGINT",y),process.on("SIGTERM",y),console.log(e.gray(` Fetching logs for workflow ${e.cyan(s)}...`)),console.log(o?e.gray(` Press Ctrl+C to stop.
|
|
24
|
+
`):"");!l;)try{let w=new URLSearchParams({limit:String(f)});u&&w.set("nextToken",u);let t=await E(`${a}?${w}`,i);g=0,t.message&&t.lines?.length===0&&n===0&&console.log(e.gray(` ${t.message}`)),t.status==="starting"&&t.lines?.length===0&&n===0&&console.log(e.gray(" Container starting..."));for(let p of t.lines||[]){let $=`${p.timestamp}:${p.message}`;if(d.has($))continue;d.add($);let I=e.gray(k(p.timestamp)),P=t.taskId?e.gray(`(${t.taskId.slice(-8)}) `):"";console.log(`${I} ${P}${p.message.replace(/\n$/,"")}`)}if(n=t.lines?.length>0?0:n+1,u=t.nextForwardToken||null,t.status==="completed"||t.status==="failed"){let p=t.status==="completed"?e.green:e.red;console.log(p(`
|
|
25
25
|
Job ${t.status}.`)),process.exit(t.status==="completed"?0:1)}if(!o){t.status&&console.log(e.gray(`
|
|
26
|
-
Status: ${t.status}`));break}let S=t.lines?.length>0?500:
|
|
26
|
+
Status: ${t.status}`));break}let S=t.lines?.length>0?500:n>5?5e3:2e3;await new Promise(p=>setTimeout(p,S))}catch(w){if(w.name==="AbortError")break;w.message.match(/API (400|401|403|404):/)&&(console.error(e.red(`
|
|
27
27
|
${w.message}
|
|
28
|
-
`)),process.exit(1)),
|
|
28
|
+
`)),process.exit(1)),g++,console.error(e.red(` Error: ${w.message}`)),g>=c&&(console.error(e.red(`
|
|
29
29
|
Too many consecutive errors (${c}). Stopping.
|
|
30
|
-
`)),process.exit(1)),o||process.exit(1),await new Promise(m=>setTimeout(m,3e3))}}async function U({token:i,projectId:r,workflow:
|
|
30
|
+
`)),process.exit(1)),o||process.exit(1),await new Promise(m=>setTimeout(m,3e3))}}async function U({token:i,projectId:r,workflow:s,follow:o,limit:f}){let a=`${x}/all/${r}`,u=null,n=0,d=new Set,l=null,g=!1,c=0,y=5,w=()=>{g=!0,console.log(e.gray(`
|
|
31
31
|
Stopped tailing.
|
|
32
32
|
`)),process.exit(0)};for(process.on("SIGINT",w),process.on("SIGTERM",w),console.log(e.gray(`
|
|
33
|
-
Tailing all runs for ${e.cyan(
|
|
34
|
-
`):"");!
|
|
35
|
-
... more logs available. Run again or use --follow to stream.`)),m.jobCount&&console.log(e.gray(` ${m.jobCount} job(s) found.`));break}if(!m.hasRunning&&!u&&
|
|
36
|
-
No running jobs. All caught up.`));break}let S=m.lines?.length>0?500:
|
|
33
|
+
Tailing all runs for ${e.cyan(s)}...`)),console.log(o?e.gray(` Press Ctrl+C to stop.
|
|
34
|
+
`):"");!g;)try{let t=new URLSearchParams({workflow:s,limit:String(f)});u&&t.set("nextToken",u);let m=await E(`${a}?${t}`,i);c=0,m.message&&m.lines?.length===0&&n===0&&console.log(e.gray(` ${m.message}`));for(let p of m.lines||[]){let $=`${p.timestamp}:${p.jobId}:${p.message}`;if(d.has($))continue;d.add($),p.jobId!==l&&(l!==null&&console.log(""),console.log(e.dim(` \u2500\u2500 ${p.jobId} \u2500\u2500`)),l=p.jobId);let I=e.gray(k(p.timestamp));console.log(`${I} ${p.message.replace(/\n$/,"")}`)}if(n=m.lines?.length>0?0:n+1,u=m.nextToken||null,!o){u&&console.log(e.gray(`
|
|
35
|
+
... more logs available. Run again or use --follow to stream.`)),m.jobCount&&console.log(e.gray(` ${m.jobCount} job(s) found.`));break}if(!m.hasRunning&&!u&&n>2){console.log(e.gray(`
|
|
36
|
+
No running jobs. All caught up.`));break}let S=m.lines?.length>0?500:n>5?5e3:2e3;await new Promise(p=>setTimeout(p,S))}catch(t){if(t.name==="AbortError")break;t.message.match(/API (400|401|403|404):/)&&(console.error(e.red(`
|
|
37
37
|
${t.message}
|
|
38
38
|
`)),process.exit(1)),c++,console.error(e.red(` Error: ${t.message}`)),c>=y&&(console.error(e.red(`
|
|
39
39
|
Too many consecutive errors (${y}). Stopping.
|
|
40
|
-
`)),process.exit(1)),o||process.exit(1),await new Promise(h=>setTimeout(h,3e3))}}async function
|
|
40
|
+
`)),process.exit(1)),o||process.exit(1),await new Promise(h=>setTimeout(h,3e3))}}async function V(i,r){let{token:s,projectId:o}=L(r),f=r.follow===!0,a=r.lines?parseInt(r.lines,10):1e5;if(r.all){let n=r.workflow;return n||(console.log(e.red(`
|
|
41
41
|
--workflow is required with --all`)),console.log(e.gray(` Example: zibby workflow logs --workflow ticket-triage --all --project <id>
|
|
42
|
-
`)),process.exit(1)),U({token:
|
|
42
|
+
`)),process.exit(1)),U({token:s,projectId:o,workflow:n,follow:f,limit:a})}let u=await O(i,r,s,o);return f?W({token:s,jobId:u,follow:f,projectId:o}):T({token:s,projectId:o,jobId:u,follow:!1,limit:a})}export{V as logsCommand,F as parseSseChunk,z as runReconnectLoop,M as sseBackoffMs};
|
|
@@ -0,0 +1,19 @@
|
|
|
1
|
+
import{existsSync as d}from"fs";import{readFile as U}from"fs/promises";import{join as g}from"path";import{pathToFileURL as W}from"url";import s from"chalk";import D from"ora";import{existsSync as F}from"fs";import{join as N}from"path";import{pathToFileURL as S}from"url";async function k(o){let r=N(o,".zibby.config.mjs");if(!F(r))throw new Error(".zibby.config.mjs not found");try{let e=await import(S(r).href);return e.default||e}catch(e){throw new Error(`Failed to load .zibby.config.mjs: ${e.message}`,{cause:e})}}import{existsSync as B,readFileSync as I}from"fs";import{resolve as P}from"path";function z(o){return o==="true"?!0:o==="false"?!1:o==="null"?null:o!==""&&!isNaN(Number(o))?Number(o):o}function L(o){let r={};for(let e of o||[]){let t=e.indexOf("=");if(t===-1){console.warn(` Warning: ignored param "${e}" \u2014 expected key=value format`);continue}let l=e.slice(0,t).trim(),a=z(e.slice(t+1)),n=l.split("."),c=r;for(let i=0;i<n.length-1;i++)(typeof c[n[i]]!="object"||c[n[i]]===null)&&(c[n[i]]={}),c=c[n[i]];c[n[n.length-1]]=a}return r}function O(o){let r=P(o);B(r)||(console.log(`
|
|
2
|
+
Error: --input-file not found: ${o}
|
|
3
|
+
`),process.exit(1));try{return JSON.parse(I(r,"utf-8"))}catch(e){console.log(`
|
|
4
|
+
Error: --input-file is not valid JSON: ${e.message}
|
|
5
|
+
`),process.exit(1)}}function x(o){let r={};if(o.inputFile&&(r={...O(o.inputFile)}),o.input)try{r={...r,...JSON.parse(o.input)}}catch(e){console.log(`
|
|
6
|
+
Error: --input is not valid JSON`),console.log(` ${e.message}
|
|
7
|
+
`),process.exit(1)}return o.param?.length&&(r={...r,...L(o.param)}),r}function j({workflowType:o,jobId:r,projectId:e,agentType:t,model:l}){let a="\u2500".repeat(60),n=`${t||"default"} (model: ${l||"auto"})`;return["",a,` Workflow: ${o}`,` Job: ${r||"local"}`,` Project: ${e||"none"}`,` Agent: ${n}`,a].join(`
|
|
8
|
+
`)}async function G(o){try{let r=await k(o);return{userConfig:r,workflowsBasePath:r?.paths?.workflows||".zibby/workflows"}}catch{return{userConfig:null,workflowsBasePath:".zibby/workflows"}}}async function J(o,r,e){let t=g(o,"graph.mjs");if(!d(t))throw new Error(`graph.mjs not found in ${e}/${r}/`);let l=await T(o,r),a=await import(W(t).href),n=l.entryClass,c=n&&a[n]||a.default||Object.values(a).find(i=>typeof i=="function"&&i.prototype?.buildGraph);if(!c)throw new Error("No WorkflowAgent class found in graph.mjs. Export a class with buildGraph() method.");return{AgentClass:c,manifest:l}}async function T(o,r){let e=g(o,"workflow.json");if(!d(e))return{name:r,triggers:{api:!0}};let t=await U(e,"utf-8");return JSON.parse(t)}async function so(o,r={}){o||(console.log(s.red(`
|
|
9
|
+
Workflow name is required`)),console.log(s.gray(" Usage: zibby workflow run <workflow-name>")),console.log(s.gray(` Example: zibby workflow run my-pipeline -p ticket=BUG-123
|
|
10
|
+
`)),process.exit(1));let e=o.toLowerCase(),t=process.cwd(),{userConfig:l,workflowsBasePath:a}=await G(t),n=g(t,a,e);d(n)||(console.log(s.red(`
|
|
11
|
+
Workflow not found: ${a}/${e}/`)),console.log(s.gray(" Create one first:")),console.log(s.cyan(` zibby workflow new ${e}
|
|
12
|
+
`)),process.exit(1));let c=x(r),i=D(` Loading workflow "${e}"...`).start(),m,w;try{({AgentClass:m,manifest:w}=await J(n,e,a)),i.succeed(` Loaded ${s.bold(w.entryClass||m.name)} (${e})`)}catch(u){i.fail(" Failed to load workflow"),console.log(s.red(`
|
|
13
|
+
${u.message}
|
|
14
|
+
`)),process.exit(1)}let y=`local-${Date.now()}`,C=j({workflowType:e,jobId:y,projectId:l?.projectId,agentType:l?.agent?.provider||process.env.AGENT_TYPE,model:l?.agent?.model||process.env.MODEL});console.log(C);let h=Date.now(),E={input:c,cwd:t,runId:y,config:l||{}},f,p;try{p=new m({workflow:e}),f=await p.buildGraph().run(p,E)}catch(u){let v=((Date.now()-h)/1e3).toFixed(1);console.log(s.red(`
|
|
15
|
+
\u2716 ${e} failed after ${v}s`)),console.log(s.red(` ${u.message}
|
|
16
|
+
`)),process.env.ZIBBY_DEBUG&&console.error(u.stack),process.exit(1)}let $=((Date.now()-h)/1e3).toFixed(1),b=f?.success!==!1;b?console.log(s.green(`
|
|
17
|
+
\u2714 ${e} completed in ${$}s`)):(console.log(s.red(`
|
|
18
|
+
\u2716 ${e} failed after ${$}s`)),f?.error&&console.log(s.red(` ${f.error}
|
|
19
|
+
`))),p.onComplete&&await p.onComplete(f),b||process.exit(1)}export{so as runLocalWorkflowCommand};
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
#!/usr/bin/env node
|
|
2
|
-
import{mkdirSync as
|
|
3
|
-
`)}import"@zibby/core";var T=process.env.WORKSPACE||"/workspace";async function
|
|
4
|
-
Workflow execution failed: ${
|
|
5
|
-
[done] ${
|
|
6
|
-
[done] ${
|
|
2
|
+
import{mkdirSync as j,writeFileSync as oe,existsSync as w,readFileSync as K}from"fs";import{join as g,dirname as te,resolve as G}from"path";import{pathToFileURL as x}from"url";import{execSync as ne,spawn as W}from"node:child_process";import{SQSClient as Y,SendMessageCommand as X}from"@aws-sdk/client-sqs";var P=null;function V(){return P||(P=new Y({region:process.env.AWS_REGION||"ap-southeast-2"})),P}async function L(s,{status:e,error:o}){let{EXECUTION_ID:t,SQS_AUTH_TOKEN:c,PROGRESS_API_URL:i,PROGRESS_QUEUE_URL:r,PROJECT_API_TOKEN:a}=s;if(!t)return;let f={executionId:t,...c&&{sqsAuthToken:c},status:e,...o&&{error:o},timestamp:new Date().toISOString()},l=i?"HTTP":r?"SQS":"NONE",u=JSON.stringify(f).length;console.log(`Sending final status: ${e} via ${l} (${(u/1024).toFixed(1)}KB)`);try{if(i)await Z(i,t,f,a);else if(r){let d=["completed","failed","insufficient_context","blocked"].includes(e)?"execution_completed":"progress_update";await ee(r,t,f,d)}else{console.warn("No transport configured for final status \u2014 neither PROGRESS_API_URL nor PROGRESS_QUEUE_URL set");return}console.log(`Final status ${e} sent via ${l}`)}catch(d){console.error(`Failed to send final status (${e}) via ${l}:`),console.error(` Payload: ${(u/1024).toFixed(1)}KB`),console.error(` Error: ${d.message}`),d.name&&console.error(` Error type: ${d.name}`),d.code&&console.error(` Error code: ${d.code}`)}}async function Z(s,e,o,t){let c=`${s}/${e}/progress`,i={"Content-Type":"application/json"};t&&(i.Authorization=`Bearer ${t}`);let r=await fetch(c,{method:"POST",headers:i,body:JSON.stringify(o)});if(!r.ok){let a=await r.text();throw new Error(`HTTP ${r.status}: ${a}`)}}async function ee(s,e,o,t="progress_update"){let c=JSON.stringify(o),i=(c.length/1024).toFixed(1);c.length>256*1024&&console.error(`\u274C SQS message too large: ${i}KB (limit 256KB) for ${e} [${t}]`),await V().send(new X({QueueUrl:s,MessageBody:c,MessageGroupId:e,MessageAttributes:{executionId:{DataType:"String",StringValue:e},messageType:{DataType:"String",StringValue:t}}}))}function D({workflowType:s,jobId:e,projectId:o,agentType:t,model:c}){let i="\u2500".repeat(60),r=`${t||"default"} (model: ${c||"auto"})`;return["",i,` Workflow: ${s}`,` Job: ${e||"local"}`,` Project: ${o||"none"}`,` Agent: ${r}`,i].join(`
|
|
3
|
+
`)}import"@zibby/core";var T=process.env.WORKSPACE||"/workspace";async function se(s,e){j(e,{recursive:!0});let o=Date.now();console.log("[setup] Fetching bundle...");let t=setInterval(()=>{let i=((Date.now()-o)/1e3).toFixed(1);console.log(`[setup] still fetching (${i}s elapsed)`)},3e3);try{await new Promise((i,r)=>{let a=W("curl",["-fsSL",s],{stdio:["ignore","pipe","inherit"]}),f=W("tar",["-xzf","-","-C",e],{stdio:["pipe","inherit","inherit"]});a.stdout.pipe(f.stdin);let l,u,d=()=>{if(l!==void 0&&u!==void 0){if(l!==0)return r(new Error(`curl exited ${l}`));if(u!==0)return r(new Error(`tar exited ${u}`));i()}};a.on("close",S=>{l=S,d()}),f.on("close",S=>{u=S,d()}),a.on("error",r),f.on("error",r)})}finally{clearInterval(t)}let c=((Date.now()-o)/1e3).toFixed(1);return console.log(`[setup] Bundle extracted (${c}s)`),e}async function z(){let s=process.env.WORKFLOW_SOURCES_URL;if(!s)throw new Error("WORKFLOW_SOURCES_URL env var is required");let e=await fetch(s);if(!e.ok)throw new Error(`Failed to fetch sources: ${e.status} ${e.statusText}`);let o=await e.json();if(!o.sources||typeof o.sources!="object")throw new Error('Invalid sources payload \u2014 missing "sources" map');return o}function re(s,e){let o=G(e),t=0;for(let[c,i]of Object.entries(s)){let r=G(e,c);if(!r.startsWith(`${o}/`)&&r!==o){console.error(` \u26D4 Skipping unsafe path: ${c}`);continue}j(te(r),{recursive:!0}),oe(r,i,"utf-8"),t++}return t}async function ie(){let s=process.env.ZIBBY_EGRESS_PROXY_URL,e=process.env.ZIBBY_EGRESS_TOKEN;if(!(!s||!e))try{let o=await import("undici"),t=new o.ProxyAgent({uri:s,token:`Bearer ${e}`});o.setGlobalDispatcher(t),console.log(`[setup] Egress proxy active \u2192 ${s}`)}catch(o){console.warn(`[setup] Failed to install egress proxy dispatcher: ${o.message}`)}}async function ae(s,e){let o=g(s,"graph.mjs");if(!w(o))throw new Error(`graph.mjs not found at ${o}`);let t=await import(x(o).href),c=e?.entryClass,i=c&&t[c]||t.default||Object.values(t).find(r=>typeof r=="function"&&r.prototype?.buildGraph);if(!i)throw new Error("No WorkflowAgent class found in graph.mjs");return i}async function me(){if(!process.env.NODE_PATH){process.env.NODE_PATH="/opt/zibby/packages";let n=await import("module");n.default._initPaths&&n.default._initPaths()}await ie();let{WORKFLOW_JOB_ID:s,WORKFLOW_TYPE:e,PROJECT_ID:o,AGENT_TYPE:t,MODEL:c}=process.env;e||(console.error("Missing WORKFLOW_TYPE env var"),process.exit(1));let i=D({workflowType:e,jobId:s,projectId:o,agentType:t,model:c});console.log(i);let r=process.env.WORKFLOW_BUNDLE_URL,a,f={},l,u;if(r){l=e,a=g(T,".zibby","workflows",l);try{await se(r,a);try{let n=await z();f=n.input||{},u=n.version}catch(n){console.warn(`[setup] Could not fetch input payload: ${n.message}`)}}catch(n){console.warn(`[setup] Bundle extract failed (${n.message}); falling back to source install`),a=null}}if(!a){let n=await z(),{sources:p,input:m,workflowType:O,version:E}=n;f=m||{},l=O||e,u=E,console.log(`[setup] Workflow v${u||"?"} (${Object.keys(p).length} files)`),a=g(T,".zibby","workflows",l);let b=re(p,a);console.log(`[setup] Wrote ${b} files`),console.log("[setup] Installing dependencies...");try{ne("npm install --silent --no-audit --no-fund",{cwd:a,stdio:"inherit"}),console.log("[setup] Dependencies installed")}catch($){console.warn(`[setup] npm install failed: ${$.message}`)}}let d={},S=g(a,"workflow.json");w(S)&&(d=JSON.parse(K(S,"utf-8")));let A={},I=g(a,"zibby.config.json");if(w(I))try{A=JSON.parse(K(I,"utf-8")),console.log("[setup] Loaded user config from zibby.config.json")}catch(n){console.warn(`[setup] Failed to parse zibby.config.json: ${n.message} \u2014 falling back to defaults`)}let k=await ae(a,d);console.log(`[setup] Loaded ${k.name}`);let y=[],C=g(a,"node_modules","@zibby","agent-workflow"),v=g(a,"node_modules","@zibby","core","node_modules","@zibby","agent-workflow");w(C)&&y.push({kind:"hoisted",path:C}),w(v)&&y.push({kind:"nested",path:v});let R=process.env.ZIBBY_RUN_DIAG==="1";if(R){let{readdirSync:n}=await import("fs");console.log(` [diag] @zibby/agent-workflow copies in bundle: ${y.length}`);for(let p of y)console.log(` [diag] ${p.kind}: ${p.path}`);try{let p=g(a,"node_modules","@zibby");w(p)&&console.log(` [diag] node_modules/@zibby/ contents: [${n(p).join(", ")}]`)}catch{}}let N=g(a,"node_modules","@zibby","core","dist","index.js");if(w(N)&&y.length>0)try{let n=await import(x(N).href),p=[n.AssistantStrategy,n.CursorAgentStrategy,n.ClaudeAgentStrategy,n.CodexAgentStrategy,n.GeminiAgentStrategy].filter(Boolean);for(let m of y){let O=g(m.path,"dist","index.js");if(!w(O))continue;let E=await import(x(O).href),b=R?E.listStrategies():null;for(let $ of p)try{E.registerStrategy(new $)}catch(q){console.warn(` register ${$.name} into ${m.kind} failed: ${q.message}`)}R&&console.log(` [diag] ${m.kind} registry: before=[${b.join(",")||"empty"}] after=[${E.listStrategies().join(",")||"empty"}]`)}console.log("[setup] Registered 5 agent strategies (assistant, cursor, claude, codex, gemini)")}catch(n){console.warn(`[setup] Failed to bridge strategies: ${n.message}`)}else console.warn("[setup] No @zibby/core or @zibby/agent-workflow in bundle \u2014 agent strategies may be unavailable");let Q=Date.now(),_=new k({workflow:l||e}),J=_.buildGraph(),H={input:f||{},cwd:T,runId:s||`run-${Date.now()}`,config:A};console.log("");let h;try{h=await J.run(_,H)}catch(n){console.error(`
|
|
4
|
+
Workflow execution failed: ${n.message}`),console.error(n.stack),await U("failed",n.message),process.exit(1)}let F=((Date.now()-Q)/1e3).toFixed(1),M=h?.success!==!1,B=l||e;M?(console.log(`
|
|
5
|
+
[done] ${B} completed in ${F}s`),await U("completed")):(console.error(`
|
|
6
|
+
[done] ${B} failed after ${F}s`),await U("failed",h?.error||"Workflow execution failed"),process.exit(1)),_.onComplete&&await _.onComplete(h)}async function U(s,e=null){let o={EXECUTION_ID:process.env.WORKFLOW_JOB_ID,PROGRESS_API_URL:process.env.PROGRESS_API_URL,PROGRESS_QUEUE_URL:process.env.PROGRESS_QUEUE_URL,PROJECT_API_TOKEN:process.env.PROJECT_API_TOKEN,SQS_AUTH_TOKEN:process.env.SQS_AUTH_TOKEN};if(o.EXECUTION_ID)try{await L(o,{status:s,...e&&{error:e}})}catch(t){console.error(`\u26A0\uFE0F Failed to report status: ${t.message}`)}}export{me as runWorkflowCommand};
|
|
@@ -1,16 +1,16 @@
|
|
|
1
|
-
import{existsSync as h}from"fs";import{readFile as
|
|
1
|
+
import{existsSync as h}from"fs";import{readFile as R}from"fs/promises";import{join as k}from"path";import{pathToFileURL as W}from"url";import o from"chalk";import L from"ora";import{existsSync as z}from"fs";import{join as S}from"path";import{pathToFileURL as F}from"url";async function x(e){let n=S(e,".zibby.config.mjs");if(!z(n))throw new Error(".zibby.config.mjs not found");try{let t=await import(F(n).href);return t.default||t}catch(t){throw new Error(`Failed to load .zibby.config.mjs: ${t.message}`,{cause:t})}}var O=3848;async function U(e){try{let n=await x(e);return{userConfig:n,workflowsBasePath:n?.paths?.workflows||".zibby/workflows"}}catch{return{userConfig:null,workflowsBasePath:".zibby/workflows"}}}async function A(e,n,t){let s=k(e,"graph.mjs");if(!h(s))throw new Error(`graph.mjs not found in ${t}/${n}/`);let w=await D(e,n),a=await import(W(s).href),c=w.entryClass,i=c&&a[c]||a.default||Object.values(a).find(r=>typeof r=="function"&&r.prototype?.buildGraph);if(!i)throw new Error("No WorkflowAgent class found in graph.mjs. Export a class with buildGraph() method.");return{AgentClass:i,manifest:w}}async function D(e,n){let t=k(e,"workflow.json");if(!h(t))return{name:n,triggers:{api:!0}};let s=await R(t,"utf-8");return JSON.parse(s)}async function Z(e,n){e||(console.log(o.red(`
|
|
2
2
|
Workflow name is required`)),console.log(o.gray(" Usage: zibby workflow start <workflow-name>")),console.log(o.gray(` Example: zibby workflow start ticket-triage
|
|
3
|
-
`)),process.exit(1));let t=
|
|
3
|
+
`)),process.exit(1));let t=e.toLowerCase(),s=process.cwd(),{userConfig:w,workflowsBasePath:a}=await U(s),c=k(s,a,t);h(c)||(console.log(o.red(`
|
|
4
4
|
Workflow not found: ${a}/${t}/`)),console.log(o.gray(" Create one first:")),console.log(o.cyan(` zibby workflow new ${t}
|
|
5
|
-
`)),process.exit(1));let i=
|
|
6
|
-
${
|
|
7
|
-
`)),process.exit(1)}let
|
|
5
|
+
`)),process.exit(1));let i=L(` Loading workflow "${t}"...`).start(),r,b;try{({AgentClass:r,manifest:b}=await A(c,t,a)),i.succeed(` Loaded ${o.bold(b.entryClass||r.name)} (${t})`)}catch(g){i.fail(" Failed to load workflow"),console.log(o.red(`
|
|
6
|
+
${g.message}
|
|
7
|
+
`)),process.exit(1)}let f=parseInt(n.port,10)||O,u;try{u=(await import("express")).default}catch{console.log(o.red(`
|
|
8
8
|
express is required for local workflow server`)),console.log(o.gray(` npm install express
|
|
9
|
-
`)),process.exit(1)}let p=u();p.use(u.json({limit:"1mb"})),p.get("/health",(
|
|
10
|
-
\u25B6 Run ${
|
|
9
|
+
`)),process.exit(1)}let p=u();p.use(u.json({limit:"1mb"})),p.get("/health",(g,y)=>{y.json({status:"ok",workflow:t,class:r.name})}),p.post("/trigger",async(g,y)=>{let l=`local-${Date.now()}`,$=g.body.input||g.body||{};console.log(o.cyan(`
|
|
10
|
+
\u25B6 Run ${l} triggered`)),console.log(o.gray(` input: ${JSON.stringify($).slice(0,200)}`)),y.status(202).json({runId:l,status:"running",workflow:t});try{let m=Date.now(),d=new r({workflow:t}),E=d.buildGraph(),P={input:$,cwd:process.cwd(),runId:l,config:w||{}},C=await E.run(d,P),j=((Date.now()-m)/1e3).toFixed(1),T=C?.success!==!1;console.log(T?o.green(` \u2714 Run ${l} succeeded (${j}s)`):o.red(` \u2716 Run ${l} failed (${j}s)`)),d.onComplete&&await d.onComplete(C)}catch(m){console.log(o.red(` \u2716 Run ${l} error: ${m.message}`))}}),p.listen(f,()=>{console.log(o.bold.cyan(`
|
|
11
11
|
Zibby Workflow Server \u2014 ${t}
|
|
12
|
-
`)),console.log(o.gray(" ".padEnd(56,"-"))),console.log(o.white(` Workflow: ${o.cyan(t)}`)),console.log(o.white(` Class: ${o.cyan(r.name)}`)),console.log(o.white(` Port: ${o.cyan(
|
|
13
|
-
Endpoints:`)),console.log(o.gray(` GET http://localhost:${
|
|
14
|
-
Test with:`)),console.log(o.gray(` curl -X POST http://localhost:${
|
|
12
|
+
`)),console.log(o.gray(" ".padEnd(56,"-"))),console.log(o.white(` Workflow: ${o.cyan(t)}`)),console.log(o.white(` Class: ${o.cyan(r.name)}`)),console.log(o.white(` Port: ${o.cyan(f)}`)),console.log(o.gray(" ".padEnd(56,"-"))),console.log(o.white(`
|
|
13
|
+
Endpoints:`)),console.log(o.gray(` GET http://localhost:${f}/health`)),console.log(o.cyan(` POST http://localhost:${f}/trigger`)),console.log(o.white(`
|
|
14
|
+
Test with:`)),console.log(o.gray(` curl -X POST http://localhost:${f}/trigger \\`)),console.log(o.gray(' -H "Content-Type: application/json" \\')),console.log(o.gray(` -d '{"input": {"key": "value"}}'
|
|
15
15
|
`)),console.log(o.gray(` Ctrl+C to stop
|
|
16
|
-
`))})}export{
|
|
16
|
+
`))})}export{Z as startWorkflowCommand};
|
|
@@ -1,22 +1,22 @@
|
|
|
1
|
-
import h from"ora";import{select as
|
|
1
|
+
import h from"ora";import{select as v}from"@inquirer/prompts";import{readFileSync as x,existsSync as N}from"fs";import{homedir as R}from"os";import{join as T}from"path";var u={local:{name:"Local Development",apiUrl:"http://localhost:3001",accountApiUrl:"http://localhost:3001",frontendUrl:"http://localhost:3000",description:"Local backend running on port 3001"},prod:{name:"Production",apiUrl:process.env.ZIBBY_PROD_API_URL||"https://api-prod.zibby.app",accountApiUrl:process.env.ZIBBY_PROD_ACCOUNT_API_URL||"https://account-api-prod.zibby.app",frontendUrl:process.env.ZIBBY_PROD_FRONTEND_URL||"https://studio.zibby.app",description:"Production environment"}};function g(){let o;if(process.env.ZIBBY_API_URL)o=process.env.ZIBBY_API_URL;else{let e=process.env.ZIBBY_ENV||"prod";u[e]?o=u[e].apiUrl:o=u.prod.apiUrl}try{let e=new URL(o);return e.protocol!=="http:"&&e.protocol!=="https:"?(console.error(`\u26A0\uFE0F Invalid API URL protocol: ${e.protocol} (only http/https allowed)`),u.prod.apiUrl):o}catch{return console.error(`\u26A0\uFE0F Invalid API URL: ${o}`),u.prod.apiUrl}}var b=/^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$/i;function y(o){return o?b.test(o)?{ok:!0}:{ok:!1,error:`'${o}' is not a UUID. Cloud workflows are identified by UUID only. Run \`zibby workflow list\` to find yours, or run \`zibby workflow trigger\` with no argument for interactive selection.`}:{ok:!0}}import{existsSync as j,readFileSync as _}from"fs";import{resolve as B}from"path";function I(o){return o==="true"?!0:o==="false"?!1:o==="null"?null:o!==""&&!isNaN(Number(o))?Number(o):o}function $(o){let e={};for(let r of o||[]){let i=r.indexOf("=");if(i===-1){console.warn(` Warning: ignored param "${r}" \u2014 expected key=value format`);continue}let t=r.slice(0,i).trim(),n=I(r.slice(i+1)),l=t.split("."),c=e;for(let s=0;s<l.length-1;s++)(typeof c[l[s]]!="object"||c[l[s]]===null)&&(c[l[s]]={}),c=c[l[s]];c[l[l.length-1]]=n}return e}function E(o){let e=B(o);j(e)||(console.log(`
|
|
2
2
|
Error: --input-file not found: ${o}
|
|
3
|
-
`),process.exit(1));try{return JSON.parse(
|
|
3
|
+
`),process.exit(1));try{return JSON.parse(_(e,"utf-8"))}catch(r){console.log(`
|
|
4
4
|
Error: --input-file is not valid JSON: ${r.message}
|
|
5
|
-
`),process.exit(1)}}function
|
|
5
|
+
`),process.exit(1)}}function U(o){let e={};if(o.inputFile&&(e={...E(o.inputFile)}),o.input)try{e={...e,...JSON.parse(o.input)}}catch(r){console.log(`
|
|
6
6
|
Error: --input is not valid JSON`),console.log(` ${r.message}
|
|
7
|
-
`),process.exit(1)}return o.param?.length&&(e={...e
|
|
7
|
+
`),process.exit(1)}return o.param?.length&&(e={...e,...$(o.param)}),e}function A(){let o=T(R(),".zibby","config.json");if(N(o))try{let r=JSON.parse(x(o,"utf-8"));if(r.sessionToken)return r.sessionToken}catch{}let e=process.env.ZIBBY_API_KEY;if(e)return e;console.log(`
|
|
8
8
|
Not authenticated`),console.log(" Run: zibby login"),console.log(` OR set ZIBBY_API_KEY env var (for CI/CD)
|
|
9
|
-
`),process.exit(1)}async function
|
|
9
|
+
`),process.exit(1)}async function D(o){let e=g(),r=h("Fetching projects...").start();try{let i=await fetch(`${e}/projects`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${o}`}});i.ok||(r.fail("Failed to fetch projects"),process.exit(1));let t=await i.json();Array.isArray(t)||(t.projects?t=t.projects:t.data&&(t=t.data)),(!t||t.length===0)&&(r.fail("No projects found"),process.exit(1)),r.succeed(`Found ${t.length} project${t.length===1?"":"s"}`),console.log("");let n=t.map(l=>({name:`${l.name||"Unnamed"} (${l.projectId||l.id})`,value:l.projectId||l.id}));return await v({message:"Select a project:",choices:n})}catch(i){r.fail(`Error: ${i.message}`),process.exit(1)}}async function P(o,e){let r=g(),i=h("Fetching deployed workflows...").start();try{let t=["analysis","implementation","run_test"],n=[];for(let c of t){let s=await fetch(`${r}/projects/${o}/workflows/${c}`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${e}`}});if(s.ok){let a=await s.json();a.graph&&n.push({name:c,version:a.version||0,isDefault:a.isDefault!==!1})}}n.length===0&&(i.fail("No deployed workflows found for this project"),process.exit(1)),i.succeed(`Found ${n.length} deployed workflow${n.length===1?"":"s"}`),console.log("");let l=n.map(c=>({name:`${c.name} (v${c.version})${c.isDefault?" [default]":""}`,value:c.name}));return await v({message:"Select a workflow to trigger:",choices:l})}catch(t){i.fail(`Error: ${t.message}`),process.exit(1)}}async function M(o,e={}){let r=y(o);r.ok||(console.log(`
|
|
10
10
|
Error: ${r.error}
|
|
11
|
-
`),process.exit(1));let i=
|
|
11
|
+
`),process.exit(1));let i=A(),t=e.project||process.env.ZIBBY_PROJECT_ID,n;if(o){let s=g();try{let a=await fetch(`${s}/projects`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`}});if(a.ok){let d=(await a.json()).projects||[];for(let p of d){let m=await fetch(`${s}/projects/${p.projectId}/workflows`,{method:"GET",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`}});if(m.ok){let w=(await m.json()).find(k=>k.uuid===o);if(w){t=p.projectId,n=w.workflowType||w.name,console.log(`
|
|
12
12
|
\u2713 Found workflow "${n}" (UUID: ${o})
|
|
13
13
|
`);break}}}(!n||n===o)&&(console.log(`
|
|
14
14
|
Error: Workflow with UUID "${o}" not found`),console.log(` Check: zibby workflow list
|
|
15
15
|
`),process.exit(1))}}catch(a){console.log(`
|
|
16
16
|
Error looking up workflow UUID: ${a.message}
|
|
17
|
-
`),process.exit(1)}}t||(console.log(""),t=await
|
|
17
|
+
`),process.exit(1)}}t||(console.log(""),t=await D(i)),n||(console.log(""),n=await P(t,i));let l=U(e);if(console.log(`
|
|
18
18
|
Triggering Workflow
|
|
19
|
-
`),console.log(" ".padEnd(60,"-")),console.log(` Workflow: ${n}`),console.log(` Project: ${t}`),Object.keys(l).length>0){let s=JSON.stringify(l);console.log(` Input: ${s.length>60?`${s.substring(0,57)}...`:s}`)}e.idempotencyKey&&console.log(` Idempotency: ${e.idempotencyKey}`),console.log(" ".padEnd(60,"-")),console.log("");let c=h("Triggering workflow execution...").start();try{let s=
|
|
19
|
+
`),console.log(" ".padEnd(60,"-")),console.log(` Workflow: ${n}`),console.log(` Project: ${t}`),Object.keys(l).length>0){let s=JSON.stringify(l);console.log(` Input: ${s.length>60?`${s.substring(0,57)}...`:s}`)}e.idempotencyKey&&console.log(` Idempotency: ${e.idempotencyKey}`),console.log(" ".padEnd(60,"-")),console.log("");let c=h("Triggering workflow execution...").start();try{let s=g(),a={input:l};e.idempotencyKey&&(a.idempotencyKey=e.idempotencyKey);let f=await fetch(`${s}/projects/${t}/workflows/${n}/trigger`,{method:"POST",headers:{"Content-Type":"application/json",Authorization:`Bearer ${i}`},body:JSON.stringify(a)});if(!f.ok){let p=await f.json().catch(()=>({}));f.status===429&&(c.fail("Quota exceeded"),console.log(`
|
|
20
20
|
Your workflow execution quota has been exceeded`),p.quotaInfo&&(console.log(` Used: ${p.quotaInfo.used}/${p.quotaInfo.limit} executions`),console.log(` Plan: ${p.quotaInfo.planId}`),p.quotaInfo.periodEnd&&console.log(` Resets: ${new Date(p.quotaInfo.periodEnd).toLocaleDateString()}`)),console.log(""),process.exit(1)),c.fail("Trigger failed"),console.log(` Error: ${p.message||f.statusText}
|
|
21
21
|
`),process.exit(1)}let d=await f.json();c.succeed("Workflow triggered successfully"),console.log(""),console.log(" Job Details:"),console.log(` Job ID: ${d.jobId}`),console.log(` Status: ${d.status}`),console.log(` Version: ${d.version}`),console.log(` Triggered: ${new Date(d.triggeredAt).toLocaleString()}`),console.log(""),console.log(" Monitor execution:"),o?(console.log(` zibby workflow logs ${o}`),console.log(` zibby workflow logs ${o} -t`)):(console.log(` zibby workflow logs --workflow ${n} --project ${t}`),console.log(` zibby workflow logs --workflow ${n} --project ${t} -t`)),console.log("")}catch(s){c.fail("Trigger failed"),console.log(` Error: ${s.message}
|
|
22
|
-
`),process.exit(1)}}export{
|
|
22
|
+
`),process.exit(1)}}export{I as coerceValue,$ as parseParams,U as resolveInput,M as triggerWorkflowCommand};
|