@aurodesignsystem-dev/auro-cli 0.0.0-pr286.3 → 0.0.0-pr286.5

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/auro-cli.js CHANGED
@@ -1,36 +1,36 @@
1
1
  #!/usr/bin/env node
2
- import{program as Le}from"commander";import Lt from"figlet";import{mind as Wt}from"gradient-string";var We=()=>Wt(Lt.textSync("Auro CLI"));import Ue from"node:fs";import ze from"node:path";import{fileURLToPath as Ut}from"node:url";function J(r){process.env.DEBUG&&console.log(`[DEBUG] ${r}`)}function he(){try{let r=Ut(import.meta.url),t=ze.dirname(r);J(`Current module path: ${t}`);let e=ze.resolve(t,"..","package.json");return J(`Checking package.json at: ${e}`),Ue.existsSync(e)?(J(`Found package.json at: ${e}`),JSON.parse(Ue.readFileSync(e,"utf8")).version):(J("Could not find package.json in the standard installed module location, using default version"),"0.0.0")}catch(r){return console.error("Error retrieving package version:",r),"0.0.0"}}import{program as Hr}from"commander";import yt from"ora";function K(r){return r.option("-m, --module-paths [paths...]","Path(s) to node_modules folder").option("-w, --watch","Watches for changes").option("--skip-docs","Skip documentation generation",!1).option("-r, --readme-template <url>","URL to the README template file").option("--wca-input [files...]","Source file(s) to analyze for API documentation").option("--wca-output [files...]","Output file(s) for API documentation")}function Y(r){return r.option("-s, --serve","Starts a server").option("-p, --port <number>","Port for the server").option("-o, --open","Open the browser after starting the server")}import Fr from"@rollup/plugin-terser";import{watch as Br}from"rollup";import{existsSync as lt,readFileSync as Cr,rmSync as Ir,writeFileSync as Rr}from"node:fs";import{basename as Sr,dirname as Dr,join as M,resolve as se}from"node:path";import{pathToFileURL as mt}from"node:url";import{glob as Pr}from"glob";import Ce from"ora";import{rollup as Te}from"rollup";import*as ut from"sass";import B from"ora";import{spawn as zt}from"node:child_process";import qt from"ora";var A=(r,t)=>{let e=`${r} ${t?t.join(" "):""}`,n=qt(),o=r,s=t||[];if(!t&&typeof r=="string"){let m=r.split(" ");o=m[0],s=m.slice(1)}let i=e.includes("--watch")||e.includes(" -w"),l=zt(o,s,{stdio:i?"inherit":["inherit","pipe","pipe"],shell:!0});if(!i){let m=[];l.stdout?.on("data",u=>{let c=u.toString();m.push(c),process.stdout.write(c)}),l.stderr?.on("data",u=>{let c=u.toString();m.push(c),process.stderr.write(c)})}return new Promise((m,u)=>{l.on("close",c=>{c!==0?i?(n.info(`Watch mode terminated with code ${c}`),m()):(n.fail(`${e} failed (code ${c})`),u(new Error(`Command failed with exit code ${c}`))):(n.succeed(`${e} completed successfully`),m())})})};import X from"node:fs";import Vt from"node:path";import{markdownTable as qe}from"markdown-table";var O=class r{static{this.manifest={schemaVersion:"1.0.0",readme:"",modules:[]}}static generate(t={}){let{outDir:e="./docs",outFile:n="api.md",manifestPath:o="./custom-elements.json"}=t,{getElements:s,renderAllElements:i}=r;if(o)try{let c=X.readFileSync(o,"utf8");r.manifest=JSON.parse(c)}catch(c){throw console.error(`Error reading manifest file at ${o}:`,c),c}let a=s(),l=e;X.existsSync(l)||X.mkdirSync(l,{recursive:!0});let m=i(a),u=Vt.join(l,n);X.writeFileSync(u,m),console.log(`Generated combined API documentation at ${u}`)}static getElements(){let t=r.manifest.modules.filter(r.isWcaModule);return r.manifest.modules.reduce((e,n)=>e.concat(n.declarations?.filter(o=>"customElement"in o&&o.customElement===!0&&"tagName"in o&&(t.length>0?r.isWcaModule(n):!0))??[]),[])}static isWcaModule(t){let{path:e}=t;return e?e.startsWith("scripts/wca/auro-")&&e.endsWith(".js"):!1}static renderAllElements(t){return`${t.sort((e,n)=>(e.tagName||"").localeCompare(n.tagName||"")).map(e=>r.renderElement(e,!0)).join(`
2
+ import{program as We}from"commander";import Ut from"figlet";import{mind as zt}from"gradient-string";var Ue=()=>zt(Ut.textSync("Auro CLI"));import ze from"node:fs";import qe from"node:path";import{fileURLToPath as qt}from"node:url";function K(r){process.env.DEBUG&&console.log(`[DEBUG] ${r}`)}function ye(){try{let r=qt(import.meta.url),t=qe.dirname(r);K(`Current module path: ${t}`);let e=qe.resolve(t,"..","package.json");return K(`Checking package.json at: ${e}`),ze.existsSync(e)?(K(`Found package.json at: ${e}`),JSON.parse(ze.readFileSync(e,"utf8")).version):(K("Could not find package.json in the standard installed module location, using default version"),"0.0.0")}catch(r){return console.error("Error retrieving package version:",r),"0.0.0"}}import{program as Lr}from"commander";import bt from"ora";function Y(r){return r.option("-m, --module-paths [paths...]","Path(s) to node_modules folder").option("-w, --watch","Watches for changes").option("--skip-docs","Skip documentation generation",!1).option("-r, --readme-template <url>","URL to the README template file").option("--wca-input [files...]","Source file(s) to analyze for API documentation").option("--wca-output [files...]","Output file(s) for API documentation")}function X(r){return r.option("-s, --serve","Starts a server").option("-p, --port <number>","Port for the server").option("-o, --open","Open the browser after starting the server")}import Br from"@rollup/plugin-terser";import{watch as Gr}from"rollup";import{existsSync as ut,readFileSync as Ir,rmSync as Rr,writeFileSync as Sr}from"node:fs";import{basename as Dr,dirname as xr,join as H,resolve as ie}from"node:path";import{pathToFileURL as dt}from"node:url";import{glob as Pr}from"glob";import Ie from"ora";import{rollup as Ce}from"rollup";import*as pt from"sass";import G from"ora";import{spawn as Vt}from"node:child_process";import Jt from"ora";var A=(r,t)=>{let e=`${r} ${t?t.join(" "):""}`,n=Jt(),o=r,s=t||[];if(!t&&typeof r=="string"){let d=r.split(" ");o=d[0],s=d.slice(1)}let i=e.includes("--watch")||e.includes(" -w"),l=Vt(o,s,{stdio:i?"inherit":["inherit","pipe","pipe"],shell:!0});if(!i){let d=[];l.stdout?.on("data",m=>{let c=m.toString();d.push(c),process.stdout.write(c)}),l.stderr?.on("data",m=>{let c=m.toString();d.push(c),process.stderr.write(c)})}return new Promise((d,m)=>{l.on("close",c=>{c!==0?i?(n.info(`Watch mode terminated with code ${c}`),d()):(n.fail(`${e} failed (code ${c})`),m(new Error(`Command failed with exit code ${c}`))):(n.succeed(`${e} completed successfully`),d())})})};import Z from"node:fs";import Kt from"node:path";import{markdownTable as Ve}from"markdown-table";var N=class r{static{this.manifest={schemaVersion:"1.0.0",readme:"",modules:[]}}static generate(t={}){let{outDir:e="./docs",outFile:n="api.md",manifestPath:o="./custom-elements.json"}=t,{getElements:s,renderAllElements:i}=r;if(o)try{let c=Z.readFileSync(o,"utf8");r.manifest=JSON.parse(c)}catch(c){throw console.error(`Error reading manifest file at ${o}:`,c),c}let a=s(),l=e;Z.existsSync(l)||Z.mkdirSync(l,{recursive:!0});let d=i(a),m=Kt.join(l,n);Z.writeFileSync(m,d),console.log(`Generated combined API documentation at ${m}`)}static getElements(){let t=r.manifest.modules.filter(r.isWcaModule);return r.manifest.modules.reduce((e,n)=>e.concat(n.declarations?.filter(o=>"customElement"in o&&o.customElement===!0&&"tagName"in o&&(t.length>0?r.isWcaModule(n):!0))??[]),[])}static isWcaModule(t){let{path:e}=t;return e?e.startsWith("scripts/wca/auro-")&&e.endsWith(".js"):!1}static renderAllElements(t){return`${t.sort((e,n)=>(e.tagName||"").localeCompare(n.tagName||"")).map(e=>r.renderElement(e,!0)).join(`
3
3
 
4
- `)}`}static renderElement(t,e=!0){let n=[],{renderTable:o,renderPropertiesAttributesTable:s,renderParameters:i,getType:a}=r;n.push(e?`# ${t.tagName}`:""),t.description&&n.push(t.description);let l=s(t);l&&n.push(l.trim());let m=o("Methods",["name","parameters","return","description"],(t.members||[]).filter(f=>f.kind==="method"&&("privacy"in f?f.privacy!=="private":!0)&&f.name[0]!=="_").map(f=>({...f,parameters:i("parameters"in f?f.parameters:void 0),return:"return"in f&&f.return?a(f.return):""})));m&&n.push(m.trim());let u=o("Events",["name","description"],t.events);u&&n.push(u.trim());let c=o("Slots",[["name","(default)"],"description"],t.slots);c&&n.push(c.trim());let d=o("CSS Shadow Parts",["name","description"],t.cssParts);d&&n.push(d.trim());let p=o("CSS Custom Properties",["name","description"],t.cssProperties);return p&&n.push(p.trim()),n.join(`
4
+ `)}`}static renderElement(t,e=!0){let n=[],{renderTable:o,renderPropertiesAttributesTable:s,renderParameters:i,getType:a}=r;n.push(e?`# ${t.tagName}`:""),t.description&&n.push(t.description);let l=s(t);l&&n.push(l.trim());let d=o("Methods",["name","parameters","return","description"],(t.members||[]).filter(p=>p.kind==="method"&&("privacy"in p?p.privacy!=="private":!0)&&p.name[0]!=="_").map(p=>({...p,parameters:i("parameters"in p?p.parameters:void 0),return:"return"in p&&p.return?a(p.return):""})));d&&n.push(d.trim());let m=o("Events",["name","description"],t.events);m&&n.push(m.trim());let c=o("Slots",[["name","(default)"],"description"],t.slots);c&&n.push(c.trim());let u=o("CSS Shadow Parts",["name","description"],t.cssParts);u&&n.push(u.trim());let f=o("CSS Custom Properties",["name","description"],t.cssProperties);return f&&n.push(f.trim()),n.join(`
5
5
 
6
- `)}static renderPropertiesAttributesTable(t){let{getType:e,escapeMarkdown:n}=r,o=t.members?.filter(c=>c.kind==="field"&&("privacy"in c?c.privacy!=="private":!0)&&c.name[0]!=="_")||[],s=t.attributes||[],i=[],a=new Set;if(o.forEach(c=>{if(c.description?.trim()){let d=e(c)||"",f=("return"in c&&c.return?e(c.return):"")||d;i.push({name:c.name,properties:c.name,attributes:("attribute"in c?c.attribute:"")||"",modifiers:"readonly"in c&&c.readonly?"readonly":"",type:f,default:("default"in c?c.default:"")||"",description:c.description||""})}a.add(c.name),"attribute"in c&&c.attribute&&a.add(c.attribute)}),s.forEach(c=>{!a.has(c.name)&&c.description?.trim()&&i.push({name:c.name,properties:"",attributes:c.name,modifiers:"",type:e(c)||"",default:c.default||"",description:c.description||""})}),i.length===0)return"";let l=["Properties","Attributes","Modifiers","Type","Default","Description"],m=i.map(c=>{let g=(c.default||"").trim().replace(/^'([^']+)'$/,"$1").replace(/^"([^"]+)"$/,"$1"),b=g?g.startsWith("`")&&g.endsWith("`")?g:`\`${g}\``:"";return[n(c.properties),n(c.attributes),n(c.modifiers),n(c.type),n(b),n(c.description)]});return`### Properties & Attributes
6
+ `)}static renderPropertiesAttributesTable(t){let{getType:e,escapeMarkdown:n}=r,o=t.members?.filter(c=>c.kind==="field"&&("privacy"in c?c.privacy!=="private":!0)&&c.name[0]!=="_")||[],s=t.attributes||[],i=[],a=new Set;if(o.forEach(c=>{if(c.description?.trim()){let u=e(c)||"",p=("return"in c&&c.return?e(c.return):"")||u;i.push({name:c.name,properties:c.name,attributes:("attribute"in c?c.attribute:"")||"",modifiers:"readonly"in c&&c.readonly?"readonly":"",type:p,default:("default"in c?c.default:"")||"",description:c.description||""})}a.add(c.name),"attribute"in c&&c.attribute&&a.add(c.attribute)}),s.forEach(c=>{!a.has(c.name)&&c.description?.trim()&&i.push({name:c.name,properties:"",attributes:c.name,modifiers:"",type:e(c)||"",default:c.default||"",description:c.description||""})}),i.length===0)return"";let l=["Properties","Attributes","Modifiers","Type","Default","Description"],d=i.map(c=>{let g=(c.default||"").trim().replace(/^'([^']+)'$/,"$1").replace(/^"([^"]+)"$/,"$1"),b=g?g.startsWith("`")&&g.endsWith("`")?g:`\`${g}\``:"";return[n(c.properties),n(c.attributes),n(c.modifiers),n(c.type),n(b),n(c.description)]});return`### Properties & Attributes
7
7
 
8
- ${qe([l,...m])}
9
- `}static renderParameters(t){let{escapeMarkdown:e,getType:n}=r;return!t||t.length===0?"None":t.map(o=>{let s=n(o)||"any",i=o.description?` - ${o.description}`:"";return`\`${o.name}\` (${e(s)})${e(i)}`}).join("<br>")}static renderTable(t,e,n){let{escapeMarkdown:o,get:s,capitalize:i}=r;if(n===void 0||n.length===0)return"";let a=n.filter(c=>{let{description:d}=c;return typeof d=="string"&&d.trim()});if(a.length===0)return"";let l=e.map(c=>i((Array.isArray(c)?c[0]:c).split(".")[0])),m=a.map(c=>e.map(d=>{let p=s(c,d);return o(String(p||""))})),u=qe([l,...m]);return`### ${t}
8
+ ${Ve([l,...d])}
9
+ `}static renderParameters(t){let{escapeMarkdown:e,getType:n}=r;return!t||t.length===0?"None":t.map(o=>{let s=n(o)||"any",i=o.description?` - ${o.description}`:"";return`\`${o.name}\` (${e(s)})${e(i)}`}).join("<br>")}static renderTable(t,e,n){let{escapeMarkdown:o,get:s,capitalize:i}=r;if(n===void 0||n.length===0)return"";let a=n.filter(c=>{let{description:u}=c;return typeof u=="string"&&u.trim()});if(a.length===0)return"";let l=e.map(c=>i((Array.isArray(c)?c[0]:c).split(".")[0])),d=a.map(c=>e.map(u=>{let f=s(c,u);return o(String(f||""))})),m=Ve([l,...d]);return`### ${t}
10
10
 
11
- ${u}
12
- `}static escapeMarkdown(t){return t.replace(/\\/g,"\\\\").replace(/\n/g,"<br>").replace(/\|/g,"\\|")}static getType(t){if(!t||!t.type)return"";let{type:e}=t,n=s=>s.replace(/\s*\|\s*/g," | ").replace(/'([^']+)'/g,"`$1`");if(typeof e=="string")return n(e);if(e.text)return n(e.text);if(Array.isArray(e))return e.map(s=>typeof s=="string"?s:s.text?s.text:s.name?s.name:String(s)).join(" \\| ");if(e.name)return n(e.name);if(e.references&&Array.isArray(e.references))return e.references.map(s=>s.name||String(s)).join(" \\| ");let o=String(e);return n(o)}static get(t,e){let n="",o=e;Array.isArray(e)&&([o,n]=e);let s=o.split("."),i=t;for(;i&&s.length;)i=i[s.shift()];return i==null||i===""?n:String(i)}static capitalize(t){return t.replace(/([A-Z])/g," $1").replace(/^./,e=>e.toUpperCase()).trim()}};import Jt from"node:fs";import Kt from"node:os";import Z from"node:path";import ye from"node:process";function Yt(){let r=Kt.homedir()||ye.env.HOME||ye.env.USERPROFILE;if(!r)throw new Error("Unable to determine user home directory");return Z.join(r,".auro")}function we(...r){return Z.join(Yt(),...r)}function be(...r){let t=Jt.realpathSync(ye.argv[1]),e=Z.dirname(t);return Z.resolve(e,...r)}var Ve=r=>be("configs",r);import{copyFileSync as Xt,existsSync as Je,mkdirSync as Zt}from"node:fs";import{join as Qt,resolve as ke}from"node:path";function N(){let r=process.cwd(),t=ke(r,"README.md"),e=Qt(ke(r,"demo"),"readme.md");if(!Je(t))return;let n=ke(r,"demo");Je(n)||Zt(n,{recursive:!0}),Xt(t,e)}import er from"ora";var Ye=[];function j(r){Ye.push(r)}var Ke=!1;function Q(){Ke||(Ke=!0,process.on("SIGINT",()=>{let r=er("Wrapping up...").start();for(let t of Ye)t.close();r.succeed("All done! See you next time. \u2728"),process.exit(0)}))}import{Logger as tr}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{generateReadmeUrl as rr,processContentForFile as nr,templateFiller as $e}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";import T from"node:fs";import{readFileSync as or,existsSync as Xe}from"node:fs";import D from"node:path";var sr="/docs/pages",Qe={overwriteLocalCopies:!0,remoteReadmeVersion:"master",remoteReadmeVariant:"_updated_paths"};function E(r){return`${process.cwd()}/${r}`}async function ir(r,t=!1){let e=[];if(!t){let o=r.localReadmePath?r.localReadmePath:{remoteUrl:r.remoteReadmeUrl||rr(r.remoteReadmeVersion,r.remoteReadmeVariant),fileName:E("/docTemplates/README.md"),overwrite:r.overwriteLocalCopies};e.push({identifier:"README.md",input:o,output:E("/README.md")})}Ze("/docs/partials/index.md")&&e.push({identifier:"index.md",input:E("/docs/partials/index.md"),output:E("/demo/index.md"),mdMagicConfig:{output:{directory:E("/demo")}}}),Ze("/docs/partials/api.md")&&e.push({identifier:"api.md",input:E("/docs/partials/api.md"),output:E("/demo/api.md"),preProcessors:[$e.formatApiTable]});let n=E(sr);if(T.existsSync(n)){let s=(await T.promises.readdir(n)).map(i=>({identifier:i,input:D.join(n,i),output:E(`/demo/${i}`)}));e.push(...s)}return e}async function ar(r=Qe,t=!1){await $e.extractNames();let e=await ir(r,t);for(let n of e)try{await nr(n),n.output.endsWith(".md")&&await cr(n.output)}catch(o){tr.error(`Error processing ${n.identifier}: ${o.message}`)}}async function cr(r){let t=D.dirname(r),e=await T.promises.readFile(r,"utf8"),n=/^[ \t]*<!-- AURO-GENERATED-CONTENT:START \((FILE|CODE):src=([^)]+)\) -->\n[ \t]*<!-- AURO-GENERATED-CONTENT:END -->/gm,o,s=!1,i=E("demo");for(;(o=n.exec(e))!==null;){let[c,d,p]=o,f=D.resolve(t,p),g=D.resolve(i,p),b=Xe(f)?f:Xe(g)?g:null;if(b){let x=or(b,"utf8"),k;if(d==="FILE")k=`<!-- AURO-GENERATED-CONTENT:START (FILE:src=${p}) -->
11
+ ${m}
12
+ `}static escapeMarkdown(t){return t.replace(/\\/g,"\\\\").replace(/\n/g,"<br>").replace(/\|/g,"\\|")}static getType(t){if(!t||!t.type)return"";let{type:e}=t,n=s=>s.replace(/\s*\|\s*/g," | ").replace(/'([^']+)'/g,"`$1`");if(typeof e=="string")return n(e);if(e.text)return n(e.text);if(Array.isArray(e))return e.map(s=>typeof s=="string"?s:s.text?s.text:s.name?s.name:String(s)).join(" \\| ");if(e.name)return n(e.name);if(e.references&&Array.isArray(e.references))return e.references.map(s=>s.name||String(s)).join(" \\| ");let o=String(e);return n(o)}static get(t,e){let n="",o=e;Array.isArray(e)&&([o,n]=e);let s=o.split("."),i=t;for(;i&&s.length;)i=i[s.shift()];return i==null||i===""?n:String(i)}static capitalize(t){return t.replace(/([A-Z])/g," $1").replace(/^./,e=>e.toUpperCase()).trim()}};import Yt from"node:fs";import Xt from"node:os";import Q from"node:path";import we from"node:process";function Zt(){let r=Xt.homedir()||we.env.HOME||we.env.USERPROFILE;if(!r)throw new Error("Unable to determine user home directory");return Q.join(r,".auro")}function be(...r){return Q.join(Zt(),...r)}function ke(...r){let t=Yt.realpathSync(we.argv[1]),e=Q.dirname(t);return Q.resolve(e,...r)}var Je=r=>ke("configs",r);import{copyFileSync as Qt,existsSync as Ke,mkdirSync as er}from"node:fs";import{join as tr,resolve as $e}from"node:path";function j(){let r=process.cwd(),t=$e(r,"README.md"),e=tr($e(r,"demo"),"readme.md");if(!Ke(t))return;let n=$e(r,"demo");Ke(n)||er(n,{recursive:!0}),Qt(t,e)}import rr from"ora";var Xe=[];function _(r){Xe.push(r)}var Ye=!1;function ee(){Ye||(Ye=!0,process.on("SIGINT",()=>{let r=rr("Wrapping up...").start();for(let t of Xe)t.close();r.succeed("All done! See you next time. \u2728"),process.exit(0)}))}import{Logger as nr}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{generateReadmeUrl as or,processContentForFile as sr,templateFiller as Ee}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";import T from"node:fs";import{readFileSync as et,existsSync as Ze}from"node:fs";import D from"node:path";var ir="/docs/pages",tt={overwriteLocalCopies:!0,remoteReadmeVersion:"master",remoteReadmeVariant:"_updated_paths",monorepoName:void 0,extraVars:{}};function k(r){return`${process.cwd()}/${r}`}async function ar(r,t=!1){let e=[];if(!t){let o=r.localReadmePath?r.localReadmePath:{remoteUrl:r.remoteReadmeUrl||or(r.remoteReadmeVersion,r.remoteReadmeVariant),fileName:k("/docTemplates/README.md"),overwrite:r.overwriteLocalCopies};e.push({identifier:"README.md",input:o,output:k("/README.md")})}Qe("/docs/partials/index.md")&&e.push({identifier:"index.md",input:k("/docs/partials/index.md"),output:k("/demo/index.md"),mdMagicConfig:{output:{directory:k("/demo")}}}),Qe("/docs/partials/api.md")&&e.push({identifier:"api.md",input:k("/docs/partials/api.md"),output:k("/demo/api.md"),preProcessors:[Ee.formatApiTable]});let n=k(ir);if(T.existsSync(n)){let s=(await T.promises.readdir(n)).map(i=>({identifier:i,input:D.join(n,i),output:k(`/demo/${i}`)}));e.push(...s)}return e}async function cr(r=tt,t=!1){await Ee.extractNames();let e=await ar(r,t),n=r.monorepoName;if(!n)try{n=JSON.parse(et(k("package.json"),"utf8")).name?.replace(/^@[^/]+\//,"")}catch{}let o={...n?{monorepoName:n}:{},...r.extraVars||{}};for(let s of e)try{await sr({...s,extraVars:o}),s.output.endsWith(".md")&&await lr(s.output,o)}catch(i){nr.error(`Error processing ${s.identifier}: ${i.message}`)}}async function lr(r,t={}){let e=D.dirname(r),n=await T.promises.readFile(r,"utf8"),o=/^[ \t]*<!-- AURO-GENERATED-CONTENT:START \((FILE|CODE):src=([^)]+)\) -->\n[ \t]*<!-- AURO-GENERATED-CONTENT:END -->/gm,s,i=!1,a=k("demo");for(;(s=o.exec(n))!==null;){let[u,f,p]=s,g=D.resolve(e,p),b=D.resolve(a,p),P=Ze(g)?g:Ze(b)?b:null;if(P){let O=et(P,"utf8"),$;if(f==="FILE")$=`<!-- AURO-GENERATED-CONTENT:START (FILE:src=${p}) -->
13
13
  <!-- The below content is automatically added from ${p} -->
14
- ${x.trimEnd()}
15
- <!-- AURO-GENERATED-CONTENT:END -->`;else{let $=D.extname(p).slice(1)||"html",V=x.trimEnd().replace(/&/g,"&amp;").replace(/</g,"&lt;").replace(/>/g,"&gt;");k=`<!-- AURO-GENERATED-CONTENT:START (CODE:src=${p}) -->
14
+ ${O.trimEnd()}
15
+ <!-- AURO-GENERATED-CONTENT:END -->`;else{let E=D.extname(p).slice(1)||"html",J=O.trimEnd().replace(/&/g,"&amp;").replace(/</g,"&lt;").replace(/>/g,"&gt;");$=`<!-- AURO-GENERATED-CONTENT:START (CODE:src=${p}) -->
16
16
  <!-- The below code snippet is automatically added from ${p} -->
17
- <pre class="language-${$}"><code class="language-${$}">${V}
17
+ <pre class="language-${E}"><code class="language-${E}">${J}
18
18
  </code></pre>
19
- <!-- AURO-GENERATED-CONTENT:END -->`}e=e.replace(c,k),n.lastIndex=0,s=!0}}s&&(e=$e.replaceTemplateValues(e),await T.promises.writeFile(r,e)),e=await T.promises.readFile(r,"utf8");let a=/^[ \t]*```(\w*)\n([\s\S]*?)^[ \t]*```[ \t]*$/gm,l=e.replace(a,(c,d,p)=>{let f=d||"html",g=p.trimEnd().replace(/&/g,"&amp;").replace(/</g,"&lt;").replace(/>/g,"&gt;");return`<pre class="language-${f}"><code class="language-${f}">${g}
20
- </code></pre>`});l!==e&&await T.promises.writeFile(r,l),e=await T.promises.readFile(r,"utf8"),e=e.replace(/(<pre[^>]*><code[^>]*>)([\s\S]*?)(<\/code><\/pre>)/g,(c,d,p,f)=>{let g=p.split(`
21
- `),b=g.filter($=>$.trim().length>0);if(b.length===0)return c;let x=Math.min(...b.map($=>{let V=$.match(/^[ \t]*/);return V?V[0].length:0})),k=g.map($=>$===""||$.trim().length===0?"\u200B":x>0?$.substring(x):$);for(;k.length>0&&(k[k.length-1]==="\u200B"||k[k.length-1]==="");)k.pop();return d+k.join(`
22
- `)+f});let m=e.split(`
23
- `),u=!1;for(let c=0;c<m.length;c++)/<pre[\s>]/i.test(m[c])&&(u=!0),u||(m[c]=m[c].replace(/^[ \t]+/,"")),/<\/pre>/i.test(m[c])&&(u=!1);await T.promises.writeFile(r,m.join(`
24
- `))}async function _(r={}){let t=r.readmeTemplate,e=t&&!t.startsWith("http");await ar({...Qe,...e?{localReadmePath:D.resolve(process.cwd(),t)}:{remoteReadmeUrl:t||"https://raw.githubusercontent.com/AlaskaAirlines/auro-templates/main/templates/default/README.md"}},r.skipReadme)}function Ze(r){return T.existsSync(E(r))}import{existsSync as et,readFileSync as lr}from"node:fs";import{builtinModules as mr}from"node:module";import{resolve as ee}from"node:path";import{startDevServer as ur}from"@web/dev-server";import{hmrPlugin as dr}from"@web/dev-server-hmr";import*as tt from"esbuild";import pr from"ora";var P=["node_modules","../node_modules","../../node_modules","../../../node_modules"];var fr=/^\/__wds-outside-root__\/(\d+)\/(.+)$/;function gr(){return{name:"node-modules-css",serve(r){if(!r.path.endsWith(".css")||!r.path.startsWith("/@")&&!/^\/[a-z]/i.test(r.path))return;let t=r.path.slice(1),e=process.cwd();for(let n of P){let o=ee(e,n,t);if(et(o))return{body:lr(o,"utf-8"),type:"css"}}}}}function hr(r,t){let e=r.match(fr);return e?ee(t,"../".repeat(Number.parseInt(e[1],10)),e[2]):ee(t,`.${r}`)}function yr(){let r=/\b(require\s*\(|module\.exports\b|exports\.\w)/,t=new Map,e;return{name:"cjs-to-esm",async serverStart({config:n}){e=ee(n.rootDir)},async transform(n){if(!n.path.includes("node_modules")||!n.path.endsWith(".js")&&!n.path.endsWith(".cjs")||typeof n.body!="string"||!r.test(n.body))return;let o=hr(n.path,e);if(t.has(o))return t.get(o);if(et(o))try{let i={body:(await tt.build({entryPoints:[o],bundle:!0,format:"esm",platform:"browser",write:!1,logLevel:"silent",external:mr})).outputFiles[0].text};return t.set(o,i),i}catch(s){console.error(`CJS-to-ESM bundling failed for ${n.path}:`,s.message)}}}}var F={watch:!0,nodeResolve:!0,basePath:"/",rootDir:"./demo",hmrInclude:["src/**/*","demo/**/*","apiExamples/**/*","docs/**/*"]};async function te(r={}){if(!r.serve)return;let t=pr(`Firing up dev server...
25
- `).start();try{let e={port:Number(r.port)||void 0,open:r.open?"/":void 0,watch:r.watch??F.watch,nodeResolve:r.nodeResolve??F.nodeResolve,basePath:r.basePath??F.basePath,rootDir:r.rootDir??F.rootDir,middleware:[function(s,i){return!s.url.endsWith("/")&&!s.url.includes(".")&&(s.url+=".html"),i()}],plugins:[gr(),yr(),dr({include:r.hmrInclude??F.hmrInclude})]},n=await ur({config:e,readCliArgs:!1,readFileConfig:!1});return t.stop(),n}catch(e){throw t.fail("Server snag! Couldn't start dev server."),console.error("Error starting development server:",e),new Error(`Development server failed to start: ${e.message}`)}}async function re(){let r=B("Generating Custom Elements Manifest...").start();try{await A(`npx --package=@custom-elements-manifest/analyzer -y -- cem analyze --config '${Ve("custom-elements-manifest.config.mjs")}'`),r.succeed("Custom Elements Manifest generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);r.warn("CEM analyzer completed with warnings: "+e)}}async function ne(){let r=B("Generating API md file...").start();try{await O.generate(),r.succeed("API md file generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);throw r.fail("Failed to generate API md file: "+e),t}}async function rt(r={}){let t=B("Compiling documentation...").start();try{await _(r),t.succeed("Documentation compiled successfully!")}catch(e){let n=e instanceof Error?e.message:String(e);throw t.fail("Failed to compile MD documentation: "+n),e}N(),await C(),await Ee(r)}async function nt(r={}){await te(r)}async function ot(r={}){let t=await import("chokidar"),e=["./src/**/*","./docs/**/*","./docTemplates/**/*","./apiExamples/**/*","./demo/**/*.scss"],n=["**/demo/*.min.js","**/demo/*.min.css","**/demo/*.md","**/demo/readme.md","**/docs/api.md","**/custom-elements.json","**/node_modules/**","**/dist/**","**/.git/**"],o=t.watch(e,{ignoreInitial:!0,ignored:n,awaitWriteFinish:{stabilityThreshold:1e3,pollInterval:100}}),s=B("Waiting for changes...");s.spinner="bouncingBar",s.color="green",s.start();let i=null,a=!1,l=!1;async function m(u){if(a){l=!0;return}a=!0,o.unwatch(e);let c=B(`Change detected: ${u}`).start();try{await _(r),N(),await C(),await Ee(r),c.succeed("Docs rebuilt!")}catch(d){let p=d instanceof Error?d.message:String(d);c.fail("Rebuild failed: "+p)}finally{setTimeout(()=>{o.add(e),a=!1,l&&(l=!1,m("queued changes"))},1e3)}}o.on("all",(u,c)=>{a||(i&&clearTimeout(i),i=setTimeout(()=>{m(c)},1e3))}),j(o),Q()}async function oe(){await re(),await ne()}import{basename as kr,join as it}from"node:path";import $r from"@rollup/plugin-commonjs";import{nodeResolve as Er}from"@rollup/plugin-node-resolve";import{glob as Ar}from"glob";import{litScss as vr}from"rollup-plugin-scss-lit";import wr from"node:path";import{glob as br}from"glob";function st(r){return{name:"watch-globs",buildStart(){let t=Array.isArray(r)?r:[r];for(let e of t)try{for(let n of br.sync(wr.resolve(e)))this.addWatchFile(n)}catch(n){this.error(`Error watching glob pattern "${e}": ${n.message}`)}}}}var Ae={moduleDirectories:["node_modules"],modulePaths:["../../node_modules","../node_modules","node_modules"],watchPatterns:["./apiExamples/**/*","./docs/**/*"]};function at(r=[],t={}){let{watchPatterns:e=Ae.watchPatterns,dedupe:n=["lit","lit-element","lit-html"],dev:o=!1}=t,s=[...Ae.modulePaths,...r];return[Er({dedupe:n,preferBuiltins:!1,moduleDirectories:Ae.moduleDirectories}),$r(),vr({minify:o?!1:{fast:!0},options:{loadPaths:[...s,it(process.cwd(),"src","styles"),it(process.cwd(),"src")]}}),st(e)]}function ve(r={}){let{modulePaths:t=[],watch:e=!1,input:n=["./src/index.js","./src/registered.js"],outputDir:o="./dist",format:s="esm",dev:i=!1}=r;return{name:"Main",config:{input:n,output:{format:s,dir:o,entryFileNames:a=>i||["index","registered"].includes(a.name)?"[name].js":"[name]-[hash].js",chunkFileNames:i?"[name].js":"[name]-[hash].js",assetFileNames:i?"[name][extname]":"[name]-[hash][extname]"},external:Tr(),plugins:at(t,{dev:i}),watch:ct(e)}}}function G(r={}){let{modulePaths:t=[],watch:e=!1,globPattern:n="./demo/*.js",ignorePattern:o=["./demo/*.min.js"],outputDir:s="./demo",dev:i=!1}=r;return{name:"Demo",config:{input:Object.fromEntries(Ar.sync(n,{ignore:o}).map(a=>[kr(a,".js"),a])),output:{format:"esm",dir:s,entryFileNames:"[name].min.js",chunkFileNames:"[name].min.js",assetFileNames:i?"[name][extname]":"[name]-[hash][extname]"},plugins:at(t,{dev:i}),watch:ct(e)}}}function ct(r){if(!r)return!1;let t=typeof r=="object"?r:{};return{clearScreen:t.clearScreen??!0,buildDelay:t.buildDelay??500,chokidar:{ignoreInitial:!0,ignored:t.ignored??["**/dist/**/*.d.ts","**/custom-elements.json","**/demo/*.md","**/demo/**/*.min.js","**/demo/**/*.min.css","**/docs/api.md","**/node_modules/**","**/.git/**"],awaitWriteFinish:t.awaitWriteFinish??{stabilityThreshold:1e3,pollInterval:100}},include:t.include??["./src/**/*.scss","./src/**/*.js","./src/**/*.ts","./demo/**/*.js","./apiExamples/**/*","./docs/**/*.md"],exclude:t.exclude??["./dist/**/*","./node_modules/**/*"]}}function Tr(r=[]){return[...[/node_modules\/lit/,/node_modules\/lit-element/,/node_modules\/lit-html/,/node_modules\/@lit/],...r]}function dt(){let r=M("./dist"),t=Ce("Cleaning dist folder...").start();try{return Ir(r,{recursive:!0,force:!0}),t.succeed("All clean! Dist folder wiped."),!0}catch(e){return t.fail(`Oops! Couldn't clean dist/ folder: ${e.message}`),console.error(e),!1}}async function ie(r,t,e,n){let o=Ce(r).start();try{let s=await t();return o.succeed(e),s}catch(s){throw o.fail(n),console.error(`Error: ${s.message}`),s}}async function pt(r,t){return ie(`Bundling ${r.name||"main"} and ${t.name||"demo"}...`,async()=>{let e=await Te(r);await e.write(r.output),await e.close();let n=await Te(t);await n.write(t.output),await n.close()},`Bundles ready! ${r.name||"Main"} and ${t.name||"demo"} built.`,"Bundle hiccup! Build failed.")}async function ae(r){let{wcaInput:t,wcaOutput:e,skipDocs:n}=r;if(n){let o=Ce("Skipping docs generation...").start();setTimeout(()=>{o.succeed("Docs generation skipped.")},0);return}return ie("Analyzing components and making docs...",async()=>{await oe(t,e),await _(r),N()},"Docs ready! Looking good.","Doc troubles!")}function xr(){let r=process.cwd();function t(o){return[o,`${o}.scss`,`${o}.css`,M(Dr(o),`_${Sr(o)}.scss`),M(o,"_index.scss"),M(o,"index.scss")].find(i=>lt(i))}function e(o){for(let i of P){let a=t(se(r,i,o));if(a)return a}let s=n(o);return s||null}function n(o){let s,i;if(o.startsWith("@")){let a=o.split("/");if(a.length<3)return null;s=`${a[0]}/${a[1]}`,i=`./${a.slice(2).join("/")}`}else{let a=o.indexOf("/");if(a===-1)return null;s=o.slice(0,a),i=`./${o.slice(a+1)}`}for(let a of P){let l=se(r,a,s,"package.json");if(lt(l))try{let u=JSON.parse(Cr(l,"utf-8")).exports;if(!u||typeof u!="object")continue;let c=u[i];if(!c)continue;let d=typeof c=="string"?c:c.default||c.import;if(!d)continue;let p=t(se(r,a,s,d));if(p)return p}catch{continue}}return null}return{findFileUrl(o){if(o.includes("/node_modules/")){let s=o.lastIndexOf("/node_modules/"),i=o.slice(s+14),a=e(i);if(a)return mt(a)}if(!o.startsWith(".")&&!o.startsWith("/")&&!o.startsWith("file:")){let s=e(o);if(s)return mt(s)}return null}}}async function C(r="./demo"){return ie("Compiling demo SCSS...",async()=>{let t=Pr.sync(M(r,"**/*.scss")),e=xr(),n=process.cwd(),o=P.map(s=>se(n,s));for(let s of t){let i=ut.compile(s,{importers:[e],loadPaths:o,silenceDeprecations:["import"],style:"compressed"}),a=s.replace(/\.scss$/,".min.css");Rr(a,i.css)}return t.length},"Demo SCSS compiled.","SCSS compilation failed.")}async function Ee(r={}){let t=G(r);return ie("Bundling demo JS...",async()=>{let e=await Te(t.config);await e.write(t.config.output),await e.close()},"Demo JS bundled.","Demo JS bundling failed.")}import ft from"node:path";import R from"ora";import"rollup";var H=!1,Or={analyze:{active:!1,lastTime:0},docs:{active:!1,lastTime:0},scss:{active:!1,lastTime:0}},Nr=5e3,ce=new Set,jr=["/dist/index.d.ts","/custom-elements.json","/demo/api.md","/docs/api.md","/demo/index.min.js"];function _r(r){if(!r||typeof r!="string")return!1;try{let t=ft.normalize(r);return jr.some(e=>t.endsWith(e))||t.includes("/dist/")||t.endsWith(".min.js")||t.endsWith(".min.css")||t.endsWith(".d.ts")}catch(t){return console.error(`Error checking path (${typeof r}):`,t.message),!1}}async function Ie(r,t){let e=Or[r];if(e.active||Date.now()-e.lastTime<Nr)return!1;try{return e.active=!0,e.lastTime=Date.now(),await t()}catch(n){return console.error(`Error in ${r} task:`,n),!1}finally{e.active=!1}}async function gt(r,t,e){let n=!0,o={analyze:!1,docs:!1,scss:!1},s=null,i,a=R("Activating watch mode...").start(),l={analyze:async()=>{let{wcaInput:c,wcaOutput:d,skipDocs:p}=t;if(p){let g=R("Skipping component analysis...").start();return setTimeout(()=>{g.succeed("Component analysis skipped.")},0),!0}let f=R("Detective work: analyzing components...").start();try{return await oe(c,d),f.succeed("Component analysis complete! API generated."),!0}catch(g){return f.fail("Analysis hiccup! Something went wrong."),console.error("Component analysis error:",g),!1}},docs:async()=>{if(H)return!1;if(t.skipDocs){let d=R("Skipping docs generation...").start();return setTimeout(()=>{d.succeed("Docs generation skipped.")},0),!0}let c=R("Refreshing docs...").start();try{return await ae(t),c.succeed("Documentation refreshed!"),!0}catch(d){c.fail("Docs stumble! Couldn't refresh."),console.error("Documentation rebuild error:",d)}},scss:async()=>{if(H)return!1;try{return await C(),!0}catch(c){return console.error("Demo SCSS compilation error:",c),!1}}},m=()=>{n&&o.analyze&&o.docs&&o.scss&&typeof e=="function"&&(n=!1,e())};function u(c=1e3){s&&clearTimeout(s),s=setTimeout(async()=>{setTimeout(async()=>{o.analyze=await Ie("analyze",l.analyze),setTimeout(async()=>{o.docs=await Ie("docs",l.docs),setTimeout(async()=>{o.scss=await Ie("scss",l.scss),m()},500)},1e3)},1e3)},c)}r.on("event",async c=>{switch(c.code){case"START":a.succeed("Watch mode active! Eyes peeled.");break;case"BUNDLE_START":if(ce.clear(),c.input)try{let d=Array.isArray(c.input)?c.input:typeof c.input=="string"?[c.input]:typeof c.input=="object"&&c.input!==null?Object.values(c.input):[];for(let p of d)typeof p=="string"&&!_r(p)&&ce.add(ft.normalize(p))}catch(d){console.error("Error processing input paths:",d)}i=R("Weaving bundles...").start(),H=!0;break;case"BUNDLE_END":i&&i.succeed(`Bundle ${Array.isArray(c.input)?`of ${c.input.join("& ")} `:""}done in ${c.duration}ms! \u{1F680}`),H=!1,ce.size>0&&u();break;case"END":break;case"ERROR":H=!1,i?i.fail(`Oops! Bundle hit a snag: ${c.error.message}`):R().fail(`Watch mode hiccup: ${c.error.message}`),ce.clear();break}})}function ht(r,t){return j(r),t&&j(t),Q(),r}async function Gr(r){let t=ve(r),e=G(r);r.dev||t.config.plugins.push(Fr()),await ae(r),await C(),await pt(t.config,e.config)}async function Mr(r){let{dev:t}=r,e=ve({...r,watch:!0}),n=G({...r,watch:!0}),o=Br([e.config,n.config]);gt(o,r,t?async()=>te(r):void 0);let s=await import("chokidar"),i=!1,a=!1,l=null,m=s.watch("./demo/**/*.scss",{ignoreInitial:!0,ignored:["**/demo/**/*.min.css"],awaitWriteFinish:{stabilityThreshold:500,pollInterval:100}});return m.on("all",()=>{l&&clearTimeout(l),l=setTimeout(async()=>{if(i){a=!0;return}i=!0;try{await C()}catch(u){console.error("Demo SCSS watch compilation error:",u)}finally{i=!1,a&&(a=!1,m.emit("all"))}},500)}),ht(o,m),o}async function le(r={}){try{let{watch:t}=r;return dt(),t?await Mr(r):await Gr(r)}catch(t){throw new Error(`Build failed: ${t.message}`)}}var L=Hr.command("dev").description("Runs development server for auro components");L=K(L);L=Y(L);var Ps=L.action(async r=>{try{let t=yt("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),r.watch||t.succeed("Build completed!"),await le({...r,dev:!0,watch:r.watch})}catch(t){yt().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{program as Lr}from"commander";import wt from"ora";var Re=Lr.command("build").description("Builds auro components");Re=K(Re);var Fs=Re.action(async r=>{try{let t=wt("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),await le(r),r.watch||t.succeed("Build completed!")}catch(t){wt().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{exec as Wr}from"node:child_process";import bt from"node:path";import Ur from"node:process";import{fileURLToPath as zr}from"node:url";import qr from"node:util";import{program as Vr}from"commander";import Jr from"inquirer";var Vs=Vr.command("migrate").description("Script runner to perform repetitive code change tasks").requiredOption("-i, --id <string>","Select the migration you would like to run by id").option("-m, --multi-gitter","Run the migration on all repositories in the multi-gitter config").action(async r=>{let t=zr(import.meta.url),e=bt.dirname(t),n=bt.resolve(e,"migrations",r.id);if(r.multiGitter){let o=qr.promisify(Wr);try{await o("command -v multi-gitter")}catch{console.error("multi-gitter is not installed."),Ur.exit(1)}(await Jr.prompt([{type:"confirm",name:"dryRun",message:"Run migration in dry-run mode? (no changes will be committed)",default:!0}])).dryRun?A(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml" --dry-run`):A(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml"`)}else A(`${n}/script.sh`)});import vt from"node:process";import{program as Qr}from"commander";import{readFile as en,writeFile as tn}from"node:fs/promises";import{Logger as W}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import Se from"node:fs/promises";import De from"node:path";import{Octokit as Kr}from"@octokit/rest";import v from"ora";import{processContentForFile as Yr,templateFiller as Xr}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";async function kt(r,t){let e=new Kr({auth:process.env.GITHUB_TOKEN||""});try{let o=(await e.request("GET /repos/{owner}/{repo}/contents/{path}",{ref:t,owner:"AlaskaAirlines",repo:"auro-templates",path:r,headers:{"X-GitHub-Api-Version":"2022-11-28"}})).data;if(typeof o!="object"||!Array.isArray(o)){let s=`Unexpected response format: ${JSON.stringify(o)}`;throw v().start().fail(s),new Error("Failed to retrieve folder items")}return o}catch(n){let o=v().start();throw n.status===404?o.fail(`Template '${r.split("/")[1]}' not found`):o.fail(`Error accessing template: ${n.message}`),n}}async function $t({folderItems:r,templatePathToReplace:t,rootDir:e,ref:n}){let o=[];for(let s of r){if(s.type=="dir"){let m=v(`Processing directory: ${s.path}`).start(),u=await kt(s.path,n);m.succeed(`Found ${u.length} additional items in ${s.path}`);let c=await $t({folderItems:u,templatePathToReplace:t,rootDir:e,ref:n});o.push(...c);continue}let i=s.path.replace(`${t}/`,""),a=`${e}/.github/${i}`,l={identifier:s.name,input:{remoteUrl:s.download_url||"",fileName:a,overwrite:!0},output:a};o.push(l)}return o}async function Zr(r){try{await Se.rm(r,{recursive:!0,force:!0}),v().start().succeed(`Successfully removed directory: ${r}`)}catch(t){throw v().start().fail(`Error removing directory ${r}: ${t.message}`),t}}async function Et(r,t="",e=!0){try{let n=await Se.stat(r),o=De.basename(r);if(!n.isDirectory())return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}
19
+ <!-- AURO-GENERATED-CONTENT:END -->`}n=n.replace(u,$),o.lastIndex=0,i=!0}}i&&(n=Ee.replaceTemplateValues(n,t),await T.promises.writeFile(r,n)),n=await T.promises.readFile(r,"utf8");let l=/^[ \t]*```(\w*)\n([\s\S]*?)^[ \t]*```[ \t]*$/gm,d=n.replace(l,(u,f,p)=>{let g=f||"html",b=p.trimEnd().replace(/&/g,"&amp;").replace(/</g,"&lt;").replace(/>/g,"&gt;");return`<pre class="language-${g}"><code class="language-${g}">${b}
20
+ </code></pre>`});d!==n&&await T.promises.writeFile(r,d),n=await T.promises.readFile(r,"utf8"),n=n.replace(/(<pre[^>]*><code[^>]*>)([\s\S]*?)(<\/code><\/pre>)/g,(u,f,p,g)=>{let b=p.split(`
21
+ `),P=b.filter(E=>E.trim().length>0);if(P.length===0)return u;let O=Math.min(...P.map(E=>{let J=E.match(/^[ \t]*/);return J?J[0].length:0})),$=b.map(E=>E===""||E.trim().length===0?"\u200B":O>0?E.substring(O):E);for(;$.length>0&&($[$.length-1]==="\u200B"||$[$.length-1]==="");)$.pop();return f+$.join(`
22
+ `)+g});let m=n.split(`
23
+ `),c=!1;for(let u=0;u<m.length;u++)/<pre[\s>]/i.test(m[u])&&(c=!0),c||(m[u]=m[u].replace(/^[ \t]+(?=<)/,"")),/<\/pre>/i.test(m[u])&&(c=!1);await T.promises.writeFile(r,m.join(`
24
+ `))}async function F(r={}){let t=r.readmeTemplate,e=t&&!t.startsWith("http");await cr({...tt,...e?{localReadmePath:D.resolve(process.cwd(),t)}:{remoteReadmeUrl:t||"https://raw.githubusercontent.com/AlaskaAirlines/auro-templates/main/templates/default/README.md"}},r.skipReadme)}function Qe(r){return T.existsSync(k(r))}import{existsSync as rt,readFileSync as mr}from"node:fs";import{builtinModules as ur}from"node:module";import{resolve as te}from"node:path";import{startDevServer as dr}from"@web/dev-server";import{hmrPlugin as pr}from"@web/dev-server-hmr";import*as nt from"esbuild";import fr from"ora";var x=["node_modules","../node_modules","../../node_modules","../../../node_modules"];var gr=/^\/__wds-outside-root__\/(\d+)\/(.+)$/;function hr(){return{name:"node-modules-css",serve(r){if(!r.path.endsWith(".css")||!r.path.startsWith("/@")&&!/^\/[a-z]/i.test(r.path))return;let t=r.path.slice(1),e=process.cwd();for(let n of x){let o=te(e,n,t);if(rt(o))return{body:mr(o,"utf-8"),type:"css"}}}}}function yr(r,t){let e=r.match(gr);return e?te(t,"../".repeat(Number.parseInt(e[1],10)),e[2]):te(t,`.${r}`)}function wr(){let r=/\b(require\s*\(|module\.exports\b|exports\.\w)/,t=new Map,e;return{name:"cjs-to-esm",async serverStart({config:n}){e=te(n.rootDir)},async transform(n){if(!n.path.includes("node_modules")||!n.path.endsWith(".js")&&!n.path.endsWith(".cjs")||typeof n.body!="string"||!r.test(n.body))return;let o=yr(n.path,e);if(t.has(o))return t.get(o);if(rt(o))try{let i={body:(await nt.build({entryPoints:[o],bundle:!0,format:"esm",platform:"browser",write:!1,logLevel:"silent",external:ur})).outputFiles[0].text};return t.set(o,i),i}catch(s){console.error(`CJS-to-ESM bundling failed for ${n.path}:`,s.message)}}}}var B={watch:!0,nodeResolve:!0,basePath:"/",rootDir:"./demo",hmrInclude:["src/**/*","demo/**/*","apiExamples/**/*","docs/**/*"]};async function re(r={}){if(!r.serve)return;let t=fr(`Firing up dev server...
25
+ `).start();try{let e={port:Number(r.port)||void 0,open:r.open?"/":void 0,watch:r.watch??B.watch,nodeResolve:r.nodeResolve??B.nodeResolve,basePath:r.basePath??B.basePath,rootDir:r.rootDir??B.rootDir,middleware:[function(s,i){return!s.url.endsWith("/")&&!s.url.includes(".")&&(s.url+=".html"),i()}],plugins:[hr(),wr(),pr({include:r.hmrInclude??B.hmrInclude})]},n=await dr({config:e,readCliArgs:!1,readFileConfig:!1});return t.stop(),n}catch(e){throw t.fail("Server snag! Couldn't start dev server."),console.error("Error starting development server:",e),new Error(`Development server failed to start: ${e.message}`)}}async function ne(){let r=G("Generating Custom Elements Manifest...").start();try{await A(`npx --package=@custom-elements-manifest/analyzer -y -- cem analyze --config '${Je("custom-elements-manifest.config.mjs")}'`),r.succeed("Custom Elements Manifest generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);r.warn("CEM analyzer completed with warnings: "+e)}}async function oe(){let r=G("Generating API md file...").start();try{await N.generate(),r.succeed("API md file generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);throw r.fail("Failed to generate API md file: "+e),t}}async function ot(r={}){let t=G("Compiling documentation...").start();try{await F(r),t.succeed("Documentation compiled successfully!")}catch(e){let n=e instanceof Error?e.message:String(e);throw t.fail("Failed to compile MD documentation: "+n),e}j(),await C(),await Ae(r)}async function st(r={}){await re(r)}async function it(r={}){let t=await import("chokidar"),e=["./src/**/*","./docs/**/*","./docTemplates/**/*","./apiExamples/**/*","./demo/**/*.scss"],n=["**/demo/*.min.js","**/demo/*.min.css","**/demo/*.md","**/demo/readme.md","**/docs/api.md","**/custom-elements.json","**/node_modules/**","**/dist/**","**/.git/**"],o=t.watch(e,{ignoreInitial:!0,ignored:n,awaitWriteFinish:{stabilityThreshold:1e3,pollInterval:100}}),s=G("Waiting for changes...");s.spinner="bouncingBar",s.color="green",s.start();let i=null,a=!1,l=!1;async function d(m){if(a){l=!0;return}a=!0,o.unwatch(e);let c=G(`Change detected: ${m}`).start();try{await F(r),j(),await C(),await Ae(r),c.succeed("Docs rebuilt!")}catch(u){let f=u instanceof Error?u.message:String(u);c.fail("Rebuild failed: "+f)}finally{setTimeout(()=>{o.add(e),a=!1,l&&(l=!1,d("queued changes"))},1e3)}}o.on("all",(m,c)=>{a||(i&&clearTimeout(i),i=setTimeout(()=>{d(c)},1e3))}),_(o),ee()}async function se(){await ne(),await oe()}import{basename as $r,join as ct}from"node:path";import Er from"@rollup/plugin-commonjs";import{nodeResolve as Ar}from"@rollup/plugin-node-resolve";import{glob as vr}from"glob";import{litScss as Tr}from"rollup-plugin-scss-lit";import br from"node:path";import{glob as kr}from"glob";function at(r){return{name:"watch-globs",buildStart(){let t=Array.isArray(r)?r:[r];for(let e of t)try{for(let n of kr.sync(br.resolve(e)))this.addWatchFile(n)}catch(n){this.error(`Error watching glob pattern "${e}": ${n.message}`)}}}}var ve={moduleDirectories:["node_modules"],modulePaths:["../../node_modules","../node_modules","node_modules"],watchPatterns:["./apiExamples/**/*","./docs/**/*"]};function lt(r=[],t={}){let{watchPatterns:e=ve.watchPatterns,dedupe:n=["lit","lit-element","lit-html"],dev:o=!1}=t,s=[...ve.modulePaths,...r];return[Ar({dedupe:n,preferBuiltins:!1,moduleDirectories:ve.moduleDirectories}),Er(),Tr({minify:o?!1:{fast:!0},options:{loadPaths:[...s,ct(process.cwd(),"src","styles"),ct(process.cwd(),"src")]}}),at(e)]}function Te(r={}){let{modulePaths:t=[],watch:e=!1,input:n=["./src/index.js","./src/registered.js"],outputDir:o="./dist",format:s="esm",dev:i=!1}=r;return{name:"Main",config:{input:n,output:{format:s,dir:o,entryFileNames:a=>i||["index","registered"].includes(a.name)?"[name].js":"[name]-[hash].js",chunkFileNames:i?"[name].js":"[name]-[hash].js",assetFileNames:i?"[name][extname]":"[name]-[hash][extname]"},external:Cr(),plugins:lt(t,{dev:i}),watch:mt(e)}}}function M(r={}){let{modulePaths:t=[],watch:e=!1,globPattern:n="./demo/*.js",ignorePattern:o=["./demo/*.min.js"],outputDir:s="./demo",dev:i=!1}=r;return{name:"Demo",config:{input:Object.fromEntries(vr.sync(n,{ignore:o}).map(a=>[$r(a,".js"),a])),output:{format:"esm",dir:s,entryFileNames:"[name].min.js",chunkFileNames:"[name].min.js",assetFileNames:i?"[name][extname]":"[name]-[hash][extname]"},plugins:lt(t,{dev:i}),watch:mt(e)}}}function mt(r){if(!r)return!1;let t=typeof r=="object"?r:{};return{clearScreen:t.clearScreen??!0,buildDelay:t.buildDelay??500,chokidar:{ignoreInitial:!0,ignored:t.ignored??["**/dist/**/*.d.ts","**/custom-elements.json","**/demo/*.md","**/demo/**/*.min.js","**/demo/**/*.min.css","**/docs/api.md","**/node_modules/**","**/.git/**"],awaitWriteFinish:t.awaitWriteFinish??{stabilityThreshold:1e3,pollInterval:100}},include:t.include??["./src/**/*.scss","./src/**/*.js","./src/**/*.ts","./demo/**/*.js","./apiExamples/**/*","./docs/**/*.md"],exclude:t.exclude??["./dist/**/*","./node_modules/**/*"]}}function Cr(r=[]){return[...[/node_modules\/lit/,/node_modules\/lit-element/,/node_modules\/lit-html/,/node_modules\/@lit/],...r]}function ft(){let r=H("./dist"),t=Ie("Cleaning dist folder...").start();try{return Rr(r,{recursive:!0,force:!0}),t.succeed("All clean! Dist folder wiped."),!0}catch(e){return t.fail(`Oops! Couldn't clean dist/ folder: ${e.message}`),console.error(e),!1}}async function ae(r,t,e,n){let o=Ie(r).start();try{let s=await t();return o.succeed(e),s}catch(s){throw o.fail(n),console.error(`Error: ${s.message}`),s}}async function gt(r,t){return ae(`Bundling ${r.name||"main"} and ${t.name||"demo"}...`,async()=>{let e=await Ce(r);await e.write(r.output),await e.close();let n=await Ce(t);await n.write(t.output),await n.close()},`Bundles ready! ${r.name||"Main"} and ${t.name||"demo"} built.`,"Bundle hiccup! Build failed.")}async function ce(r){let{wcaInput:t,wcaOutput:e,skipDocs:n}=r;if(n){let o=Ie("Skipping docs generation...").start();setTimeout(()=>{o.succeed("Docs generation skipped.")},0);return}return ae("Analyzing components and making docs...",async()=>{await se(t,e),await F(r),j()},"Docs ready! Looking good.","Doc troubles!")}function Or(){let r=process.cwd();function t(o){return[o,`${o}.scss`,`${o}.css`,H(xr(o),`_${Dr(o)}.scss`),H(o,"_index.scss"),H(o,"index.scss")].find(i=>ut(i))}function e(o){for(let i of x){let a=t(ie(r,i,o));if(a)return a}let s=n(o);return s||null}function n(o){let s,i;if(o.startsWith("@")){let a=o.split("/");if(a.length<3)return null;s=`${a[0]}/${a[1]}`,i=`./${a.slice(2).join("/")}`}else{let a=o.indexOf("/");if(a===-1)return null;s=o.slice(0,a),i=`./${o.slice(a+1)}`}for(let a of x){let l=ie(r,a,s,"package.json");if(ut(l))try{let m=JSON.parse(Ir(l,"utf-8")).exports;if(!m||typeof m!="object")continue;let c=m[i];if(!c)continue;let u=typeof c=="string"?c:c.default||c.import;if(!u)continue;let f=t(ie(r,a,s,u));if(f)return f}catch{continue}}return null}return{findFileUrl(o){if(o.includes("/node_modules/")){let s=o.lastIndexOf("/node_modules/"),i=o.slice(s+14),a=e(i);if(a)return dt(a)}if(!o.startsWith(".")&&!o.startsWith("/")&&!o.startsWith("file:")){let s=e(o);if(s)return dt(s)}return null}}}async function C(r="./demo"){return ae("Compiling demo SCSS...",async()=>{let t=Pr.sync(H(r,"**/*.scss")),e=Or(),n=process.cwd(),o=x.map(s=>ie(n,s));for(let s of t){let i=pt.compile(s,{importers:[e],loadPaths:o,silenceDeprecations:["import"],style:"compressed"}),a=s.replace(/\.scss$/,".min.css");Sr(a,i.css)}return t.length},"Demo SCSS compiled.","SCSS compilation failed.")}async function Ae(r={}){let t=M(r);return ae("Bundling demo JS...",async()=>{let e=await Ce(t.config);await e.write(t.config.output),await e.close()},"Demo JS bundled.","Demo JS bundling failed.")}import ht from"node:path";import R from"ora";import"rollup";var L=!1,Nr={analyze:{active:!1,lastTime:0},docs:{active:!1,lastTime:0},scss:{active:!1,lastTime:0}},jr=5e3,le=new Set,_r=["/dist/index.d.ts","/custom-elements.json","/demo/api.md","/docs/api.md","/demo/index.min.js"];function Fr(r){if(!r||typeof r!="string")return!1;try{let t=ht.normalize(r);return _r.some(e=>t.endsWith(e))||t.includes("/dist/")||t.endsWith(".min.js")||t.endsWith(".min.css")||t.endsWith(".d.ts")}catch(t){return console.error(`Error checking path (${typeof r}):`,t.message),!1}}async function Re(r,t){let e=Nr[r];if(e.active||Date.now()-e.lastTime<jr)return!1;try{return e.active=!0,e.lastTime=Date.now(),await t()}catch(n){return console.error(`Error in ${r} task:`,n),!1}finally{e.active=!1}}async function yt(r,t,e){let n=!0,o={analyze:!1,docs:!1,scss:!1},s=null,i,a=R("Activating watch mode...").start(),l={analyze:async()=>{let{wcaInput:c,wcaOutput:u,skipDocs:f}=t;if(f){let g=R("Skipping component analysis...").start();return setTimeout(()=>{g.succeed("Component analysis skipped.")},0),!0}let p=R("Detective work: analyzing components...").start();try{return await se(c,u),p.succeed("Component analysis complete! API generated."),!0}catch(g){return p.fail("Analysis hiccup! Something went wrong."),console.error("Component analysis error:",g),!1}},docs:async()=>{if(L)return!1;if(t.skipDocs){let u=R("Skipping docs generation...").start();return setTimeout(()=>{u.succeed("Docs generation skipped.")},0),!0}let c=R("Refreshing docs...").start();try{return await ce(t),c.succeed("Documentation refreshed!"),!0}catch(u){c.fail("Docs stumble! Couldn't refresh."),console.error("Documentation rebuild error:",u)}},scss:async()=>{if(L)return!1;try{return await C(),!0}catch(c){return console.error("Demo SCSS compilation error:",c),!1}}},d=()=>{n&&o.analyze&&o.docs&&o.scss&&typeof e=="function"&&(n=!1,e())};function m(c=1e3){s&&clearTimeout(s),s=setTimeout(async()=>{setTimeout(async()=>{o.analyze=await Re("analyze",l.analyze),setTimeout(async()=>{o.docs=await Re("docs",l.docs),setTimeout(async()=>{o.scss=await Re("scss",l.scss),d()},500)},1e3)},1e3)},c)}r.on("event",async c=>{switch(c.code){case"START":a.succeed("Watch mode active! Eyes peeled.");break;case"BUNDLE_START":if(le.clear(),c.input)try{let u=Array.isArray(c.input)?c.input:typeof c.input=="string"?[c.input]:typeof c.input=="object"&&c.input!==null?Object.values(c.input):[];for(let f of u)typeof f=="string"&&!Fr(f)&&le.add(ht.normalize(f))}catch(u){console.error("Error processing input paths:",u)}i=R("Weaving bundles...").start(),L=!0;break;case"BUNDLE_END":i&&i.succeed(`Bundle ${Array.isArray(c.input)?`of ${c.input.join("& ")} `:""}done in ${c.duration}ms! \u{1F680}`),L=!1,le.size>0&&m();break;case"END":break;case"ERROR":L=!1,i?i.fail(`Oops! Bundle hit a snag: ${c.error.message}`):R().fail(`Watch mode hiccup: ${c.error.message}`),le.clear();break}})}function wt(r,t){return _(r),t&&_(t),ee(),r}async function Mr(r){let t=Te(r),e=M(r);r.dev||t.config.plugins.push(Br()),await ce(r),await C(),await gt(t.config,e.config)}async function Hr(r){let{dev:t}=r,e=Te({...r,watch:!0}),n=M({...r,watch:!0}),o=Gr([e.config,n.config]);yt(o,r,t?async()=>re(r):void 0);let s=await import("chokidar"),i=!1,a=!1,l=null,d=s.watch("./demo/**/*.scss",{ignoreInitial:!0,ignored:["**/demo/**/*.min.css"],awaitWriteFinish:{stabilityThreshold:500,pollInterval:100}});return d.on("all",()=>{l&&clearTimeout(l),l=setTimeout(async()=>{if(i){a=!0;return}i=!0;try{await C()}catch(m){console.error("Demo SCSS watch compilation error:",m)}finally{i=!1,a&&(a=!1,d.emit("all"))}},500)}),wt(o,d),o}async function me(r={}){try{let{watch:t}=r;return ft(),t?await Hr(r):await Mr(r)}catch(t){throw new Error(`Build failed: ${t.message}`)}}var W=Lr.command("dev").description("Runs development server for auro components");W=Y(W);W=X(W);var Ps=W.action(async r=>{try{let t=bt("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),r.watch||t.succeed("Build completed!"),await me({...r,dev:!0,watch:r.watch})}catch(t){bt().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{program as Wr}from"commander";import kt from"ora";var Se=Wr.command("build").description("Builds auro components");Se=Y(Se);var Bs=Se.action(async r=>{try{let t=kt("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),await me(r),r.watch||t.succeed("Build completed!")}catch(t){kt().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{exec as Ur}from"node:child_process";import $t from"node:path";import zr from"node:process";import{fileURLToPath as qr}from"node:url";import Vr from"node:util";import{program as Jr}from"commander";import Kr from"inquirer";var Js=Jr.command("migrate").description("Script runner to perform repetitive code change tasks").requiredOption("-i, --id <string>","Select the migration you would like to run by id").option("-m, --multi-gitter","Run the migration on all repositories in the multi-gitter config").action(async r=>{let t=qr(import.meta.url),e=$t.dirname(t),n=$t.resolve(e,"migrations",r.id);if(r.multiGitter){let o=Vr.promisify(Ur);try{await o("command -v multi-gitter")}catch{console.error("multi-gitter is not installed."),zr.exit(1)}(await Kr.prompt([{type:"confirm",name:"dryRun",message:"Run migration in dry-run mode? (no changes will be committed)",default:!0}])).dryRun?A(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml" --dry-run`):A(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml"`)}else A(`${n}/script.sh`)});import Ct from"node:process";import{program as en}from"commander";import{readFile as tn,writeFile as rn}from"node:fs/promises";import{Logger as U}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import De from"node:fs/promises";import xe from"node:path";import{Octokit as Yr}from"@octokit/rest";import v from"ora";import{processContentForFile as Xr,templateFiller as Zr}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";async function Et(r,t){let e=new Yr({auth:process.env.GITHUB_TOKEN||""});try{let o=(await e.request("GET /repos/{owner}/{repo}/contents/{path}",{ref:t,owner:"AlaskaAirlines",repo:"auro-templates",path:r,headers:{"X-GitHub-Api-Version":"2022-11-28"}})).data;if(typeof o!="object"||!Array.isArray(o)){let s=`Unexpected response format: ${JSON.stringify(o)}`;throw v().start().fail(s),new Error("Failed to retrieve folder items")}return o}catch(n){let o=v().start();throw n.status===404?o.fail(`Template '${r.split("/")[1]}' not found`):o.fail(`Error accessing template: ${n.message}`),n}}async function At({folderItems:r,templatePathToReplace:t,rootDir:e,ref:n}){let o=[];for(let s of r){if(s.type=="dir"){let d=v(`Processing directory: ${s.path}`).start(),m=await Et(s.path,n);d.succeed(`Found ${m.length} additional items in ${s.path}`);let c=await At({folderItems:m,templatePathToReplace:t,rootDir:e,ref:n});o.push(...c);continue}let i=s.path.replace(`${t}/`,""),a=`${e}/.github/${i}`,l={identifier:s.name,input:{remoteUrl:s.download_url||"",fileName:a,overwrite:!0},output:a};o.push(l)}return o}async function Qr(r){try{await De.rm(r,{recursive:!0,force:!0}),v().start().succeed(`Successfully removed directory: ${r}`)}catch(t){throw v().start().fail(`Error removing directory ${r}: ${t.message}`),t}}async function vt(r,t="",e=!0){try{let n=await De.stat(r),o=xe.basename(r);if(!n.isDirectory())return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}
26
26
  `;let s=`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}/
27
- `;try{let a=(await Se.readdir(r)).sort();for(let l=0;l<a.length;l++){let m=a[l],u=De.join(r,m),c=l===a.length-1,d=t+(e?" ":"\u2502 ");s+=await Et(u,d,c)}}catch{s+=`${t}${e?" ":"\u2502 "}\u2514\u2500\u2500 [Permission denied or error reading directory]
27
+ `;try{let a=(await De.readdir(r)).sort();for(let l=0;l<a.length;l++){let d=a[l],m=xe.join(r,d),c=l===a.length-1,u=t+(e?" ":"\u2502 ");s+=await vt(m,u,c)}}catch{s+=`${t}${e?" ":"\u2502 "}\u2514\u2500\u2500 [Permission denied or error reading directory]
28
28
  `}return s}catch(n){return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}[Error: ${n}]
29
- `}}async function At(r,t="main",e="default"){if(!r)throw v().start().fail("Root directory must be specified"),new Error("Root directory must be specified");if(await Xr.extractNames(),!process.env.GITHUB_TOKEN)throw v().start().fail("GITHUB_TOKEN environment variable is not set."),new Error("GITHUB_TOKEN environment variable is not set");let n=`templates/${e}/.github`,o=await kt(n,t),s=".github",i=v("Removing existing .github directory...").start();try{await Zr(s),i.succeed(".github directory removed successfully")}catch(m){throw i.fail(`Error removing .github directory: ${m.message}`),new Error(`Failed to remove .github directory: ${m.message}`)}let a=await $t({folderItems:o,templatePathToReplace:n,rootDir:r,ref:t}),l=v("Processing all files...").start();try{await Promise.all(a.map(u=>Yr(u))),l.succeed("All files processed.");let m=v("Generating directory tree...").start();try{let u=De.join(r,".github"),c=await Et(u);m.succeed("Synced .github directory structure:"),console.log(c)}catch(u){m.fail(`Error generating directory tree: ${u.message}`)}}catch(m){throw l.fail(`Error processing files: ${m.message}`),new Error(`Failed to process files: ${m.message}`)}}var ii=Qr.command("sync").option("-r, --ref <branch/tag/commit>","Git reference (branch/tag/commit) to use","main").option("-t, --template <name>","Template based on which to sync","default").description("Script runner to synchronize local repository configuration files").action(async r=>{W.info("Synchronizing repository configuration files..."),W.warn("Note: sync does not create a new git branch. Changes are added to the current branch.");let t=vt.cwd();try{await At(t,r.ref,r.template);try{let e=`${t}/.github/CODEOWNERS`,o=(await en(e,{encoding:"utf-8"})).replace(/\r\n/gu,`
29
+ `}}async function Tt(r,t="main",e="default"){if(!r)throw v().start().fail("Root directory must be specified"),new Error("Root directory must be specified");if(await Zr.extractNames(),!process.env.GITHUB_TOKEN)throw v().start().fail("GITHUB_TOKEN environment variable is not set."),new Error("GITHUB_TOKEN environment variable is not set");let n=`templates/${e}/.github`,o=await Et(n,t),s=".github",i=v("Removing existing .github directory...").start();try{await Qr(s),i.succeed(".github directory removed successfully")}catch(d){throw i.fail(`Error removing .github directory: ${d.message}`),new Error(`Failed to remove .github directory: ${d.message}`)}let a=await At({folderItems:o,templatePathToReplace:n,rootDir:r,ref:t}),l=v("Processing all files...").start();try{await Promise.all(a.map(m=>Xr(m))),l.succeed("All files processed.");let d=v("Generating directory tree...").start();try{let m=xe.join(r,".github"),c=await vt(m);d.succeed("Synced .github directory structure:"),console.log(c)}catch(m){d.fail(`Error generating directory tree: ${m.message}`)}}catch(d){throw l.fail(`Error processing files: ${d.message}`),new Error(`Failed to process files: ${d.message}`)}}var ai=en.command("sync").option("-r, --ref <branch/tag/commit>","Git reference (branch/tag/commit) to use","main").option("-t, --template <name>","Template based on which to sync","default").description("Script runner to synchronize local repository configuration files").action(async r=>{U.info("Synchronizing repository configuration files..."),U.warn("Note: sync does not create a new git branch. Changes are added to the current branch.");let t=Ct.cwd();try{await Tt(t,r.ref,r.template);try{let e=`${t}/.github/CODEOWNERS`,o=(await tn(e,{encoding:"utf-8"})).replace(/\r\n/gu,`
30
30
  `).replace(/\n\n/gu,`
31
- `);await tn(e,o,{encoding:"utf-8"}),(o.includes("\r")||o.includes(`
31
+ `);await rn(e,o,{encoding:"utf-8"}),(o.includes("\r")||o.includes(`
32
32
 
33
- `))&&W.error("CODEOWNERS file still has Windows line endings.")}catch(e){throw W.error(`Error processing CODEOWNERS file: ${e.message}`),e}}catch(e){W.error(`Sync operation failed: ${e.message}`),vt.exit(1)}});import me from"node:fs";import U from"node:path";import{Logger as Ct}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{program as rn}from"commander";import{glob as nn}from"glob";var Tt=(r,t)=>{let e=(r.match(/static register\(name \= (.+)\)/)||r.match(/customElements.get\((.+?)\)/))[1],n=r.match(/export class (.+) extends/)?.[1],o=r.match(/\/\*\*((.|\n)*?)(\*\n|\*\/|[@])/)?.[1]||"";return!e||!n?r:`
33
+ `))&&U.error("CODEOWNERS file still has Windows line endings.")}catch(e){throw U.error(`Error processing CODEOWNERS file: ${e.message}`),e}}catch(e){U.error(`Sync operation failed: ${e.message}`),Ct.exit(1)}});import ue from"node:fs";import z from"node:path";import{Logger as Rt}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{program as nn}from"commander";import{glob as on}from"glob";var It=(r,t)=>{let e=(r.match(/static register\(name \= (.+)\)/)||r.match(/customElements.get\((.+?)\)/))[1],n=r.match(/export class (.+) extends/)?.[1],o=r.match(/\/\*\*((.|\n)*?)(\*\n|\*\/|[@])/)?.[1]||"";return!e||!n?r:`
34
34
  import { ${n} } from '${t}';
35
35
 
36
36
  /**${o}*/
@@ -39,30 +39,30 @@ class ${n}WCA extends ${n} {}
39
39
  if (!customElements.get(${e})) {
40
40
  customElements.define(${e}, ${n}WCA);
41
41
  }
42
- `};var ue=U.resolve(process.cwd(),"./scripts/wca");async function on(r){try{return(await Promise.all(r.map(e=>nn(e)))).flat()}catch(t){throw console.error("Error processing glob patterns:",t),t}}async function sn(r){me.existsSync(ue)||await me.promises.mkdir(ue,{recursive:!0});for(let t of r){let e=U.resolve(process.cwd(),t),n=await me.promises.readFile(e,"utf-8"),o=U.resolve(ue,`${U.basename(t)}`),s=Tt(n,U.relative(ue,t));await me.promises.writeFile(o,s)}}async function an(){let r=await on(["./src/auro-*.js"]);await sn(r)}var gi=rn.command("wca-setup").description("Set up WCA (Web Component Analyzer) for the project").action(()=>{an().then(()=>{Ct.success("WCA setup completed successfully.")}).catch(r=>{Ct.error(`WCA setup failed: ${r.message}`)})});import{program as gn}from"commander";import Dt from"chalk";import Pt from"ora";import{appendFile as cn,readFile as ln}from"node:fs/promises";import{Logger as h}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{simpleGit as mn}from"simple-git";var y;try{y=mn({baseDir:process.cwd(),binary:"git",maxConcurrentProcesses:1})}catch(r){h.error(`Failed to initialize git: ${r}`),y={}}var I=class r{static async checkGitignore(t){if(t==="")return!1;try{return(await ln(".gitignore","utf-8")).includes(t)}catch(e){return h.error(`Error reading file: ${e}`),!1}}static async getCommitMessages(t=""){try{let e=t;e||(e=(await y.branchLocal()).current);let n="main",o="";if(!!process.env.GITHUB_ACTIONS){n=process.env.GITHUB_BASE_REF||"main";try{if(await y.fetch("origin",n),e!=="HEAD")try{await y.raw(["rev-parse","--verify",`origin/${e}`])}catch{await y.fetch("origin",e)}let i=e==="HEAD"?"HEAD":`origin/${e}`;o=`${(await y.raw(["merge-base",`origin/${n}`,i])).trim()}..${i}`}catch(i){h.warn(`Error setting up commit range in CI: ${i}`);let a=e==="HEAD"?"HEAD":`origin/${e}`;o=`origin/${n}..${a}`}}else try{try{await y.raw(["rev-parse","--verify",`origin/${n}`])}catch{h.info(`Fetching ${n} from origin`),await y.fetch("origin",n)}if(e!=="HEAD")try{await y.raw(["rev-parse","--verify",e])}catch{await y.fetch("origin",e)}o=`${(await y.raw(["merge-base",`origin/${n}`,e])).trim()}..${e}`}catch(i){h.warn(`Error determining commits locally: ${i}`),o=`${e}~10..${e}`}return await r.getFormattedCommits(o)}catch(e){return h.error(`Error getting commit messages: ${e}`),[]}}static async getRepoOwnerAndName(){try{let t=await y.getRemotes(!0);if(t.length===0)return h.warn("No remotes found"),null;let e=t.find(o=>o.name==="origin")||t[0],n=e.refs.fetch||e.refs.push;return r.parseGitUrl(n)}catch(t){return h.error(`Error getting repo owner and name: ${t}`),null}}static async getCurrentBranchName(){try{return(await y.branchLocal()).current||null}catch(t){return h.error(`Error getting current branch name: ${t}`),null}}static parseGitUrl(t){let e;return t.includes("@")&&t.includes(":")&&(e=t.match(/@([^:]+):([^/]+)\/(.+?)(?:\.git)?$/),e)?{owner:e[2],repo:e[3]}:(e=t.match(/https?:\/\/(?:[^@]+@)?[^/]+\/([^/]+)\/(.+?)(?:\.git)?$/),e?{owner:e[1],repo:e[2]}:(h.warn(`Could not parse git URL: ${t}`),null))}static async getFormattedCommits(t){let n=(await y.raw(["log","--pretty=format:COMMIT_START%n%H%n%ad%n%an%n%s%n%b%nCOMMIT_END","--date=short",t])).split(`COMMIT_START
42
+ `};var de=z.resolve(process.cwd(),"./scripts/wca");async function sn(r){try{return(await Promise.all(r.map(e=>on(e)))).flat()}catch(t){throw console.error("Error processing glob patterns:",t),t}}async function an(r){ue.existsSync(de)||await ue.promises.mkdir(de,{recursive:!0});for(let t of r){let e=z.resolve(process.cwd(),t),n=await ue.promises.readFile(e,"utf-8"),o=z.resolve(de,`${z.basename(t)}`),s=It(n,z.relative(de,t));await ue.promises.writeFile(o,s)}}async function cn(){let r=await sn(["./src/auro-*.js"]);await an(r)}var hi=nn.command("wca-setup").description("Set up WCA (Web Component Analyzer) for the project").action(()=>{cn().then(()=>{Rt.success("WCA setup completed successfully.")}).catch(r=>{Rt.error(`WCA setup failed: ${r.message}`)})});import{program as hn}from"commander";import Pt from"chalk";import Ot from"ora";import{appendFile as ln,readFile as mn}from"node:fs/promises";import{Logger as h}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{simpleGit as un}from"simple-git";var y;try{y=un({baseDir:process.cwd(),binary:"git",maxConcurrentProcesses:1})}catch(r){h.error(`Failed to initialize git: ${r}`),y={}}var I=class r{static async checkGitignore(t){if(t==="")return!1;try{return(await mn(".gitignore","utf-8")).includes(t)}catch(e){return h.error(`Error reading file: ${e}`),!1}}static async getCommitMessages(t=""){try{let e=t;e||(e=(await y.branchLocal()).current);let n="main",o="";if(!!process.env.GITHUB_ACTIONS){n=process.env.GITHUB_BASE_REF||"main";try{if(await y.fetch("origin",n),e!=="HEAD")try{await y.raw(["rev-parse","--verify",`origin/${e}`])}catch{await y.fetch("origin",e)}let i=e==="HEAD"?"HEAD":`origin/${e}`;o=`${(await y.raw(["merge-base",`origin/${n}`,i])).trim()}..${i}`}catch(i){h.warn(`Error setting up commit range in CI: ${i}`);let a=e==="HEAD"?"HEAD":`origin/${e}`;o=`origin/${n}..${a}`}}else try{try{await y.raw(["rev-parse","--verify",`origin/${n}`])}catch{h.info(`Fetching ${n} from origin`),await y.fetch("origin",n)}if(e!=="HEAD")try{await y.raw(["rev-parse","--verify",e])}catch{await y.fetch("origin",e)}o=`${(await y.raw(["merge-base",`origin/${n}`,e])).trim()}..${e}`}catch(i){h.warn(`Error determining commits locally: ${i}`),o=`${e}~10..${e}`}return await r.getFormattedCommits(o)}catch(e){return h.error(`Error getting commit messages: ${e}`),[]}}static async getRepoOwnerAndName(){try{let t=await y.getRemotes(!0);if(t.length===0)return h.warn("No remotes found"),null;let e=t.find(o=>o.name==="origin")||t[0],n=e.refs.fetch||e.refs.push;return r.parseGitUrl(n)}catch(t){return h.error(`Error getting repo owner and name: ${t}`),null}}static async getCurrentBranchName(){try{return(await y.branchLocal()).current||null}catch(t){return h.error(`Error getting current branch name: ${t}`),null}}static parseGitUrl(t){let e;return t.includes("@")&&t.includes(":")&&(e=t.match(/@([^:]+):([^/]+)\/(.+?)(?:\.git)?$/),e)?{owner:e[2],repo:e[3]}:(e=t.match(/https?:\/\/(?:[^@]+@)?[^/]+\/([^/]+)\/(.+?)(?:\.git)?$/),e?{owner:e[1],repo:e[2]}:(h.warn(`Could not parse git URL: ${t}`),null))}static async getFormattedCommits(t){let n=(await y.raw(["log","--pretty=format:COMMIT_START%n%H%n%ad%n%an%n%s%n%b%nCOMMIT_END","--date=short",t])).split(`COMMIT_START
43
43
  `).filter(s=>s.trim()!==""),o=[];for(let s of n){let i=s.split(`
44
- `);if(i.length>=4){let a=i[0],l=i[1],m=i[2],u=i[3],c=i.slice(4).filter(b=>b!=="COMMIT_END"),d=c.length>0?c.join(""):"",p=a.substring(0,7),f=u.match(/^(feat|fix|docs|style|refactor|perf|test|build|ci|chore)(\(.+\))?:/),g=f?f[1]:"unknown";d.includes("BREAKING CHANGE")&&(g="breaking"),o.push({type:g,hash:p,date:l,subject:u,body:d,message:`${u}${d?`
44
+ `);if(i.length>=4){let a=i[0],l=i[1],d=i[2],m=i[3],c=i.slice(4).filter(b=>b!=="COMMIT_END"),u=c.length>0?c.join(""):"",f=a.substring(0,7),p=m.match(/^(feat|fix|docs|style|refactor|perf|test|build|ci|chore)(\(.+\))?:/),g=p?p[1]:"unknown";u.includes("BREAKING CHANGE")&&(g="breaking"),o.push({type:g,hash:f,date:l,subject:m,body:u,message:`${m}${u?`
45
45
 
46
- ${d}`:""}`,author_name:m})}}return o}static async addToGitignore(t,e=!0){await r.checkGitignore(t).then(async n=>{if(n)h.warn(`${t} already exists`);else try{await cn(".gitignore",`
47
- ${t}`),e&&h.success(`${t} added to .gitignore`)}catch(o){h.error(o)}})}static async removeFromGitCache(t){try{await y.rmKeepLocal(t),h.success(`${t.join(", ")} are removed from git cache`)}catch(e){h.error(e)}}static async createBranch(t){try{await y.checkoutLocalBranch(t),h.success(`Created and switched to ${t} branch`)}catch(e){h.error(e)}}static async commitStagedFiles(t){try{await y.add("."),await y.commit(t),h.success(`Committed with message: ${t}`)}catch(e){h.error(e)}}};import w from"chalk";var un=60,dn=100;function z(r){switch(r){case"breaking":return w.bold.red(r);case"feat":return w.bold.green(r);case"fix":return w.bold.green(r);case"perf":return w.bold.green(r);case"docs":return w.bold.cyan(r);case"style":return w.bold.cyan(r);case"refactor":return w.bold.cyan(r);case"test":return w.bold.cyan(r);case"build":return w.bold.cyan(r);case"ci":return w.bold.cyan(r);case"chore":return w.bold.cyan(r);default:return w.bold.white(r)}}function It(r,t){if(!r)return"";if(r.length<=t)return r;let e=r.split(" "),n="",o="";for(let s of e)(o+s).length>t&&o.length>0&&(n+=`${o.trim()}
48
- `,o=""),o=`${o}${s} `;return o.length>0&&(n+=o.trim()),n}function Rt(r){for(let t of r){console.log("\u2500".repeat(60));let e=It(t.subject,un),n=It(t.body,dn);console.log(w.bold(`${z(t.type)}`)),console.log(w.dim(`${t.hash} | ${t.date} | ${t.author_name}`)),console.log(w.bold(`${w.white(e)}`)),t.body&&console.log(w.dim(n))}console.log("\u2500".repeat(60)),console.log(`
49
- `)}import de from"@actions/github";async function Pe(){try{let r=process.env.GITHUB_TOKEN;if(!r)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let t=de.getOctokit(r),{context:e}=de;if(!e.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[n,o]=process.env.GITHUB_REPOSITORY.split("/"),s=e.payload.pull_request.number,{data:i}=await t.rest.issues.listLabelsOnIssue({owner:n,repo:o,issue_number:s});return i.map(a=>a.name)}catch(r){throw r instanceof Error?new Error(`Failed to get existing labels: ${r.message}`):r}}async function St(r){try{let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let e=de.getOctokit(t),{context:n}=de;if(!n.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[o,s]=process.env.GITHUB_REPOSITORY.split("/"),i=n.payload.pull_request.number,a=`semantic-status: ${r}`,l=await Pe();if(l.includes(a))return;let m=l.filter(u=>u.startsWith("semantic-status:")&&u!==a);for(let u of m)await e.rest.issues.removeLabel({owner:o,repo:s,issue_number:i,name:u});await e.rest.issues.addLabels({owner:o,repo:s,issue_number:i,labels:[a]});return}catch(t){throw t instanceof Error?new Error(`Failed to apply label: ${t.message}`):t}}var pn=["feat","fix","breaking","perf"];function xe(r,t=!0){let e=`### In this release
46
+ ${u}`:""}`,author_name:d})}}return o}static async addToGitignore(t,e=!0){await r.checkGitignore(t).then(async n=>{if(n)h.warn(`${t} already exists`);else try{await ln(".gitignore",`
47
+ ${t}`),e&&h.success(`${t} added to .gitignore`)}catch(o){h.error(o)}})}static async removeFromGitCache(t){try{await y.rmKeepLocal(t),h.success(`${t.join(", ")} are removed from git cache`)}catch(e){h.error(e)}}static async createBranch(t){try{await y.checkoutLocalBranch(t),h.success(`Created and switched to ${t} branch`)}catch(e){h.error(e)}}static async commitStagedFiles(t){try{await y.add("."),await y.commit(t),h.success(`Committed with message: ${t}`)}catch(e){h.error(e)}}};import w from"chalk";var dn=60,pn=100;function q(r){switch(r){case"breaking":return w.bold.red(r);case"feat":return w.bold.green(r);case"fix":return w.bold.green(r);case"perf":return w.bold.green(r);case"docs":return w.bold.cyan(r);case"style":return w.bold.cyan(r);case"refactor":return w.bold.cyan(r);case"test":return w.bold.cyan(r);case"build":return w.bold.cyan(r);case"ci":return w.bold.cyan(r);case"chore":return w.bold.cyan(r);default:return w.bold.white(r)}}function St(r,t){if(!r)return"";if(r.length<=t)return r;let e=r.split(" "),n="",o="";for(let s of e)(o+s).length>t&&o.length>0&&(n+=`${o.trim()}
48
+ `,o=""),o=`${o}${s} `;return o.length>0&&(n+=o.trim()),n}function Dt(r){for(let t of r){console.log("\u2500".repeat(60));let e=St(t.subject,dn),n=St(t.body,pn);console.log(w.bold(`${q(t.type)}`)),console.log(w.dim(`${t.hash} | ${t.date} | ${t.author_name}`)),console.log(w.bold(`${w.white(e)}`)),t.body&&console.log(w.dim(n))}console.log("\u2500".repeat(60)),console.log(`
49
+ `)}import pe from"@actions/github";async function Pe(){try{let r=process.env.GITHUB_TOKEN;if(!r)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let t=pe.getOctokit(r),{context:e}=pe;if(!e.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[n,o]=process.env.GITHUB_REPOSITORY.split("/"),s=e.payload.pull_request.number,{data:i}=await t.rest.issues.listLabelsOnIssue({owner:n,repo:o,issue_number:s});return i.map(a=>a.name)}catch(r){throw r instanceof Error?new Error(`Failed to get existing labels: ${r.message}`):r}}async function xt(r){try{let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let e=pe.getOctokit(t),{context:n}=pe;if(!n.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[o,s]=process.env.GITHUB_REPOSITORY.split("/"),i=n.payload.pull_request.number,a=`semantic-status: ${r}`,l=await Pe();if(l.includes(a))return;let d=l.filter(m=>m.startsWith("semantic-status:")&&m!==a);for(let m of d)await e.rest.issues.removeLabel({owner:o,repo:s,issue_number:i,name:m});await e.rest.issues.addLabels({owner:o,repo:s,issue_number:i,labels:[a]});return}catch(t){throw t instanceof Error?new Error(`Failed to apply label: ${t.message}`):t}}var fn=["feat","fix","breaking","perf"];function Oe(r,t=!0){let e=`### In this release
50
50
  `;for(let n of r)if(e+=`- ${n.hash} ${n.subject}
51
- `,n.body?.trim()){let s=n.body.trim().split(/\n+/).map(i=>i.trim()).filter(i=>i.length>0);for(let i of s){let a=i;a=a.replace(/([^\s])(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)/g,"$1 $2"),a=a.replace(/(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)([^\s])/g,"$1 $2"),e+=` - ${a}`}}return r.length===0?"":(t&&console.log(Dt.green(`\u2713 Generating release notes for ${r.length} commits`)),e)}function Oe(r,t=!0){let e=r.filter(o=>pn.includes(o.type)),n;return t?n=e.length>0?e:r:n=e,n.length===0&&console.log(`No commits found to include in release notes.
52
- `),n}async function xt(r=!1,t=!1,e=!1){let n=Pt(`Checking commits...
53
- `).start();try{let o=await I.getCommitMessages();if(e){n.succeed(`Total commits analyzed: ${o.length}`);let s=Oe(o);console.log(xe(s));return}if(r&&Rt(o),n.succeed(`Total commits analyzed: ${o.length}`),o.length!==0){let s=o.map(l=>l.type),a=Array.from(new Set(s)).map(l=>z(l)).join(", ");n.succeed(`Found commit types: ${a}`)}else n.info(`The list of commits is created by comparing the current branch
51
+ `,n.body?.trim()){let s=n.body.trim().split(/\n+/).map(i=>i.trim()).filter(i=>i.length>0);for(let i of s){let a=i;a=a.replace(/([^\s])(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)/g,"$1 $2"),a=a.replace(/(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)([^\s])/g,"$1 $2"),e+=` - ${a}`}}return r.length===0?"":(t&&console.log(Pt.green(`\u2713 Generating release notes for ${r.length} commits`)),e)}function Ne(r,t=!0){let e=r.filter(o=>fn.includes(o.type)),n;return t?n=e.length>0?e:r:n=e,n.length===0&&console.log(`No commits found to include in release notes.
52
+ `),n}async function Nt(r=!1,t=!1,e=!1){let n=Ot(`Checking commits...
53
+ `).start();try{let o=await I.getCommitMessages();if(e){n.succeed(`Total commits analyzed: ${o.length}`);let s=Ne(o);console.log(Oe(s));return}if(r&&Dt(o),n.succeed(`Total commits analyzed: ${o.length}`),o.length!==0){let s=o.map(l=>l.type),a=Array.from(new Set(s)).map(l=>q(l)).join(", ");n.succeed(`Found commit types: ${a}`)}else n.info(`The list of commits is created by comparing the current branch
54
54
  with the main branch. If you are on a new branch, please
55
- make sure to commit some changes before running this command.`);t&&await fn(o,n)}catch(o){n.fail("Error getting commit messages"),console.error(o)}}async function fn(r,t){let e=["breaking","feat","fix","perf","docs","style","refactor","test","build","ci","chore"],n=r.map(i=>i.type).filter(i=>e.includes(i)),o=null,s=Number.POSITIVE_INFINITY;for(let i of n){let a=e.indexOf(i);a<s&&(s=a,o=i)}if(o){let i=Pt("Checking existing labels on pull request...").start();try{if((await Pe()).includes(`semantic-status: ${o}`)){i.info(`Label "semantic-status: ${z(o)}" already exists on the pull request.`);return}i.text="Applying label to pull request...",await St(o),i.succeed(`Label "semantic-status: ${z(o)}" applied to the pull request.`)}catch(a){let l=a instanceof Error?a.message:String(a);i.fail(l)}}else t.warn(Dt.yellow("No semantic commit type found to apply as label."))}var Oi=gn.command("check-commits").alias("cc").option("-l, --set-label","Set label on the pull request based on the commit message type").option("-d, --debug","Display detailed commit information for debugging").option("-r, --release-notes","Generate release notes based on commit messages").description("Check commits in the local repository for the types of semantic commit messages made and return the results.").action(async r=>{await xt(r.debug,r.setLabel,r.releaseNotes)});import Ot from"node:fs";import{get as Nt}from"node:https";import S from"chalk";import{program as hn}from"commander";import yn from"ora";var Mi=hn.command("pr-release").option("-n, --namespace <package-namespace>","Set namespace of the package release","@aurodesignsystem-dev").option("-p, --pr-number <number>","Set pull request number for the release","0").description("Generate the package version based off of PR number then update the package.json file. Note: this does not publish the package.").action(async r=>{await wn(r)}),wn=async r=>{let{namespace:t,prNumber:e}=r,n=yn("Updating package.json").start();try{let o="package.json",s=JSON.parse(Ot.readFileSync(o,"utf8"));n.text="Checking npm registry for version information...";let i=`0.0.0-pr${e}`,a=s.name.split("/")[1],l=`${t}/${a}`,m=await bn(i,l,n),u=`${i}.${m}`;s.name=l,s.version=u,n.text="Writing updated package.json...",Ot.writeFileSync(o,`${JSON.stringify(s,null,2)}
56
- `,"utf8"),n.succeed(`Package.json updated to use ${S.green(u)} and ${S.green(l)}`),process.exit(0)}catch(o){n.fail(`Failed to update package.json: ${o}`),process.exit(1)}},bn=(r,t,e)=>new Promise(n=>{try{let a=function(l){if(l.statusCode!==200){e.info(`Package not found. Status code: ${S.red(l.statusCode)}, defaulting to version 0`),n(0);return}e.text="Processing version information...";let m="";l.on("data",u=>{m+=u}),l.on("end",()=>{try{let u=JSON.parse(m),c=u.versions?Object.keys(u.versions):[];e.text="Calculating next version number...";let d=-1,p=new RegExp(`^${r}\\.(\\d+)$`);for(let f of c){let g=f.match(p);if(g){let b=Number.parseInt(g[1],10);d=Math.max(d,b)}}d>=0?e.info(`Found existing version ${S.green(`${r}.${d}`)}. Incrementing to ${S.green(`${r}.${d+1}`)}`):e.info(`No existing version found for ${S.green(r)}. Starting with ${S.green(`${r}.0`)}`),n(d+1)}catch(u){e.warn(`Failed to parse NPM registry response: ${u instanceof Error?u.message:"Unknown error"}, defaulting to version 0`),n(0)}})};var o=a;let s=`https://registry.npmjs.org/${t}`,i=Nt(s,{headers:{Accept:"application/json"}},l=>{if((l.statusCode===301||l.statusCode===302)&&l.headers.location){e.info(`Following redirect to ${l.headers.location}...`);try{Nt(l.headers.location,{headers:{Accept:"application/json"}},a).on("error",m=>{e.warn(`Error following redirect: ${m.message}, defaulting to version 0`),n(0)}).end()}catch(m){e.warn(`Redirect request failed: ${m instanceof Error?m.message:"Unknown error"}, defaulting to version 0`),n(0)}return}a(l)});i.on("error",l=>{e.warn(`Request error: ${l.message}, defaulting to version 0`),n(0)}),i.end()}catch{e.warn("Error checking version in npm registry, defaulting to version 0"),n(0)}});import Ne from"node:path";import{fileURLToPath as kn}from"node:url";import{program as $n}from"commander";import En from"open";var An=kn(import.meta.url),vn=Ne.resolve(Ne.dirname(An),".."),Vi=$n.command("test").option("-w, --watch","Set watch number for the test").option("-c, --coverage-report","Generate coverage report").option("-o, --open","Open the coverage report in the browser").option("-f, --files <String|String[]>","Test files glob pattern").description("Run the web test runner to test the component library").action(async r=>{let e=`npx wtr --config "${Ne.join(vn,"dist","configs","web-test-runner.config.mjs")}"`,n=`${process.cwd()}/coverage/index.html`;if(r.coverageReport&&(e+=" --coverage"),r.watch&&(e+=" --watch"),r.files){let o=Array.isArray(r.files)?r.files.join(" "):r.files;e+=` --files "${o}"`}A(e),r.open&&await En(n)});import Be from"node:fs/promises";import Bt from"node:path";import{program as Pn}from"commander";import xn from"inquirer";import Ge from"ora";import Tn from"node:fs/promises";import Cn from"node:path";import In from"ora";var Rn={"auth-type":"workspace-token","author-email":null,"author-name":null,"base-branch":"main","base-url":null,"clone-dir":".gitter-temp","code-search":null,concurrent:4,"conflict-strategy":"replace",draft:!1,"dry-run":!0,"fetch-depth":1,fork:!1,"fork-owner":null,"git-type":"go",group:null,"include-subgroups":!1,insecure:!1,interactive:!1,labels:null,"log-file":"'-'","log-format":"'text'","log-level":"'error'","max-reviewers":0,"max-team-reviewers":0,org:null,output:"'-'","plain-output":!1,platform:"github",project:null,"push-only":!1,repo:["AlaskaAirlines/auro-accordion","AlaskaAirlines/auro-alert","AlaskaAirlines/auro-avatar","AlaskaAirlines/auro-background","AlaskaAirlines/auro-backtotop","AlaskaAirlines/auro-button","AlaskaAirlines/auro-badge","AlaskaAirlines/auro-banner","AlaskaAirlines/auro-card","AlaskaAirlines/auro-carousel","AlaskaAirlines/auro-datetime","AlaskaAirlines/auro-dialog","AlaskaAirlines/auro-drawer","AlaskaAirlines/auro-flight","AlaskaAirlines/auro-flightline","AlaskaAirlines/auro-header","AlaskaAirlines/auro-hyperlink","AlaskaAirlines/auro-icon","AlaskaAirlines/auro-loader","AlaskaAirlines/auro-lockup","AlaskaAirlines/auro-nav","AlaskaAirlines/auro-pane","AlaskaAirlines/auro-popover","AlaskaAirlines/auro-sidenav","AlaskaAirlines/auro-skeleton","AlaskaAirlines/auro-slideshow","AlaskaAirlines/auro-table","AlaskaAirlines/auro-tabs","AlaskaAirlines/auro-toast"],"repo-exclude":null,"repo-include":null,"repo-search":null,reviewers:null,"skip-forks":!1,"skip-pr":!1,"skip-repo":null,"ssh-auth":!1,"team-reviewers":null};function Sn(r){return Object.entries(r).map(([t,e])=>Array.isArray(e)?`${t}:
55
+ make sure to commit some changes before running this command.`);t&&await gn(o,n)}catch(o){n.fail("Error getting commit messages"),console.error(o)}}async function gn(r,t){let e=["breaking","feat","fix","perf","docs","style","refactor","test","build","ci","chore"],n=r.map(i=>i.type).filter(i=>e.includes(i)),o=null,s=Number.POSITIVE_INFINITY;for(let i of n){let a=e.indexOf(i);a<s&&(s=a,o=i)}if(o){let i=Ot("Checking existing labels on pull request...").start();try{if((await Pe()).includes(`semantic-status: ${o}`)){i.info(`Label "semantic-status: ${q(o)}" already exists on the pull request.`);return}i.text="Applying label to pull request...",await xt(o),i.succeed(`Label "semantic-status: ${q(o)}" applied to the pull request.`)}catch(a){let l=a instanceof Error?a.message:String(a);i.fail(l)}}else t.warn(Pt.yellow("No semantic commit type found to apply as label."))}var Ni=hn.command("check-commits").alias("cc").option("-l, --set-label","Set label on the pull request based on the commit message type").option("-d, --debug","Display detailed commit information for debugging").option("-r, --release-notes","Generate release notes based on commit messages").description("Check commits in the local repository for the types of semantic commit messages made and return the results.").action(async r=>{await Nt(r.debug,r.setLabel,r.releaseNotes)});import jt from"node:fs";import{get as _t}from"node:https";import S from"chalk";import{program as yn}from"commander";import wn from"ora";var Hi=yn.command("pr-release").option("-n, --namespace <package-namespace>","Set namespace of the package release","@aurodesignsystem-dev").option("-p, --pr-number <number>","Set pull request number for the release","0").description("Generate the package version based off of PR number then update the package.json file. Note: this does not publish the package.").action(async r=>{await bn(r)}),bn=async r=>{let{namespace:t,prNumber:e}=r,n=wn("Updating package.json").start();try{let o="package.json",s=JSON.parse(jt.readFileSync(o,"utf8"));n.text="Checking npm registry for version information...";let i=`0.0.0-pr${e}`,a=s.name.split("/")[1],l=`${t}/${a}`,d=await kn(i,l,n),m=`${i}.${d}`;s.name=l,s.version=m,n.text="Writing updated package.json...",jt.writeFileSync(o,`${JSON.stringify(s,null,2)}
56
+ `,"utf8"),n.succeed(`Package.json updated to use ${S.green(m)} and ${S.green(l)}`),process.exit(0)}catch(o){n.fail(`Failed to update package.json: ${o}`),process.exit(1)}},kn=(r,t,e)=>new Promise(n=>{try{let a=function(l){if(l.statusCode!==200){e.info(`Package not found. Status code: ${S.red(l.statusCode)}, defaulting to version 0`),n(0);return}e.text="Processing version information...";let d="";l.on("data",m=>{d+=m}),l.on("end",()=>{try{let m=JSON.parse(d),c=m.versions?Object.keys(m.versions):[];e.text="Calculating next version number...";let u=-1,f=new RegExp(`^${r}\\.(\\d+)$`);for(let p of c){let g=p.match(f);if(g){let b=Number.parseInt(g[1],10);u=Math.max(u,b)}}u>=0?e.info(`Found existing version ${S.green(`${r}.${u}`)}. Incrementing to ${S.green(`${r}.${u+1}`)}`):e.info(`No existing version found for ${S.green(r)}. Starting with ${S.green(`${r}.0`)}`),n(u+1)}catch(m){e.warn(`Failed to parse NPM registry response: ${m instanceof Error?m.message:"Unknown error"}, defaulting to version 0`),n(0)}})};var o=a;let s=`https://registry.npmjs.org/${t}`,i=_t(s,{headers:{Accept:"application/json"}},l=>{if((l.statusCode===301||l.statusCode===302)&&l.headers.location){e.info(`Following redirect to ${l.headers.location}...`);try{_t(l.headers.location,{headers:{Accept:"application/json"}},a).on("error",d=>{e.warn(`Error following redirect: ${d.message}, defaulting to version 0`),n(0)}).end()}catch(d){e.warn(`Redirect request failed: ${d instanceof Error?d.message:"Unknown error"}, defaulting to version 0`),n(0)}return}a(l)});i.on("error",l=>{e.warn(`Request error: ${l.message}, defaulting to version 0`),n(0)}),i.end()}catch{e.warn("Error checking version in npm registry, defaulting to version 0"),n(0)}});import je from"node:path";import{fileURLToPath as $n}from"node:url";import{program as En}from"commander";import An from"open";var vn=$n(import.meta.url),Tn=je.resolve(je.dirname(vn),".."),Ji=En.command("test").option("-w, --watch","Set watch number for the test").option("-c, --coverage-report","Generate coverage report").option("-o, --open","Open the coverage report in the browser").option("-f, --files <String|String[]>","Test files glob pattern").description("Run the web test runner to test the component library").action(async r=>{let e=`npx wtr --config "${je.join(Tn,"dist","configs","web-test-runner.config.mjs")}"`,n=`${process.cwd()}/coverage/index.html`;if(r.coverageReport&&(e+=" --coverage"),r.watch&&(e+=" --watch"),r.files){let o=Array.isArray(r.files)?r.files.join(" "):r.files;e+=` --files "${o}"`}A(e),r.open&&await An(n)});import Ge from"node:fs/promises";import Mt from"node:path";import{program as Pn}from"commander";import On from"inquirer";import Me from"ora";import Cn from"node:fs/promises";import In from"node:path";import Rn from"ora";var Sn={"auth-type":"workspace-token","author-email":null,"author-name":null,"base-branch":"main","base-url":null,"clone-dir":".gitter-temp","code-search":null,concurrent:4,"conflict-strategy":"replace",draft:!1,"dry-run":!0,"fetch-depth":1,fork:!1,"fork-owner":null,"git-type":"go",group:null,"include-subgroups":!1,insecure:!1,interactive:!1,labels:null,"log-file":"'-'","log-format":"'text'","log-level":"'error'","max-reviewers":0,"max-team-reviewers":0,org:null,output:"'-'","plain-output":!1,platform:"github",project:null,"push-only":!1,repo:["AlaskaAirlines/auro-accordion","AlaskaAirlines/auro-alert","AlaskaAirlines/auro-avatar","AlaskaAirlines/auro-background","AlaskaAirlines/auro-backtotop","AlaskaAirlines/auro-button","AlaskaAirlines/auro-badge","AlaskaAirlines/auro-banner","AlaskaAirlines/auro-card","AlaskaAirlines/auro-carousel","AlaskaAirlines/auro-datetime","AlaskaAirlines/auro-dialog","AlaskaAirlines/auro-drawer","AlaskaAirlines/auro-flight","AlaskaAirlines/auro-flightline","AlaskaAirlines/auro-header","AlaskaAirlines/auro-hyperlink","AlaskaAirlines/auro-icon","AlaskaAirlines/auro-loader","AlaskaAirlines/auro-lockup","AlaskaAirlines/auro-nav","AlaskaAirlines/auro-pane","AlaskaAirlines/auro-popover","AlaskaAirlines/auro-sidenav","AlaskaAirlines/auro-skeleton","AlaskaAirlines/auro-slideshow","AlaskaAirlines/auro-table","AlaskaAirlines/auro-tabs","AlaskaAirlines/auro-toast"],"repo-exclude":null,"repo-include":null,"repo-search":null,reviewers:null,"skip-forks":!1,"skip-pr":!1,"skip-repo":null,"ssh-auth":!1,"team-reviewers":null};function Dn(r){return Object.entries(r).map(([t,e])=>Array.isArray(e)?`${t}:
57
57
  - ${e.join(`
58
58
  - `)}`:typeof e=="object"&&e!==null?`${t}:
59
59
  ${Object.entries(e).map(([n,o])=>` ${n}: ${o}`).join(`
60
60
  `)}`:`${t}: ${e}`).join(`
61
- `)}async function jt(r){let t=In("Writing multi-gitter configuration...").start(),e=Sn(Rn),n=Cn.join(r,"multi-gitter_DEPENDENCY_TREE.yml");try{await Tn.writeFile(n,e,"utf8"),t.succeed(`Multi-gitter configuration written to ${n}`)}catch(o){t.fail("Error writing multi-gitter configuration:"),console.error(o)}}import je from"node:fs";function _t(r){let t={},e=[],n=[],o=[];for(let s in r)t[s]=r[s].dependsOn.length;for(let s in t)t[s]===0&&o.push(s);for(;o.length>0;){n=[];let s=o.length;for(let i=0;i<s;i++){let a=o.shift();n.push(a);for(let l of r[a].dependentPackages)t[l]--,t[l]===0&&o.push(l)}e.push(n)}if(e.flat().length!==Object.keys(r).length)throw new Error("Circular dependency detected!");return e}function Dn(r){return je.readdirSync(r).filter(t=>t.endsWith(".json"))}async function Ft(r,t=[]){console.log(t);let e={},n=Dn(r);for(let o of n){if(o==="dependencyTree.json")continue;let s=je.readFileSync(`${r}/${o}`,"utf-8"),i=JSON.parse(s),a=i.name,l=Object.keys(i.peerDependencies),m=Object.keys(i.devDependencies),u=Object.keys(i.dependencies);e[a]||(e[a]={dependsOn:[],dependentPackages:[]});let c=[...l,...m,...u];e[a].dependsOn=[...new Set(c)];for(let d of c)e[d]||(e[d]={dependsOn:[],dependentPackages:[]}),e[d].dependentPackages.includes(a)||e[d].dependentPackages.push(a)}if(t.length){let o=new Set;for(let[i,a]of Object.entries(e))a.dependsOn.some(l=>t.includes(l))&&o.add(i);for(let i of t)e[i]&&o.add(i);let s={};for(let i of o)s[i]={dependsOn:e[i].dependsOn.filter(a=>o.has(a)),dependentPackages:e[i].dependentPackages.filter(a=>o.has(a))};e=s}else console.log("No target dependencies provided - using all packages.");return je.writeFileSync(`${r}/dependencyTree.json`,JSON.stringify(e,null,2)),e}var _e=we("run-migrations","config"),Fe=we("run-migrations","outputs");var Gt=["@aurodesignsystem/auro-accordion","@aurodesignsystem/auro-alert","@aurodesignsystem/auro-avatar","@aurodesignsystem/auro-background","@aurodesignsystem/auro-backtotop","@aurodesignsystem/auro-button","@aurodesignsystem/auro-badge","@aurodesignsystem/auro-banner","@aurodesignsystem/auro-card","@aurodesignsystem/auro-carousel","@aurodesignsystem/auro-datetime","@aurodesignsystem/auro-dialog","@aurodesignsystem/auro-drawer","@aurodesignsystem/auro-formkit","@aurodesignsystem/auro-flight","@aurodesignsystem/auro-flightline","@aurodesignsystem/auro-header","@aurodesignsystem/auro-hyperlink","@aurodesignsystem/auro-icon","@aurodesignsystem/auro-loader","@aurodesignsystem/auro-lockup","@aurodesignsystem/auro-nav","@aurodesignsystem/auro-pane","@aurodesignsystem/auro-popover","@aurodesignsystem/auro-sidenav","@aurodesignsystem/auro-skeleton","@aurodesignsystem/auro-slideshow","@aurodesignsystem/auro-table","@aurodesignsystem/auro-tabs","@aurodesignsystem/auro-toast"],ma=[...Gt,"@aurodesignsystem/auro-library","@aurodesignsystem/WebCoreStyleSheets","@aurodesignsystem/AuroDesignTokens","@aurodesignsystem/auro-cli","@alaskaairux/icons"];async function On(r){try{await Be.mkdir(Fe,{recursive:!0}),await Be.mkdir(_e,{recursive:!0})}catch(o){console.error("Failed to create output or config directories:",o),process.exit(1)}let t=Ge("Creating dependency tree...").start();t.text="Creating multi-gitter dependency tree configuration...",await jt(_e),t.text="Scraping dependencies from Auro packages...";let n=`multi-gitter run "node ${be("static","getAuroDeps.js")}" --config ${Bt.join(_e,"multi-gitter_DEPENDENCY_TREE.yml")}`;try{await A(n)}catch(o){t.fail("Failed to generate dependency tree:"),console.error(o),process.exit(1)}return t.text="Generating dependency tree JSON file using packages...",await Ft(Fe,r),t.succeed("Dependency tree generated successfully."),Bt.join(Fe,"dependencyTree.json")}var Nn=async r=>{let t=Ge("Loading dependency tree...").start(),e=JSON.parse(await Be.readFile(r,"utf-8"));t.text="Processing dependency tree...";let n=_t(e);return t.succeed("Dependency batches created successfully."),n},ua=Pn.command("agent").action(async r=>{let t=await xn.prompt([{type:"select",name:"agentAction",message:"What agent action do you want to perform?",choices:[{name:"Run a migration on auro components",value:"run-migration"}],default:["run-migration"]},{type:"input",name:"migrationId",message:"What migration id do you want to run?",when:e=>e.agentAction==="run-migration",validate:e=>e.trim()!==""||"Migration id cannot be empty."},{type:"confirm",name:"useExisting",message:"Would you like to specify starting packages?",default:!0,transformer:e=>e?"Yes = Packages related to selections":"No = All packages",when:e=>e.agentAction==="run-migration"},{type:"checkbox",name:"startWithComponents",message:"Enter the components to start with (comma-separated, blank for all):",choices:Gt.map(e=>({name:e.replace("@aurodesignsystem/",""),value:e})),when:e=>e.agentAction==="run-migration"&&e.useExisting}]);switch(t.agentAction){case"run-migration":{let e=Ge("Running migration...").start(),n=await On(t.startWithComponents);e.text="Getting dependency batches from tree...";let s=(await Nn(n)).map((i,a)=>`Batch ${a+1}
61
+ `)}async function Ft(r){let t=Rn("Writing multi-gitter configuration...").start(),e=Dn(Sn),n=In.join(r,"multi-gitter_DEPENDENCY_TREE.yml");try{await Cn.writeFile(n,e,"utf8"),t.succeed(`Multi-gitter configuration written to ${n}`)}catch(o){t.fail("Error writing multi-gitter configuration:"),console.error(o)}}import _e from"node:fs";function Bt(r){let t={},e=[],n=[],o=[];for(let s in r)t[s]=r[s].dependsOn.length;for(let s in t)t[s]===0&&o.push(s);for(;o.length>0;){n=[];let s=o.length;for(let i=0;i<s;i++){let a=o.shift();n.push(a);for(let l of r[a].dependentPackages)t[l]--,t[l]===0&&o.push(l)}e.push(n)}if(e.flat().length!==Object.keys(r).length)throw new Error("Circular dependency detected!");return e}function xn(r){return _e.readdirSync(r).filter(t=>t.endsWith(".json"))}async function Gt(r,t=[]){console.log(t);let e={},n=xn(r);for(let o of n){if(o==="dependencyTree.json")continue;let s=_e.readFileSync(`${r}/${o}`,"utf-8"),i=JSON.parse(s),a=i.name,l=Object.keys(i.peerDependencies),d=Object.keys(i.devDependencies),m=Object.keys(i.dependencies);e[a]||(e[a]={dependsOn:[],dependentPackages:[]});let c=[...l,...d,...m];e[a].dependsOn=[...new Set(c)];for(let u of c)e[u]||(e[u]={dependsOn:[],dependentPackages:[]}),e[u].dependentPackages.includes(a)||e[u].dependentPackages.push(a)}if(t.length){let o=new Set;for(let[i,a]of Object.entries(e))a.dependsOn.some(l=>t.includes(l))&&o.add(i);for(let i of t)e[i]&&o.add(i);let s={};for(let i of o)s[i]={dependsOn:e[i].dependsOn.filter(a=>o.has(a)),dependentPackages:e[i].dependentPackages.filter(a=>o.has(a))};e=s}else console.log("No target dependencies provided - using all packages.");return _e.writeFileSync(`${r}/dependencyTree.json`,JSON.stringify(e,null,2)),e}var Fe=be("run-migrations","config"),Be=be("run-migrations","outputs");var Ht=["@aurodesignsystem/auro-accordion","@aurodesignsystem/auro-alert","@aurodesignsystem/auro-avatar","@aurodesignsystem/auro-background","@aurodesignsystem/auro-backtotop","@aurodesignsystem/auro-button","@aurodesignsystem/auro-badge","@aurodesignsystem/auro-banner","@aurodesignsystem/auro-card","@aurodesignsystem/auro-carousel","@aurodesignsystem/auro-datetime","@aurodesignsystem/auro-dialog","@aurodesignsystem/auro-drawer","@aurodesignsystem/auro-formkit","@aurodesignsystem/auro-flight","@aurodesignsystem/auro-flightline","@aurodesignsystem/auro-header","@aurodesignsystem/auro-hyperlink","@aurodesignsystem/auro-icon","@aurodesignsystem/auro-loader","@aurodesignsystem/auro-lockup","@aurodesignsystem/auro-nav","@aurodesignsystem/auro-pane","@aurodesignsystem/auro-popover","@aurodesignsystem/auro-sidenav","@aurodesignsystem/auro-skeleton","@aurodesignsystem/auro-slideshow","@aurodesignsystem/auro-table","@aurodesignsystem/auro-tabs","@aurodesignsystem/auro-toast"],ua=[...Ht,"@aurodesignsystem/auro-library","@aurodesignsystem/WebCoreStyleSheets","@aurodesignsystem/AuroDesignTokens","@aurodesignsystem/auro-cli","@alaskaairux/icons"];async function Nn(r){try{await Ge.mkdir(Be,{recursive:!0}),await Ge.mkdir(Fe,{recursive:!0})}catch(o){console.error("Failed to create output or config directories:",o),process.exit(1)}let t=Me("Creating dependency tree...").start();t.text="Creating multi-gitter dependency tree configuration...",await Ft(Fe),t.text="Scraping dependencies from Auro packages...";let n=`multi-gitter run "node ${ke("static","getAuroDeps.js")}" --config ${Mt.join(Fe,"multi-gitter_DEPENDENCY_TREE.yml")}`;try{await A(n)}catch(o){t.fail("Failed to generate dependency tree:"),console.error(o),process.exit(1)}return t.text="Generating dependency tree JSON file using packages...",await Gt(Be,r),t.succeed("Dependency tree generated successfully."),Mt.join(Be,"dependencyTree.json")}var jn=async r=>{let t=Me("Loading dependency tree...").start(),e=JSON.parse(await Ge.readFile(r,"utf-8"));t.text="Processing dependency tree...";let n=Bt(e);return t.succeed("Dependency batches created successfully."),n},da=Pn.command("agent").action(async r=>{let t=await On.prompt([{type:"select",name:"agentAction",message:"What agent action do you want to perform?",choices:[{name:"Run a migration on auro components",value:"run-migration"}],default:["run-migration"]},{type:"input",name:"migrationId",message:"What migration id do you want to run?",when:e=>e.agentAction==="run-migration",validate:e=>e.trim()!==""||"Migration id cannot be empty."},{type:"confirm",name:"useExisting",message:"Would you like to specify starting packages?",default:!0,transformer:e=>e?"Yes = Packages related to selections":"No = All packages",when:e=>e.agentAction==="run-migration"},{type:"checkbox",name:"startWithComponents",message:"Enter the components to start with (comma-separated, blank for all):",choices:Ht.map(e=>({name:e.replace("@aurodesignsystem/",""),value:e})),when:e=>e.agentAction==="run-migration"&&e.useExisting}]);switch(t.agentAction){case"run-migration":{let e=Me("Running migration...").start(),n=await Nn(t.startWithComponents);e.text="Getting dependency batches from tree...";let s=(await jn(n)).map((i,a)=>`Batch ${a+1}
62
62
  ${i.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alaskaairux/icons","AlaskaAirlines/Icons")}`).join(`
63
63
  `)}`).join(`
64
64
 
65
- `);console.log(s),e.text="Running migrations on dependency batches...",new Promise(i=>setTimeout(i,2e3)),e.succeed("Migration process completed successfully.");break}default:console.error("Unknown action selected.")}});import{program as jn}from"commander";var Me=jn.command("docs").description("Generate API documentation").option("-c, --cem","Generate Custom Elements Manifest (CEM) file",!1).option("-a, --api","Creates api md file from CEM",!1).option("-w, --watch","Watch for changes and rebuild docs",!1).option("-r, --readme-template <url>","URL to the README template file").option("--skip-readme","Skip README.md processing",!1);Me=Y(Me);var ha=Me.action(async r=>{r.cem&&await re(),r.api&&await ne(),await rt(r),r.serve&&await nt(r),r.watch&&await ot(r)});import{program as Mn}from"commander";import{Octokit as He}from"@octokit/rest";import*as fe from"azure-devops-node-api";import pe from"ora";var _n=async r=>{let t=process.env.GH_TOKEN;if(!t)throw new Error("GH_TOKEN environment variable is required");let e=new He({auth:t}),n,o,s;if(r.includes("github.com")){let a=r.match(/github\.com\/([^\/]+)\/([^\/]+)\/issues\/(\d+)/);if(!a)throw new Error("Invalid GitHub issue URL format");[,n,o,s]=a}else if(r.includes("#")){let a=r.match(/([^\/]+)\/([^#]+)#(\d+)/);if(!a)throw new Error("Invalid GitHub issue reference format");[,n,o,s]=a}else throw new Error("Issue must be provided as full URL or in format 'owner/repo#number'");let i=Number.parseInt(s,10);try{let{data:a}=await e.rest.issues.get({owner:n,repo:o,issue_number:i});return{title:a.title,body:a.body??null,html_url:a.html_url,number:a.number,repository:{owner:{login:n},name:o}}}catch(a){throw new Error(`Failed to fetch GitHub issue: ${a}`)}},Fn=async r=>{let t=process.env.GH_TOKEN;if(!t)return null;let e=new He({auth:t});try{let n=`
65
+ `);console.log(s),e.text="Running migrations on dependency batches...",new Promise(i=>setTimeout(i,2e3)),e.succeed("Migration process completed successfully.");break}default:console.error("Unknown action selected.")}});import{program as _n}from"commander";var He=_n.command("docs").description("Generate API documentation").option("-c, --cem","Generate Custom Elements Manifest (CEM) file",!1).option("-a, --api","Creates api md file from CEM",!1).option("-w, --watch","Watch for changes and rebuild docs",!1).option("-r, --readme-template <url>","URL to the README template file").option("--skip-readme","Skip README.md processing",!1);He=X(He);var ya=He.action(async r=>{r.cem&&await ne(),r.api&&await oe(),await ot(r),r.serve&&await st(r),r.watch&&await it(r)});import{program as Hn}from"commander";import{Octokit as Le}from"@octokit/rest";import*as ge from"azure-devops-node-api";import fe from"ora";var Fn=async r=>{let t=process.env.GH_TOKEN;if(!t)throw new Error("GH_TOKEN environment variable is required");let e=new Le({auth:t}),n,o,s;if(r.includes("github.com")){let a=r.match(/github\.com\/([^\/]+)\/([^\/]+)\/issues\/(\d+)/);if(!a)throw new Error("Invalid GitHub issue URL format");[,n,o,s]=a}else if(r.includes("#")){let a=r.match(/([^\/]+)\/([^#]+)#(\d+)/);if(!a)throw new Error("Invalid GitHub issue reference format");[,n,o,s]=a}else throw new Error("Issue must be provided as full URL or in format 'owner/repo#number'");let i=Number.parseInt(s,10);try{let{data:a}=await e.rest.issues.get({owner:n,repo:o,issue_number:i});return{title:a.title,body:a.body??null,html_url:a.html_url,number:a.number,repository:{owner:{login:n},name:o}}}catch(a){throw new Error(`Failed to fetch GitHub issue: ${a}`)}},Bn=async r=>{let t=process.env.GH_TOKEN;if(!t)return null;let e=new Le({auth:t});try{let n=`
66
66
  query($owner: String!, $repo: String!, $issueNumber: Int!) {
67
67
  repository(owner: $owner, name: $repo) {
68
68
  issue(number: $issueNumber) {
@@ -88,7 +88,7 @@ ${i.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
88
88
  }
89
89
  }
90
90
  }
91
- `,o={owner:r.repository.owner.login,repo:r.repository.name,issueNumber:r.number},i=(await e.graphql(n,o)).repository.issue.projectItems.nodes.find(a=>a.project.number===19);if(i){let a=i.fieldValues.nodes.find(l=>l.field?.name?.toLowerCase()==="ado"&&l.text?.trim());if(a?.text?.trim())return a.text.trim()}return null}catch(n){return console.error(`Failed to check existing ADO link: ${n}`),null}},Bn=async(r,t)=>{let e=process.env.GH_TOKEN;if(!e)throw new Error("GH_TOKEN environment variable is required");let n=new He({auth:e}),o=19;try{let s=`
91
+ `,o={owner:r.repository.owner.login,repo:r.repository.name,issueNumber:r.number},i=(await e.graphql(n,o)).repository.issue.projectItems.nodes.find(a=>a.project.number===19);if(i){let a=i.fieldValues.nodes.find(l=>l.field?.name?.toLowerCase()==="ado"&&l.text?.trim());if(a?.text?.trim())return a.text.trim()}return null}catch(n){return console.error(`Failed to check existing ADO link: ${n}`),null}},Gn=async(r,t)=>{let e=process.env.GH_TOKEN;if(!e)throw new Error("GH_TOKEN environment variable is required");let n=new Le({auth:e}),o=19;try{let s=`
92
92
  query($org: String!, $projectNumber: Int!, $owner: String!, $repo: String!, $issueNumber: Int!) {
93
93
  organization(login: $org) {
94
94
  projectV2(number: $projectNumber) {
@@ -125,7 +125,7 @@ ${i.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
125
125
  }
126
126
  }
127
127
  }
128
- `,i={org:"AlaskaAirlines",projectNumber:o,owner:r.repository.owner.login,repo:r.repository.name,issueNumber:r.number},a=await n.graphql(s,i),l=a.organization.projectV2.id,m=a.repository.issue.id,u=a.organization.projectV2.fields.nodes.find(d=>d.name?.toLowerCase()==="ado"),c=a.repository.issue.projectItems.nodes.find(d=>d.project.number===o)?.id;if(c||(c=(await n.graphql(`
128
+ `,i={org:"AlaskaAirlines",projectNumber:o,owner:r.repository.owner.login,repo:r.repository.name,issueNumber:r.number},a=await n.graphql(s,i),l=a.organization.projectV2.id,d=a.repository.issue.id,m=a.organization.projectV2.fields.nodes.find(u=>u.name?.toLowerCase()==="ado"),c=a.repository.issue.projectItems.nodes.find(u=>u.project.number===o)?.id;if(c||(c=(await n.graphql(`
129
129
  mutation($projectId: ID!, $contentId: ID!) {
130
130
  addProjectV2ItemById(
131
131
  input: {
@@ -138,7 +138,7 @@ ${i.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
138
138
  }
139
139
  }
140
140
  }
141
- `,{projectId:l,contentId:m})).addProjectV2ItemById.item.id),u&&c)await n.graphql(`
141
+ `,{projectId:l,contentId:d})).addProjectV2ItemById.item.id),m&&c)await n.graphql(`
142
142
  mutation($projectId: ID!, $itemId: ID!, $fieldId: ID!, $value: String!) {
143
143
  updateProjectV2ItemFieldValue(
144
144
  input: {
@@ -155,4 +155,4 @@ ${i.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
155
155
  }
156
156
  }
157
157
  }
158
- `,{projectId:l,itemId:c,fieldId:u.id,value:t});else if(!u)throw new Error("No 'ado' field found in GitHub project")}catch(s){console.error(`Failed to update GitHub project: ${s}`)}},Gn=async r=>{let t=process.env.ADO_TOKEN;if(!t)throw new Error("ADO_TOKEN environment variable is required");let e="https://dev.azure.com/itsals",n="E_Retain_Content",o="E_Retain_Content\\Auro Design System",s=fe.getPersonalAccessTokenHandler(t),a=await new fe.WebApi(e,s).getWorkItemTrackingApi();try{let l=[{op:"add",path:"/fields/System.Title",value:r.title},{op:"add",path:"/fields/System.Description",value:`GitHub Issue: <a href="${r.html_url}">${r.html_url}</a>`},{op:"add",path:"/fields/System.AreaPath",value:o}];return await a.createWorkItem(null,l,n,"User Story")}catch(l){throw new Error(`Failed to create ADO work item: ${l}`)}},Mt=async r=>{let t=pe(`Processing GitHub issue: ${r}`).start();try{if(!process.env.GH_TOKEN)throw new Error("GH_TOKEN environment variable is required");if(!process.env.ADO_TOKEN)throw new Error("ADO_TOKEN environment variable is required");t.text="Fetching GitHub issue details...";let e=await _n(r);t.succeed(`Found issue: "${e.title}"`);let n=pe("Checking for existing ADO work item...").start(),o=await Fn(e);if(o){n.succeed("ADO work item already exists for this issue!"),console.log(`${o}`);return}n.succeed("No existing ADO work item found");let s=pe("Creating new ADO work item...").start(),i=await Gn(e);if(s.succeed(`Successfully created ADO work item #${i.id}`),console.log(`Work item: ${i._links?.html?.href||"N/A"}`),i._links?.html?.href){let a=pe("Adding to GitHub project and updating ADO field...").start();await Bn(e,i._links.html.href),a.succeed("Updated GitHub project with ADO link")}}catch(e){t.fail(`Error: ${e instanceof Error?e.message:e}`),process.exit(1)}};var Aa=Mn.command("ado").description("Generate ADO item from GitHub issue").option("-g, --gh-issue <issue>","What GitHub issue to use").action(async r=>{r.ghIssue&&await Mt(r.ghIssue)});import{program as Un}from"commander";import{Octokit as Hn}from"@octokit/rest";import{simpleGit as Ln}from"simple-git";var Ht="Release Candidate",q="dev",Wn="main",ge=class r{constructor(t,e,n){this.filteredCommits=null;this.repoInfo={owner:t,repo:e},this.octokit=n}static async create(){let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN is required to run RC workflow.");let e=await I.getRepoOwnerAndName(),n=new Hn({auth:t});if(!e)throw new Error("Failed to retrieve repository information. Ensure you're in a valid git repository.");let o=await r.getTriggerBranchName();return o&&o!==q&&(console.log(`Switching from ${o} to ${q} branch...`),await Ln().checkout(q)),new r(e.owner,e.repo,n)}get owner(){return this.repoInfo.owner}get repo(){return this.repoInfo.repo}get repoData(){return{...this.repoInfo}}async createReleaseCandidate(){await this.hasCommitsReadyInDev()||console.log("No filtered commits found. Continuing to update RC issue/branch/PR.");let e=await this.getLatestOpenRcIssue(),n=e?await this.getLinkedPrByHead(e.number):null;if(n?.multipleOpen)throw new Error("Multiple open RC PRs found for the same rc/<issueNumber> branch.");if(n?.state==="closed"&&(console.log("Linked RC PR is closed. Creating a new RC issue and PR."),e=await this.createRcIssue(),n=null),e?await this.updateRcIssue(e.number):e=await this.createRcIssue(),!e)throw new Error("Failed to resolve RC issue.");await this.createOrUpdateRcBranch(e.number),n?await this.updateRcPr(e.number,n.number):n=await this.createRcPr(e.number)}async getFilteredCommits(){if(this.filteredCommits===null){let t=await I.getCommitMessages(q);this.filteredCommits=Oe(t)}return this.filteredCommits}async hasCommitsReadyInDev(){return(await this.getFilteredCommits()).length>0}async getLatestOpenRcIssue(){let{data:t}=await this.octokit.rest.issues.listForRepo({owner:this.repoInfo.owner,repo:this.repoInfo.repo,labels:Ht,state:"open",sort:"updated",direction:"desc",per_page:30}),e=t.filter(o=>!o.pull_request);if(e.length===0)return console.log(`No open Release Candidate issues found in ${this.repoInfo.repo}`),null;let n=e[0];return console.log(`Using latest open Release Candidate issue: #${n.number}`),{number:n.number,title:n.title||""}}async updateRcIssue(t){let e=await this.getReleaseNotes(),n=`RC ${this.getCurrentDate()}`;await this.octokit.rest.issues.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,issue_number:t,title:n,body:e})}async createRcIssue(){let t=await this.getReleaseNotes(),{data:e}=await this.octokit.rest.issues.create({owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC ${this.getCurrentDate()}`,labels:[Ht],body:t});return console.log(`Created Release Candidate issue: #${e.number} (${e.html_url})`),{number:e.number,html_url:e.html_url}}async createOrUpdateRcBranch(t){let e=`heads/rc/${t}`,n=`rc/${t}`,{data:o}=await this.octokit.rest.repos.getBranch({owner:this.repoInfo.owner,repo:this.repoInfo.repo,branch:q}),{data:s}=await this.octokit.rest.git.listMatchingRefs({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e}),i=s.length>0;try{i?(console.log(`Updating existing RC branch: ${n}`),await this.octokit.rest.git.updateRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e,sha:o.commit.sha,force:!0})):(console.log(`Creating new RC branch: ${n}`),await this.octokit.rest.git.createRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:`refs/${e}`,sha:o.commit.sha}))}catch(a){throw new Error(`Failed to create or update ${n} branch: ${a}`)}}async getLinkedPrByHead(t){let e=`${this.repoInfo.owner}:rc/${t}`,{data:n}=await this.octokit.rest.pulls.list({owner:this.repoInfo.owner,repo:this.repoInfo.repo,state:"all",head:e,per_page:30}),o=n.filter(i=>i.state==="open");if(o.length>1)return{state:"open",multipleOpen:!0};if(o.length===1)return{state:"open",html_url:o[0].html_url,number:o[0].number};let s=n.filter(i=>i.state==="closed");return s.length>0?{state:"closed",html_url:s[0].html_url,number:s[0].number}:null}async fetchPrTemplate(t){try{let{data:e}=await this.octokit.rest.repos.getContent({owner:this.repoInfo.owner,repo:this.repoInfo.repo,path:".github/PULL_REQUEST_TEMPLATE.md"});if("content"in e&&e.type==="file"){let n=Buffer.from(e.content,"base64").toString("utf-8");return n=n.replace("Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.",`Release candidate pull request. See issue #${t} for details.`),n=n.replace(/<details>/g,"<details open>"),n}}catch(e){e&&typeof e=="object"&&"status"in e&&e.status===404?console.log("No PR template found in repo, using default message."):console.warn("Failed to fetch PR template:",e)}return`Release candidate pull request. See issue #${t} for details.`}async createRcPr(t){try{let e=await this.fetchPrTemplate(t),{data:n}=await this.octokit.request(`POST /repos/${this.repoInfo.owner}/${this.repoInfo.repo}/pulls`,{owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC #${t}`,body:e,head:`rc/${t}`,base:Wn,headers:{"X-GitHub-Api-Version":"2022-11-28"}});return console.log(`Created Release Candidate pull request: #${n.number} (${n.html_url})`),{state:"open",html_url:n.html_url,number:n.number}}catch(e){throw console.error("Failed to create RC PR:",e),e}}async updateRcPr(t,e){try{let n=await this.fetchPrTemplate(t);await this.octokit.rest.pulls.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,pull_number:e,body:n}),console.log(`Updated Release Candidate pull request: #${e}`)}catch(n){throw console.error("Failed to update RC PR:",n),n}}async getReleaseNotes(){let t=await this.getFilteredCommits();return xe(t,!1)}static async getTriggerBranchName(){return process.env.GITHUB_REF_NAME?process.env.GITHUB_REF_NAME:process.env.GITHUB_REF?.startsWith("refs/heads/")?process.env.GITHUB_REF.replace("refs/heads/",""):I.getCurrentBranchName()}getCurrentDate(){return new Date().toISOString().split("T")[0]}};var xa=Un.command("rc-workflow").description("Generate RC issue and pull request").action(async()=>{await(await ge.create()).createReleaseCandidate()});Le.name("auro").version(he()).description("A cli tool to support the Auro Design System");Le.addHelpText("beforeAll",We());Le.parse();
158
+ `,{projectId:l,itemId:c,fieldId:m.id,value:t});else if(!m)throw new Error("No 'ado' field found in GitHub project")}catch(s){console.error(`Failed to update GitHub project: ${s}`)}},Mn=async r=>{let t=process.env.ADO_TOKEN;if(!t)throw new Error("ADO_TOKEN environment variable is required");let e="https://dev.azure.com/itsals",n="E_Retain_Content",o="E_Retain_Content\\Auro Design System",s=ge.getPersonalAccessTokenHandler(t),a=await new ge.WebApi(e,s).getWorkItemTrackingApi();try{let l=[{op:"add",path:"/fields/System.Title",value:r.title},{op:"add",path:"/fields/System.Description",value:`GitHub Issue: <a href="${r.html_url}">${r.html_url}</a>`},{op:"add",path:"/fields/System.AreaPath",value:o}];return await a.createWorkItem(null,l,n,"User Story")}catch(l){throw new Error(`Failed to create ADO work item: ${l}`)}},Lt=async r=>{let t=fe(`Processing GitHub issue: ${r}`).start();try{if(!process.env.GH_TOKEN)throw new Error("GH_TOKEN environment variable is required");if(!process.env.ADO_TOKEN)throw new Error("ADO_TOKEN environment variable is required");t.text="Fetching GitHub issue details...";let e=await Fn(r);t.succeed(`Found issue: "${e.title}"`);let n=fe("Checking for existing ADO work item...").start(),o=await Bn(e);if(o){n.succeed("ADO work item already exists for this issue!"),console.log(`${o}`);return}n.succeed("No existing ADO work item found");let s=fe("Creating new ADO work item...").start(),i=await Mn(e);if(s.succeed(`Successfully created ADO work item #${i.id}`),console.log(`Work item: ${i._links?.html?.href||"N/A"}`),i._links?.html?.href){let a=fe("Adding to GitHub project and updating ADO field...").start();await Gn(e,i._links.html.href),a.succeed("Updated GitHub project with ADO link")}}catch(e){t.fail(`Error: ${e instanceof Error?e.message:e}`),process.exit(1)}};var va=Hn.command("ado").description("Generate ADO item from GitHub issue").option("-g, --gh-issue <issue>","What GitHub issue to use").action(async r=>{r.ghIssue&&await Lt(r.ghIssue)});import{program as zn}from"commander";import{Octokit as Ln}from"@octokit/rest";import{simpleGit as Wn}from"simple-git";var Wt="Release Candidate",V="dev",Un="main",he=class r{constructor(t,e,n){this.filteredCommits=null;this.repoInfo={owner:t,repo:e},this.octokit=n}static async create(){let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN is required to run RC workflow.");let e=await I.getRepoOwnerAndName(),n=new Ln({auth:t});if(!e)throw new Error("Failed to retrieve repository information. Ensure you're in a valid git repository.");let o=await r.getTriggerBranchName();return o&&o!==V&&(console.log(`Switching from ${o} to ${V} branch...`),await Wn().checkout(V)),new r(e.owner,e.repo,n)}get owner(){return this.repoInfo.owner}get repo(){return this.repoInfo.repo}get repoData(){return{...this.repoInfo}}async createReleaseCandidate(){await this.hasCommitsReadyInDev()||console.log("No filtered commits found. Continuing to update RC issue/branch/PR.");let e=await this.getLatestOpenRcIssue(),n=e?await this.getLinkedPrByHead(e.number):null;if(n?.multipleOpen)throw new Error("Multiple open RC PRs found for the same rc/<issueNumber> branch.");if(n?.state==="closed"&&(console.log("Linked RC PR is closed. Creating a new RC issue and PR."),e=await this.createRcIssue(),n=null),e?await this.updateRcIssue(e.number):e=await this.createRcIssue(),!e)throw new Error("Failed to resolve RC issue.");await this.createOrUpdateRcBranch(e.number),n?await this.updateRcPr(e.number,n.number):n=await this.createRcPr(e.number)}async getFilteredCommits(){if(this.filteredCommits===null){let t=await I.getCommitMessages(V);this.filteredCommits=Ne(t)}return this.filteredCommits}async hasCommitsReadyInDev(){return(await this.getFilteredCommits()).length>0}async getLatestOpenRcIssue(){let{data:t}=await this.octokit.rest.issues.listForRepo({owner:this.repoInfo.owner,repo:this.repoInfo.repo,labels:Wt,state:"open",sort:"updated",direction:"desc",per_page:30}),e=t.filter(o=>!o.pull_request);if(e.length===0)return console.log(`No open Release Candidate issues found in ${this.repoInfo.repo}`),null;let n=e[0];return console.log(`Using latest open Release Candidate issue: #${n.number}`),{number:n.number,title:n.title||""}}async updateRcIssue(t){let e=await this.getReleaseNotes(),n=`RC ${this.getCurrentDate()}`;await this.octokit.rest.issues.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,issue_number:t,title:n,body:e})}async createRcIssue(){let t=await this.getReleaseNotes(),{data:e}=await this.octokit.rest.issues.create({owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC ${this.getCurrentDate()}`,labels:[Wt],body:t});return console.log(`Created Release Candidate issue: #${e.number} (${e.html_url})`),{number:e.number,html_url:e.html_url}}async createOrUpdateRcBranch(t){let e=`heads/rc/${t}`,n=`rc/${t}`,{data:o}=await this.octokit.rest.repos.getBranch({owner:this.repoInfo.owner,repo:this.repoInfo.repo,branch:V}),{data:s}=await this.octokit.rest.git.listMatchingRefs({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e}),i=s.length>0;try{i?(console.log(`Updating existing RC branch: ${n}`),await this.octokit.rest.git.updateRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e,sha:o.commit.sha,force:!0})):(console.log(`Creating new RC branch: ${n}`),await this.octokit.rest.git.createRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:`refs/${e}`,sha:o.commit.sha}))}catch(a){throw new Error(`Failed to create or update ${n} branch: ${a}`)}}async getLinkedPrByHead(t){let e=`${this.repoInfo.owner}:rc/${t}`,{data:n}=await this.octokit.rest.pulls.list({owner:this.repoInfo.owner,repo:this.repoInfo.repo,state:"all",head:e,per_page:30}),o=n.filter(i=>i.state==="open");if(o.length>1)return{state:"open",multipleOpen:!0};if(o.length===1)return{state:"open",html_url:o[0].html_url,number:o[0].number};let s=n.filter(i=>i.state==="closed");return s.length>0?{state:"closed",html_url:s[0].html_url,number:s[0].number}:null}async fetchPrTemplate(t){try{let{data:e}=await this.octokit.rest.repos.getContent({owner:this.repoInfo.owner,repo:this.repoInfo.repo,path:".github/PULL_REQUEST_TEMPLATE.md"});if("content"in e&&e.type==="file"){let n=Buffer.from(e.content,"base64").toString("utf-8");return n=n.replace("Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.",`Release candidate pull request. See issue #${t} for details.`),n=n.replace(/<details>/g,"<details open>"),n}}catch(e){e&&typeof e=="object"&&"status"in e&&e.status===404?console.log("No PR template found in repo, using default message."):console.warn("Failed to fetch PR template:",e)}return`Release candidate pull request. See issue #${t} for details.`}async createRcPr(t){try{let e=await this.fetchPrTemplate(t),{data:n}=await this.octokit.request(`POST /repos/${this.repoInfo.owner}/${this.repoInfo.repo}/pulls`,{owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC #${t}`,body:e,head:`rc/${t}`,base:Un,headers:{"X-GitHub-Api-Version":"2022-11-28"}});return console.log(`Created Release Candidate pull request: #${n.number} (${n.html_url})`),{state:"open",html_url:n.html_url,number:n.number}}catch(e){throw console.error("Failed to create RC PR:",e),e}}async updateRcPr(t,e){try{let n=await this.fetchPrTemplate(t);await this.octokit.rest.pulls.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,pull_number:e,body:n}),console.log(`Updated Release Candidate pull request: #${e}`)}catch(n){throw console.error("Failed to update RC PR:",n),n}}async getReleaseNotes(){let t=await this.getFilteredCommits();return Oe(t,!1)}static async getTriggerBranchName(){return process.env.GITHUB_REF_NAME?process.env.GITHUB_REF_NAME:process.env.GITHUB_REF?.startsWith("refs/heads/")?process.env.GITHUB_REF.replace("refs/heads/",""):I.getCurrentBranchName()}getCurrentDate(){return new Date().toISOString().split("T")[0]}};var Oa=zn.command("rc-workflow").description("Generate RC issue and pull request").action(async()=>{await(await he.create()).createReleaseCandidate()});We.name("auro").version(ye()).description("A cli tool to support the Auro Design System");We.addHelpText("beforeAll",Ue());We.parse();