@aurodesignsystem-dev/auro-cli 0.0.0-pr286.0 → 0.0.0-pr286.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/auro-cli.js CHANGED
@@ -1,24 +1,24 @@
1
1
  #!/usr/bin/env node
2
- import{program as je}from"commander";import Rt from"figlet";import{mind as xt}from"gradient-string";var _e=()=>xt(Rt.textSync("Auro CLI"));import Ne from"node:fs";import Fe from"node:path";import{fileURLToPath as Ot}from"node:url";function G(r){process.env.DEBUG&&console.log(`[DEBUG] ${r}`)}function ie(){try{let r=Ot(import.meta.url),t=Fe.dirname(r);G(`Current module path: ${t}`);let e=Fe.resolve(t,"..","package.json");return G(`Checking package.json at: ${e}`),Ne.existsSync(e)?(G(`Found package.json at: ${e}`),JSON.parse(Ne.readFileSync(e,"utf8")).version):(G("Could not find package.json in the standard installed module location, using default version"),"0.0.0")}catch(r){return console.error("Error retrieving package version:",r),"0.0.0"}}import{program as Pr}from"commander";import it from"ora";function M(r){return r.option("-m, --module-paths [paths...]","Path(s) to node_modules folder").option("-w, --watch","Watches for changes").option("--skip-docs","Skip documentation generation",!1).option("--wca-input [files...]","Source file(s) to analyze for API documentation").option("--wca-output [files...]","Output file(s) for API documentation")}function H(r){return r.option("-s, --serve","Starts a server").option("-p, --port <number>","Port for the server").option("-o, --open","Open the browser after starting the server")}import Cr from"@rollup/plugin-terser";import{watch as Ir}from"rollup";import{copyFileSync as mr,existsSync as ge,mkdirSync as dr,rmSync as pr,writeFileSync as fr}from"node:fs";import{basename as gr,dirname as hr,join as T,resolve as he}from"node:path";import{pathToFileURL as Qe}from"node:url";import{glob as yr}from"glob";import we from"ora";import{rollup as ye}from"rollup";import*as et from"sass";import I from"ora";import q from"node:fs";import me from"node:path";import{spawn as jt}from"node:child_process";import _t from"ora";var b=(r,t)=>{let e=`${r} ${t?t.join(" "):""}`,n=_t(),o=r,s=t||[];if(!t&&typeof r=="string"){let u=r.split(" ");o=u[0],s=u.slice(1)}let i=e.includes("--watch")||e.includes(" -w"),l=jt(o,s,{stdio:i?"inherit":["inherit","pipe","pipe"],shell:!0});if(!i){let u=[];l.stdout?.on("data",m=>{let c=m.toString();u.push(c),process.stdout.write(c)}),l.stderr?.on("data",m=>{let c=m.toString();u.push(c),process.stderr.write(c)})}return new Promise((u,m)=>{l.on("close",c=>{c!==0?i?(n.info(`Watch mode terminated with code ${c}`),u()):(n.fail(`${e} failed (code ${c})`),m(new Error(`Command failed with exit code ${c}`))):(n.succeed(`${e} completed successfully`),u())})})};import W from"node:fs";import Nt from"node:path";import{markdownTable as Be}from"markdown-table";var P=class r{static{this.manifest={schemaVersion:"1.0.0",readme:"",modules:[]}}static generate(t={}){let{outDir:e="./docs",outFile:n="api.md",manifestPath:o="./custom-elements.json"}=t,{getElements:s,renderAllElements:i}=r;if(o)try{let c=W.readFileSync(o,"utf8");r.manifest=JSON.parse(c)}catch(c){throw console.error(`Error reading manifest file at ${o}:`,c),c}let a=s(),l=e;W.existsSync(l)||W.mkdirSync(l,{recursive:!0});let u=i(a),m=Nt.join(l,n);W.writeFileSync(m,u),console.log(`Generated combined API documentation at ${m}`)}static getElements(){let t=r.manifest.modules.filter(r.isWcaModule);return r.manifest.modules.reduce((e,n)=>e.concat(n.declarations?.filter(o=>"customElement"in o&&o.customElement===!0&&"tagName"in o&&(t.length>0?r.isWcaModule(n):!0))??[]),[])}static isWcaModule(t){let{path:e}=t;return e?e.startsWith("scripts/wca/auro-")&&e.endsWith(".js"):!1}static renderAllElements(t){return`${t.sort((e,n)=>(e.tagName||"").localeCompare(n.tagName||"")).map(e=>r.renderElement(e,!0)).join(`
2
+ import{program as _e}from"commander";import _t from"figlet";import{mind as Nt}from"gradient-string";var Ne=()=>Nt(_t.textSync("Auro CLI"));import Fe from"node:fs";import Be from"node:path";import{fileURLToPath as Ft}from"node:url";function H(r){process.env.DEBUG&&console.log(`[DEBUG] ${r}`)}function le(){try{let r=Ft(import.meta.url),t=Be.dirname(r);H(`Current module path: ${t}`);let e=Be.resolve(t,"..","package.json");return H(`Checking package.json at: ${e}`),Fe.existsSync(e)?(H(`Found package.json at: ${e}`),JSON.parse(Fe.readFileSync(e,"utf8")).version):(H("Could not find package.json in the standard installed module location, using default version"),"0.0.0")}catch(r){return console.error("Error retrieving package version:",r),"0.0.0"}}import{program as Or}from"commander";import mt from"ora";function W(r){return r.option("-m, --module-paths [paths...]","Path(s) to node_modules folder").option("-w, --watch","Watches for changes").option("--skip-docs","Skip documentation generation",!1).option("-r, --readme-template <url>","URL to the README template file").option("--wca-input [files...]","Source file(s) to analyze for API documentation").option("--wca-output [files...]","Output file(s) for API documentation")}function L(r){return r.option("-s, --serve","Starts a server").option("-p, --port <number>","Port for the server").option("-o, --open","Open the browser after starting the server")}import Pr from"@rollup/plugin-terser";import{watch as Rr}from"rollup";import{existsSync as yr,rmSync as wr,writeFileSync as br}from"node:fs";import{basename as kr,dirname as $r,join as j,resolve as vr}from"node:path";import{pathToFileURL as nt}from"node:url";import{glob as Ar}from"glob";import be from"ora";import{rollup as we}from"rollup";import*as ot from"sass";import x from"ora";import{spawn as Bt}from"node:child_process";import Mt from"ora";var b=(r,t)=>{let e=`${r} ${t?t.join(" "):""}`,n=Mt(),o=r,s=t||[];if(!t&&typeof r=="string"){let m=r.split(" ");o=m[0],s=m.slice(1)}let i=e.includes("--watch")||e.includes(" -w"),l=Bt(o,s,{stdio:i?"inherit":["inherit","pipe","pipe"],shell:!0});if(!i){let m=[];l.stdout?.on("data",u=>{let c=u.toString();m.push(c),process.stdout.write(c)}),l.stderr?.on("data",u=>{let c=u.toString();m.push(c),process.stderr.write(c)})}return new Promise((m,u)=>{l.on("close",c=>{c!==0?i?(n.info(`Watch mode terminated with code ${c}`),m()):(n.fail(`${e} failed (code ${c})`),u(new Error(`Command failed with exit code ${c}`))):(n.succeed(`${e} completed successfully`),m())})})};import U from"node:fs";import Gt from"node:path";import{markdownTable as Me}from"markdown-table";var I=class r{static{this.manifest={schemaVersion:"1.0.0",readme:"",modules:[]}}static generate(t={}){let{outDir:e="./docs",outFile:n="api.md",manifestPath:o="./custom-elements.json"}=t,{getElements:s,renderAllElements:i}=r;if(o)try{let c=U.readFileSync(o,"utf8");r.manifest=JSON.parse(c)}catch(c){throw console.error(`Error reading manifest file at ${o}:`,c),c}let a=s(),l=e;U.existsSync(l)||U.mkdirSync(l,{recursive:!0});let m=i(a),u=Gt.join(l,n);U.writeFileSync(u,m),console.log(`Generated combined API documentation at ${u}`)}static getElements(){let t=r.manifest.modules.filter(r.isWcaModule);return r.manifest.modules.reduce((e,n)=>e.concat(n.declarations?.filter(o=>"customElement"in o&&o.customElement===!0&&"tagName"in o&&(t.length>0?r.isWcaModule(n):!0))??[]),[])}static isWcaModule(t){let{path:e}=t;return e?e.startsWith("scripts/wca/auro-")&&e.endsWith(".js"):!1}static renderAllElements(t){return`${t.sort((e,n)=>(e.tagName||"").localeCompare(n.tagName||"")).map(e=>r.renderElement(e,!0)).join(`
3
3
 
4
- `)}`}static renderElement(t,e=!0){let n=[],{renderTable:o,renderPropertiesAttributesTable:s,renderParameters:i,getType:a}=r;n.push(e?`# ${t.tagName}`:""),t.description&&n.push(t.description);let l=s(t);l&&n.push(l.trim());let u=o("Methods",["name","parameters","return","description"],(t.members||[]).filter(p=>p.kind==="method"&&("privacy"in p?p.privacy!=="private":!0)&&p.name[0]!=="_").map(p=>({...p,parameters:i("parameters"in p?p.parameters:void 0),return:"return"in p&&p.return?a(p.return):""})));u&&n.push(u.trim());let m=o("Events",["name","description"],t.events);m&&n.push(m.trim());let c=o("Slots",[["name","(default)"],"description"],t.slots);c&&n.push(c.trim());let d=o("CSS Shadow Parts",["name","description"],t.cssParts);d&&n.push(d.trim());let f=o("CSS Custom Properties",["name","description"],t.cssProperties);return f&&n.push(f.trim()),n.join(`
4
+ `)}`}static renderElement(t,e=!0){let n=[],{renderTable:o,renderPropertiesAttributesTable:s,renderParameters:i,getType:a}=r;n.push(e?`# ${t.tagName}`:""),t.description&&n.push(t.description);let l=s(t);l&&n.push(l.trim());let m=o("Methods",["name","parameters","return","description"],(t.members||[]).filter(p=>p.kind==="method"&&("privacy"in p?p.privacy!=="private":!0)&&p.name[0]!=="_").map(p=>({...p,parameters:i("parameters"in p?p.parameters:void 0),return:"return"in p&&p.return?a(p.return):""})));m&&n.push(m.trim());let u=o("Events",["name","description"],t.events);u&&n.push(u.trim());let c=o("Slots",[["name","(default)"],"description"],t.slots);c&&n.push(c.trim());let d=o("CSS Shadow Parts",["name","description"],t.cssParts);d&&n.push(d.trim());let f=o("CSS Custom Properties",["name","description"],t.cssProperties);return f&&n.push(f.trim()),n.join(`
5
5
 
6
- `)}static renderPropertiesAttributesTable(t){let{getType:e,escapeMarkdown:n}=r,o=t.members?.filter(c=>c.kind==="field"&&("privacy"in c?c.privacy!=="private":!0)&&c.name[0]!=="_")||[],s=t.attributes||[],i=[],a=new Set;if(o.forEach(c=>{if(c.description?.trim()){let d=e(c)||"",p=("return"in c&&c.return?e(c.return):"")||d;i.push({name:c.name,properties:c.name,attributes:("attribute"in c?c.attribute:"")||"",modifiers:"readonly"in c&&c.readonly?"readonly":"",type:p,default:("default"in c?c.default:"")||"",description:c.description||""})}a.add(c.name),"attribute"in c&&c.attribute&&a.add(c.attribute)}),s.forEach(c=>{!a.has(c.name)&&c.description?.trim()&&i.push({name:c.name,properties:"",attributes:c.name,modifiers:"",type:e(c)||"",default:c.default||"",description:c.description||""})}),i.length===0)return"";let l=["Properties","Attributes","Modifiers","Type","Default","Description"],u=i.map(c=>{let w=(c.default||"").trim().replace(/^'([^']+)'$/,"$1").replace(/^"([^"]+)"$/,"$1"),S=w?w.startsWith("`")&&w.endsWith("`")?w:`\`${w}\``:"";return[n(c.properties),n(c.attributes),n(c.modifiers),n(c.type),n(S),n(c.description)]});return`### Properties & Attributes
6
+ `)}static renderPropertiesAttributesTable(t){let{getType:e,escapeMarkdown:n}=r,o=t.members?.filter(c=>c.kind==="field"&&("privacy"in c?c.privacy!=="private":!0)&&c.name[0]!=="_")||[],s=t.attributes||[],i=[],a=new Set;if(o.forEach(c=>{if(c.description?.trim()){let d=e(c)||"",p=("return"in c&&c.return?e(c.return):"")||d;i.push({name:c.name,properties:c.name,attributes:("attribute"in c?c.attribute:"")||"",modifiers:"readonly"in c&&c.readonly?"readonly":"",type:p,default:("default"in c?c.default:"")||"",description:c.description||""})}a.add(c.name),"attribute"in c&&c.attribute&&a.add(c.attribute)}),s.forEach(c=>{!a.has(c.name)&&c.description?.trim()&&i.push({name:c.name,properties:"",attributes:c.name,modifiers:"",type:e(c)||"",default:c.default||"",description:c.description||""})}),i.length===0)return"";let l=["Properties","Attributes","Modifiers","Type","Default","Description"],m=i.map(c=>{let w=(c.default||"").trim().replace(/^'([^']+)'$/,"$1").replace(/^"([^"]+)"$/,"$1"),T=w?w.startsWith("`")&&w.endsWith("`")?w:`\`${w}\``:"";return[n(c.properties),n(c.attributes),n(c.modifiers),n(c.type),n(T),n(c.description)]});return`### Properties & Attributes
7
7
 
8
- ${Be([l,...u])}
9
- `}static renderParameters(t){let{escapeMarkdown:e,getType:n}=r;return!t||t.length===0?"None":t.map(o=>{let s=n(o)||"any",i=o.description?` - ${o.description}`:"";return`\`${o.name}\` (${e(s)})${e(i)}`}).join("<br>")}static renderTable(t,e,n){let{escapeMarkdown:o,get:s,capitalize:i}=r;if(n===void 0||n.length===0)return"";let a=n.filter(c=>{let{description:d}=c;return typeof d=="string"&&d.trim()});if(a.length===0)return"";let l=e.map(c=>i((Array.isArray(c)?c[0]:c).split(".")[0])),u=a.map(c=>e.map(d=>{let f=s(c,d);return o(String(f||""))})),m=Be([l,...u]);return`### ${t}
8
+ ${Me([l,...m])}
9
+ `}static renderParameters(t){let{escapeMarkdown:e,getType:n}=r;return!t||t.length===0?"None":t.map(o=>{let s=n(o)||"any",i=o.description?` - ${o.description}`:"";return`\`${o.name}\` (${e(s)})${e(i)}`}).join("<br>")}static renderTable(t,e,n){let{escapeMarkdown:o,get:s,capitalize:i}=r;if(n===void 0||n.length===0)return"";let a=n.filter(c=>{let{description:d}=c;return typeof d=="string"&&d.trim()});if(a.length===0)return"";let l=e.map(c=>i((Array.isArray(c)?c[0]:c).split(".")[0])),m=a.map(c=>e.map(d=>{let f=s(c,d);return o(String(f||""))})),u=Me([l,...m]);return`### ${t}
10
10
 
11
- ${m}
12
- `}static escapeMarkdown(t){return t.replace(/\\/g,"\\\\").replace(/\n/g,"<br>").replace(/\|/g,"\\|")}static getType(t){if(!t||!t.type)return"";let{type:e}=t,n=s=>s.replace(/\s*\|\s*/g," | ").replace(/'([^']+)'/g,"`$1`");if(typeof e=="string")return n(e);if(e.text)return n(e.text);if(Array.isArray(e))return e.map(s=>typeof s=="string"?s:s.text?s.text:s.name?s.name:String(s)).join(" \\| ");if(e.name)return n(e.name);if(e.references&&Array.isArray(e.references))return e.references.map(s=>s.name||String(s)).join(" \\| ");let o=String(e);return n(o)}static get(t,e){let n="",o=e;Array.isArray(e)&&([o,n]=e);let s=o.split("."),i=t;for(;i&&s.length;)i=i[s.shift()];return i==null||i===""?n:String(i)}static capitalize(t){return t.replace(/([A-Z])/g," $1").replace(/^./,e=>e.toUpperCase()).trim()}};import Ft from"node:fs";import Bt from"node:os";import L from"node:path";import ae from"node:process";function Gt(){let r=Bt.homedir()||ae.env.HOME||ae.env.USERPROFILE;if(!r)throw new Error("Unable to determine user home directory");return L.join(r,".auro")}function ce(...r){return L.join(Gt(),...r)}function le(...r){let t=Ft.realpathSync(ae.argv[1]),e=L.dirname(t);return L.resolve(e,...r)}var Ge=r=>le("configs",r);import{Logger as Mt}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{generateReadmeUrl as Ht,processContentForFile as Wt,templateFiller as He}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";import ue from"node:fs";import Lt from"node:path";var Ut="/docs/pages",We={overwriteLocalCopies:!0,remoteReadmeVersion:"master",remoteReadmeVariant:"_updated_paths"};function k(r){return`${process.cwd()}/${r}`}async function zt(r,t=!1){let e=[];t||e.push({identifier:"README.md",input:{remoteUrl:r.remoteReadmeUrl||Ht(r.remoteReadmeVersion,r.remoteReadmeVariant),fileName:k("/docTemplates/README.md"),overwrite:r.overwriteLocalCopies},output:k("/README.md")}),Me("/docs/partials/index.md")&&e.push({identifier:"index.md",input:k("/docs/partials/index.md"),output:k("/demo/index.md"),mdMagicConfig:{output:{directory:k("/demo")}}}),Me("/docs/partials/api.md")&&e.push({identifier:"api.md",input:k("/docs/partials/api.md"),output:k("/demo/api.md"),preProcessors:[He.formatApiTable]});let n=k(Ut);if(ue.existsSync(n)){let s=(await ue.promises.readdir(n)).map(i=>({identifier:i,input:Lt.join(n,i),output:k(`/demo/${i}`)}));e.push(...s)}return e}async function qt(r=We,t=!1){await He.extractNames();let e=await zt(r,t);for(let n of e)try{await Wt(n)}catch(o){Mt.error(`Error processing ${n.identifier}: ${o.message}`)}}async function D(r={}){await qt({...We,remoteReadmeUrl:"https://raw.githubusercontent.com/AlaskaAirlines/auro-templates/main/templates/default/README.md"},r.skipReadme)}function Me(r){return ue.existsSync(k(r))}import{existsSync as Le,readFileSync as Vt}from"node:fs";import{builtinModules as Jt}from"node:module";import{resolve as U}from"node:path";import{startDevServer as Kt}from"@web/dev-server";import{hmrPlugin as Yt}from"@web/dev-server-hmr";import*as Ue from"esbuild";import Xt from"ora";var Zt=["node_modules","../node_modules","../../node_modules","../../../node_modules"],Qt=/^\/__wds-outside-root__\/(\d+)\/(.+)$/;function er(){return{name:"node-modules-css",serve(r){if(!r.path.endsWith(".css")||!r.path.startsWith("/@")&&!/^\/[a-z]/i.test(r.path))return;let t=r.path.slice(1),e=process.cwd();for(let n of Zt){let o=U(e,n,t);if(Le(o))return{body:Vt(o,"utf-8"),type:"css"}}}}}function tr(r,t){let e=r.match(Qt);return e?U(t,"../".repeat(Number.parseInt(e[1],10)),e[2]):U(t,`.${r}`)}function rr(){let r=/\b(require\s*\(|module\.exports\b|exports\.\w)/,t=new Map,e;return{name:"cjs-to-esm",async serverStart({config:n}){e=U(n.rootDir)},async transform(n){if(!n.path.includes("node_modules")||!n.path.endsWith(".js")&&!n.path.endsWith(".cjs")||typeof n.body!="string"||!r.test(n.body))return;let o=tr(n.path,e);if(t.has(o))return t.get(o);if(Le(o))try{let i={body:(await Ue.build({entryPoints:[o],bundle:!0,format:"esm",platform:"browser",write:!1,logLevel:"silent",external:Jt})).outputFiles[0].text};return t.set(o,i),i}catch(s){console.error(`CJS-to-ESM bundling failed for ${n.path}:`,s.message)}}}}var R={watch:!0,nodeResolve:!0,basePath:"/",rootDir:"./demo",hmrInclude:["src/**/*","demo/**/*","apiExamples/**/*","docs/**/*"]};async function z(r={}){if(!r.serve)return;let t=Xt(`Firing up dev server...
13
- `).start();try{let e={port:Number(r.port)||void 0,open:r.open?"/":void 0,watch:r.watch??R.watch,nodeResolve:r.nodeResolve??R.nodeResolve,basePath:r.basePath??R.basePath,rootDir:r.rootDir??R.rootDir,middleware:[function(s,i){return!s.url.endsWith("/")&&!s.url.includes(".")&&(s.url+=".html"),i()}],plugins:[er(),rr(),Yt({include:r.hmrInclude??R.hmrInclude})]},n=await Kt({config:e,readCliArgs:!1,readFileConfig:!1});return t.stop(),n}catch(e){throw t.fail("Server snag! Couldn't start dev server."),console.error("Error starting development server:",e),new Error(`Development server failed to start: ${e.message}`)}}async function V(){let r=I("Generating Custom Elements Manifest...").start();try{await b(`npx --package=@custom-elements-manifest/analyzer -y -- cem analyze --config '${Ge("custom-elements-manifest.config.mjs")}'`),r.succeed("Custom Elements Manifest generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);r.warn("CEM analyzer completed with warnings: "+e)}}async function J(){let r=I("Generating API md file...").start();try{await P.generate(),r.succeed("API md file generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);throw r.fail("Failed to generate API md file: "+e),t}}async function ze(r={}){let t=I("Compiling documentation...").start();try{await D(r),t.succeed("Documentation compiled successfully!")}catch(e){let n=e instanceof Error?e.message:String(e);throw t.fail("Failed to compile MD documentation: "+n),e}qe(),await v(),await de(r)}function qe(){let r=me.resolve(process.cwd(),"README.md"),t=me.resolve(process.cwd(),"demo"),e=me.join(t,"readme.md");q.existsSync(r)&&(q.existsSync(t)||q.mkdirSync(t,{recursive:!0}),q.copyFileSync(r,e))}async function Ve(r={}){await z(r)}async function Je(r={}){let t=await import("chokidar"),e=["./src/**/*","./docs/**/*","./docTemplates/**/*","./apiExamples/**/*","./demo/**/*.scss"],n=["**/demo/*.min.js","**/demo/*.min.css","**/demo/*.md","**/demo/readme.md","**/docs/api.md","**/custom-elements.json","**/node_modules/**","**/dist/**","**/.git/**"],o=t.watch(e,{ignoreInitial:!0,ignored:n,awaitWriteFinish:{stabilityThreshold:1e3,pollInterval:100}}),s=I("Waiting for changes...");s.spinner="bouncingBar",s.color="green",s.start();let i=null,a=!1,l=!1;async function u(m){if(a){l=!0;return}a=!0,o.unwatch(e);let c=I(`Change detected: ${m}`).start();try{await D(r),qe(),await v(),await de(r),c.succeed("Docs rebuilt!")}catch(d){let f=d instanceof Error?d.message:String(d);c.fail("Rebuild failed: "+f)}finally{setTimeout(()=>{o.add(e),a=!1,l&&(l=!1,u("queued changes"))},1e3)}}o.on("all",(m,c)=>{a||(i&&clearTimeout(i),i=setTimeout(()=>{u(c)},1e3))}),process.on("SIGINT",()=>{o.close(),I().succeed("Doc watch stopped."),process.exit(0)})}async function K(){await V(),await J()}import{basename as sr,join as Ye}from"node:path";import ir from"@rollup/plugin-commonjs";import{nodeResolve as ar}from"@rollup/plugin-node-resolve";import{glob as cr}from"glob";import{litScss as lr}from"rollup-plugin-scss-lit";import nr from"node:path";import{glob as or}from"glob";function Ke(r){return{name:"watch-globs",buildStart(){let t=Array.isArray(r)?r:[r];for(let e of t)try{for(let n of or.sync(nr.resolve(e)))this.addWatchFile(n)}catch(n){this.error(`Error watching glob pattern "${e}": ${n.message}`)}}}}var pe={moduleDirectories:["node_modules"],modulePaths:["../../node_modules","../node_modules","node_modules"],watchPatterns:["./apiExamples/**/*","./docs/**/*"]};function Xe(r=[],t={}){let{watchPatterns:e=pe.watchPatterns,dedupe:n=["lit","lit-element","lit-html"],dev:o=!1}=t,s=[...pe.modulePaths,...r];return[ar({dedupe:n,preferBuiltins:!1,moduleDirectories:pe.moduleDirectories}),ir(),lr({minify:o?!1:{fast:!0},options:{loadPaths:[...s,Ye(process.cwd(),"src","styles"),Ye(process.cwd(),"src")]}}),Ke(e)]}function fe(r={}){let{modulePaths:t=[],watch:e=!1,input:n=["./src/index.js","./src/registered.js"],outputDir:o="./dist",format:s="esm",dev:i=!1}=r;return{name:"Main",config:{input:n,output:{format:s,dir:o,entryFileNames:a=>i||["index","registered"].includes(a.name)?"[name].js":"[name]-[hash].js",chunkFileNames:i?"[name].js":"[name]-[hash].js",assetFileNames:i?"[name][extname]":"[name]-[hash][extname]"},external:ur(),plugins:Xe(t,{dev:i}),watch:Ze(e)}}}function x(r={}){let{modulePaths:t=[],watch:e=!1,globPattern:n="./demo/*.js",ignorePattern:o=["./demo/*.min.js"],outputDir:s="./demo",dev:i=!1}=r;return{name:"Demo",config:{input:Object.fromEntries(cr.sync(n,{ignore:o}).map(a=>[sr(a,".js"),a])),output:{format:"esm",dir:s,entryFileNames:"[name].min.js",chunkFileNames:"[name].min.js",assetFileNames:i?"[name][extname]":"[name]-[hash][extname]"},plugins:Xe(t,{dev:i}),watch:Ze(e)}}}function Ze(r){if(!r)return!1;let t=typeof r=="object"?r:{};return{clearScreen:t.clearScreen??!0,buildDelay:t.buildDelay??500,chokidar:{ignoreInitial:!0,ignored:t.ignored??["**/dist/**/*.d.ts","**/custom-elements.json","**/demo/*.md","**/demo/**/*.min.js","**/demo/**/*.min.css","**/docs/api.md","**/node_modules/**","**/.git/**"],awaitWriteFinish:t.awaitWriteFinish??{stabilityThreshold:1e3,pollInterval:100}},include:t.include??["./src/**/*.scss","./src/**/*.js","./src/**/*.ts","./demo/**/*.js","./apiExamples/**/*","./docs/**/*.md"],exclude:t.exclude??["./dist/**/*","./node_modules/**/*"]}}function ur(r=[]){return[...[/node_modules\/lit/,/node_modules\/lit-element/,/node_modules\/lit-html/,/node_modules\/@lit/],...r]}var wr=["node_modules","../node_modules","../../node_modules","../../../node_modules"];function tt(){let r=T("./dist"),t=we("Cleaning dist folder...").start();try{return pr(r,{recursive:!0,force:!0}),t.succeed("All clean! Dist folder wiped."),!0}catch(e){return t.fail(`Oops! Couldn't clean dist/ folder: ${e.message}`),console.error(e),!1}}async function Y(r,t,e,n){let o=we(r).start();try{let s=await t();return o.succeed(e),s}catch(s){throw o.fail(n),console.error(`Error: ${s.message}`),s}}async function rt(r,t){return Y(`Bundling ${r.name||"main"} and ${t.name||"demo"}...`,async()=>{let e=await ye(r);await e.write(r.output),await e.close();let n=await ye(t);await n.write(t.output),await n.close()},`Bundles ready! ${r.name||"Main"} and ${t.name||"demo"} built.`,"Bundle hiccup! Build failed.")}async function X(r){let{wcaInput:t,wcaOutput:e,skipDocs:n}=r;if(n){let o=we("Skipping docs generation...").start();setTimeout(()=>{o.succeed("Docs generation skipped.")},0);return}return Y("Analyzing components and making docs...",async()=>{await K(t,e),await D(),br()},"Docs ready! Looking good.","Doc troubles!")}function br(){let r=process.cwd(),t=he(r,"README.md"),e=he(r,"demo"),n=T(e,"readme.md");ge(t)&&(ge(e)||dr(e,{recursive:!0}),mr(t,n))}function kr(){let r=process.cwd();function t(n){return[n,`${n}.scss`,`${n}.css`,T(hr(n),`_${gr(n)}.scss`),T(n,"_index.scss"),T(n,"index.scss")].find(s=>ge(s))}function e(n){for(let o of wr){let s=t(he(r,o,n));if(s)return s}return null}return{findFileUrl(n){if(n.includes("/node_modules/")){let o=n.lastIndexOf("/node_modules/"),s=n.slice(o+14),i=e(s);if(i)return Qe(i)}if(!n.startsWith(".")&&!n.startsWith("/")&&!n.startsWith("file:")){let o=e(n);if(o)return Qe(o)}return null}}}async function v(r="./demo"){return Y("Compiling demo SCSS...",async()=>{let t=yr.sync(T(r,"**/*.scss")),e=kr();for(let n of t){let o=et.compile(n,{importers:[e],silenceDeprecations:["import"],style:"compressed"}),s=n.replace(/\.scss$/,".min.css");fr(s,o.css)}return t.length},"Demo SCSS compiled.","SCSS compilation failed.")}async function de(r={}){let t=x(r);return Y("Bundling demo JS...",async()=>{let e=await ye(t.config);await e.write(t.config.output),await e.close()},"Demo JS bundled.","Demo JS bundling failed.")}import nt from"node:path";import A from"ora";import"rollup";var O=!1,$r={analyze:{active:!1,lastTime:0},docs:{active:!1,lastTime:0},scss:{active:!1,lastTime:0}},vr=5e3,Z=new Set,Ar=["/dist/index.d.ts","/custom-elements.json","/demo/api.md","/docs/api.md","/demo/index.min.js"];function Er(r){if(!r||typeof r!="string")return!1;try{let t=nt.normalize(r);return Ar.some(e=>t.endsWith(e))||t.includes("/dist/")||t.endsWith(".min.js")||t.endsWith(".min.css")||t.endsWith(".d.ts")}catch(t){return console.error(`Error checking path (${typeof r}):`,t.message),!1}}async function be(r,t){let e=$r[r];if(e.active||Date.now()-e.lastTime<vr)return!1;try{return e.active=!0,e.lastTime=Date.now(),await t()}catch(n){return console.error(`Error in ${r} task:`,n),!1}finally{e.active=!1}}async function ot(r,t,e){let n=!0,o={analyze:!1,docs:!1,scss:!1},s=null,i,a=A("Activating watch mode...").start(),l={analyze:async()=>{let{wcaInput:c,wcaOutput:d,skipDocs:f}=t;if(f){let w=A("Skipping component analysis...").start();return setTimeout(()=>{w.succeed("Component analysis skipped.")},0),!0}let p=A("Detective work: analyzing components...").start();try{return await K(c,d),p.succeed("Component analysis complete! API generated."),!0}catch(w){return p.fail("Analysis hiccup! Something went wrong."),console.error("Component analysis error:",w),!1}},docs:async()=>{if(O)return!1;if(t.skipDocs){let d=A("Skipping docs generation...").start();return setTimeout(()=>{d.succeed("Docs generation skipped.")},0),!0}let c=A("Refreshing docs...").start();try{return await X(t),c.succeed("Documentation refreshed!"),!0}catch(d){c.fail("Docs stumble! Couldn't refresh."),console.error("Documentation rebuild error:",d)}},scss:async()=>{if(O)return!1;try{return await v(),!0}catch(c){return console.error("Demo SCSS compilation error:",c),!1}}},u=()=>{n&&o.analyze&&o.docs&&o.scss&&typeof e=="function"&&(n=!1,e())};function m(c=1e3){s&&clearTimeout(s),s=setTimeout(async()=>{setTimeout(async()=>{o.analyze=await be("analyze",l.analyze),setTimeout(async()=>{o.docs=await be("docs",l.docs),setTimeout(async()=>{o.scss=await be("scss",l.scss),u()},500)},1e3)},1e3)},c)}r.on("event",async c=>{switch(c.code){case"START":a.succeed("Watch mode active! Eyes peeled.");break;case"BUNDLE_START":if(Z.clear(),c.input)try{let d=Array.isArray(c.input)?c.input:typeof c.input=="string"?[c.input]:typeof c.input=="object"&&c.input!==null?Object.values(c.input):[];for(let f of d)typeof f=="string"&&!Er(f)&&Z.add(nt.normalize(f))}catch(d){console.error("Error processing input paths:",d)}i=A("Weaving bundles...").start(),O=!0;break;case"BUNDLE_END":i&&i.succeed(`Bundle ${Array.isArray(c.input)?`of ${c.input.join("& ")} `:""}done in ${c.duration}ms! \u{1F680}`),O=!1,Z.size>0&&m();break;case"END":break;case"ERROR":O=!1,i?i.fail(`Oops! Bundle hit a snag: ${c.error.message}`):A().fail(`Watch mode hiccup: ${c.error.message}`),Z.clear();break}})}function st(r,t){return process.on("SIGINT",()=>{let e=A("Wrapping up...").start();r.close(),t&&t.close(),e.succeed("All done! See you next time. \u2728"),process.exit(0)}),r}async function Tr(r){let t=fe(r),e=x(r);r.dev||t.config.plugins.push(Cr()),await X(r),await v(),await rt(t.config,e.config)}async function Sr(r){let{dev:t}=r,e=fe({...r,watch:!0}),n=x({...r,watch:!0}),o=Ir([e.config,n.config]);ot(o,r,t?async()=>z(r):void 0);let s=await import("chokidar"),i=!1,a=!1,l=null,u=s.watch("./demo/**/*.scss",{ignoreInitial:!0,ignored:["**/demo/**/*.min.css"],awaitWriteFinish:{stabilityThreshold:500,pollInterval:100}});return u.on("all",()=>{l&&clearTimeout(l),l=setTimeout(async()=>{if(i){a=!0;return}i=!0;try{await v()}catch(m){console.error("Demo SCSS watch compilation error:",m)}finally{i=!1,a&&(a=!1,u.emit("all"))}},500)}),st(o,u),o}async function Q(r={}){try{let{watch:t}=r;return tt(),t?await Sr(r):await Tr(r)}catch(t){throw new Error(`Build failed: ${t.message}`)}}var j=Pr.command("dev").description("Runs development server for auro components");j=M(j);j=H(j);var cs=j.action(async r=>{try{let t=it("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),r.watch||t.succeed("Build completed!"),await Q({...r,dev:!0,watch:r.watch})}catch(t){it().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{program as Dr}from"commander";import at from"ora";var ke=Dr.command("build").description("Builds auro components");ke=M(ke);var fs=ke.action(async r=>{try{let t=at("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),await Q(r),r.watch||t.succeed("Build completed!")}catch(t){at().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{exec as Rr}from"node:child_process";import ct from"node:path";import xr from"node:process";import{fileURLToPath as Or}from"node:url";import jr from"node:util";import{program as _r}from"commander";import Nr from"inquirer";var Es=_r.command("migrate").description("Script runner to perform repetitive code change tasks").requiredOption("-i, --id <string>","Select the migration you would like to run by id").option("-m, --multi-gitter","Run the migration on all repositories in the multi-gitter config").action(async r=>{let t=Or(import.meta.url),e=ct.dirname(t),n=ct.resolve(e,"migrations",r.id);if(r.multiGitter){let o=jr.promisify(Rr);try{await o("command -v multi-gitter")}catch{console.error("multi-gitter is not installed."),xr.exit(1)}(await Nr.prompt([{type:"confirm",name:"dryRun",message:"Run migration in dry-run mode? (no changes will be committed)",default:!0}])).dryRun?b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml" --dry-run`):b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml"`)}else b(`${n}/script.sh`)});import pt from"node:process";import{program as Hr}from"commander";import{readFile as Wr,writeFile as Lr}from"node:fs/promises";import{Logger as _}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import $e from"node:fs/promises";import ve from"node:path";import{Octokit as Fr}from"@octokit/rest";import $ from"ora";import{processContentForFile as Br,templateFiller as Gr}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";async function lt(r,t){let e=new Fr({auth:process.env.GITHUB_TOKEN||""});try{let o=(await e.request("GET /repos/{owner}/{repo}/contents/{path}",{ref:t,owner:"AlaskaAirlines",repo:"auro-templates",path:r,headers:{"X-GitHub-Api-Version":"2022-11-28"}})).data;if(typeof o!="object"||!Array.isArray(o)){let s=`Unexpected response format: ${JSON.stringify(o)}`;throw $().start().fail(s),new Error("Failed to retrieve folder items")}return o}catch(n){let o=$().start();throw n.status===404?o.fail(`Template '${r.split("/")[1]}' not found`):o.fail(`Error accessing template: ${n.message}`),n}}async function ut({folderItems:r,templatePathToReplace:t,rootDir:e,ref:n}){let o=[];for(let s of r){if(s.type=="dir"){let u=$(`Processing directory: ${s.path}`).start(),m=await lt(s.path,n);u.succeed(`Found ${m.length} additional items in ${s.path}`);let c=await ut({folderItems:m,templatePathToReplace:t,rootDir:e,ref:n});o.push(...c);continue}let i=s.path.replace(`${t}/`,""),a=`${e}/.github/${i}`,l={identifier:s.name,input:{remoteUrl:s.download_url||"",fileName:a,overwrite:!0},output:a};o.push(l)}return o}async function Mr(r){try{await $e.rm(r,{recursive:!0,force:!0}),$().start().succeed(`Successfully removed directory: ${r}`)}catch(t){throw $().start().fail(`Error removing directory ${r}: ${t.message}`),t}}async function mt(r,t="",e=!0){try{let n=await $e.stat(r),o=ve.basename(r);if(!n.isDirectory())return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}
11
+ ${u}
12
+ `}static escapeMarkdown(t){return t.replace(/\\/g,"\\\\").replace(/\n/g,"<br>").replace(/\|/g,"\\|")}static getType(t){if(!t||!t.type)return"";let{type:e}=t,n=s=>s.replace(/\s*\|\s*/g," | ").replace(/'([^']+)'/g,"`$1`");if(typeof e=="string")return n(e);if(e.text)return n(e.text);if(Array.isArray(e))return e.map(s=>typeof s=="string"?s:s.text?s.text:s.name?s.name:String(s)).join(" \\| ");if(e.name)return n(e.name);if(e.references&&Array.isArray(e.references))return e.references.map(s=>s.name||String(s)).join(" \\| ");let o=String(e);return n(o)}static get(t,e){let n="",o=e;Array.isArray(e)&&([o,n]=e);let s=o.split("."),i=t;for(;i&&s.length;)i=i[s.shift()];return i==null||i===""?n:String(i)}static capitalize(t){return t.replace(/([A-Z])/g," $1").replace(/^./,e=>e.toUpperCase()).trim()}};import Ht from"node:fs";import Wt from"node:os";import z from"node:path";import me from"node:process";function Lt(){let r=Wt.homedir()||me.env.HOME||me.env.USERPROFILE;if(!r)throw new Error("Unable to determine user home directory");return z.join(r,".auro")}function ue(...r){return z.join(Lt(),...r)}function de(...r){let t=Ht.realpathSync(me.argv[1]),e=z.dirname(t);return z.resolve(e,...r)}var Ge=r=>de("configs",r);import{copyFileSync as Ut,existsSync as He,mkdirSync as zt}from"node:fs";import{join as qt,resolve as pe}from"node:path";function S(){let r=process.cwd(),t=pe(r,"README.md"),e=qt(pe(r,"demo"),"readme.md");if(!He(t))return;let n=pe(r,"demo");He(n)||zt(n,{recursive:!0}),Ut(t,e)}import Vt from"ora";var Le=[];function P(r){Le.push(r)}var We=!1;function q(){We||(We=!0,process.on("SIGINT",()=>{let r=Vt("Wrapping up...").start();for(let t of Le)t.close();r.succeed("All done! See you next time. \u2728"),process.exit(0)}))}import{Logger as Jt}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{generateReadmeUrl as Kt,processContentForFile as Yt,templateFiller as ze}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";import fe from"node:fs";import qe from"node:path";var Xt="/docs/pages",Ve={overwriteLocalCopies:!0,remoteReadmeVersion:"master",remoteReadmeVariant:"_updated_paths"};function k(r){return`${process.cwd()}/${r}`}async function Zt(r,t=!1){let e=[];if(!t){let o=r.localReadmePath?r.localReadmePath:{remoteUrl:r.remoteReadmeUrl||Kt(r.remoteReadmeVersion,r.remoteReadmeVariant),fileName:k("/docTemplates/README.md"),overwrite:r.overwriteLocalCopies};e.push({identifier:"README.md",input:o,output:k("/README.md")})}Ue("/docs/partials/index.md")&&e.push({identifier:"index.md",input:k("/docs/partials/index.md"),output:k("/demo/index.md"),mdMagicConfig:{output:{directory:k("/demo")}}}),Ue("/docs/partials/api.md")&&e.push({identifier:"api.md",input:k("/docs/partials/api.md"),output:k("/demo/api.md"),preProcessors:[ze.formatApiTable]});let n=k(Xt);if(fe.existsSync(n)){let s=(await fe.promises.readdir(n)).map(i=>({identifier:i,input:qe.join(n,i),output:k(`/demo/${i}`)}));e.push(...s)}return e}async function Qt(r=Ve,t=!1){await ze.extractNames();let e=await Zt(r,t);for(let n of e)try{await Yt(n)}catch(o){Jt.error(`Error processing ${n.identifier}: ${o.message}`)}}async function R(r={}){let t=r.readmeTemplate,e=t&&!t.startsWith("http");await Qt({...Ve,...e?{localReadmePath:qe.resolve(process.cwd(),t)}:{remoteReadmeUrl:t||"https://raw.githubusercontent.com/AlaskaAirlines/auro-templates/main/templates/default/README.md"}},r.skipReadme)}function Ue(r){return fe.existsSync(k(r))}import{existsSync as Je,readFileSync as er}from"node:fs";import{builtinModules as tr}from"node:module";import{resolve as J}from"node:path";import{startDevServer as rr}from"@web/dev-server";import{hmrPlugin as nr}from"@web/dev-server-hmr";import*as Ke from"esbuild";import or from"ora";var V=["node_modules","../node_modules","../../node_modules","../../../node_modules"];var sr=/^\/__wds-outside-root__\/(\d+)\/(.+)$/;function ir(){return{name:"node-modules-css",serve(r){if(!r.path.endsWith(".css")||!r.path.startsWith("/@")&&!/^\/[a-z]/i.test(r.path))return;let t=r.path.slice(1),e=process.cwd();for(let n of V){let o=J(e,n,t);if(Je(o))return{body:er(o,"utf-8"),type:"css"}}}}}function ar(r,t){let e=r.match(sr);return e?J(t,"../".repeat(Number.parseInt(e[1],10)),e[2]):J(t,`.${r}`)}function cr(){let r=/\b(require\s*\(|module\.exports\b|exports\.\w)/,t=new Map,e;return{name:"cjs-to-esm",async serverStart({config:n}){e=J(n.rootDir)},async transform(n){if(!n.path.includes("node_modules")||!n.path.endsWith(".js")&&!n.path.endsWith(".cjs")||typeof n.body!="string"||!r.test(n.body))return;let o=ar(n.path,e);if(t.has(o))return t.get(o);if(Je(o))try{let i={body:(await Ke.build({entryPoints:[o],bundle:!0,format:"esm",platform:"browser",write:!1,logLevel:"silent",external:tr})).outputFiles[0].text};return t.set(o,i),i}catch(s){console.error(`CJS-to-ESM bundling failed for ${n.path}:`,s.message)}}}}var D={watch:!0,nodeResolve:!0,basePath:"/",rootDir:"./demo",hmrInclude:["src/**/*","demo/**/*","apiExamples/**/*","docs/**/*"]};async function K(r={}){if(!r.serve)return;let t=or(`Firing up dev server...
13
+ `).start();try{let e={port:Number(r.port)||void 0,open:r.open?"/":void 0,watch:r.watch??D.watch,nodeResolve:r.nodeResolve??D.nodeResolve,basePath:r.basePath??D.basePath,rootDir:r.rootDir??D.rootDir,middleware:[function(s,i){return!s.url.endsWith("/")&&!s.url.includes(".")&&(s.url+=".html"),i()}],plugins:[ir(),cr(),nr({include:r.hmrInclude??D.hmrInclude})]},n=await rr({config:e,readCliArgs:!1,readFileConfig:!1});return t.stop(),n}catch(e){throw t.fail("Server snag! Couldn't start dev server."),console.error("Error starting development server:",e),new Error(`Development server failed to start: ${e.message}`)}}async function Y(){let r=x("Generating Custom Elements Manifest...").start();try{await b(`npx --package=@custom-elements-manifest/analyzer -y -- cem analyze --config '${Ge("custom-elements-manifest.config.mjs")}'`),r.succeed("Custom Elements Manifest generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);r.warn("CEM analyzer completed with warnings: "+e)}}async function X(){let r=x("Generating API md file...").start();try{await I.generate(),r.succeed("API md file generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);throw r.fail("Failed to generate API md file: "+e),t}}async function Ye(r={}){let t=x("Compiling documentation...").start();try{await R(r),t.succeed("Documentation compiled successfully!")}catch(e){let n=e instanceof Error?e.message:String(e);throw t.fail("Failed to compile MD documentation: "+n),e}S(),await v(),await ge(r)}async function Xe(r={}){await K(r)}async function Ze(r={}){let t=await import("chokidar"),e=["./src/**/*","./docs/**/*","./docTemplates/**/*","./apiExamples/**/*","./demo/**/*.scss"],n=["**/demo/*.min.js","**/demo/*.min.css","**/demo/*.md","**/demo/readme.md","**/docs/api.md","**/custom-elements.json","**/node_modules/**","**/dist/**","**/.git/**"],o=t.watch(e,{ignoreInitial:!0,ignored:n,awaitWriteFinish:{stabilityThreshold:1e3,pollInterval:100}}),s=x("Waiting for changes...");s.spinner="bouncingBar",s.color="green",s.start();let i=null,a=!1,l=!1;async function m(u){if(a){l=!0;return}a=!0,o.unwatch(e);let c=x(`Change detected: ${u}`).start();try{await R(r),S(),await v(),await ge(r),c.succeed("Docs rebuilt!")}catch(d){let f=d instanceof Error?d.message:String(d);c.fail("Rebuild failed: "+f)}finally{setTimeout(()=>{o.add(e),a=!1,l&&(l=!1,m("queued changes"))},1e3)}}o.on("all",(u,c)=>{a||(i&&clearTimeout(i),i=setTimeout(()=>{m(c)},1e3))}),P(o),q()}async function Z(){await Y(),await X()}import{basename as ur,join as et}from"node:path";import dr from"@rollup/plugin-commonjs";import{nodeResolve as pr}from"@rollup/plugin-node-resolve";import{glob as fr}from"glob";import{litScss as gr}from"rollup-plugin-scss-lit";import lr from"node:path";import{glob as mr}from"glob";function Qe(r){return{name:"watch-globs",buildStart(){let t=Array.isArray(r)?r:[r];for(let e of t)try{for(let n of mr.sync(lr.resolve(e)))this.addWatchFile(n)}catch(n){this.error(`Error watching glob pattern "${e}": ${n.message}`)}}}}var he={moduleDirectories:["node_modules"],modulePaths:["../../node_modules","../node_modules","node_modules"],watchPatterns:["./apiExamples/**/*","./docs/**/*"]};function tt(r=[],t={}){let{watchPatterns:e=he.watchPatterns,dedupe:n=["lit","lit-element","lit-html"],dev:o=!1}=t,s=[...he.modulePaths,...r];return[pr({dedupe:n,preferBuiltins:!1,moduleDirectories:he.moduleDirectories}),dr(),gr({minify:o?!1:{fast:!0},options:{loadPaths:[...s,et(process.cwd(),"src","styles"),et(process.cwd(),"src")]}}),Qe(e)]}function ye(r={}){let{modulePaths:t=[],watch:e=!1,input:n=["./src/index.js","./src/registered.js"],outputDir:o="./dist",format:s="esm",dev:i=!1}=r;return{name:"Main",config:{input:n,output:{format:s,dir:o,entryFileNames:a=>i||["index","registered"].includes(a.name)?"[name].js":"[name]-[hash].js",chunkFileNames:i?"[name].js":"[name]-[hash].js",assetFileNames:i?"[name][extname]":"[name]-[hash][extname]"},external:hr(),plugins:tt(t,{dev:i}),watch:rt(e)}}}function O(r={}){let{modulePaths:t=[],watch:e=!1,globPattern:n="./demo/*.js",ignorePattern:o=["./demo/*.min.js"],outputDir:s="./demo",dev:i=!1}=r;return{name:"Demo",config:{input:Object.fromEntries(fr.sync(n,{ignore:o}).map(a=>[ur(a,".js"),a])),output:{format:"esm",dir:s,entryFileNames:"[name].min.js",chunkFileNames:"[name].min.js",assetFileNames:i?"[name][extname]":"[name]-[hash][extname]"},plugins:tt(t,{dev:i}),watch:rt(e)}}}function rt(r){if(!r)return!1;let t=typeof r=="object"?r:{};return{clearScreen:t.clearScreen??!0,buildDelay:t.buildDelay??500,chokidar:{ignoreInitial:!0,ignored:t.ignored??["**/dist/**/*.d.ts","**/custom-elements.json","**/demo/*.md","**/demo/**/*.min.js","**/demo/**/*.min.css","**/docs/api.md","**/node_modules/**","**/.git/**"],awaitWriteFinish:t.awaitWriteFinish??{stabilityThreshold:1e3,pollInterval:100}},include:t.include??["./src/**/*.scss","./src/**/*.js","./src/**/*.ts","./demo/**/*.js","./apiExamples/**/*","./docs/**/*.md"],exclude:t.exclude??["./dist/**/*","./node_modules/**/*"]}}function hr(r=[]){return[...[/node_modules\/lit/,/node_modules\/lit-element/,/node_modules\/lit-html/,/node_modules\/@lit/],...r]}function st(){let r=j("./dist"),t=be("Cleaning dist folder...").start();try{return wr(r,{recursive:!0,force:!0}),t.succeed("All clean! Dist folder wiped."),!0}catch(e){return t.fail(`Oops! Couldn't clean dist/ folder: ${e.message}`),console.error(e),!1}}async function Q(r,t,e,n){let o=be(r).start();try{let s=await t();return o.succeed(e),s}catch(s){throw o.fail(n),console.error(`Error: ${s.message}`),s}}async function it(r,t){return Q(`Bundling ${r.name||"main"} and ${t.name||"demo"}...`,async()=>{let e=await we(r);await e.write(r.output),await e.close();let n=await we(t);await n.write(t.output),await n.close()},`Bundles ready! ${r.name||"Main"} and ${t.name||"demo"} built.`,"Bundle hiccup! Build failed.")}async function ee(r){let{wcaInput:t,wcaOutput:e,skipDocs:n}=r;if(n){let o=be("Skipping docs generation...").start();setTimeout(()=>{o.succeed("Docs generation skipped.")},0);return}return Q("Analyzing components and making docs...",async()=>{await Z(t,e),await R(r),S()},"Docs ready! Looking good.","Doc troubles!")}function Er(){let r=process.cwd();function t(n){return[n,`${n}.scss`,`${n}.css`,j($r(n),`_${kr(n)}.scss`),j(n,"_index.scss"),j(n,"index.scss")].find(s=>yr(s))}function e(n){for(let o of V){let s=t(vr(r,o,n));if(s)return s}return null}return{findFileUrl(n){if(n.includes("/node_modules/")){let o=n.lastIndexOf("/node_modules/"),s=n.slice(o+14),i=e(s);if(i)return nt(i)}if(!n.startsWith(".")&&!n.startsWith("/")&&!n.startsWith("file:")){let o=e(n);if(o)return nt(o)}return null}}}async function v(r="./demo"){return Q("Compiling demo SCSS...",async()=>{let t=Ar.sync(j(r,"**/*.scss")),e=Er();for(let n of t){let o=ot.compile(n,{importers:[e],silenceDeprecations:["import"],style:"compressed"}),s=n.replace(/\.scss$/,".min.css");br(s,o.css)}return t.length},"Demo SCSS compiled.","SCSS compilation failed.")}async function ge(r={}){let t=O(r);return Q("Bundling demo JS...",async()=>{let e=await we(t.config);await e.write(t.config.output),await e.close()},"Demo JS bundled.","Demo JS bundling failed.")}import at from"node:path";import E from"ora";import"rollup";var _=!1,Cr={analyze:{active:!1,lastTime:0},docs:{active:!1,lastTime:0},scss:{active:!1,lastTime:0}},Tr=5e3,te=new Set,Ir=["/dist/index.d.ts","/custom-elements.json","/demo/api.md","/docs/api.md","/demo/index.min.js"];function Sr(r){if(!r||typeof r!="string")return!1;try{let t=at.normalize(r);return Ir.some(e=>t.endsWith(e))||t.includes("/dist/")||t.endsWith(".min.js")||t.endsWith(".min.css")||t.endsWith(".d.ts")}catch(t){return console.error(`Error checking path (${typeof r}):`,t.message),!1}}async function ke(r,t){let e=Cr[r];if(e.active||Date.now()-e.lastTime<Tr)return!1;try{return e.active=!0,e.lastTime=Date.now(),await t()}catch(n){return console.error(`Error in ${r} task:`,n),!1}finally{e.active=!1}}async function ct(r,t,e){let n=!0,o={analyze:!1,docs:!1,scss:!1},s=null,i,a=E("Activating watch mode...").start(),l={analyze:async()=>{let{wcaInput:c,wcaOutput:d,skipDocs:f}=t;if(f){let w=E("Skipping component analysis...").start();return setTimeout(()=>{w.succeed("Component analysis skipped.")},0),!0}let p=E("Detective work: analyzing components...").start();try{return await Z(c,d),p.succeed("Component analysis complete! API generated."),!0}catch(w){return p.fail("Analysis hiccup! Something went wrong."),console.error("Component analysis error:",w),!1}},docs:async()=>{if(_)return!1;if(t.skipDocs){let d=E("Skipping docs generation...").start();return setTimeout(()=>{d.succeed("Docs generation skipped.")},0),!0}let c=E("Refreshing docs...").start();try{return await ee(t),c.succeed("Documentation refreshed!"),!0}catch(d){c.fail("Docs stumble! Couldn't refresh."),console.error("Documentation rebuild error:",d)}},scss:async()=>{if(_)return!1;try{return await v(),!0}catch(c){return console.error("Demo SCSS compilation error:",c),!1}}},m=()=>{n&&o.analyze&&o.docs&&o.scss&&typeof e=="function"&&(n=!1,e())};function u(c=1e3){s&&clearTimeout(s),s=setTimeout(async()=>{setTimeout(async()=>{o.analyze=await ke("analyze",l.analyze),setTimeout(async()=>{o.docs=await ke("docs",l.docs),setTimeout(async()=>{o.scss=await ke("scss",l.scss),m()},500)},1e3)},1e3)},c)}r.on("event",async c=>{switch(c.code){case"START":a.succeed("Watch mode active! Eyes peeled.");break;case"BUNDLE_START":if(te.clear(),c.input)try{let d=Array.isArray(c.input)?c.input:typeof c.input=="string"?[c.input]:typeof c.input=="object"&&c.input!==null?Object.values(c.input):[];for(let f of d)typeof f=="string"&&!Sr(f)&&te.add(at.normalize(f))}catch(d){console.error("Error processing input paths:",d)}i=E("Weaving bundles...").start(),_=!0;break;case"BUNDLE_END":i&&i.succeed(`Bundle ${Array.isArray(c.input)?`of ${c.input.join("& ")} `:""}done in ${c.duration}ms! \u{1F680}`),_=!1,te.size>0&&u();break;case"END":break;case"ERROR":_=!1,i?i.fail(`Oops! Bundle hit a snag: ${c.error.message}`):E().fail(`Watch mode hiccup: ${c.error.message}`),te.clear();break}})}function lt(r,t){return P(r),t&&P(t),q(),r}async function Dr(r){let t=ye(r),e=O(r);r.dev||t.config.plugins.push(Pr()),await ee(r),await v(),await it(t.config,e.config)}async function xr(r){let{dev:t}=r,e=ye({...r,watch:!0}),n=O({...r,watch:!0}),o=Rr([e.config,n.config]);ct(o,r,t?async()=>K(r):void 0);let s=await import("chokidar"),i=!1,a=!1,l=null,m=s.watch("./demo/**/*.scss",{ignoreInitial:!0,ignored:["**/demo/**/*.min.css"],awaitWriteFinish:{stabilityThreshold:500,pollInterval:100}});return m.on("all",()=>{l&&clearTimeout(l),l=setTimeout(async()=>{if(i){a=!0;return}i=!0;try{await v()}catch(u){console.error("Demo SCSS watch compilation error:",u)}finally{i=!1,a&&(a=!1,m.emit("all"))}},500)}),lt(o,m),o}async function re(r={}){try{let{watch:t}=r;return st(),t?await xr(r):await Dr(r)}catch(t){throw new Error(`Build failed: ${t.message}`)}}var N=Or.command("dev").description("Runs development server for auro components");N=W(N);N=L(N);var vs=N.action(async r=>{try{let t=mt("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),r.watch||t.succeed("Build completed!"),await re({...r,dev:!0,watch:r.watch})}catch(t){mt().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{program as jr}from"commander";import ut from"ora";var $e=jr.command("build").description("Builds auro components");$e=W($e);var Ss=$e.action(async r=>{try{let t=ut("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),await re(r),r.watch||t.succeed("Build completed!")}catch(t){ut().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{exec as _r}from"node:child_process";import dt from"node:path";import Nr from"node:process";import{fileURLToPath as Fr}from"node:url";import Br from"node:util";import{program as Mr}from"commander";import Gr from"inquirer";var Bs=Mr.command("migrate").description("Script runner to perform repetitive code change tasks").requiredOption("-i, --id <string>","Select the migration you would like to run by id").option("-m, --multi-gitter","Run the migration on all repositories in the multi-gitter config").action(async r=>{let t=Fr(import.meta.url),e=dt.dirname(t),n=dt.resolve(e,"migrations",r.id);if(r.multiGitter){let o=Br.promisify(_r);try{await o("command -v multi-gitter")}catch{console.error("multi-gitter is not installed."),Nr.exit(1)}(await Gr.prompt([{type:"confirm",name:"dryRun",message:"Run migration in dry-run mode? (no changes will be committed)",default:!0}])).dryRun?b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml" --dry-run`):b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml"`)}else b(`${n}/script.sh`)});import yt from"node:process";import{program as zr}from"commander";import{readFile as qr,writeFile as Vr}from"node:fs/promises";import{Logger as F}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import ve from"node:fs/promises";import Ae from"node:path";import{Octokit as Hr}from"@octokit/rest";import $ from"ora";import{processContentForFile as Wr,templateFiller as Lr}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";async function pt(r,t){let e=new Hr({auth:process.env.GITHUB_TOKEN||""});try{let o=(await e.request("GET /repos/{owner}/{repo}/contents/{path}",{ref:t,owner:"AlaskaAirlines",repo:"auro-templates",path:r,headers:{"X-GitHub-Api-Version":"2022-11-28"}})).data;if(typeof o!="object"||!Array.isArray(o)){let s=`Unexpected response format: ${JSON.stringify(o)}`;throw $().start().fail(s),new Error("Failed to retrieve folder items")}return o}catch(n){let o=$().start();throw n.status===404?o.fail(`Template '${r.split("/")[1]}' not found`):o.fail(`Error accessing template: ${n.message}`),n}}async function ft({folderItems:r,templatePathToReplace:t,rootDir:e,ref:n}){let o=[];for(let s of r){if(s.type=="dir"){let m=$(`Processing directory: ${s.path}`).start(),u=await pt(s.path,n);m.succeed(`Found ${u.length} additional items in ${s.path}`);let c=await ft({folderItems:u,templatePathToReplace:t,rootDir:e,ref:n});o.push(...c);continue}let i=s.path.replace(`${t}/`,""),a=`${e}/.github/${i}`,l={identifier:s.name,input:{remoteUrl:s.download_url||"",fileName:a,overwrite:!0},output:a};o.push(l)}return o}async function Ur(r){try{await ve.rm(r,{recursive:!0,force:!0}),$().start().succeed(`Successfully removed directory: ${r}`)}catch(t){throw $().start().fail(`Error removing directory ${r}: ${t.message}`),t}}async function gt(r,t="",e=!0){try{let n=await ve.stat(r),o=Ae.basename(r);if(!n.isDirectory())return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}
14
14
  `;let s=`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}/
15
- `;try{let a=(await $e.readdir(r)).sort();for(let l=0;l<a.length;l++){let u=a[l],m=ve.join(r,u),c=l===a.length-1,d=t+(e?" ":"\u2502 ");s+=await mt(m,d,c)}}catch{s+=`${t}${e?" ":"\u2502 "}\u2514\u2500\u2500 [Permission denied or error reading directory]
15
+ `;try{let a=(await ve.readdir(r)).sort();for(let l=0;l<a.length;l++){let m=a[l],u=Ae.join(r,m),c=l===a.length-1,d=t+(e?" ":"\u2502 ");s+=await gt(u,d,c)}}catch{s+=`${t}${e?" ":"\u2502 "}\u2514\u2500\u2500 [Permission denied or error reading directory]
16
16
  `}return s}catch(n){return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}[Error: ${n}]
17
- `}}async function dt(r,t="main",e="default"){if(!r)throw $().start().fail("Root directory must be specified"),new Error("Root directory must be specified");if(await Gr.extractNames(),!process.env.GITHUB_TOKEN)throw $().start().fail("GITHUB_TOKEN environment variable is not set."),new Error("GITHUB_TOKEN environment variable is not set");let n=`templates/${e}/.github`,o=await lt(n,t),s=".github",i=$("Removing existing .github directory...").start();try{await Mr(s),i.succeed(".github directory removed successfully")}catch(u){throw i.fail(`Error removing .github directory: ${u.message}`),new Error(`Failed to remove .github directory: ${u.message}`)}let a=await ut({folderItems:o,templatePathToReplace:n,rootDir:r,ref:t}),l=$("Processing all files...").start();try{await Promise.all(a.map(m=>Br(m))),l.succeed("All files processed.");let u=$("Generating directory tree...").start();try{let m=ve.join(r,".github"),c=await mt(m);u.succeed("Synced .github directory structure:"),console.log(c)}catch(m){u.fail(`Error generating directory tree: ${m.message}`)}}catch(u){throw l.fail(`Error processing files: ${u.message}`),new Error(`Failed to process files: ${u.message}`)}}var Fs=Hr.command("sync").option("-r, --ref <branch/tag/commit>","Git reference (branch/tag/commit) to use","main").option("-t, --template <name>","Template based on which to sync","default").description("Script runner to synchronize local repository configuration files").action(async r=>{_.info("Synchronizing repository configuration files..."),_.warn("Note: sync does not create a new git branch. Changes are added to the current branch.");let t=pt.cwd();try{await dt(t,r.ref,r.template);try{let e=`${t}/.github/CODEOWNERS`,o=(await Wr(e,{encoding:"utf-8"})).replace(/\r\n/gu,`
17
+ `}}async function ht(r,t="main",e="default"){if(!r)throw $().start().fail("Root directory must be specified"),new Error("Root directory must be specified");if(await Lr.extractNames(),!process.env.GITHUB_TOKEN)throw $().start().fail("GITHUB_TOKEN environment variable is not set."),new Error("GITHUB_TOKEN environment variable is not set");let n=`templates/${e}/.github`,o=await pt(n,t),s=".github",i=$("Removing existing .github directory...").start();try{await Ur(s),i.succeed(".github directory removed successfully")}catch(m){throw i.fail(`Error removing .github directory: ${m.message}`),new Error(`Failed to remove .github directory: ${m.message}`)}let a=await ft({folderItems:o,templatePathToReplace:n,rootDir:r,ref:t}),l=$("Processing all files...").start();try{await Promise.all(a.map(u=>Wr(u))),l.succeed("All files processed.");let m=$("Generating directory tree...").start();try{let u=Ae.join(r,".github"),c=await gt(u);m.succeed("Synced .github directory structure:"),console.log(c)}catch(u){m.fail(`Error generating directory tree: ${u.message}`)}}catch(m){throw l.fail(`Error processing files: ${m.message}`),new Error(`Failed to process files: ${m.message}`)}}var Xs=zr.command("sync").option("-r, --ref <branch/tag/commit>","Git reference (branch/tag/commit) to use","main").option("-t, --template <name>","Template based on which to sync","default").description("Script runner to synchronize local repository configuration files").action(async r=>{F.info("Synchronizing repository configuration files..."),F.warn("Note: sync does not create a new git branch. Changes are added to the current branch.");let t=yt.cwd();try{await ht(t,r.ref,r.template);try{let e=`${t}/.github/CODEOWNERS`,o=(await qr(e,{encoding:"utf-8"})).replace(/\r\n/gu,`
18
18
  `).replace(/\n\n/gu,`
19
- `);await Lr(e,o,{encoding:"utf-8"}),(o.includes("\r")||o.includes(`
19
+ `);await Vr(e,o,{encoding:"utf-8"}),(o.includes("\r")||o.includes(`
20
20
 
21
- `))&&_.error("CODEOWNERS file still has Windows line endings.")}catch(e){throw _.error(`Error processing CODEOWNERS file: ${e.message}`),e}}catch(e){_.error(`Sync operation failed: ${e.message}`),pt.exit(1)}});import ee from"node:fs";import N from"node:path";import{Logger as gt}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{program as Ur}from"commander";import{glob as zr}from"glob";var ft=(r,t)=>{let e=(r.match(/static register\(name \= (.+)\)/)||r.match(/customElements.get\((.+?)\)/))[1],n=r.match(/export class (.+) extends/)?.[1],o=r.match(/\/\*\*((.|\n)*?)(\*\n|\*\/|[@])/)?.[1]||"";return!e||!n?r:`
21
+ `))&&F.error("CODEOWNERS file still has Windows line endings.")}catch(e){throw F.error(`Error processing CODEOWNERS file: ${e.message}`),e}}catch(e){F.error(`Sync operation failed: ${e.message}`),yt.exit(1)}});import ne from"node:fs";import B from"node:path";import{Logger as bt}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{program as Jr}from"commander";import{glob as Kr}from"glob";var wt=(r,t)=>{let e=(r.match(/static register\(name \= (.+)\)/)||r.match(/customElements.get\((.+?)\)/))[1],n=r.match(/export class (.+) extends/)?.[1],o=r.match(/\/\*\*((.|\n)*?)(\*\n|\*\/|[@])/)?.[1]||"";return!e||!n?r:`
22
22
  import { ${n} } from '${t}';
23
23
 
24
24
  /**${o}*/
@@ -27,30 +27,30 @@ class ${n}WCA extends ${n} {}
27
27
  if (!customElements.get(${e})) {
28
28
  customElements.define(${e}, ${n}WCA);
29
29
  }
30
- `};var te=N.resolve(process.cwd(),"./scripts/wca");async function qr(r){try{return(await Promise.all(r.map(e=>zr(e)))).flat()}catch(t){throw console.error("Error processing glob patterns:",t),t}}async function Vr(r){ee.existsSync(te)||await ee.promises.mkdir(te,{recursive:!0});for(let t of r){let e=N.resolve(process.cwd(),t),n=await ee.promises.readFile(e,"utf-8"),o=N.resolve(te,`${N.basename(t)}`),s=ft(n,N.relative(te,t));await ee.promises.writeFile(o,s)}}async function Jr(){let r=await qr(["./src/auro-*.js"]);await Vr(r)}var qs=Ur.command("wca-setup").description("Set up WCA (Web Component Analyzer) for the project").action(()=>{Jr().then(()=>{gt.success("WCA setup completed successfully.")}).catch(r=>{gt.error(`WCA setup failed: ${r.message}`)})});import{program as rn}from"commander";import bt from"chalk";import kt from"ora";import{appendFile as Kr,readFile as Yr}from"node:fs/promises";import{Logger as g}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{simpleGit as Xr}from"simple-git";var h;try{h=Xr({baseDir:process.cwd(),binary:"git",maxConcurrentProcesses:1})}catch(r){g.error(`Failed to initialize git: ${r}`),h={}}var E=class r{static async checkGitignore(t){if(t==="")return!1;try{return(await Yr(".gitignore","utf-8")).includes(t)}catch(e){return g.error(`Error reading file: ${e}`),!1}}static async getCommitMessages(t=""){try{let e=t;e||(e=(await h.branchLocal()).current);let n="main",o="";if(!!process.env.GITHUB_ACTIONS){n=process.env.GITHUB_BASE_REF||"main";try{if(await h.fetch("origin",n),e!=="HEAD")try{await h.raw(["rev-parse","--verify",`origin/${e}`])}catch{await h.fetch("origin",e)}let i=e==="HEAD"?"HEAD":`origin/${e}`;o=`${(await h.raw(["merge-base",`origin/${n}`,i])).trim()}..${i}`}catch(i){g.warn(`Error setting up commit range in CI: ${i}`);let a=e==="HEAD"?"HEAD":`origin/${e}`;o=`origin/${n}..${a}`}}else try{try{await h.raw(["rev-parse","--verify",`origin/${n}`])}catch{g.info(`Fetching ${n} from origin`),await h.fetch("origin",n)}if(e!=="HEAD")try{await h.raw(["rev-parse","--verify",e])}catch{await h.fetch("origin",e)}o=`${(await h.raw(["merge-base",`origin/${n}`,e])).trim()}..${e}`}catch(i){g.warn(`Error determining commits locally: ${i}`),o=`${e}~10..${e}`}return await r.getFormattedCommits(o)}catch(e){return g.error(`Error getting commit messages: ${e}`),[]}}static async getRepoOwnerAndName(){try{let t=await h.getRemotes(!0);if(t.length===0)return g.warn("No remotes found"),null;let e=t.find(o=>o.name==="origin")||t[0],n=e.refs.fetch||e.refs.push;return r.parseGitUrl(n)}catch(t){return g.error(`Error getting repo owner and name: ${t}`),null}}static async getCurrentBranchName(){try{return(await h.branchLocal()).current||null}catch(t){return g.error(`Error getting current branch name: ${t}`),null}}static parseGitUrl(t){let e;return t.includes("@")&&t.includes(":")&&(e=t.match(/@([^:]+):([^/]+)\/(.+?)(?:\.git)?$/),e)?{owner:e[2],repo:e[3]}:(e=t.match(/https?:\/\/(?:[^@]+@)?[^/]+\/([^/]+)\/(.+?)(?:\.git)?$/),e?{owner:e[1],repo:e[2]}:(g.warn(`Could not parse git URL: ${t}`),null))}static async getFormattedCommits(t){let n=(await h.raw(["log","--pretty=format:COMMIT_START%n%H%n%ad%n%an%n%s%n%b%nCOMMIT_END","--date=short",t])).split(`COMMIT_START
30
+ `};var oe=B.resolve(process.cwd(),"./scripts/wca");async function Yr(r){try{return(await Promise.all(r.map(e=>Kr(e)))).flat()}catch(t){throw console.error("Error processing glob patterns:",t),t}}async function Xr(r){ne.existsSync(oe)||await ne.promises.mkdir(oe,{recursive:!0});for(let t of r){let e=B.resolve(process.cwd(),t),n=await ne.promises.readFile(e,"utf-8"),o=B.resolve(oe,`${B.basename(t)}`),s=wt(n,B.relative(oe,t));await ne.promises.writeFile(o,s)}}async function Zr(){let r=await Yr(["./src/auro-*.js"]);await Xr(r)}var ii=Jr.command("wca-setup").description("Set up WCA (Web Component Analyzer) for the project").action(()=>{Zr().then(()=>{bt.success("WCA setup completed successfully.")}).catch(r=>{bt.error(`WCA setup failed: ${r.message}`)})});import{program as an}from"commander";import At from"chalk";import Et from"ora";import{appendFile as Qr,readFile as en}from"node:fs/promises";import{Logger as g}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{simpleGit as tn}from"simple-git";var h;try{h=tn({baseDir:process.cwd(),binary:"git",maxConcurrentProcesses:1})}catch(r){g.error(`Failed to initialize git: ${r}`),h={}}var A=class r{static async checkGitignore(t){if(t==="")return!1;try{return(await en(".gitignore","utf-8")).includes(t)}catch(e){return g.error(`Error reading file: ${e}`),!1}}static async getCommitMessages(t=""){try{let e=t;e||(e=(await h.branchLocal()).current);let n="main",o="";if(!!process.env.GITHUB_ACTIONS){n=process.env.GITHUB_BASE_REF||"main";try{if(await h.fetch("origin",n),e!=="HEAD")try{await h.raw(["rev-parse","--verify",`origin/${e}`])}catch{await h.fetch("origin",e)}let i=e==="HEAD"?"HEAD":`origin/${e}`;o=`${(await h.raw(["merge-base",`origin/${n}`,i])).trim()}..${i}`}catch(i){g.warn(`Error setting up commit range in CI: ${i}`);let a=e==="HEAD"?"HEAD":`origin/${e}`;o=`origin/${n}..${a}`}}else try{try{await h.raw(["rev-parse","--verify",`origin/${n}`])}catch{g.info(`Fetching ${n} from origin`),await h.fetch("origin",n)}if(e!=="HEAD")try{await h.raw(["rev-parse","--verify",e])}catch{await h.fetch("origin",e)}o=`${(await h.raw(["merge-base",`origin/${n}`,e])).trim()}..${e}`}catch(i){g.warn(`Error determining commits locally: ${i}`),o=`${e}~10..${e}`}return await r.getFormattedCommits(o)}catch(e){return g.error(`Error getting commit messages: ${e}`),[]}}static async getRepoOwnerAndName(){try{let t=await h.getRemotes(!0);if(t.length===0)return g.warn("No remotes found"),null;let e=t.find(o=>o.name==="origin")||t[0],n=e.refs.fetch||e.refs.push;return r.parseGitUrl(n)}catch(t){return g.error(`Error getting repo owner and name: ${t}`),null}}static async getCurrentBranchName(){try{return(await h.branchLocal()).current||null}catch(t){return g.error(`Error getting current branch name: ${t}`),null}}static parseGitUrl(t){let e;return t.includes("@")&&t.includes(":")&&(e=t.match(/@([^:]+):([^/]+)\/(.+?)(?:\.git)?$/),e)?{owner:e[2],repo:e[3]}:(e=t.match(/https?:\/\/(?:[^@]+@)?[^/]+\/([^/]+)\/(.+?)(?:\.git)?$/),e?{owner:e[1],repo:e[2]}:(g.warn(`Could not parse git URL: ${t}`),null))}static async getFormattedCommits(t){let n=(await h.raw(["log","--pretty=format:COMMIT_START%n%H%n%ad%n%an%n%s%n%b%nCOMMIT_END","--date=short",t])).split(`COMMIT_START
31
31
  `).filter(s=>s.trim()!==""),o=[];for(let s of n){let i=s.split(`
32
- `);if(i.length>=4){let a=i[0],l=i[1],u=i[2],m=i[3],c=i.slice(4).filter(S=>S!=="COMMIT_END"),d=c.length>0?c.join(""):"",f=a.substring(0,7),p=m.match(/^(feat|fix|docs|style|refactor|perf|test|build|ci|chore)(\(.+\))?:/),w=p?p[1]:"unknown";d.includes("BREAKING CHANGE")&&(w="breaking"),o.push({type:w,hash:f,date:l,subject:m,body:d,message:`${m}${d?`
32
+ `);if(i.length>=4){let a=i[0],l=i[1],m=i[2],u=i[3],c=i.slice(4).filter(T=>T!=="COMMIT_END"),d=c.length>0?c.join(""):"",f=a.substring(0,7),p=u.match(/^(feat|fix|docs|style|refactor|perf|test|build|ci|chore)(\(.+\))?:/),w=p?p[1]:"unknown";d.includes("BREAKING CHANGE")&&(w="breaking"),o.push({type:w,hash:f,date:l,subject:u,body:d,message:`${u}${d?`
33
33
 
34
- ${d}`:""}`,author_name:u})}}return o}static async addToGitignore(t,e=!0){await r.checkGitignore(t).then(async n=>{if(n)g.warn(`${t} already exists`);else try{await Kr(".gitignore",`
35
- ${t}`),e&&g.success(`${t} added to .gitignore`)}catch(o){g.error(o)}})}static async removeFromGitCache(t){try{await h.rmKeepLocal(t),g.success(`${t.join(", ")} are removed from git cache`)}catch(e){g.error(e)}}static async createBranch(t){try{await h.checkoutLocalBranch(t),g.success(`Created and switched to ${t} branch`)}catch(e){g.error(e)}}static async commitStagedFiles(t){try{await h.add("."),await h.commit(t),g.success(`Committed with message: ${t}`)}catch(e){g.error(e)}}};import y from"chalk";var Zr=60,Qr=100;function F(r){switch(r){case"breaking":return y.bold.red(r);case"feat":return y.bold.green(r);case"fix":return y.bold.green(r);case"perf":return y.bold.green(r);case"docs":return y.bold.cyan(r);case"style":return y.bold.cyan(r);case"refactor":return y.bold.cyan(r);case"test":return y.bold.cyan(r);case"build":return y.bold.cyan(r);case"ci":return y.bold.cyan(r);case"chore":return y.bold.cyan(r);default:return y.bold.white(r)}}function ht(r,t){if(!r)return"";if(r.length<=t)return r;let e=r.split(" "),n="",o="";for(let s of e)(o+s).length>t&&o.length>0&&(n+=`${o.trim()}
36
- `,o=""),o=`${o}${s} `;return o.length>0&&(n+=o.trim()),n}function yt(r){for(let t of r){console.log("\u2500".repeat(60));let e=ht(t.subject,Zr),n=ht(t.body,Qr);console.log(y.bold(`${F(t.type)}`)),console.log(y.dim(`${t.hash} | ${t.date} | ${t.author_name}`)),console.log(y.bold(`${y.white(e)}`)),t.body&&console.log(y.dim(n))}console.log("\u2500".repeat(60)),console.log(`
37
- `)}import re from"@actions/github";async function Ae(){try{let r=process.env.GITHUB_TOKEN;if(!r)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let t=re.getOctokit(r),{context:e}=re;if(!e.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[n,o]=process.env.GITHUB_REPOSITORY.split("/"),s=e.payload.pull_request.number,{data:i}=await t.rest.issues.listLabelsOnIssue({owner:n,repo:o,issue_number:s});return i.map(a=>a.name)}catch(r){throw r instanceof Error?new Error(`Failed to get existing labels: ${r.message}`):r}}async function wt(r){try{let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let e=re.getOctokit(t),{context:n}=re;if(!n.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[o,s]=process.env.GITHUB_REPOSITORY.split("/"),i=n.payload.pull_request.number,a=`semantic-status: ${r}`,l=await Ae();if(l.includes(a))return;let u=l.filter(m=>m.startsWith("semantic-status:")&&m!==a);for(let m of u)await e.rest.issues.removeLabel({owner:o,repo:s,issue_number:i,name:m});await e.rest.issues.addLabels({owner:o,repo:s,issue_number:i,labels:[a]});return}catch(t){throw t instanceof Error?new Error(`Failed to apply label: ${t.message}`):t}}var en=["feat","fix","breaking","perf"];function Ee(r,t=!0){let e=`### In this release
34
+ ${d}`:""}`,author_name:m})}}return o}static async addToGitignore(t,e=!0){await r.checkGitignore(t).then(async n=>{if(n)g.warn(`${t} already exists`);else try{await Qr(".gitignore",`
35
+ ${t}`),e&&g.success(`${t} added to .gitignore`)}catch(o){g.error(o)}})}static async removeFromGitCache(t){try{await h.rmKeepLocal(t),g.success(`${t.join(", ")} are removed from git cache`)}catch(e){g.error(e)}}static async createBranch(t){try{await h.checkoutLocalBranch(t),g.success(`Created and switched to ${t} branch`)}catch(e){g.error(e)}}static async commitStagedFiles(t){try{await h.add("."),await h.commit(t),g.success(`Committed with message: ${t}`)}catch(e){g.error(e)}}};import y from"chalk";var rn=60,nn=100;function M(r){switch(r){case"breaking":return y.bold.red(r);case"feat":return y.bold.green(r);case"fix":return y.bold.green(r);case"perf":return y.bold.green(r);case"docs":return y.bold.cyan(r);case"style":return y.bold.cyan(r);case"refactor":return y.bold.cyan(r);case"test":return y.bold.cyan(r);case"build":return y.bold.cyan(r);case"ci":return y.bold.cyan(r);case"chore":return y.bold.cyan(r);default:return y.bold.white(r)}}function kt(r,t){if(!r)return"";if(r.length<=t)return r;let e=r.split(" "),n="",o="";for(let s of e)(o+s).length>t&&o.length>0&&(n+=`${o.trim()}
36
+ `,o=""),o=`${o}${s} `;return o.length>0&&(n+=o.trim()),n}function $t(r){for(let t of r){console.log("\u2500".repeat(60));let e=kt(t.subject,rn),n=kt(t.body,nn);console.log(y.bold(`${M(t.type)}`)),console.log(y.dim(`${t.hash} | ${t.date} | ${t.author_name}`)),console.log(y.bold(`${y.white(e)}`)),t.body&&console.log(y.dim(n))}console.log("\u2500".repeat(60)),console.log(`
37
+ `)}import se from"@actions/github";async function Ee(){try{let r=process.env.GITHUB_TOKEN;if(!r)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let t=se.getOctokit(r),{context:e}=se;if(!e.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[n,o]=process.env.GITHUB_REPOSITORY.split("/"),s=e.payload.pull_request.number,{data:i}=await t.rest.issues.listLabelsOnIssue({owner:n,repo:o,issue_number:s});return i.map(a=>a.name)}catch(r){throw r instanceof Error?new Error(`Failed to get existing labels: ${r.message}`):r}}async function vt(r){try{let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let e=se.getOctokit(t),{context:n}=se;if(!n.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[o,s]=process.env.GITHUB_REPOSITORY.split("/"),i=n.payload.pull_request.number,a=`semantic-status: ${r}`,l=await Ee();if(l.includes(a))return;let m=l.filter(u=>u.startsWith("semantic-status:")&&u!==a);for(let u of m)await e.rest.issues.removeLabel({owner:o,repo:s,issue_number:i,name:u});await e.rest.issues.addLabels({owner:o,repo:s,issue_number:i,labels:[a]});return}catch(t){throw t instanceof Error?new Error(`Failed to apply label: ${t.message}`):t}}var on=["feat","fix","breaking","perf"];function Ce(r,t=!0){let e=`### In this release
38
38
  `;for(let n of r)if(e+=`- ${n.hash} ${n.subject}
39
- `,n.body?.trim()){let s=n.body.trim().split(/\n+/).map(i=>i.trim()).filter(i=>i.length>0);for(let i of s){let a=i;a=a.replace(/([^\s])(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)/g,"$1 $2"),a=a.replace(/(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)([^\s])/g,"$1 $2"),e+=` - ${a}`}}return r.length===0?"":(t&&console.log(bt.green(`\u2713 Generating release notes for ${r.length} commits`)),e)}function Ce(r,t=!0){let e=r.filter(o=>en.includes(o.type)),n;return t?n=e.length>0?e:r:n=e,n.length===0&&console.log(`No commits found to include in release notes.
40
- `),n}async function $t(r=!1,t=!1,e=!1){let n=kt(`Checking commits...
41
- `).start();try{let o=await E.getCommitMessages();if(e){n.succeed(`Total commits analyzed: ${o.length}`);let s=Ce(o);console.log(Ee(s));return}if(r&&yt(o),n.succeed(`Total commits analyzed: ${o.length}`),o.length!==0){let s=o.map(l=>l.type),a=Array.from(new Set(s)).map(l=>F(l)).join(", ");n.succeed(`Found commit types: ${a}`)}else n.info(`The list of commits is created by comparing the current branch
39
+ `,n.body?.trim()){let s=n.body.trim().split(/\n+/).map(i=>i.trim()).filter(i=>i.length>0);for(let i of s){let a=i;a=a.replace(/([^\s])(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)/g,"$1 $2"),a=a.replace(/(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)([^\s])/g,"$1 $2"),e+=` - ${a}`}}return r.length===0?"":(t&&console.log(At.green(`\u2713 Generating release notes for ${r.length} commits`)),e)}function Te(r,t=!0){let e=r.filter(o=>on.includes(o.type)),n;return t?n=e.length>0?e:r:n=e,n.length===0&&console.log(`No commits found to include in release notes.
40
+ `),n}async function Ct(r=!1,t=!1,e=!1){let n=Et(`Checking commits...
41
+ `).start();try{let o=await A.getCommitMessages();if(e){n.succeed(`Total commits analyzed: ${o.length}`);let s=Te(o);console.log(Ce(s));return}if(r&&$t(o),n.succeed(`Total commits analyzed: ${o.length}`),o.length!==0){let s=o.map(l=>l.type),a=Array.from(new Set(s)).map(l=>M(l)).join(", ");n.succeed(`Found commit types: ${a}`)}else n.info(`The list of commits is created by comparing the current branch
42
42
  with the main branch. If you are on a new branch, please
43
- make sure to commit some changes before running this command.`);t&&await tn(o,n)}catch(o){n.fail("Error getting commit messages"),console.error(o)}}async function tn(r,t){let e=["breaking","feat","fix","perf","docs","style","refactor","test","build","ci","chore"],n=r.map(i=>i.type).filter(i=>e.includes(i)),o=null,s=Number.POSITIVE_INFINITY;for(let i of n){let a=e.indexOf(i);a<s&&(s=a,o=i)}if(o){let i=kt("Checking existing labels on pull request...").start();try{if((await Ae()).includes(`semantic-status: ${o}`)){i.info(`Label "semantic-status: ${F(o)}" already exists on the pull request.`);return}i.text="Applying label to pull request...",await wt(o),i.succeed(`Label "semantic-status: ${F(o)}" applied to the pull request.`)}catch(a){let l=a instanceof Error?a.message:String(a);i.fail(l)}}else t.warn(bt.yellow("No semantic commit type found to apply as label."))}var ui=rn.command("check-commits").alias("cc").option("-l, --set-label","Set label on the pull request based on the commit message type").option("-d, --debug","Display detailed commit information for debugging").option("-r, --release-notes","Generate release notes based on commit messages").description("Check commits in the local repository for the types of semantic commit messages made and return the results.").action(async r=>{await $t(r.debug,r.setLabel,r.releaseNotes)});import vt from"node:fs";import{get as At}from"node:https";import C from"chalk";import{program as nn}from"commander";import on from"ora";var yi=nn.command("pr-release").option("-n, --namespace <package-namespace>","Set namespace of the package release","@aurodesignsystem-dev").option("-p, --pr-number <number>","Set pull request number for the release","0").description("Generate the package version based off of PR number then update the package.json file. Note: this does not publish the package.").action(async r=>{await sn(r)}),sn=async r=>{let{namespace:t,prNumber:e}=r,n=on("Updating package.json").start();try{let o="package.json",s=JSON.parse(vt.readFileSync(o,"utf8"));n.text="Checking npm registry for version information...";let i=`0.0.0-pr${e}`,a=s.name.split("/")[1],l=`${t}/${a}`,u=await an(i,l,n),m=`${i}.${u}`;s.name=l,s.version=m,n.text="Writing updated package.json...",vt.writeFileSync(o,`${JSON.stringify(s,null,2)}
44
- `,"utf8"),n.succeed(`Package.json updated to use ${C.green(m)} and ${C.green(l)}`),process.exit(0)}catch(o){n.fail(`Failed to update package.json: ${o}`),process.exit(1)}},an=(r,t,e)=>new Promise(n=>{try{let a=function(l){if(l.statusCode!==200){e.info(`Package not found. Status code: ${C.red(l.statusCode)}, defaulting to version 0`),n(0);return}e.text="Processing version information...";let u="";l.on("data",m=>{u+=m}),l.on("end",()=>{try{let m=JSON.parse(u),c=m.versions?Object.keys(m.versions):[];e.text="Calculating next version number...";let d=-1,f=new RegExp(`^${r}\\.(\\d+)$`);for(let p of c){let w=p.match(f);if(w){let S=Number.parseInt(w[1],10);d=Math.max(d,S)}}d>=0?e.info(`Found existing version ${C.green(`${r}.${d}`)}. Incrementing to ${C.green(`${r}.${d+1}`)}`):e.info(`No existing version found for ${C.green(r)}. Starting with ${C.green(`${r}.0`)}`),n(d+1)}catch(m){e.warn(`Failed to parse NPM registry response: ${m instanceof Error?m.message:"Unknown error"}, defaulting to version 0`),n(0)}})};var o=a;let s=`https://registry.npmjs.org/${t}`,i=At(s,{headers:{Accept:"application/json"}},l=>{if((l.statusCode===301||l.statusCode===302)&&l.headers.location){e.info(`Following redirect to ${l.headers.location}...`);try{At(l.headers.location,{headers:{Accept:"application/json"}},a).on("error",u=>{e.warn(`Error following redirect: ${u.message}, defaulting to version 0`),n(0)}).end()}catch(u){e.warn(`Redirect request failed: ${u instanceof Error?u.message:"Unknown error"}, defaulting to version 0`),n(0)}return}a(l)});i.on("error",l=>{e.warn(`Request error: ${l.message}, defaulting to version 0`),n(0)}),i.end()}catch{e.warn("Error checking version in npm registry, defaulting to version 0"),n(0)}});import Ie from"node:path";import{fileURLToPath as cn}from"node:url";import{program as ln}from"commander";import un from"open";var mn=cn(import.meta.url),dn=Ie.resolve(Ie.dirname(mn),".."),Ei=ln.command("test").option("-w, --watch","Set watch number for the test").option("-c, --coverage-report","Generate coverage report").option("-o, --open","Open the coverage report in the browser").option("-f, --files <String|String[]>","Test files glob pattern").description("Run the web test runner to test the component library").action(async r=>{let e=`npx wtr --config "${Ie.join(dn,"dist","configs","web-test-runner.config.mjs")}"`,n=`${process.cwd()}/coverage/index.html`;if(r.coverageReport&&(e+=" --coverage"),r.watch&&(e+=" --watch"),r.files){let o=Array.isArray(r.files)?r.files.join(" "):r.files;e+=` --files "${o}"`}b(e),r.open&&await un(n)});import De from"node:fs/promises";import Tt from"node:path";import{program as bn}from"commander";import kn from"inquirer";import Re from"ora";import pn from"node:fs/promises";import fn from"node:path";import gn from"ora";var hn={"auth-type":"workspace-token","author-email":null,"author-name":null,"base-branch":"main","base-url":null,"clone-dir":".gitter-temp","code-search":null,concurrent:4,"conflict-strategy":"replace",draft:!1,"dry-run":!0,"fetch-depth":1,fork:!1,"fork-owner":null,"git-type":"go",group:null,"include-subgroups":!1,insecure:!1,interactive:!1,labels:null,"log-file":"'-'","log-format":"'text'","log-level":"'error'","max-reviewers":0,"max-team-reviewers":0,org:null,output:"'-'","plain-output":!1,platform:"github",project:null,"push-only":!1,repo:["AlaskaAirlines/auro-accordion","AlaskaAirlines/auro-alert","AlaskaAirlines/auro-avatar","AlaskaAirlines/auro-background","AlaskaAirlines/auro-backtotop","AlaskaAirlines/auro-button","AlaskaAirlines/auro-badge","AlaskaAirlines/auro-banner","AlaskaAirlines/auro-card","AlaskaAirlines/auro-carousel","AlaskaAirlines/auro-datetime","AlaskaAirlines/auro-dialog","AlaskaAirlines/auro-drawer","AlaskaAirlines/auro-flight","AlaskaAirlines/auro-flightline","AlaskaAirlines/auro-header","AlaskaAirlines/auro-hyperlink","AlaskaAirlines/auro-icon","AlaskaAirlines/auro-loader","AlaskaAirlines/auro-lockup","AlaskaAirlines/auro-nav","AlaskaAirlines/auro-pane","AlaskaAirlines/auro-popover","AlaskaAirlines/auro-sidenav","AlaskaAirlines/auro-skeleton","AlaskaAirlines/auro-slideshow","AlaskaAirlines/auro-table","AlaskaAirlines/auro-tabs","AlaskaAirlines/auro-toast"],"repo-exclude":null,"repo-include":null,"repo-search":null,reviewers:null,"skip-forks":!1,"skip-pr":!1,"skip-repo":null,"ssh-auth":!1,"team-reviewers":null};function yn(r){return Object.entries(r).map(([t,e])=>Array.isArray(e)?`${t}:
43
+ make sure to commit some changes before running this command.`);t&&await sn(o,n)}catch(o){n.fail("Error getting commit messages"),console.error(o)}}async function sn(r,t){let e=["breaking","feat","fix","perf","docs","style","refactor","test","build","ci","chore"],n=r.map(i=>i.type).filter(i=>e.includes(i)),o=null,s=Number.POSITIVE_INFINITY;for(let i of n){let a=e.indexOf(i);a<s&&(s=a,o=i)}if(o){let i=Et("Checking existing labels on pull request...").start();try{if((await Ee()).includes(`semantic-status: ${o}`)){i.info(`Label "semantic-status: ${M(o)}" already exists on the pull request.`);return}i.text="Applying label to pull request...",await vt(o),i.succeed(`Label "semantic-status: ${M(o)}" applied to the pull request.`)}catch(a){let l=a instanceof Error?a.message:String(a);i.fail(l)}}else t.warn(At.yellow("No semantic commit type found to apply as label."))}var Ei=an.command("check-commits").alias("cc").option("-l, --set-label","Set label on the pull request based on the commit message type").option("-d, --debug","Display detailed commit information for debugging").option("-r, --release-notes","Generate release notes based on commit messages").description("Check commits in the local repository for the types of semantic commit messages made and return the results.").action(async r=>{await Ct(r.debug,r.setLabel,r.releaseNotes)});import Tt from"node:fs";import{get as It}from"node:https";import C from"chalk";import{program as cn}from"commander";import ln from"ora";var Di=cn.command("pr-release").option("-n, --namespace <package-namespace>","Set namespace of the package release","@aurodesignsystem-dev").option("-p, --pr-number <number>","Set pull request number for the release","0").description("Generate the package version based off of PR number then update the package.json file. Note: this does not publish the package.").action(async r=>{await mn(r)}),mn=async r=>{let{namespace:t,prNumber:e}=r,n=ln("Updating package.json").start();try{let o="package.json",s=JSON.parse(Tt.readFileSync(o,"utf8"));n.text="Checking npm registry for version information...";let i=`0.0.0-pr${e}`,a=s.name.split("/")[1],l=`${t}/${a}`,m=await un(i,l,n),u=`${i}.${m}`;s.name=l,s.version=u,n.text="Writing updated package.json...",Tt.writeFileSync(o,`${JSON.stringify(s,null,2)}
44
+ `,"utf8"),n.succeed(`Package.json updated to use ${C.green(u)} and ${C.green(l)}`),process.exit(0)}catch(o){n.fail(`Failed to update package.json: ${o}`),process.exit(1)}},un=(r,t,e)=>new Promise(n=>{try{let a=function(l){if(l.statusCode!==200){e.info(`Package not found. Status code: ${C.red(l.statusCode)}, defaulting to version 0`),n(0);return}e.text="Processing version information...";let m="";l.on("data",u=>{m+=u}),l.on("end",()=>{try{let u=JSON.parse(m),c=u.versions?Object.keys(u.versions):[];e.text="Calculating next version number...";let d=-1,f=new RegExp(`^${r}\\.(\\d+)$`);for(let p of c){let w=p.match(f);if(w){let T=Number.parseInt(w[1],10);d=Math.max(d,T)}}d>=0?e.info(`Found existing version ${C.green(`${r}.${d}`)}. Incrementing to ${C.green(`${r}.${d+1}`)}`):e.info(`No existing version found for ${C.green(r)}. Starting with ${C.green(`${r}.0`)}`),n(d+1)}catch(u){e.warn(`Failed to parse NPM registry response: ${u instanceof Error?u.message:"Unknown error"}, defaulting to version 0`),n(0)}})};var o=a;let s=`https://registry.npmjs.org/${t}`,i=It(s,{headers:{Accept:"application/json"}},l=>{if((l.statusCode===301||l.statusCode===302)&&l.headers.location){e.info(`Following redirect to ${l.headers.location}...`);try{It(l.headers.location,{headers:{Accept:"application/json"}},a).on("error",m=>{e.warn(`Error following redirect: ${m.message}, defaulting to version 0`),n(0)}).end()}catch(m){e.warn(`Redirect request failed: ${m instanceof Error?m.message:"Unknown error"}, defaulting to version 0`),n(0)}return}a(l)});i.on("error",l=>{e.warn(`Request error: ${l.message}, defaulting to version 0`),n(0)}),i.end()}catch{e.warn("Error checking version in npm registry, defaulting to version 0"),n(0)}});import Ie from"node:path";import{fileURLToPath as dn}from"node:url";import{program as pn}from"commander";import fn from"open";var gn=dn(import.meta.url),hn=Ie.resolve(Ie.dirname(gn),".."),Bi=pn.command("test").option("-w, --watch","Set watch number for the test").option("-c, --coverage-report","Generate coverage report").option("-o, --open","Open the coverage report in the browser").option("-f, --files <String|String[]>","Test files glob pattern").description("Run the web test runner to test the component library").action(async r=>{let e=`npx wtr --config "${Ie.join(hn,"dist","configs","web-test-runner.config.mjs")}"`,n=`${process.cwd()}/coverage/index.html`;if(r.coverageReport&&(e+=" --coverage"),r.watch&&(e+=" --watch"),r.files){let o=Array.isArray(r.files)?r.files.join(" "):r.files;e+=` --files "${o}"`}b(e),r.open&&await fn(n)});import De from"node:fs/promises";import Dt from"node:path";import{program as An}from"commander";import En from"inquirer";import xe from"ora";import yn from"node:fs/promises";import wn from"node:path";import bn from"ora";var kn={"auth-type":"workspace-token","author-email":null,"author-name":null,"base-branch":"main","base-url":null,"clone-dir":".gitter-temp","code-search":null,concurrent:4,"conflict-strategy":"replace",draft:!1,"dry-run":!0,"fetch-depth":1,fork:!1,"fork-owner":null,"git-type":"go",group:null,"include-subgroups":!1,insecure:!1,interactive:!1,labels:null,"log-file":"'-'","log-format":"'text'","log-level":"'error'","max-reviewers":0,"max-team-reviewers":0,org:null,output:"'-'","plain-output":!1,platform:"github",project:null,"push-only":!1,repo:["AlaskaAirlines/auro-accordion","AlaskaAirlines/auro-alert","AlaskaAirlines/auro-avatar","AlaskaAirlines/auro-background","AlaskaAirlines/auro-backtotop","AlaskaAirlines/auro-button","AlaskaAirlines/auro-badge","AlaskaAirlines/auro-banner","AlaskaAirlines/auro-card","AlaskaAirlines/auro-carousel","AlaskaAirlines/auro-datetime","AlaskaAirlines/auro-dialog","AlaskaAirlines/auro-drawer","AlaskaAirlines/auro-flight","AlaskaAirlines/auro-flightline","AlaskaAirlines/auro-header","AlaskaAirlines/auro-hyperlink","AlaskaAirlines/auro-icon","AlaskaAirlines/auro-loader","AlaskaAirlines/auro-lockup","AlaskaAirlines/auro-nav","AlaskaAirlines/auro-pane","AlaskaAirlines/auro-popover","AlaskaAirlines/auro-sidenav","AlaskaAirlines/auro-skeleton","AlaskaAirlines/auro-slideshow","AlaskaAirlines/auro-table","AlaskaAirlines/auro-tabs","AlaskaAirlines/auro-toast"],"repo-exclude":null,"repo-include":null,"repo-search":null,reviewers:null,"skip-forks":!1,"skip-pr":!1,"skip-repo":null,"ssh-auth":!1,"team-reviewers":null};function $n(r){return Object.entries(r).map(([t,e])=>Array.isArray(e)?`${t}:
45
45
  - ${e.join(`
46
46
  - `)}`:typeof e=="object"&&e!==null?`${t}:
47
47
  ${Object.entries(e).map(([n,o])=>` ${n}: ${o}`).join(`
48
48
  `)}`:`${t}: ${e}`).join(`
49
- `)}async function Et(r){let t=gn("Writing multi-gitter configuration...").start(),e=yn(hn),n=fn.join(r,"multi-gitter_DEPENDENCY_TREE.yml");try{await pn.writeFile(n,e,"utf8"),t.succeed(`Multi-gitter configuration written to ${n}`)}catch(o){t.fail("Error writing multi-gitter configuration:"),console.error(o)}}import Te from"node:fs";function Ct(r){let t={},e=[],n=[],o=[];for(let s in r)t[s]=r[s].dependsOn.length;for(let s in t)t[s]===0&&o.push(s);for(;o.length>0;){n=[];let s=o.length;for(let i=0;i<s;i++){let a=o.shift();n.push(a);for(let l of r[a].dependentPackages)t[l]--,t[l]===0&&o.push(l)}e.push(n)}if(e.flat().length!==Object.keys(r).length)throw new Error("Circular dependency detected!");return e}function wn(r){return Te.readdirSync(r).filter(t=>t.endsWith(".json"))}async function It(r,t=[]){console.log(t);let e={},n=wn(r);for(let o of n){if(o==="dependencyTree.json")continue;let s=Te.readFileSync(`${r}/${o}`,"utf-8"),i=JSON.parse(s),a=i.name,l=Object.keys(i.peerDependencies),u=Object.keys(i.devDependencies),m=Object.keys(i.dependencies);e[a]||(e[a]={dependsOn:[],dependentPackages:[]});let c=[...l,...u,...m];e[a].dependsOn=[...new Set(c)];for(let d of c)e[d]||(e[d]={dependsOn:[],dependentPackages:[]}),e[d].dependentPackages.includes(a)||e[d].dependentPackages.push(a)}if(t.length){let o=new Set;for(let[i,a]of Object.entries(e))a.dependsOn.some(l=>t.includes(l))&&o.add(i);for(let i of t)e[i]&&o.add(i);let s={};for(let i of o)s[i]={dependsOn:e[i].dependsOn.filter(a=>o.has(a)),dependentPackages:e[i].dependentPackages.filter(a=>o.has(a))};e=s}else console.log("No target dependencies provided - using all packages.");return Te.writeFileSync(`${r}/dependencyTree.json`,JSON.stringify(e,null,2)),e}var Se=ce("run-migrations","config"),Pe=ce("run-migrations","outputs");var St=["@aurodesignsystem/auro-accordion","@aurodesignsystem/auro-alert","@aurodesignsystem/auro-avatar","@aurodesignsystem/auro-background","@aurodesignsystem/auro-backtotop","@aurodesignsystem/auro-button","@aurodesignsystem/auro-badge","@aurodesignsystem/auro-banner","@aurodesignsystem/auro-card","@aurodesignsystem/auro-carousel","@aurodesignsystem/auro-datetime","@aurodesignsystem/auro-dialog","@aurodesignsystem/auro-drawer","@aurodesignsystem/auro-formkit","@aurodesignsystem/auro-flight","@aurodesignsystem/auro-flightline","@aurodesignsystem/auro-header","@aurodesignsystem/auro-hyperlink","@aurodesignsystem/auro-icon","@aurodesignsystem/auro-loader","@aurodesignsystem/auro-lockup","@aurodesignsystem/auro-nav","@aurodesignsystem/auro-pane","@aurodesignsystem/auro-popover","@aurodesignsystem/auro-sidenav","@aurodesignsystem/auro-skeleton","@aurodesignsystem/auro-slideshow","@aurodesignsystem/auro-table","@aurodesignsystem/auro-tabs","@aurodesignsystem/auro-toast"],Hi=[...St,"@aurodesignsystem/auro-library","@aurodesignsystem/WebCoreStyleSheets","@aurodesignsystem/AuroDesignTokens","@aurodesignsystem/auro-cli","@alaskaairux/icons"];async function $n(r){try{await De.mkdir(Pe,{recursive:!0}),await De.mkdir(Se,{recursive:!0})}catch(o){console.error("Failed to create output or config directories:",o),process.exit(1)}let t=Re("Creating dependency tree...").start();t.text="Creating multi-gitter dependency tree configuration...",await Et(Se),t.text="Scraping dependencies from Auro packages...";let n=`multi-gitter run "node ${le("static","getAuroDeps.js")}" --config ${Tt.join(Se,"multi-gitter_DEPENDENCY_TREE.yml")}`;try{await b(n)}catch(o){t.fail("Failed to generate dependency tree:"),console.error(o),process.exit(1)}return t.text="Generating dependency tree JSON file using packages...",await It(Pe,r),t.succeed("Dependency tree generated successfully."),Tt.join(Pe,"dependencyTree.json")}var vn=async r=>{let t=Re("Loading dependency tree...").start(),e=JSON.parse(await De.readFile(r,"utf-8"));t.text="Processing dependency tree...";let n=Ct(e);return t.succeed("Dependency batches created successfully."),n},Wi=bn.command("agent").action(async r=>{let t=await kn.prompt([{type:"select",name:"agentAction",message:"What agent action do you want to perform?",choices:[{name:"Run a migration on auro components",value:"run-migration"}],default:["run-migration"]},{type:"input",name:"migrationId",message:"What migration id do you want to run?",when:e=>e.agentAction==="run-migration",validate:e=>e.trim()!==""||"Migration id cannot be empty."},{type:"confirm",name:"useExisting",message:"Would you like to specify starting packages?",default:!0,transformer:e=>e?"Yes = Packages related to selections":"No = All packages",when:e=>e.agentAction==="run-migration"},{type:"checkbox",name:"startWithComponents",message:"Enter the components to start with (comma-separated, blank for all):",choices:St.map(e=>({name:e.replace("@aurodesignsystem/",""),value:e})),when:e=>e.agentAction==="run-migration"&&e.useExisting}]);switch(t.agentAction){case"run-migration":{let e=Re("Running migration...").start(),n=await $n(t.startWithComponents);e.text="Getting dependency batches from tree...";let s=(await vn(n)).map((i,a)=>`Batch ${a+1}
49
+ `)}async function St(r){let t=bn("Writing multi-gitter configuration...").start(),e=$n(kn),n=wn.join(r,"multi-gitter_DEPENDENCY_TREE.yml");try{await yn.writeFile(n,e,"utf8"),t.succeed(`Multi-gitter configuration written to ${n}`)}catch(o){t.fail("Error writing multi-gitter configuration:"),console.error(o)}}import Se from"node:fs";function Pt(r){let t={},e=[],n=[],o=[];for(let s in r)t[s]=r[s].dependsOn.length;for(let s in t)t[s]===0&&o.push(s);for(;o.length>0;){n=[];let s=o.length;for(let i=0;i<s;i++){let a=o.shift();n.push(a);for(let l of r[a].dependentPackages)t[l]--,t[l]===0&&o.push(l)}e.push(n)}if(e.flat().length!==Object.keys(r).length)throw new Error("Circular dependency detected!");return e}function vn(r){return Se.readdirSync(r).filter(t=>t.endsWith(".json"))}async function Rt(r,t=[]){console.log(t);let e={},n=vn(r);for(let o of n){if(o==="dependencyTree.json")continue;let s=Se.readFileSync(`${r}/${o}`,"utf-8"),i=JSON.parse(s),a=i.name,l=Object.keys(i.peerDependencies),m=Object.keys(i.devDependencies),u=Object.keys(i.dependencies);e[a]||(e[a]={dependsOn:[],dependentPackages:[]});let c=[...l,...m,...u];e[a].dependsOn=[...new Set(c)];for(let d of c)e[d]||(e[d]={dependsOn:[],dependentPackages:[]}),e[d].dependentPackages.includes(a)||e[d].dependentPackages.push(a)}if(t.length){let o=new Set;for(let[i,a]of Object.entries(e))a.dependsOn.some(l=>t.includes(l))&&o.add(i);for(let i of t)e[i]&&o.add(i);let s={};for(let i of o)s[i]={dependsOn:e[i].dependsOn.filter(a=>o.has(a)),dependentPackages:e[i].dependentPackages.filter(a=>o.has(a))};e=s}else console.log("No target dependencies provided - using all packages.");return Se.writeFileSync(`${r}/dependencyTree.json`,JSON.stringify(e,null,2)),e}var Pe=ue("run-migrations","config"),Re=ue("run-migrations","outputs");var xt=["@aurodesignsystem/auro-accordion","@aurodesignsystem/auro-alert","@aurodesignsystem/auro-avatar","@aurodesignsystem/auro-background","@aurodesignsystem/auro-backtotop","@aurodesignsystem/auro-button","@aurodesignsystem/auro-badge","@aurodesignsystem/auro-banner","@aurodesignsystem/auro-card","@aurodesignsystem/auro-carousel","@aurodesignsystem/auro-datetime","@aurodesignsystem/auro-dialog","@aurodesignsystem/auro-drawer","@aurodesignsystem/auro-formkit","@aurodesignsystem/auro-flight","@aurodesignsystem/auro-flightline","@aurodesignsystem/auro-header","@aurodesignsystem/auro-hyperlink","@aurodesignsystem/auro-icon","@aurodesignsystem/auro-loader","@aurodesignsystem/auro-lockup","@aurodesignsystem/auro-nav","@aurodesignsystem/auro-pane","@aurodesignsystem/auro-popover","@aurodesignsystem/auro-sidenav","@aurodesignsystem/auro-skeleton","@aurodesignsystem/auro-slideshow","@aurodesignsystem/auro-table","@aurodesignsystem/auro-tabs","@aurodesignsystem/auro-toast"],ta=[...xt,"@aurodesignsystem/auro-library","@aurodesignsystem/WebCoreStyleSheets","@aurodesignsystem/AuroDesignTokens","@aurodesignsystem/auro-cli","@alaskaairux/icons"];async function Cn(r){try{await De.mkdir(Re,{recursive:!0}),await De.mkdir(Pe,{recursive:!0})}catch(o){console.error("Failed to create output or config directories:",o),process.exit(1)}let t=xe("Creating dependency tree...").start();t.text="Creating multi-gitter dependency tree configuration...",await St(Pe),t.text="Scraping dependencies from Auro packages...";let n=`multi-gitter run "node ${de("static","getAuroDeps.js")}" --config ${Dt.join(Pe,"multi-gitter_DEPENDENCY_TREE.yml")}`;try{await b(n)}catch(o){t.fail("Failed to generate dependency tree:"),console.error(o),process.exit(1)}return t.text="Generating dependency tree JSON file using packages...",await Rt(Re,r),t.succeed("Dependency tree generated successfully."),Dt.join(Re,"dependencyTree.json")}var Tn=async r=>{let t=xe("Loading dependency tree...").start(),e=JSON.parse(await De.readFile(r,"utf-8"));t.text="Processing dependency tree...";let n=Pt(e);return t.succeed("Dependency batches created successfully."),n},ra=An.command("agent").action(async r=>{let t=await En.prompt([{type:"select",name:"agentAction",message:"What agent action do you want to perform?",choices:[{name:"Run a migration on auro components",value:"run-migration"}],default:["run-migration"]},{type:"input",name:"migrationId",message:"What migration id do you want to run?",when:e=>e.agentAction==="run-migration",validate:e=>e.trim()!==""||"Migration id cannot be empty."},{type:"confirm",name:"useExisting",message:"Would you like to specify starting packages?",default:!0,transformer:e=>e?"Yes = Packages related to selections":"No = All packages",when:e=>e.agentAction==="run-migration"},{type:"checkbox",name:"startWithComponents",message:"Enter the components to start with (comma-separated, blank for all):",choices:xt.map(e=>({name:e.replace("@aurodesignsystem/",""),value:e})),when:e=>e.agentAction==="run-migration"&&e.useExisting}]);switch(t.agentAction){case"run-migration":{let e=xe("Running migration...").start(),n=await Cn(t.startWithComponents);e.text="Getting dependency batches from tree...";let s=(await Tn(n)).map((i,a)=>`Batch ${a+1}
50
50
  ${i.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alaskaairux/icons","AlaskaAirlines/Icons")}`).join(`
51
51
  `)}`).join(`
52
52
 
53
- `);console.log(s),e.text="Running migrations on dependency batches...",new Promise(i=>setTimeout(i,2e3)),e.succeed("Migration process completed successfully.");break}default:console.error("Unknown action selected.")}});import{program as An}from"commander";var xe=An.command("docs").description("Generate API documentation").option("-c, --cem","Generate Custom Elements Manifest (CEM) file",!1).option("-a, --api","Creates api md file from CEM",!1).option("-w, --watch","Watch for changes and rebuild docs",!1).option("--skip-readme","Skip README.md processing",!1);xe=H(xe);var Vi=xe.action(async r=>{r.cem&&await V(),r.api&&await J(),await ze(r),r.serve&&await Ve(r),r.watch&&await Je(r)});import{program as Sn}from"commander";import{Octokit as Oe}from"@octokit/rest";import*as oe from"azure-devops-node-api";import ne from"ora";var En=async r=>{let t=process.env.GH_TOKEN;if(!t)throw new Error("GH_TOKEN environment variable is required");let e=new Oe({auth:t}),n,o,s;if(r.includes("github.com")){let a=r.match(/github\.com\/([^\/]+)\/([^\/]+)\/issues\/(\d+)/);if(!a)throw new Error("Invalid GitHub issue URL format");[,n,o,s]=a}else if(r.includes("#")){let a=r.match(/([^\/]+)\/([^#]+)#(\d+)/);if(!a)throw new Error("Invalid GitHub issue reference format");[,n,o,s]=a}else throw new Error("Issue must be provided as full URL or in format 'owner/repo#number'");let i=Number.parseInt(s,10);try{let{data:a}=await e.rest.issues.get({owner:n,repo:o,issue_number:i});return{title:a.title,body:a.body??null,html_url:a.html_url,number:a.number,repository:{owner:{login:n},name:o}}}catch(a){throw new Error(`Failed to fetch GitHub issue: ${a}`)}},Cn=async r=>{let t=process.env.GH_TOKEN;if(!t)return null;let e=new Oe({auth:t});try{let n=`
53
+ `);console.log(s),e.text="Running migrations on dependency batches...",new Promise(i=>setTimeout(i,2e3)),e.succeed("Migration process completed successfully.");break}default:console.error("Unknown action selected.")}});import{program as In}from"commander";var Oe=In.command("docs").description("Generate API documentation").option("-c, --cem","Generate Custom Elements Manifest (CEM) file",!1).option("-a, --api","Creates api md file from CEM",!1).option("-w, --watch","Watch for changes and rebuild docs",!1).option("-r, --readme-template <url>","URL to the README template file").option("--skip-readme","Skip README.md processing",!1);Oe=L(Oe);var aa=Oe.action(async r=>{r.cem&&await Y(),r.api&&await X(),await Ye(r),r.serve&&await Xe(r),r.watch&&await Ze(r)});import{program as xn}from"commander";import{Octokit as je}from"@octokit/rest";import*as ae from"azure-devops-node-api";import ie from"ora";var Sn=async r=>{let t=process.env.GH_TOKEN;if(!t)throw new Error("GH_TOKEN environment variable is required");let e=new je({auth:t}),n,o,s;if(r.includes("github.com")){let a=r.match(/github\.com\/([^\/]+)\/([^\/]+)\/issues\/(\d+)/);if(!a)throw new Error("Invalid GitHub issue URL format");[,n,o,s]=a}else if(r.includes("#")){let a=r.match(/([^\/]+)\/([^#]+)#(\d+)/);if(!a)throw new Error("Invalid GitHub issue reference format");[,n,o,s]=a}else throw new Error("Issue must be provided as full URL or in format 'owner/repo#number'");let i=Number.parseInt(s,10);try{let{data:a}=await e.rest.issues.get({owner:n,repo:o,issue_number:i});return{title:a.title,body:a.body??null,html_url:a.html_url,number:a.number,repository:{owner:{login:n},name:o}}}catch(a){throw new Error(`Failed to fetch GitHub issue: ${a}`)}},Pn=async r=>{let t=process.env.GH_TOKEN;if(!t)return null;let e=new je({auth:t});try{let n=`
54
54
  query($owner: String!, $repo: String!, $issueNumber: Int!) {
55
55
  repository(owner: $owner, name: $repo) {
56
56
  issue(number: $issueNumber) {
@@ -76,7 +76,7 @@ ${i.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
76
76
  }
77
77
  }
78
78
  }
79
- `,o={owner:r.repository.owner.login,repo:r.repository.name,issueNumber:r.number},i=(await e.graphql(n,o)).repository.issue.projectItems.nodes.find(a=>a.project.number===19);if(i){let a=i.fieldValues.nodes.find(l=>l.field?.name?.toLowerCase()==="ado"&&l.text?.trim());if(a?.text?.trim())return a.text.trim()}return null}catch(n){return console.error(`Failed to check existing ADO link: ${n}`),null}},In=async(r,t)=>{let e=process.env.GH_TOKEN;if(!e)throw new Error("GH_TOKEN environment variable is required");let n=new Oe({auth:e}),o=19;try{let s=`
79
+ `,o={owner:r.repository.owner.login,repo:r.repository.name,issueNumber:r.number},i=(await e.graphql(n,o)).repository.issue.projectItems.nodes.find(a=>a.project.number===19);if(i){let a=i.fieldValues.nodes.find(l=>l.field?.name?.toLowerCase()==="ado"&&l.text?.trim());if(a?.text?.trim())return a.text.trim()}return null}catch(n){return console.error(`Failed to check existing ADO link: ${n}`),null}},Rn=async(r,t)=>{let e=process.env.GH_TOKEN;if(!e)throw new Error("GH_TOKEN environment variable is required");let n=new je({auth:e}),o=19;try{let s=`
80
80
  query($org: String!, $projectNumber: Int!, $owner: String!, $repo: String!, $issueNumber: Int!) {
81
81
  organization(login: $org) {
82
82
  projectV2(number: $projectNumber) {
@@ -113,7 +113,7 @@ ${i.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
113
113
  }
114
114
  }
115
115
  }
116
- `,i={org:"AlaskaAirlines",projectNumber:o,owner:r.repository.owner.login,repo:r.repository.name,issueNumber:r.number},a=await n.graphql(s,i),l=a.organization.projectV2.id,u=a.repository.issue.id,m=a.organization.projectV2.fields.nodes.find(d=>d.name?.toLowerCase()==="ado"),c=a.repository.issue.projectItems.nodes.find(d=>d.project.number===o)?.id;if(c||(c=(await n.graphql(`
116
+ `,i={org:"AlaskaAirlines",projectNumber:o,owner:r.repository.owner.login,repo:r.repository.name,issueNumber:r.number},a=await n.graphql(s,i),l=a.organization.projectV2.id,m=a.repository.issue.id,u=a.organization.projectV2.fields.nodes.find(d=>d.name?.toLowerCase()==="ado"),c=a.repository.issue.projectItems.nodes.find(d=>d.project.number===o)?.id;if(c||(c=(await n.graphql(`
117
117
  mutation($projectId: ID!, $contentId: ID!) {
118
118
  addProjectV2ItemById(
119
119
  input: {
@@ -126,7 +126,7 @@ ${i.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
126
126
  }
127
127
  }
128
128
  }
129
- `,{projectId:l,contentId:u})).addProjectV2ItemById.item.id),m&&c)await n.graphql(`
129
+ `,{projectId:l,contentId:m})).addProjectV2ItemById.item.id),u&&c)await n.graphql(`
130
130
  mutation($projectId: ID!, $itemId: ID!, $fieldId: ID!, $value: String!) {
131
131
  updateProjectV2ItemFieldValue(
132
132
  input: {
@@ -143,4 +143,4 @@ ${i.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
143
143
  }
144
144
  }
145
145
  }
146
- `,{projectId:l,itemId:c,fieldId:m.id,value:t});else if(!m)throw new Error("No 'ado' field found in GitHub project")}catch(s){console.error(`Failed to update GitHub project: ${s}`)}},Tn=async r=>{let t=process.env.ADO_TOKEN;if(!t)throw new Error("ADO_TOKEN environment variable is required");let e="https://dev.azure.com/itsals",n="E_Retain_Content",o="E_Retain_Content\\Auro Design System",s=oe.getPersonalAccessTokenHandler(t),a=await new oe.WebApi(e,s).getWorkItemTrackingApi();try{let l=[{op:"add",path:"/fields/System.Title",value:r.title},{op:"add",path:"/fields/System.Description",value:`GitHub Issue: <a href="${r.html_url}">${r.html_url}</a>`},{op:"add",path:"/fields/System.AreaPath",value:o}];return await a.createWorkItem(null,l,n,"User Story")}catch(l){throw new Error(`Failed to create ADO work item: ${l}`)}},Pt=async r=>{let t=ne(`Processing GitHub issue: ${r}`).start();try{if(!process.env.GH_TOKEN)throw new Error("GH_TOKEN environment variable is required");if(!process.env.ADO_TOKEN)throw new Error("ADO_TOKEN environment variable is required");t.text="Fetching GitHub issue details...";let e=await En(r);t.succeed(`Found issue: "${e.title}"`);let n=ne("Checking for existing ADO work item...").start(),o=await Cn(e);if(o){n.succeed("ADO work item already exists for this issue!"),console.log(`${o}`);return}n.succeed("No existing ADO work item found");let s=ne("Creating new ADO work item...").start(),i=await Tn(e);if(s.succeed(`Successfully created ADO work item #${i.id}`),console.log(`Work item: ${i._links?.html?.href||"N/A"}`),i._links?.html?.href){let a=ne("Adding to GitHub project and updating ADO field...").start();await In(e,i._links.html.href),a.succeed("Updated GitHub project with ADO link")}}catch(e){t.fail(`Error: ${e instanceof Error?e.message:e}`),process.exit(1)}};var ea=Sn.command("ado").description("Generate ADO item from GitHub issue").option("-g, --gh-issue <issue>","What GitHub issue to use").action(async r=>{r.ghIssue&&await Pt(r.ghIssue)});import{program as xn}from"commander";import{Octokit as Pn}from"@octokit/rest";import{simpleGit as Dn}from"simple-git";var Dt="Release Candidate",B="dev",Rn="main",se=class r{constructor(t,e,n){this.filteredCommits=null;this.repoInfo={owner:t,repo:e},this.octokit=n}static async create(){let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN is required to run RC workflow.");let e=await E.getRepoOwnerAndName(),n=new Pn({auth:t});if(!e)throw new Error("Failed to retrieve repository information. Ensure you're in a valid git repository.");let o=await r.getTriggerBranchName();return o&&o!==B&&(console.log(`Switching from ${o} to ${B} branch...`),await Dn().checkout(B)),new r(e.owner,e.repo,n)}get owner(){return this.repoInfo.owner}get repo(){return this.repoInfo.repo}get repoData(){return{...this.repoInfo}}async createReleaseCandidate(){await this.hasCommitsReadyInDev()||console.log("No filtered commits found. Continuing to update RC issue/branch/PR.");let e=await this.getLatestOpenRcIssue(),n=e?await this.getLinkedPrByHead(e.number):null;if(n?.multipleOpen)throw new Error("Multiple open RC PRs found for the same rc/<issueNumber> branch.");if(n?.state==="closed"&&(console.log("Linked RC PR is closed. Creating a new RC issue and PR."),e=await this.createRcIssue(),n=null),e?await this.updateRcIssue(e.number):e=await this.createRcIssue(),!e)throw new Error("Failed to resolve RC issue.");await this.createOrUpdateRcBranch(e.number),n?await this.updateRcPr(e.number,n.number):n=await this.createRcPr(e.number)}async getFilteredCommits(){if(this.filteredCommits===null){let t=await E.getCommitMessages(B);this.filteredCommits=Ce(t)}return this.filteredCommits}async hasCommitsReadyInDev(){return(await this.getFilteredCommits()).length>0}async getLatestOpenRcIssue(){let{data:t}=await this.octokit.rest.issues.listForRepo({owner:this.repoInfo.owner,repo:this.repoInfo.repo,labels:Dt,state:"open",sort:"updated",direction:"desc",per_page:30}),e=t.filter(o=>!o.pull_request);if(e.length===0)return console.log(`No open Release Candidate issues found in ${this.repoInfo.repo}`),null;let n=e[0];return console.log(`Using latest open Release Candidate issue: #${n.number}`),{number:n.number,title:n.title||""}}async updateRcIssue(t){let e=await this.getReleaseNotes(),n=`RC ${this.getCurrentDate()}`;await this.octokit.rest.issues.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,issue_number:t,title:n,body:e})}async createRcIssue(){let t=await this.getReleaseNotes(),{data:e}=await this.octokit.rest.issues.create({owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC ${this.getCurrentDate()}`,labels:[Dt],body:t});return console.log(`Created Release Candidate issue: #${e.number} (${e.html_url})`),{number:e.number,html_url:e.html_url}}async createOrUpdateRcBranch(t){let e=`heads/rc/${t}`,n=`rc/${t}`,{data:o}=await this.octokit.rest.repos.getBranch({owner:this.repoInfo.owner,repo:this.repoInfo.repo,branch:B}),{data:s}=await this.octokit.rest.git.listMatchingRefs({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e}),i=s.length>0;try{i?(console.log(`Updating existing RC branch: ${n}`),await this.octokit.rest.git.updateRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e,sha:o.commit.sha,force:!0})):(console.log(`Creating new RC branch: ${n}`),await this.octokit.rest.git.createRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:`refs/${e}`,sha:o.commit.sha}))}catch(a){throw new Error(`Failed to create or update ${n} branch: ${a}`)}}async getLinkedPrByHead(t){let e=`${this.repoInfo.owner}:rc/${t}`,{data:n}=await this.octokit.rest.pulls.list({owner:this.repoInfo.owner,repo:this.repoInfo.repo,state:"all",head:e,per_page:30}),o=n.filter(i=>i.state==="open");if(o.length>1)return{state:"open",multipleOpen:!0};if(o.length===1)return{state:"open",html_url:o[0].html_url,number:o[0].number};let s=n.filter(i=>i.state==="closed");return s.length>0?{state:"closed",html_url:s[0].html_url,number:s[0].number}:null}async fetchPrTemplate(t){try{let{data:e}=await this.octokit.rest.repos.getContent({owner:this.repoInfo.owner,repo:this.repoInfo.repo,path:".github/PULL_REQUEST_TEMPLATE.md"});if("content"in e&&e.type==="file"){let n=Buffer.from(e.content,"base64").toString("utf-8");return n=n.replace("Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.",`Release candidate pull request. See issue #${t} for details.`),n=n.replace(/<details>/g,"<details open>"),n}}catch(e){e&&typeof e=="object"&&"status"in e&&e.status===404?console.log("No PR template found in repo, using default message."):console.warn("Failed to fetch PR template:",e)}return`Release candidate pull request. See issue #${t} for details.`}async createRcPr(t){try{let e=await this.fetchPrTemplate(t),{data:n}=await this.octokit.request(`POST /repos/${this.repoInfo.owner}/${this.repoInfo.repo}/pulls`,{owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC #${t}`,body:e,head:`rc/${t}`,base:Rn,headers:{"X-GitHub-Api-Version":"2022-11-28"}});return console.log(`Created Release Candidate pull request: #${n.number} (${n.html_url})`),{state:"open",html_url:n.html_url,number:n.number}}catch(e){throw console.error("Failed to create RC PR:",e),e}}async updateRcPr(t,e){try{let n=await this.fetchPrTemplate(t);await this.octokit.rest.pulls.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,pull_number:e,body:n}),console.log(`Updated Release Candidate pull request: #${e}`)}catch(n){throw console.error("Failed to update RC PR:",n),n}}async getReleaseNotes(){let t=await this.getFilteredCommits();return Ee(t,!1)}static async getTriggerBranchName(){return process.env.GITHUB_REF_NAME?process.env.GITHUB_REF_NAME:process.env.GITHUB_REF?.startsWith("refs/heads/")?process.env.GITHUB_REF.replace("refs/heads/",""):E.getCurrentBranchName()}getCurrentDate(){return new Date().toISOString().split("T")[0]}};var la=xn.command("rc-workflow").description("Generate RC issue and pull request").action(async()=>{await(await se.create()).createReleaseCandidate()});je.name("auro").version(ie()).description("A cli tool to support the Auro Design System");je.addHelpText("beforeAll",_e());je.parse();
146
+ `,{projectId:l,itemId:c,fieldId:u.id,value:t});else if(!u)throw new Error("No 'ado' field found in GitHub project")}catch(s){console.error(`Failed to update GitHub project: ${s}`)}},Dn=async r=>{let t=process.env.ADO_TOKEN;if(!t)throw new Error("ADO_TOKEN environment variable is required");let e="https://dev.azure.com/itsals",n="E_Retain_Content",o="E_Retain_Content\\Auro Design System",s=ae.getPersonalAccessTokenHandler(t),a=await new ae.WebApi(e,s).getWorkItemTrackingApi();try{let l=[{op:"add",path:"/fields/System.Title",value:r.title},{op:"add",path:"/fields/System.Description",value:`GitHub Issue: <a href="${r.html_url}">${r.html_url}</a>`},{op:"add",path:"/fields/System.AreaPath",value:o}];return await a.createWorkItem(null,l,n,"User Story")}catch(l){throw new Error(`Failed to create ADO work item: ${l}`)}},Ot=async r=>{let t=ie(`Processing GitHub issue: ${r}`).start();try{if(!process.env.GH_TOKEN)throw new Error("GH_TOKEN environment variable is required");if(!process.env.ADO_TOKEN)throw new Error("ADO_TOKEN environment variable is required");t.text="Fetching GitHub issue details...";let e=await Sn(r);t.succeed(`Found issue: "${e.title}"`);let n=ie("Checking for existing ADO work item...").start(),o=await Pn(e);if(o){n.succeed("ADO work item already exists for this issue!"),console.log(`${o}`);return}n.succeed("No existing ADO work item found");let s=ie("Creating new ADO work item...").start(),i=await Dn(e);if(s.succeed(`Successfully created ADO work item #${i.id}`),console.log(`Work item: ${i._links?.html?.href||"N/A"}`),i._links?.html?.href){let a=ie("Adding to GitHub project and updating ADO field...").start();await Rn(e,i._links.html.href),a.succeed("Updated GitHub project with ADO link")}}catch(e){t.fail(`Error: ${e instanceof Error?e.message:e}`),process.exit(1)}};var fa=xn.command("ado").description("Generate ADO item from GitHub issue").option("-g, --gh-issue <issue>","What GitHub issue to use").action(async r=>{r.ghIssue&&await Ot(r.ghIssue)});import{program as Nn}from"commander";import{Octokit as On}from"@octokit/rest";import{simpleGit as jn}from"simple-git";var jt="Release Candidate",G="dev",_n="main",ce=class r{constructor(t,e,n){this.filteredCommits=null;this.repoInfo={owner:t,repo:e},this.octokit=n}static async create(){let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN is required to run RC workflow.");let e=await A.getRepoOwnerAndName(),n=new On({auth:t});if(!e)throw new Error("Failed to retrieve repository information. Ensure you're in a valid git repository.");let o=await r.getTriggerBranchName();return o&&o!==G&&(console.log(`Switching from ${o} to ${G} branch...`),await jn().checkout(G)),new r(e.owner,e.repo,n)}get owner(){return this.repoInfo.owner}get repo(){return this.repoInfo.repo}get repoData(){return{...this.repoInfo}}async createReleaseCandidate(){await this.hasCommitsReadyInDev()||console.log("No filtered commits found. Continuing to update RC issue/branch/PR.");let e=await this.getLatestOpenRcIssue(),n=e?await this.getLinkedPrByHead(e.number):null;if(n?.multipleOpen)throw new Error("Multiple open RC PRs found for the same rc/<issueNumber> branch.");if(n?.state==="closed"&&(console.log("Linked RC PR is closed. Creating a new RC issue and PR."),e=await this.createRcIssue(),n=null),e?await this.updateRcIssue(e.number):e=await this.createRcIssue(),!e)throw new Error("Failed to resolve RC issue.");await this.createOrUpdateRcBranch(e.number),n?await this.updateRcPr(e.number,n.number):n=await this.createRcPr(e.number)}async getFilteredCommits(){if(this.filteredCommits===null){let t=await A.getCommitMessages(G);this.filteredCommits=Te(t)}return this.filteredCommits}async hasCommitsReadyInDev(){return(await this.getFilteredCommits()).length>0}async getLatestOpenRcIssue(){let{data:t}=await this.octokit.rest.issues.listForRepo({owner:this.repoInfo.owner,repo:this.repoInfo.repo,labels:jt,state:"open",sort:"updated",direction:"desc",per_page:30}),e=t.filter(o=>!o.pull_request);if(e.length===0)return console.log(`No open Release Candidate issues found in ${this.repoInfo.repo}`),null;let n=e[0];return console.log(`Using latest open Release Candidate issue: #${n.number}`),{number:n.number,title:n.title||""}}async updateRcIssue(t){let e=await this.getReleaseNotes(),n=`RC ${this.getCurrentDate()}`;await this.octokit.rest.issues.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,issue_number:t,title:n,body:e})}async createRcIssue(){let t=await this.getReleaseNotes(),{data:e}=await this.octokit.rest.issues.create({owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC ${this.getCurrentDate()}`,labels:[jt],body:t});return console.log(`Created Release Candidate issue: #${e.number} (${e.html_url})`),{number:e.number,html_url:e.html_url}}async createOrUpdateRcBranch(t){let e=`heads/rc/${t}`,n=`rc/${t}`,{data:o}=await this.octokit.rest.repos.getBranch({owner:this.repoInfo.owner,repo:this.repoInfo.repo,branch:G}),{data:s}=await this.octokit.rest.git.listMatchingRefs({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e}),i=s.length>0;try{i?(console.log(`Updating existing RC branch: ${n}`),await this.octokit.rest.git.updateRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e,sha:o.commit.sha,force:!0})):(console.log(`Creating new RC branch: ${n}`),await this.octokit.rest.git.createRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:`refs/${e}`,sha:o.commit.sha}))}catch(a){throw new Error(`Failed to create or update ${n} branch: ${a}`)}}async getLinkedPrByHead(t){let e=`${this.repoInfo.owner}:rc/${t}`,{data:n}=await this.octokit.rest.pulls.list({owner:this.repoInfo.owner,repo:this.repoInfo.repo,state:"all",head:e,per_page:30}),o=n.filter(i=>i.state==="open");if(o.length>1)return{state:"open",multipleOpen:!0};if(o.length===1)return{state:"open",html_url:o[0].html_url,number:o[0].number};let s=n.filter(i=>i.state==="closed");return s.length>0?{state:"closed",html_url:s[0].html_url,number:s[0].number}:null}async fetchPrTemplate(t){try{let{data:e}=await this.octokit.rest.repos.getContent({owner:this.repoInfo.owner,repo:this.repoInfo.repo,path:".github/PULL_REQUEST_TEMPLATE.md"});if("content"in e&&e.type==="file"){let n=Buffer.from(e.content,"base64").toString("utf-8");return n=n.replace("Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.",`Release candidate pull request. See issue #${t} for details.`),n=n.replace(/<details>/g,"<details open>"),n}}catch(e){e&&typeof e=="object"&&"status"in e&&e.status===404?console.log("No PR template found in repo, using default message."):console.warn("Failed to fetch PR template:",e)}return`Release candidate pull request. See issue #${t} for details.`}async createRcPr(t){try{let e=await this.fetchPrTemplate(t),{data:n}=await this.octokit.request(`POST /repos/${this.repoInfo.owner}/${this.repoInfo.repo}/pulls`,{owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC #${t}`,body:e,head:`rc/${t}`,base:_n,headers:{"X-GitHub-Api-Version":"2022-11-28"}});return console.log(`Created Release Candidate pull request: #${n.number} (${n.html_url})`),{state:"open",html_url:n.html_url,number:n.number}}catch(e){throw console.error("Failed to create RC PR:",e),e}}async updateRcPr(t,e){try{let n=await this.fetchPrTemplate(t);await this.octokit.rest.pulls.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,pull_number:e,body:n}),console.log(`Updated Release Candidate pull request: #${e}`)}catch(n){throw console.error("Failed to update RC PR:",n),n}}async getReleaseNotes(){let t=await this.getFilteredCommits();return Ce(t,!1)}static async getTriggerBranchName(){return process.env.GITHUB_REF_NAME?process.env.GITHUB_REF_NAME:process.env.GITHUB_REF?.startsWith("refs/heads/")?process.env.GITHUB_REF.replace("refs/heads/",""):A.getCurrentBranchName()}getCurrentDate(){return new Date().toISOString().split("T")[0]}};var Aa=Nn.command("rc-workflow").description("Generate RC issue and pull request").action(async()=>{await(await ce.create()).createReleaseCandidate()});_e.name("auro").version(le()).description("A cli tool to support the Auro Design System");_e.addHelpText("beforeAll",Ne());_e.parse();