@aurodesignsystem-dev/auro-cli 0.0.0-pr266.1 → 0.0.0-pr271.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/auro-cli.js CHANGED
@@ -1,5 +1,5 @@
1
1
  #!/usr/bin/env node
2
- import{program as ve}from"commander";import ht from"figlet";import{mind as yt}from"gradient-string";var Ee=()=>yt(ht.textSync("Auro CLI"));import Ce from"node:fs";import Ie from"node:path";import{fileURLToPath as wt}from"node:url";function S(r){process.env.DEBUG&&console.log(`[DEBUG] ${r}`)}function Z(){try{let r=wt(import.meta.url),t=Ie.dirname(r);S(`Current module path: ${t}`);let e=Ie.resolve(t,"..","package.json");return S(`Checking package.json at: ${e}`),Ce.existsSync(e)?(S(`Found package.json at: ${e}`),JSON.parse(Ce.readFileSync(e,"utf8")).version):(S("Could not find package.json in the standard installed module location, using default version"),"0.0.0")}catch(r){return console.error("Error retrieving package version:",r),"0.0.0"}}import{program as Zt}from"commander";import qe from"ora";function O(r){return r.option("-m, --module-paths [paths...]","Path(s) to node_modules folder").option("-w, --watch","Watches for changes").option("--skip-docs","Skip documentation generation",!1).option("--wca-input [files...]","Source file(s) to analyze for API documentation").option("--wca-output [files...]","Output file(s) for API documentation")}function j(r){return r.option("-s, --serve","Starts a server").option("-p, --port <number>","Port for the server").option("-o, --open","Open the browser after starting the server")}import Kt from"@rollup/plugin-terser";import{watch as Jt}from"rollup";import{rmSync as Nt}from"node:fs";import{join as _t}from"node:path";import oe from"ora";import{rollup as je}from"rollup";import ne from"ora";import{spawn as bt}from"node:child_process";import kt from"ora";var b=(r,t)=>{let e=`${r} ${t?t.join(" "):""}`,n=kt(),o=r,i=t||[];if(!t&&typeof r=="string"){let m=r.split(" ");o=m[0],i=m.slice(1)}let s=e.includes("--watch")||e.includes(" -w"),l=bt(o,i,{stdio:s?"inherit":["inherit","pipe","pipe"],shell:!0});if(!s){let m=[];l.stdout?.on("data",u=>{let c=u.toString();m.push(c),process.stdout.write(c)}),l.stderr?.on("data",u=>{let c=u.toString();m.push(c),process.stderr.write(c)})}return new Promise((m,u)=>{l.on("close",c=>{c!==0?s?(n.info(`Watch mode terminated with code ${c}`),m()):(n.fail(`${e} failed (code ${c})`),u(new Error(`Command failed with exit code ${c}`))):(n.succeed(`${e} completed successfully`),m())})})};import N from"node:fs";import $t from"node:path";import{markdownTable as Pe}from"markdown-table";var I=class r{static{this.manifest={schemaVersion:"1.0.0",readme:"",modules:[]}}static generate(t={}){let{outDir:e="./docs",outFile:n="api.md",manifestPath:o="./custom-elements.json"}=t,{getElements:i,renderAllElements:s}=r;if(o)try{let c=N.readFileSync(o,"utf8");r.manifest=JSON.parse(c)}catch(c){throw console.error(`Error reading manifest file at ${o}:`,c),c}let a=i(),l=e;N.existsSync(l)||N.mkdirSync(l,{recursive:!0});let m=s(a),u=$t.join(l,n);N.writeFileSync(u,m),console.log(`Generated combined API documentation at ${u}`)}static getElements(){let t=r.manifest.modules.filter(r.isWcaModule);return r.manifest.modules.reduce((e,n)=>e.concat(n.declarations?.filter(o=>"customElement"in o&&o.customElement===!0&&"tagName"in o&&(t.length>0?r.isWcaModule(n):!0))??[]),[])}static isWcaModule(t){let{path:e}=t;return e?e.startsWith("scripts/wca/auro-")&&e.endsWith(".js"):!1}static renderAllElements(t){return`${t.sort((e,n)=>(e.tagName||"").localeCompare(n.tagName||"")).map(e=>r.renderElement(e,!0)).join(`
2
+ import{program as ve}from"commander";import yt from"figlet";import{mind as wt}from"gradient-string";var Ee=()=>wt(yt.textSync("Auro CLI"));import Ce from"node:fs";import Ie from"node:path";import{fileURLToPath as bt}from"node:url";function O(r){process.env.DEBUG&&console.log(`[DEBUG] ${r}`)}function Q(){try{let r=bt(import.meta.url),t=Ie.dirname(r);O(`Current module path: ${t}`);let e=Ie.resolve(t,"..","package.json");return O(`Checking package.json at: ${e}`),Ce.existsSync(e)?(O(`Found package.json at: ${e}`),JSON.parse(Ce.readFileSync(e,"utf8")).version):(O("Could not find package.json in the standard installed module location, using default version"),"0.0.0")}catch(r){return console.error("Error retrieving package version:",r),"0.0.0"}}import{program as Qt}from"commander";import qe from"ora";function j(r){return r.option("-m, --module-paths [paths...]","Path(s) to node_modules folder").option("-w, --watch","Watches for changes").option("--skip-docs","Skip documentation generation",!1).option("--wca-input [files...]","Source file(s) to analyze for API documentation").option("--wca-output [files...]","Output file(s) for API documentation")}function N(r){return r.option("-s, --serve","Starts a server").option("-p, --port <number>","Port for the server").option("-o, --open","Open the browser after starting the server")}import Jt from"@rollup/plugin-terser";import{watch as Yt}from"rollup";import{rmSync as Ft}from"node:fs";import{join as _t}from"node:path";import se from"ora";import{rollup as je}from"rollup";import oe from"ora";import{spawn as kt}from"node:child_process";import $t from"ora";var b=(r,t)=>{let e=`${r} ${t?t.join(" "):""}`,n=$t(),o=r,i=t||[];if(!t&&typeof r=="string"){let m=r.split(" ");o=m[0],i=m.slice(1)}let s=e.includes("--watch")||e.includes(" -w"),l=kt(o,i,{stdio:s?"inherit":["inherit","pipe","pipe"],shell:!0});if(!s){let m=[];l.stdout?.on("data",u=>{let c=u.toString();m.push(c),process.stdout.write(c)}),l.stderr?.on("data",u=>{let c=u.toString();m.push(c),process.stderr.write(c)})}return new Promise((m,u)=>{l.on("close",c=>{c!==0?s?(n.info(`Watch mode terminated with code ${c}`),m()):(n.fail(`${e} failed (code ${c})`),u(new Error(`Command failed with exit code ${c}`))):(n.succeed(`${e} completed successfully`),m())})})};import F from"node:fs";import At from"node:path";import{markdownTable as Pe}from"markdown-table";var I=class r{static{this.manifest={schemaVersion:"1.0.0",readme:"",modules:[]}}static generate(t={}){let{outDir:e="./docs",outFile:n="api.md",manifestPath:o="./custom-elements.json"}=t,{getElements:i,renderAllElements:s}=r;if(o)try{let c=F.readFileSync(o,"utf8");r.manifest=JSON.parse(c)}catch(c){throw console.error(`Error reading manifest file at ${o}:`,c),c}let a=i(),l=e;F.existsSync(l)||F.mkdirSync(l,{recursive:!0});let m=s(a),u=At.join(l,n);F.writeFileSync(u,m),console.log(`Generated combined API documentation at ${u}`)}static getElements(){let t=r.manifest.modules.filter(r.isWcaModule);return r.manifest.modules.reduce((e,n)=>e.concat(n.declarations?.filter(o=>"customElement"in o&&o.customElement===!0&&"tagName"in o&&(t.length>0?r.isWcaModule(n):!0))??[]),[])}static isWcaModule(t){let{path:e}=t;return e?e.startsWith("scripts/wca/auro-")&&e.endsWith(".js"):!1}static renderAllElements(t){return`${t.sort((e,n)=>(e.tagName||"").localeCompare(n.tagName||"")).map(e=>r.renderElement(e,!0)).join(`
3
3
 
4
4
  `)}`}static renderElement(t,e=!0){let n=[],{renderTable:o,renderPropertiesAttributesTable:i,renderParameters:s,getType:a}=r;n.push(e?`# ${t.tagName}`:""),t.description&&n.push(t.description);let l=i(t);l&&n.push(l.trim());let m=o("Methods",["name","parameters","return","description"],(t.members||[]).filter(p=>p.kind==="method"&&("privacy"in p?p.privacy!=="private":!0)&&p.name[0]!=="_").map(p=>({...p,parameters:s("parameters"in p?p.parameters:void 0),return:"return"in p&&p.return?a(p.return):""})));m&&n.push(m.trim());let u=o("Events",["name","description"],t.events);u&&n.push(u.trim());let c=o("Slots",[["name","(default)"],"description"],t.slots);c&&n.push(c.trim());let d=o("CSS Shadow Parts",["name","description"],t.cssParts);d&&n.push(d.trim());let h=o("CSS Custom Properties",["name","description"],t.cssProperties);return h&&n.push(h.trim()),n.join(`
5
5
 
@@ -9,16 +9,16 @@ ${Pe([l,...m])}
9
9
  `}static renderParameters(t){let{escapeMarkdown:e,getType:n}=r;return!t||t.length===0?"None":t.map(o=>{let i=n(o)||"any",s=o.description?` - ${o.description}`:"";return`\`${o.name}\` (${e(i)})${e(s)}`}).join("<br>")}static renderTable(t,e,n){let{escapeMarkdown:o,get:i,capitalize:s}=r;if(n===void 0||n.length===0)return"";let a=n.filter(c=>{let{description:d}=c;return typeof d=="string"&&d.trim()});if(a.length===0)return"";let l=e.map(c=>s((Array.isArray(c)?c[0]:c).split(".")[0])),m=a.map(c=>e.map(d=>{let h=i(c,d);return o(String(h||""))})),u=Pe([l,...m]);return`### ${t}
10
10
 
11
11
  ${u}
12
- `}static escapeMarkdown(t){return t.replace(/\\/g,"\\\\").replace(/\n/g,"<br>").replace(/\|/g,"\\|")}static getType(t){if(!t||!t.type)return"";let{type:e}=t,n=i=>i.replace(/\s*\|\s*/g," | ").replace(/'([^']+)'/g,"`$1`");if(typeof e=="string")return n(e);if(e.text)return n(e.text);if(Array.isArray(e))return e.map(i=>typeof i=="string"?i:i.text?i.text:i.name?i.name:String(i)).join(" \\| ");if(e.name)return n(e.name);if(e.references&&Array.isArray(e.references))return e.references.map(i=>i.name||String(i)).join(" \\| ");let o=String(e);return n(o)}static get(t,e){let n="",o=e;Array.isArray(e)&&([o,n]=e);let i=o.split("."),s=t;for(;s&&i.length;)s=s[i.shift()];return s==null||s===""?n:String(s)}static capitalize(t){return t.replace(/([A-Z])/g," $1").replace(/^./,e=>e.toUpperCase()).trim()}};import At from"node:fs";import vt from"node:os";import _ from"node:path";import Q from"node:process";function Et(){let r=vt.homedir()||Q.env.HOME||Q.env.USERPROFILE;if(!r)throw new Error("Unable to determine user home directory");return _.join(r,".auro")}function ee(...r){return _.join(Et(),...r)}function te(...r){let t=At.realpathSync(Q.argv[1]),e=_.dirname(t);return _.resolve(e,...r)}var Te=r=>te("configs",r);import{Logger as Ct}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{generateReadmeUrl as It,processContentForFile as Pt,templateFiller as Re}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";import re from"node:fs";import Tt from"node:path";var xt="/docs/pages",De={overwriteLocalCopies:!0,remoteReadmeVersion:"master",remoteReadmeVariant:"_updated_paths"};function k(r){return`${process.cwd()}/${r}`}async function Rt(r,t=!1){let e=[];t||e.push({identifier:"README.md",input:{remoteUrl:r.remoteReadmeUrl||It(r.remoteReadmeVersion,r.remoteReadmeVariant),fileName:k("/docTemplates/README.md"),overwrite:r.overwriteLocalCopies},output:k("/README.md")}),xe("/docs/partials/index.md")&&e.push({identifier:"index.md",input:k("/docs/partials/index.md"),output:k("/demo/index.md"),mdMagicConfig:{output:{directory:k("/demo")}}}),xe("/docs/partials/api.md")&&e.push({identifier:"api.md",input:k("/docs/partials/api.md"),output:k("/demo/api.md"),preProcessors:[Re.formatApiTable]});let n=k(xt);if(re.existsSync(n)){let i=(await re.promises.readdir(n)).map(s=>({identifier:s,input:Tt.join(n,s),output:k(`/demo/${s}`)}));e.push(...i)}return e}async function Dt(r=De,t=!1){await Re.extractNames();let e=await Rt(r,t);for(let n of e)try{await Pt(n)}catch(o){Ct.error(`Error processing ${n.identifier}: ${o.message}`)}}async function F(r={}){await Dt({...De,remoteReadmeUrl:"https://raw.githubusercontent.com/AlaskaAirlines/auro-templates/main/templates/default/README.md"},r.skipReadme)}function xe(r){return re.existsSync(k(r))}import{startDevServer as St}from"@web/dev-server";import{hmrPlugin as Ot}from"@web/dev-server-hmr";import jt from"ora";var P={watch:!0,nodeResolve:!0,basePath:"/",rootDir:"./demo",hmrInclude:["src/**/*","demo/**/*","apiExamples/**/*","docs/**/*"]};async function B(r={}){if(!r.serve)return;let t=jt(`Firing up dev server...
13
- `).start();try{let e={port:Number(r.port)||void 0,open:r.open?"/":void 0,watch:r.watch??P.watch,nodeResolve:r.nodeResolve??P.nodeResolve,basePath:r.basePath??P.basePath,rootDir:r.rootDir??P.rootDir,middleware:[function(i,s){return!i.url.endsWith("/")&&!i.url.includes(".")&&(i.url+=".html"),s()}],plugins:[Ot({include:r.hmrInclude??P.hmrInclude})]},n=await St({config:e,readCliArgs:!1,readFileConfig:!1});return t.stop(),n}catch(e){throw t.fail("Server snag! Couldn't start dev server."),console.error("Error starting development server:",e),new Error(`Development server failed to start: ${e.message}`)}}async function G(){let r=ne("Generating Custom Elements Manifest...").start();try{await b(`npx --package=@custom-elements-manifest/analyzer -y -- cem analyze --config '${Te("custom-elements-manifest.config.mjs")}'`),r.succeed("Custom Elements Manifest generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);r.warn("CEM analyzer completed with warnings: "+e)}}async function M(){let r=ne("Generating API md file...").start();try{await I.generate(),r.succeed("API md file generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);throw r.fail("Failed to generate API md file: "+e),t}}async function Se(r={}){let t=ne("Compiling documentation...").start();try{await F(r),t.succeed("Documentation compiled successfully!")}catch(e){let n=e instanceof Error?e.message:String(e);throw t.fail("Failed to compile MD documentation: "+n),e}}async function Oe(r={}){await B(r)}async function H(){await G(),await M()}function Ne(){let r=_t("./dist"),t=oe("Cleaning dist folder...").start();try{return Nt(r,{recursive:!0,force:!0}),t.succeed("All clean! Dist folder wiped."),!0}catch(e){return t.fail(`Oops! Couldn't clean dist/ folder: ${e.message}`),console.error(e),!1}}async function _e(r,t,e,n){let o=oe(r).start();try{let i=await t();return o.succeed(e),i}catch(i){throw o.fail(n),console.error(`Error: ${i.message}`),i}}async function Fe(r,t){return _e(`Bundling ${r.name||"main"} and ${t.name||"demo"}...`,async()=>{let e=await je(r);await e.write(r.output),await e.close();let n=await je(t);await n.write(t.output),await n.close()},`Bundles ready! ${r.name||"Main"} and ${t.name||"demo"} built.`,"Bundle hiccup! Build failed.")}async function L(r){let{wcaInput:t,wcaOutput:e,skipDocs:n}=r;if(n){let o=oe("Skipping docs generation...").start();setTimeout(()=>{o.succeed("Docs generation skipped.")},0);return}return _e("Analyzing components and making docs...",async()=>{await H(t,e),await F()},"Docs ready! Looking good.","Doc troubles!")}import{basename as Gt,join as Ge}from"node:path";import{nodeResolve as Mt}from"@rollup/plugin-node-resolve";import{glob as Ht}from"glob";import{litScss as Lt}from"rollup-plugin-scss-lit";import Ft from"node:path";import{glob as Bt}from"glob";function Be(r){return{name:"watch-globs",buildStart(){let t=Array.isArray(r)?r:[r];for(let e of t)try{for(let n of Bt.sync(Ft.resolve(e)))this.addWatchFile(n)}catch(n){this.error(`Error watching glob pattern "${e}": ${n.message}`)}}}}var se={moduleDirectories:["node_modules"],modulePaths:["../../node_modules","../node_modules","node_modules"],watchPatterns:["./apiExamples/**/*","./docs/**/*"]};function Me(r=[],t={}){let{watchPatterns:e=se.watchPatterns,dedupe:n=["lit","lit-element","lit-html"],dev:o=!1}=t,i=[...se.modulePaths,...r];return[Mt({dedupe:n,preferBuiltins:!1,moduleDirectories:se.moduleDirectories}),Lt({minify:o?!1:{fast:!0},options:{loadPaths:[...i,Ge(process.cwd(),"src","styles"),Ge(process.cwd(),"src")]}}),Be(e)]}function ie(r={}){let{modulePaths:t=[],watch:e=!1,input:n=["./src/index.js","./src/registered.js"],outputDir:o="./dist",format:i="esm",dev:s=!1}=r;return{name:"Main",config:{input:n,output:{format:i,dir:o,entryFileNames:a=>s||["index","registered"].includes(a.name)?"[name].js":"[name]-[hash].js",chunkFileNames:s?"[name].js":"[name]-[hash].js",assetFileNames:s?"[name][extname]":"[name]-[hash][extname]"},external:Ut(),plugins:Me(t,{dev:s}),watch:He(e)}}}function ae(r={}){let{modulePaths:t=[],watch:e=!1,globPattern:n="./demo/*.js",ignorePattern:o=["./demo/*.min.js"],outputDir:i="./demo",dev:s=!1}=r;return{name:"Demo",config:{input:Object.fromEntries(Ht.sync(n,{ignore:o}).map(a=>[Gt(a,".js"),a])),output:{format:"esm",dir:i,entryFileNames:"[name].min.js",chunkFileNames:"[name].min.js",assetFileNames:s?"[name][extname]":"[name]-[hash][extname]"},plugins:Me(t,{dev:s}),watch:He(e)}}}function He(r){if(!r)return!1;let t=typeof r=="object"?r:{};return{clearScreen:t.clearScreen??!0,buildDelay:t.buildDelay??500,chokidar:{ignoreInitial:!0,ignored:t.ignored??["**/dist/**/*.d.ts","**/custom-elements.json","**/demo/*.md","**/demo/**/*.min.js","**/docs/api.md","**/node_modules/**","**/.git/**"],awaitWriteFinish:t.awaitWriteFinish??{stabilityThreshold:1e3,pollInterval:100}},include:t.include??["./src/**/*.scss","./src/**/*.js","./src/**/*.ts","./demo/**/*.js","./apiExamples/**/*","./docs/**/*.md"],exclude:t.exclude??["./dist/**/*","./node_modules/**/*"]}}function Ut(r=[]){return[...[/node_modules\/lit/,/node_modules\/lit-element/,/node_modules\/lit-html/,/node_modules\/@lit/],...r]}import Ue from"node:path";import A from"ora";import"rollup";var U=!1,Wt={analyze:{active:!1,lastTime:0},docs:{active:!1,lastTime:0}},zt=5e3,W=new Set,qt=["/dist/index.d.ts","/custom-elements.json","/demo/api.md","/docs/api.md","/demo/index.min.js"];function Vt(r){if(!r||typeof r!="string")return!1;try{let t=Ue.normalize(r);return qt.some(e=>t.endsWith(e))||t.includes("/dist/")||t.endsWith(".min.js")||t.endsWith(".d.ts")}catch(t){return console.error(`Error checking path (${typeof r}):`,t.message),!1}}async function Le(r,t){let e=Wt[r];if(e.active||Date.now()-e.lastTime<zt)return!1;try{return e.active=!0,e.lastTime=Date.now(),await t()}catch(n){return console.error(`Error in ${r} task:`,n),!1}finally{e.active=!1}}async function We(r,t,e){let n=!0,o={analyze:!1,docs:!1},i=null,s,a=A("Activating watch mode...").start(),l={analyze:async()=>{let{wcaInput:c,wcaOutput:d,skipDocs:h}=t;if(h){let w=A("Skipping component analysis...").start();return setTimeout(()=>{w.succeed("Component analysis skipped.")},0),!0}let p=A("Detective work: analyzing components...").start();try{return await H(c,d),p.succeed("Component analysis complete! API generated."),!0}catch(w){return p.fail("Analysis hiccup! Something went wrong."),console.error("Component analysis error:",w),!1}},docs:async()=>{if(U)return!1;if(t.skipDocs){let d=A("Skipping docs generation...").start();return setTimeout(()=>{d.succeed("Docs generation skipped.")},0),!0}let c=A("Refreshing docs...").start();try{return await L(t),c.succeed("Documentation refreshed!"),!0}catch(d){c.fail("Docs stumble! Couldn't refresh."),console.error("Documentation rebuild error:",d)}}},m=()=>{n&&o.analyze&&o.docs&&typeof e=="function"&&(n=!1,e())};function u(c=1e3){i&&clearTimeout(i),i=setTimeout(async()=>{setTimeout(async()=>{o.analyze=await Le("analyze",l.analyze),setTimeout(async()=>{o.docs=await Le("docs",l.docs),m()},1e3)},1e3)},c)}r.on("event",async c=>{switch(c.code){case"START":a.succeed("Watch mode active! Eyes peeled.");break;case"BUNDLE_START":if(W.clear(),c.input)try{let d=Array.isArray(c.input)?c.input:typeof c.input=="string"?[c.input]:typeof c.input=="object"&&c.input!==null?Object.values(c.input):[];for(let h of d)typeof h=="string"&&!Vt(h)&&W.add(Ue.normalize(h))}catch(d){console.error("Error processing input paths:",d)}s=A("Weaving bundles...").start(),U=!0;break;case"BUNDLE_END":s&&s.succeed(`Bundle ${Array.isArray(c.input)?`of ${c.input.join("& ")} `:""}done in ${c.duration}ms! \u{1F680}`),U=!1,W.size>0&&u();break;case"END":break;case"ERROR":U=!1,s?s.fail(`Oops! Bundle hit a snag: ${c.error.message}`):A().fail(`Watch mode hiccup: ${c.error.message}`),W.clear();break}})}function ze(r){return process.on("SIGINT",()=>{let t=A("Wrapping up...").start();r.close(),t.succeed("All done! See you next time. \u2728"),process.exit(0)}),r}async function Yt(r){let t=ie(r),e=ae(r);r.dev||t.config.plugins.push(Kt()),await L(r),await Fe(t.config,e.config)}async function Xt(r){let{dev:t}=r,e=ie({...r,watch:!0}),n=ae({...r,watch:!0}),o=Jt([e.config,n.config]);return We(o,r,t?async()=>B(r):void 0),ze(o),o}async function z(r={}){try{let{watch:t}=r;return Ne(),t?await Xt(r):await Yt(r)}catch(t){throw new Error(`Build failed: ${t.message}`)}}var T=Zt.command("dev").description("Runs development server for auro components");T=O(T);T=j(T);var Ao=T.action(async r=>{try{let t=qe("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),r.watch||t.succeed("Build completed!"),await z({...r,dev:!0,watch:r.watch})}catch(t){qe().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{program as Qt}from"commander";import Ve from"ora";var ce=Qt.command("build").description("Builds auro components");ce=O(ce);var To=ce.action(async r=>{try{let t=Ve("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),await z(r),r.watch||t.succeed("Build completed!")}catch(t){Ve().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{exec as er}from"node:child_process";import Ke from"node:path";import tr from"node:process";import{fileURLToPath as rr}from"node:url";import nr from"node:util";import{program as or}from"commander";import sr from"inquirer";var Bo=or.command("migrate").description("Script runner to perform repetitive code change tasks").requiredOption("-i, --id <string>","Select the migration you would like to run by id").option("-m, --multi-gitter","Run the migration on all repositories in the multi-gitter config").action(async r=>{let t=rr(import.meta.url),e=Ke.dirname(t),n=Ke.resolve(e,"migrations",r.id);if(r.multiGitter){let o=nr.promisify(er);try{await o("command -v multi-gitter")}catch{console.error("multi-gitter is not installed."),tr.exit(1)}(await sr.prompt([{type:"confirm",name:"dryRun",message:"Run migration in dry-run mode? (no changes will be committed)",default:!0}])).dryRun?b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml" --dry-run`):b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml"`)}else b(`${n}/script.sh`)});import ur from"node:process";import{program as mr}from"commander";import{readFile as dr,writeFile as pr}from"node:fs/promises";import{Logger as me}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import le from"node:fs/promises";import ue from"node:path";import{Octokit as ir}from"@octokit/rest";import $ from"ora";import{processContentForFile as ar,templateFiller as cr}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";async function Je(r,t){let o=(await new ir({auth:process.env.GITHUB_TOKEN||""}).request("GET /repos/{owner}/{repo}/contents/{path}",{ref:t,owner:"AlaskaAirlines",repo:"auro-templates",path:r,headers:{"X-GitHub-Api-Version":"2022-11-28"}})).data;if(typeof o!="object"||!Array.isArray(o)){let i=`Unexpected response format: ${JSON.stringify(o)}`;throw $().start().fail(i),new Error("Failed to retrieve folder items")}return o}async function Ye({folderItems:r,templatePathToReplace:t,rootDir:e,ref:n}){let o=[];for(let i of r){if(i.type=="dir"){let m=$(`Processing directory: ${i.path}`).start(),u=await Je(i.path,n);m.succeed(`Found ${u.length} additional items in ${i.path}`);let c=await Ye({folderItems:u,templatePathToReplace:t,rootDir:e,ref:n});o.push(...c);continue}let s=i.path.replace(`${t}/`,""),a=`${e}/.github/${s}`,l={identifier:i.name,input:{remoteUrl:i.download_url||"",fileName:a,overwrite:!0},output:a};o.push(l)}return o}async function lr(r){try{await le.rm(r,{recursive:!0,force:!0}),$().start().succeed(`Successfully removed directory: ${r}`)}catch(t){throw $().start().fail(`Error removing directory ${r}: ${t.message}`),t}}async function Xe(r,t="",e=!0){try{let n=await le.stat(r),o=ue.basename(r);if(!n.isDirectory())return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}
12
+ `}static escapeMarkdown(t){return t.replace(/\\/g,"\\\\").replace(/\n/g,"<br>").replace(/\|/g,"\\|")}static getType(t){if(!t||!t.type)return"";let{type:e}=t,n=i=>i.replace(/\s*\|\s*/g," | ").replace(/'([^']+)'/g,"`$1`");if(typeof e=="string")return n(e);if(e.text)return n(e.text);if(Array.isArray(e))return e.map(i=>typeof i=="string"?i:i.text?i.text:i.name?i.name:String(i)).join(" \\| ");if(e.name)return n(e.name);if(e.references&&Array.isArray(e.references))return e.references.map(i=>i.name||String(i)).join(" \\| ");let o=String(e);return n(o)}static get(t,e){let n="",o=e;Array.isArray(e)&&([o,n]=e);let i=o.split("."),s=t;for(;s&&i.length;)s=s[i.shift()];return s==null||s===""?n:String(s)}static capitalize(t){return t.replace(/([A-Z])/g," $1").replace(/^./,e=>e.toUpperCase()).trim()}};import vt from"node:fs";import Et from"node:os";import _ from"node:path";import ee from"node:process";function Ct(){let r=Et.homedir()||ee.env.HOME||ee.env.USERPROFILE;if(!r)throw new Error("Unable to determine user home directory");return _.join(r,".auro")}function te(...r){return _.join(Ct(),...r)}function re(...r){let t=vt.realpathSync(ee.argv[1]),e=_.dirname(t);return _.resolve(e,...r)}var Te=r=>re("configs",r);import{Logger as It}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{generateReadmeUrl as Pt,processContentForFile as Tt,templateFiller as Re}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";import ne from"node:fs";import xt from"node:path";var Rt="/docs/pages",De={overwriteLocalCopies:!0,remoteReadmeVersion:"master",remoteReadmeVariant:"_updated_paths"};function k(r){return`${process.cwd()}/${r}`}async function Dt(r,t=!1){let e=[];t||e.push({identifier:"README.md",input:{remoteUrl:r.remoteReadmeUrl||Pt(r.remoteReadmeVersion,r.remoteReadmeVariant),fileName:k("/docTemplates/README.md"),overwrite:r.overwriteLocalCopies},output:k("/README.md")}),xe("/docs/partials/index.md")&&e.push({identifier:"index.md",input:k("/docs/partials/index.md"),output:k("/demo/index.md"),mdMagicConfig:{output:{directory:k("/demo")}}}),xe("/docs/partials/api.md")&&e.push({identifier:"api.md",input:k("/docs/partials/api.md"),output:k("/demo/api.md"),preProcessors:[Re.formatApiTable]});let n=k(Rt);if(ne.existsSync(n)){let i=(await ne.promises.readdir(n)).map(s=>({identifier:s,input:xt.join(n,s),output:k(`/demo/${s}`)}));e.push(...i)}return e}async function St(r=De,t=!1){await Re.extractNames();let e=await Dt(r,t);for(let n of e)try{await Tt(n)}catch(o){It.error(`Error processing ${n.identifier}: ${o.message}`)}}async function B(r={}){await St({...De,remoteReadmeUrl:"https://raw.githubusercontent.com/AlaskaAirlines/auro-templates/main/templates/default/README.md"},r.skipReadme)}function xe(r){return ne.existsSync(k(r))}import{startDevServer as Ot}from"@web/dev-server";import{hmrPlugin as jt}from"@web/dev-server-hmr";import Nt from"ora";var P={watch:!0,nodeResolve:!0,basePath:"/",rootDir:"./demo",hmrInclude:["src/**/*","demo/**/*","apiExamples/**/*","docs/**/*"]};async function G(r={}){if(!r.serve)return;let t=Nt(`Firing up dev server...
13
+ `).start();try{let e={port:Number(r.port)||void 0,open:r.open?"/":void 0,watch:r.watch??P.watch,nodeResolve:r.nodeResolve??P.nodeResolve,basePath:r.basePath??P.basePath,rootDir:r.rootDir??P.rootDir,middleware:[function(i,s){return!i.url.endsWith("/")&&!i.url.includes(".")&&(i.url+=".html"),s()}],plugins:[jt({include:r.hmrInclude??P.hmrInclude})]},n=await Ot({config:e,readCliArgs:!1,readFileConfig:!1});return t.stop(),n}catch(e){throw t.fail("Server snag! Couldn't start dev server."),console.error("Error starting development server:",e),new Error(`Development server failed to start: ${e.message}`)}}async function M(){let r=oe("Generating Custom Elements Manifest...").start();try{await b(`npx --package=@custom-elements-manifest/analyzer -y -- cem analyze --config '${Te("custom-elements-manifest.config.mjs")}'`),r.succeed("Custom Elements Manifest generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);r.warn("CEM analyzer completed with warnings: "+e)}}async function H(){let r=oe("Generating API md file...").start();try{await I.generate(),r.succeed("API md file generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);throw r.fail("Failed to generate API md file: "+e),t}}async function Se(r={}){let t=oe("Compiling documentation...").start();try{await B(r),t.succeed("Documentation compiled successfully!")}catch(e){let n=e instanceof Error?e.message:String(e);throw t.fail("Failed to compile MD documentation: "+n),e}}async function Oe(r={}){await G(r)}async function L(){await M(),await H()}function Ne(){let r=_t("./dist"),t=se("Cleaning dist folder...").start();try{return Ft(r,{recursive:!0,force:!0}),t.succeed("All clean! Dist folder wiped."),!0}catch(e){return t.fail(`Oops! Couldn't clean dist/ folder: ${e.message}`),console.error(e),!1}}async function Fe(r,t,e,n){let o=se(r).start();try{let i=await t();return o.succeed(e),i}catch(i){throw o.fail(n),console.error(`Error: ${i.message}`),i}}async function _e(r,t){return Fe(`Bundling ${r.name||"main"} and ${t.name||"demo"}...`,async()=>{let e=await je(r);await e.write(r.output),await e.close();let n=await je(t);await n.write(t.output),await n.close()},`Bundles ready! ${r.name||"Main"} and ${t.name||"demo"} built.`,"Bundle hiccup! Build failed.")}async function U(r){let{wcaInput:t,wcaOutput:e,skipDocs:n}=r;if(n){let o=se("Skipping docs generation...").start();setTimeout(()=>{o.succeed("Docs generation skipped.")},0);return}return Fe("Analyzing components and making docs...",async()=>{await L(t,e),await B()},"Docs ready! Looking good.","Doc troubles!")}import{basename as Mt,join as Ge}from"node:path";import{nodeResolve as Ht}from"@rollup/plugin-node-resolve";import{glob as Lt}from"glob";import{litScss as Ut}from"rollup-plugin-scss-lit";import Bt from"node:path";import{glob as Gt}from"glob";function Be(r){return{name:"watch-globs",buildStart(){let t=Array.isArray(r)?r:[r];for(let e of t)try{for(let n of Gt.sync(Bt.resolve(e)))this.addWatchFile(n)}catch(n){this.error(`Error watching glob pattern "${e}": ${n.message}`)}}}}var ie={moduleDirectories:["node_modules"],modulePaths:["../../node_modules","../node_modules","node_modules"],watchPatterns:["./apiExamples/**/*","./docs/**/*"]};function Me(r=[],t={}){let{watchPatterns:e=ie.watchPatterns,dedupe:n=["lit","lit-element","lit-html"],dev:o=!1}=t,i=[...ie.modulePaths,...r];return[Ht({dedupe:n,preferBuiltins:!1,moduleDirectories:ie.moduleDirectories}),Ut({minify:o?!1:{fast:!0},options:{loadPaths:[...i,Ge(process.cwd(),"src","styles"),Ge(process.cwd(),"src")]}}),Be(e)]}function ae(r={}){let{modulePaths:t=[],watch:e=!1,input:n=["./src/index.js","./src/registered.js"],outputDir:o="./dist",format:i="esm",dev:s=!1}=r;return{name:"Main",config:{input:n,output:{format:i,dir:o,entryFileNames:a=>s||["index","registered"].includes(a.name)?"[name].js":"[name]-[hash].js",chunkFileNames:s?"[name].js":"[name]-[hash].js",assetFileNames:s?"[name][extname]":"[name]-[hash][extname]"},external:Wt(),plugins:Me(t,{dev:s}),watch:He(e)}}}function ce(r={}){let{modulePaths:t=[],watch:e=!1,globPattern:n="./demo/*.js",ignorePattern:o=["./demo/*.min.js"],outputDir:i="./demo",dev:s=!1}=r;return{name:"Demo",config:{input:Object.fromEntries(Lt.sync(n,{ignore:o}).map(a=>[Mt(a,".js"),a])),output:{format:"esm",dir:i,entryFileNames:"[name].min.js",chunkFileNames:"[name].min.js",assetFileNames:s?"[name][extname]":"[name]-[hash][extname]"},plugins:Me(t,{dev:s}),watch:He(e)}}}function He(r){if(!r)return!1;let t=typeof r=="object"?r:{};return{clearScreen:t.clearScreen??!0,buildDelay:t.buildDelay??500,chokidar:{ignoreInitial:!0,ignored:t.ignored??["**/dist/**/*.d.ts","**/custom-elements.json","**/demo/*.md","**/demo/**/*.min.js","**/docs/api.md","**/node_modules/**","**/.git/**"],awaitWriteFinish:t.awaitWriteFinish??{stabilityThreshold:1e3,pollInterval:100}},include:t.include??["./src/**/*.scss","./src/**/*.js","./src/**/*.ts","./demo/**/*.js","./apiExamples/**/*","./docs/**/*.md"],exclude:t.exclude??["./dist/**/*","./node_modules/**/*"]}}function Wt(r=[]){return[...[/node_modules\/lit/,/node_modules\/lit-element/,/node_modules\/lit-html/,/node_modules\/@lit/],...r]}import Ue from"node:path";import A from"ora";import"rollup";var W=!1,zt={analyze:{active:!1,lastTime:0},docs:{active:!1,lastTime:0}},qt=5e3,z=new Set,Vt=["/dist/index.d.ts","/custom-elements.json","/demo/api.md","/docs/api.md","/demo/index.min.js"];function Kt(r){if(!r||typeof r!="string")return!1;try{let t=Ue.normalize(r);return Vt.some(e=>t.endsWith(e))||t.includes("/dist/")||t.endsWith(".min.js")||t.endsWith(".d.ts")}catch(t){return console.error(`Error checking path (${typeof r}):`,t.message),!1}}async function Le(r,t){let e=zt[r];if(e.active||Date.now()-e.lastTime<qt)return!1;try{return e.active=!0,e.lastTime=Date.now(),await t()}catch(n){return console.error(`Error in ${r} task:`,n),!1}finally{e.active=!1}}async function We(r,t,e){let n=!0,o={analyze:!1,docs:!1},i=null,s,a=A("Activating watch mode...").start(),l={analyze:async()=>{let{wcaInput:c,wcaOutput:d,skipDocs:h}=t;if(h){let w=A("Skipping component analysis...").start();return setTimeout(()=>{w.succeed("Component analysis skipped.")},0),!0}let p=A("Detective work: analyzing components...").start();try{return await L(c,d),p.succeed("Component analysis complete! API generated."),!0}catch(w){return p.fail("Analysis hiccup! Something went wrong."),console.error("Component analysis error:",w),!1}},docs:async()=>{if(W)return!1;if(t.skipDocs){let d=A("Skipping docs generation...").start();return setTimeout(()=>{d.succeed("Docs generation skipped.")},0),!0}let c=A("Refreshing docs...").start();try{return await U(t),c.succeed("Documentation refreshed!"),!0}catch(d){c.fail("Docs stumble! Couldn't refresh."),console.error("Documentation rebuild error:",d)}}},m=()=>{n&&o.analyze&&o.docs&&typeof e=="function"&&(n=!1,e())};function u(c=1e3){i&&clearTimeout(i),i=setTimeout(async()=>{setTimeout(async()=>{o.analyze=await Le("analyze",l.analyze),setTimeout(async()=>{o.docs=await Le("docs",l.docs),m()},1e3)},1e3)},c)}r.on("event",async c=>{switch(c.code){case"START":a.succeed("Watch mode active! Eyes peeled.");break;case"BUNDLE_START":if(z.clear(),c.input)try{let d=Array.isArray(c.input)?c.input:typeof c.input=="string"?[c.input]:typeof c.input=="object"&&c.input!==null?Object.values(c.input):[];for(let h of d)typeof h=="string"&&!Kt(h)&&z.add(Ue.normalize(h))}catch(d){console.error("Error processing input paths:",d)}s=A("Weaving bundles...").start(),W=!0;break;case"BUNDLE_END":s&&s.succeed(`Bundle ${Array.isArray(c.input)?`of ${c.input.join("& ")} `:""}done in ${c.duration}ms! \u{1F680}`),W=!1,z.size>0&&u();break;case"END":break;case"ERROR":W=!1,s?s.fail(`Oops! Bundle hit a snag: ${c.error.message}`):A().fail(`Watch mode hiccup: ${c.error.message}`),z.clear();break}})}function ze(r){return process.on("SIGINT",()=>{let t=A("Wrapping up...").start();r.close(),t.succeed("All done! See you next time. \u2728"),process.exit(0)}),r}async function Xt(r){let t=ae(r),e=ce(r);r.dev||t.config.plugins.push(Jt()),await U(r),await _e(t.config,e.config)}async function Zt(r){let{dev:t}=r,e=ae({...r,watch:!0}),n=ce({...r,watch:!0}),o=Yt([e.config,n.config]);return We(o,r,t?async()=>G(r):void 0),ze(o),o}async function q(r={}){try{let{watch:t}=r;return Ne(),t?await Zt(r):await Xt(r)}catch(t){throw new Error(`Build failed: ${t.message}`)}}var T=Qt.command("dev").description("Runs development server for auro components");T=j(T);T=N(T);var Ao=T.action(async r=>{try{let t=qe("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),r.watch||t.succeed("Build completed!"),await q({...r,dev:!0,watch:r.watch})}catch(t){qe().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{program as er}from"commander";import Ve from"ora";var le=er.command("build").description("Builds auro components");le=j(le);var To=le.action(async r=>{try{let t=Ve("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),await q(r),r.watch||t.succeed("Build completed!")}catch(t){Ve().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{exec as tr}from"node:child_process";import Ke from"node:path";import rr from"node:process";import{fileURLToPath as nr}from"node:url";import or from"node:util";import{program as sr}from"commander";import ir from"inquirer";var Bo=sr.command("migrate").description("Script runner to perform repetitive code change tasks").requiredOption("-i, --id <string>","Select the migration you would like to run by id").option("-m, --multi-gitter","Run the migration on all repositories in the multi-gitter config").action(async r=>{let t=nr(import.meta.url),e=Ke.dirname(t),n=Ke.resolve(e,"migrations",r.id);if(r.multiGitter){let o=or.promisify(tr);try{await o("command -v multi-gitter")}catch{console.error("multi-gitter is not installed."),rr.exit(1)}(await ir.prompt([{type:"confirm",name:"dryRun",message:"Run migration in dry-run mode? (no changes will be committed)",default:!0}])).dryRun?b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml" --dry-run`):b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml"`)}else b(`${n}/script.sh`)});import Qe from"node:process";import{program as mr}from"commander";import{readFile as dr,writeFile as pr}from"node:fs/promises";import{Logger as x}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import ue from"node:fs/promises";import me from"node:path";import{Octokit as ar}from"@octokit/rest";import $ from"ora";import{processContentForFile as cr,templateFiller as lr}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";async function Je(r,t){let e=new ar({auth:process.env.GITHUB_TOKEN||""});try{let o=(await e.request("GET /repos/{owner}/{repo}/contents/{path}",{ref:t,owner:"AlaskaAirlines",repo:"auro-templates",path:r,headers:{"X-GitHub-Api-Version":"2022-11-28"}})).data;if(typeof o!="object"||!Array.isArray(o)){let i=`Unexpected response format: ${JSON.stringify(o)}`;throw $().start().fail(i),new Error("Failed to retrieve folder items")}return o}catch(n){let o=$().start();throw n.status===404?o.fail(`Template '${r.split("/")[1]}' not found`):o.fail(`Error accessing template: ${n.message}`),n}}async function Ye({folderItems:r,templatePathToReplace:t,rootDir:e,ref:n}){let o=[];for(let i of r){if(i.type=="dir"){let m=$(`Processing directory: ${i.path}`).start(),u=await Je(i.path,n);m.succeed(`Found ${u.length} additional items in ${i.path}`);let c=await Ye({folderItems:u,templatePathToReplace:t,rootDir:e,ref:n});o.push(...c);continue}let s=i.path.replace(`${t}/`,""),a=`${e}/.github/${s}`,l={identifier:i.name,input:{remoteUrl:i.download_url||"",fileName:a,overwrite:!0},output:a};o.push(l)}return o}async function ur(r){try{await ue.rm(r,{recursive:!0,force:!0}),$().start().succeed(`Successfully removed directory: ${r}`)}catch(t){throw $().start().fail(`Error removing directory ${r}: ${t.message}`),t}}async function Xe(r,t="",e=!0){try{let n=await ue.stat(r),o=me.basename(r);if(!n.isDirectory())return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}
14
14
  `;let i=`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}/
15
- `;try{let a=(await le.readdir(r)).sort();for(let l=0;l<a.length;l++){let m=a[l],u=ue.join(r,m),c=l===a.length-1,d=t+(e?" ":"\u2502 ");i+=await Xe(u,d,c)}}catch{i+=`${t}${e?" ":"\u2502 "}\u2514\u2500\u2500 [Permission denied or error reading directory]
15
+ `;try{let a=(await ue.readdir(r)).sort();for(let l=0;l<a.length;l++){let m=a[l],u=me.join(r,m),c=l===a.length-1,d=t+(e?" ":"\u2502 ");i+=await Xe(u,d,c)}}catch{i+=`${t}${e?" ":"\u2502 "}\u2514\u2500\u2500 [Permission denied or error reading directory]
16
16
  `}return i}catch(n){return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}[Error: ${n}]
17
- `}}async function Ze(r,t="main"){r||($().start().fail("Root directory must be specified"),process.exit(1));let e=".github",n=$("Removing existing .github directory...").start();try{await lr(e),n.succeed(".github directory removed successfully")}catch(l){n.fail(`Error removing .github directory: ${l.message}`),process.exit(1)}await cr.extractNames(),process.env.GITHUB_TOKEN||($().start().fail("GITHUB_TOKEN environment variable is not set."),process.exit(1));let o="templates/default/.github",i=await Je(o,t),s=await Ye({folderItems:i,templatePathToReplace:o,rootDir:r,ref:t}),a=$("Processing all files...").start();try{await Promise.all(s.map(m=>ar(m))),a.succeed("All files processed.");let l=$("Generating directory tree...").start();try{let m=ue.join(r,".github"),u=await Xe(m);l.succeed("Synced .github directory structure:"),console.log(u)}catch(m){l.fail(`Error generating directory tree: ${m.message}`)}}catch(l){a.fail(`Error processing files: ${l.message}`),process.exit(1)}}var Xo=mr.command("sync").option("-r, --ref <branch/tag/commit>","Git reference (branch/tag/commit) to use","main").description("Script runner to synchronize local repository configuration files").action(async r=>{me.info("Synchronizing repository configuration files..."),me.warn("Note: sync does not create a new git branch. Changes are added to the current branch.");let t=ur.cwd();await Ze(t,r.ref);let e=`${t}/.github/CODEOWNERS`,o=(await dr(e,{encoding:"utf-8"})).replace(/\r\n/gu,`
17
+ `}}async function Ze(r,t="main",e="default"){if(!r)throw $().start().fail("Root directory must be specified"),new Error("Root directory must be specified");if(await lr.extractNames(),!process.env.GITHUB_TOKEN)throw $().start().fail("GITHUB_TOKEN environment variable is not set."),new Error("GITHUB_TOKEN environment variable is not set");let n=`templates/${e}/.github`,o=await Je(n,t),i=".github",s=$("Removing existing .github directory...").start();try{await ur(i),s.succeed(".github directory removed successfully")}catch(m){throw s.fail(`Error removing .github directory: ${m.message}`),new Error(`Failed to remove .github directory: ${m.message}`)}let a=await Ye({folderItems:o,templatePathToReplace:n,rootDir:r,ref:t}),l=$("Processing all files...").start();try{await Promise.all(a.map(u=>cr(u))),l.succeed("All files processed.");let m=$("Generating directory tree...").start();try{let u=me.join(r,".github"),c=await Xe(u);m.succeed("Synced .github directory structure:"),console.log(c)}catch(u){m.fail(`Error generating directory tree: ${u.message}`)}}catch(m){throw l.fail(`Error processing files: ${m.message}`),new Error(`Failed to process files: ${m.message}`)}}var Xo=mr.command("sync").option("-r, --ref <branch/tag/commit>","Git reference (branch/tag/commit) to use","main").option("-t, --template <name>","Template based on which to sync","default").description("Script runner to synchronize local repository configuration files").action(async r=>{x.info("Synchronizing repository configuration files..."),x.warn("Note: sync does not create a new git branch. Changes are added to the current branch.");let t=Qe.cwd();try{await Ze(t,r.ref,r.template);try{let e=`${t}/.github/CODEOWNERS`,o=(await dr(e,{encoding:"utf-8"})).replace(/\r\n/gu,`
18
18
  `).replace(/\n\n/gu,`
19
19
  `);await pr(e,o,{encoding:"utf-8"}),(o.includes("\r")||o.includes(`
20
20
 
21
- `))&&me.error("CODEOWNERS file still has Windows line endings.")});import q from"node:fs";import x from"node:path";import{Logger as et}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{program as fr}from"commander";import{glob as gr}from"glob";var Qe=(r,t)=>{let e=(r.match(/static register\(name \= (.+)\)/)||r.match(/customElements.get\((.+?)\)/))[1],n=r.match(/export class (.+) extends/)?.[1],o=r.match(/\/\*\*((.|\n)*?)(\*\n|\*\/|[@])/)?.[1]||"";return!e||!n?r:`
21
+ `))&&x.error("CODEOWNERS file still has Windows line endings.")}catch(e){throw x.error(`Error processing CODEOWNERS file: ${e.message}`),e}}catch(e){x.error(`Sync operation failed: ${e.message}`),Qe.exit(1)}});import V from"node:fs";import R from"node:path";import{Logger as tt}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{program as fr}from"commander";import{glob as gr}from"glob";var et=(r,t)=>{let e=(r.match(/static register\(name \= (.+)\)/)||r.match(/customElements.get\((.+?)\)/))[1],n=r.match(/export class (.+) extends/)?.[1],o=r.match(/\/\*\*((.|\n)*?)(\*\n|\*\/|[@])/)?.[1]||"";return!e||!n?r:`
22
22
  import { ${n} } from '${t}';
23
23
 
24
24
  /**${o}*/
@@ -27,30 +27,30 @@ class ${n}WCA extends ${n} {}
27
27
  if (!customElements.get(${e})) {
28
28
  customElements.define(${e}, ${n}WCA);
29
29
  }
30
- `};var V=x.resolve(process.cwd(),"./scripts/wca");async function hr(r){try{return(await Promise.all(r.map(e=>gr(e)))).flat()}catch(t){throw console.error("Error processing glob patterns:",t),t}}async function yr(r){q.existsSync(V)||await q.promises.mkdir(V,{recursive:!0});for(let t of r){let e=x.resolve(process.cwd(),t),n=await q.promises.readFile(e,"utf-8"),o=x.resolve(V,`${x.basename(t)}`),i=Qe(n,x.relative(V,t));await q.promises.writeFile(o,i)}}async function wr(){let r=await hr(["./src/auro-*.js"]);await yr(r)}var is=fr.command("wca-setup").description("Set up WCA (Web Component Analyzer) for the project").action(()=>{wr().then(()=>{et.success("WCA setup completed successfully.")}).catch(r=>{et.error(`WCA setup failed: ${r.message}`)})});import{program as Ir}from"commander";import ot from"chalk";import st from"ora";import{appendFile as br,readFile as kr}from"node:fs/promises";import{Logger as f}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{simpleGit as $r}from"simple-git";var g;try{g=$r({baseDir:process.cwd(),binary:"git",maxConcurrentProcesses:1})}catch(r){f.error(`Failed to initialize git: ${r}`),g={}}var v=class r{static async checkGitignore(t){if(t==="")return!1;try{return(await kr(".gitignore","utf-8")).includes(t)}catch(e){return f.error(`Error reading file: ${e}`),!1}}static async getCommitMessages(t=""){try{let e=t;e||(e=(await g.branchLocal()).current);let n="main",o="";if(!!process.env.GITHUB_ACTIONS){n=process.env.GITHUB_BASE_REF||"main";try{if(await g.fetch("origin",n),e!=="HEAD")try{await g.raw(["rev-parse","--verify",`origin/${e}`])}catch{await g.fetch("origin",e)}let s=e==="HEAD"?"HEAD":`origin/${e}`;o=`${(await g.raw(["merge-base",`origin/${n}`,s])).trim()}..${s}`}catch(s){f.warn(`Error setting up commit range in CI: ${s}`);let a=e==="HEAD"?"HEAD":`origin/${e}`;o=`origin/${n}..${a}`}}else try{try{await g.raw(["rev-parse","--verify",`origin/${n}`])}catch{f.info(`Fetching ${n} from origin`),await g.fetch("origin",n)}if(e!=="HEAD")try{await g.raw(["rev-parse","--verify",e])}catch{await g.fetch("origin",e)}o=`${(await g.raw(["merge-base",`origin/${n}`,e])).trim()}..${e}`}catch(s){f.warn(`Error determining commits locally: ${s}`),o=`${e}~10..${e}`}return await r.getFormattedCommits(o)}catch(e){return f.error(`Error getting commit messages: ${e}`),[]}}static async getRepoOwnerAndName(){try{let t=await g.getRemotes(!0);if(t.length===0)return f.warn("No remotes found"),null;let e=t.find(o=>o.name==="origin")||t[0],n=e.refs.fetch||e.refs.push;return r.parseGitUrl(n)}catch(t){return f.error(`Error getting repo owner and name: ${t}`),null}}static async getCurrentBranchName(){try{return(await g.branchLocal()).current||null}catch(t){return f.error(`Error getting current branch name: ${t}`),null}}static parseGitUrl(t){let e;return t.includes("@")&&t.includes(":")&&(e=t.match(/@([^:]+):([^/]+)\/(.+?)(?:\.git)?$/),e)?{owner:e[2],repo:e[3]}:(e=t.match(/https?:\/\/(?:[^@]+@)?[^/]+\/([^/]+)\/(.+?)(?:\.git)?$/),e?{owner:e[1],repo:e[2]}:(f.warn(`Could not parse git URL: ${t}`),null))}static async getFormattedCommits(t){let n=(await g.raw(["log","--pretty=format:COMMIT_START%n%H%n%ad%n%an%n%s%n%b%nCOMMIT_END","--date=short",t])).split(`COMMIT_START
30
+ `};var K=R.resolve(process.cwd(),"./scripts/wca");async function hr(r){try{return(await Promise.all(r.map(e=>gr(e)))).flat()}catch(t){throw console.error("Error processing glob patterns:",t),t}}async function yr(r){V.existsSync(K)||await V.promises.mkdir(K,{recursive:!0});for(let t of r){let e=R.resolve(process.cwd(),t),n=await V.promises.readFile(e,"utf-8"),o=R.resolve(K,`${R.basename(t)}`),i=et(n,R.relative(K,t));await V.promises.writeFile(o,i)}}async function wr(){let r=await hr(["./src/auro-*.js"]);await yr(r)}var is=fr.command("wca-setup").description("Set up WCA (Web Component Analyzer) for the project").action(()=>{wr().then(()=>{tt.success("WCA setup completed successfully.")}).catch(r=>{tt.error(`WCA setup failed: ${r.message}`)})});import{program as Ir}from"commander";import st from"chalk";import it from"ora";import{appendFile as br,readFile as kr}from"node:fs/promises";import{Logger as f}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{simpleGit as $r}from"simple-git";var g;try{g=$r({baseDir:process.cwd(),binary:"git",maxConcurrentProcesses:1})}catch(r){f.error(`Failed to initialize git: ${r}`),g={}}var v=class r{static async checkGitignore(t){if(t==="")return!1;try{return(await kr(".gitignore","utf-8")).includes(t)}catch(e){return f.error(`Error reading file: ${e}`),!1}}static async getCommitMessages(t=""){try{let e=t;e||(e=(await g.branchLocal()).current);let n="main",o="";if(!!process.env.GITHUB_ACTIONS){n=process.env.GITHUB_BASE_REF||"main";try{if(await g.fetch("origin",n),e!=="HEAD")try{await g.raw(["rev-parse","--verify",`origin/${e}`])}catch{await g.fetch("origin",e)}let s=e==="HEAD"?"HEAD":`origin/${e}`;o=`${(await g.raw(["merge-base",`origin/${n}`,s])).trim()}..${s}`}catch(s){f.warn(`Error setting up commit range in CI: ${s}`);let a=e==="HEAD"?"HEAD":`origin/${e}`;o=`origin/${n}..${a}`}}else try{try{await g.raw(["rev-parse","--verify",`origin/${n}`])}catch{f.info(`Fetching ${n} from origin`),await g.fetch("origin",n)}if(e!=="HEAD")try{await g.raw(["rev-parse","--verify",e])}catch{await g.fetch("origin",e)}o=`${(await g.raw(["merge-base",`origin/${n}`,e])).trim()}..${e}`}catch(s){f.warn(`Error determining commits locally: ${s}`),o=`${e}~10..${e}`}return await r.getFormattedCommits(o)}catch(e){return f.error(`Error getting commit messages: ${e}`),[]}}static async getRepoOwnerAndName(){try{let t=await g.getRemotes(!0);if(t.length===0)return f.warn("No remotes found"),null;let e=t.find(o=>o.name==="origin")||t[0],n=e.refs.fetch||e.refs.push;return r.parseGitUrl(n)}catch(t){return f.error(`Error getting repo owner and name: ${t}`),null}}static async getCurrentBranchName(){try{return(await g.branchLocal()).current||null}catch(t){return f.error(`Error getting current branch name: ${t}`),null}}static parseGitUrl(t){let e;return t.includes("@")&&t.includes(":")&&(e=t.match(/@([^:]+):([^/]+)\/(.+?)(?:\.git)?$/),e)?{owner:e[2],repo:e[3]}:(e=t.match(/https?:\/\/(?:[^@]+@)?[^/]+\/([^/]+)\/(.+?)(?:\.git)?$/),e?{owner:e[1],repo:e[2]}:(f.warn(`Could not parse git URL: ${t}`),null))}static async getFormattedCommits(t){let n=(await g.raw(["log","--pretty=format:COMMIT_START%n%H%n%ad%n%an%n%s%n%b%nCOMMIT_END","--date=short",t])).split(`COMMIT_START
31
31
  `).filter(i=>i.trim()!==""),o=[];for(let i of n){let s=i.split(`
32
32
  `);if(s.length>=4){let a=s[0],l=s[1],m=s[2],u=s[3],c=s.slice(4).filter(C=>C!=="COMMIT_END"),d=c.length>0?c.join(""):"",h=a.substring(0,7),p=u.match(/^(feat|fix|docs|style|refactor|perf|test|build|ci|chore)(\(.+\))?:/),w=p?p[1]:"unknown";d.includes("BREAKING CHANGE")&&(w="breaking"),o.push({type:w,hash:h,date:l,subject:u,body:d,message:`${u}${d?`
33
33
 
34
34
  ${d}`:""}`,author_name:m})}}return o}static async addToGitignore(t,e=!0){await r.checkGitignore(t).then(async n=>{if(n)f.warn(`${t} already exists`);else try{await br(".gitignore",`
35
- ${t}`),e&&f.success(`${t} added to .gitignore`)}catch(o){f.error(o)}})}static async removeFromGitCache(t){try{await g.rmKeepLocal(t),f.success(`${t.join(", ")} are removed from git cache`)}catch(e){f.error(e)}}static async createBranch(t){try{await g.checkoutLocalBranch(t),f.success(`Created and switched to ${t} branch`)}catch(e){f.error(e)}}static async commitStagedFiles(t){try{await g.add("."),await g.commit(t),f.success(`Committed with message: ${t}`)}catch(e){f.error(e)}}};import y from"chalk";var Ar=60,vr=100;function R(r){switch(r){case"breaking":return y.bold.red(r);case"feat":return y.bold.green(r);case"fix":return y.bold.green(r);case"perf":return y.bold.green(r);case"docs":return y.bold.cyan(r);case"style":return y.bold.cyan(r);case"refactor":return y.bold.cyan(r);case"test":return y.bold.cyan(r);case"build":return y.bold.cyan(r);case"ci":return y.bold.cyan(r);case"chore":return y.bold.cyan(r);default:return y.bold.white(r)}}function tt(r,t){if(!r)return"";if(r.length<=t)return r;let e=r.split(" "),n="",o="";for(let i of e)(o+i).length>t&&o.length>0&&(n+=`${o.trim()}
36
- `,o=""),o=`${o}${i} `;return o.length>0&&(n+=o.trim()),n}function rt(r){for(let t of r){console.log("\u2500".repeat(60));let e=tt(t.subject,Ar),n=tt(t.body,vr);console.log(y.bold(`${R(t.type)}`)),console.log(y.dim(`${t.hash} | ${t.date} | ${t.author_name}`)),console.log(y.bold(`${y.white(e)}`)),t.body&&console.log(y.dim(n))}console.log("\u2500".repeat(60)),console.log(`
37
- `)}import K from"@actions/github";async function de(){try{let r=process.env.GITHUB_TOKEN;if(!r)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let t=K.getOctokit(r),{context:e}=K;if(!e.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[n,o]=process.env.GITHUB_REPOSITORY.split("/"),i=e.payload.pull_request.number,{data:s}=await t.rest.issues.listLabelsOnIssue({owner:n,repo:o,issue_number:i});return s.map(a=>a.name)}catch(r){throw r instanceof Error?new Error(`Failed to get existing labels: ${r.message}`):r}}async function nt(r){try{let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let e=K.getOctokit(t),{context:n}=K;if(!n.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[o,i]=process.env.GITHUB_REPOSITORY.split("/"),s=n.payload.pull_request.number,a=`semantic-status: ${r}`,l=await de();if(l.includes(a))return;let m=l.filter(u=>u.startsWith("semantic-status:")&&u!==a);for(let u of m)await e.rest.issues.removeLabel({owner:o,repo:i,issue_number:s,name:u});await e.rest.issues.addLabels({owner:o,repo:i,issue_number:s,labels:[a]});return}catch(t){throw t instanceof Error?new Error(`Failed to apply label: ${t.message}`):t}}var Er=["feat","fix","breaking","perf"];function pe(r,t=!0){let e=`### In this release
35
+ ${t}`),e&&f.success(`${t} added to .gitignore`)}catch(o){f.error(o)}})}static async removeFromGitCache(t){try{await g.rmKeepLocal(t),f.success(`${t.join(", ")} are removed from git cache`)}catch(e){f.error(e)}}static async createBranch(t){try{await g.checkoutLocalBranch(t),f.success(`Created and switched to ${t} branch`)}catch(e){f.error(e)}}static async commitStagedFiles(t){try{await g.add("."),await g.commit(t),f.success(`Committed with message: ${t}`)}catch(e){f.error(e)}}};import y from"chalk";var Ar=60,vr=100;function D(r){switch(r){case"breaking":return y.bold.red(r);case"feat":return y.bold.green(r);case"fix":return y.bold.green(r);case"perf":return y.bold.green(r);case"docs":return y.bold.cyan(r);case"style":return y.bold.cyan(r);case"refactor":return y.bold.cyan(r);case"test":return y.bold.cyan(r);case"build":return y.bold.cyan(r);case"ci":return y.bold.cyan(r);case"chore":return y.bold.cyan(r);default:return y.bold.white(r)}}function rt(r,t){if(!r)return"";if(r.length<=t)return r;let e=r.split(" "),n="",o="";for(let i of e)(o+i).length>t&&o.length>0&&(n+=`${o.trim()}
36
+ `,o=""),o=`${o}${i} `;return o.length>0&&(n+=o.trim()),n}function nt(r){for(let t of r){console.log("\u2500".repeat(60));let e=rt(t.subject,Ar),n=rt(t.body,vr);console.log(y.bold(`${D(t.type)}`)),console.log(y.dim(`${t.hash} | ${t.date} | ${t.author_name}`)),console.log(y.bold(`${y.white(e)}`)),t.body&&console.log(y.dim(n))}console.log("\u2500".repeat(60)),console.log(`
37
+ `)}import J from"@actions/github";async function de(){try{let r=process.env.GITHUB_TOKEN;if(!r)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let t=J.getOctokit(r),{context:e}=J;if(!e.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[n,o]=process.env.GITHUB_REPOSITORY.split("/"),i=e.payload.pull_request.number,{data:s}=await t.rest.issues.listLabelsOnIssue({owner:n,repo:o,issue_number:i});return s.map(a=>a.name)}catch(r){throw r instanceof Error?new Error(`Failed to get existing labels: ${r.message}`):r}}async function ot(r){try{let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let e=J.getOctokit(t),{context:n}=J;if(!n.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[o,i]=process.env.GITHUB_REPOSITORY.split("/"),s=n.payload.pull_request.number,a=`semantic-status: ${r}`,l=await de();if(l.includes(a))return;let m=l.filter(u=>u.startsWith("semantic-status:")&&u!==a);for(let u of m)await e.rest.issues.removeLabel({owner:o,repo:i,issue_number:s,name:u});await e.rest.issues.addLabels({owner:o,repo:i,issue_number:s,labels:[a]});return}catch(t){throw t instanceof Error?new Error(`Failed to apply label: ${t.message}`):t}}var Er=["feat","fix","breaking","perf"];function pe(r,t=!0){let e=`### In this release
38
38
  `;for(let n of r)if(e+=`- ${n.hash} ${n.subject}
39
- `,n.body?.trim()){let i=n.body.trim().split(/\n+/).map(s=>s.trim()).filter(s=>s.length>0);for(let s of i){let a=s;a=a.replace(/([^\s])(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)/g,"$1 $2"),a=a.replace(/(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)([^\s])/g,"$1 $2"),e+=` - ${a}`}}return r.length===0?"":(t&&console.log(ot.green(`\u2713 Generating release notes for ${r.length} commits`)),e)}function fe(r,t=!0){let e=r.filter(o=>Er.includes(o.type)),n;return t?n=e.length>0?e:r:n=e,n.length===0&&console.log(`No commits found to include in release notes.
40
- `),n}async function it(r=!1,t=!1,e=!1){let n=st(`Checking commits...
41
- `).start();try{let o=await v.getCommitMessages();if(e){n.succeed(`Total commits analyzed: ${o.length}`);let i=fe(o);console.log(pe(i));return}if(r&&rt(o),n.succeed(`Total commits analyzed: ${o.length}`),o.length!==0){let i=o.map(l=>l.type),a=Array.from(new Set(i)).map(l=>R(l)).join(", ");n.succeed(`Found commit types: ${a}`)}else n.info(`The list of commits is created by comparing the current branch
39
+ `,n.body?.trim()){let i=n.body.trim().split(/\n+/).map(s=>s.trim()).filter(s=>s.length>0);for(let s of i){let a=s;a=a.replace(/([^\s])(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)/g,"$1 $2"),a=a.replace(/(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)([^\s])/g,"$1 $2"),e+=` - ${a}`}}return r.length===0?"":(t&&console.log(st.green(`\u2713 Generating release notes for ${r.length} commits`)),e)}function fe(r,t=!0){let e=r.filter(o=>Er.includes(o.type)),n;return t?n=e.length>0?e:r:n=e,n.length===0&&console.log(`No commits found to include in release notes.
40
+ `),n}async function at(r=!1,t=!1,e=!1){let n=it(`Checking commits...
41
+ `).start();try{let o=await v.getCommitMessages();if(e){n.succeed(`Total commits analyzed: ${o.length}`);let i=fe(o);console.log(pe(i));return}if(r&&nt(o),n.succeed(`Total commits analyzed: ${o.length}`),o.length!==0){let i=o.map(l=>l.type),a=Array.from(new Set(i)).map(l=>D(l)).join(", ");n.succeed(`Found commit types: ${a}`)}else n.info(`The list of commits is created by comparing the current branch
42
42
  with the main branch. If you are on a new branch, please
43
- make sure to commit some changes before running this command.`);t&&await Cr(o,n)}catch(o){n.fail("Error getting commit messages"),console.error(o)}}async function Cr(r,t){let e=["breaking","feat","fix","perf","docs","style","refactor","test","build","ci","chore"],n=r.map(s=>s.type).filter(s=>e.includes(s)),o=null,i=Number.POSITIVE_INFINITY;for(let s of n){let a=e.indexOf(s);a<i&&(i=a,o=s)}if(o){let s=st("Checking existing labels on pull request...").start();try{if((await de()).includes(`semantic-status: ${o}`)){s.info(`Label "semantic-status: ${R(o)}" already exists on the pull request.`);return}s.text="Applying label to pull request...",await nt(o),s.succeed(`Label "semantic-status: ${R(o)}" applied to the pull request.`)}catch(a){let l=a instanceof Error?a.message:String(a);s.fail(l)}}else t.warn(ot.yellow("No semantic commit type found to apply as label."))}var Es=Ir.command("check-commits").alias("cc").option("-l, --set-label","Set label on the pull request based on the commit message type").option("-d, --debug","Display detailed commit information for debugging").option("-r, --release-notes","Generate release notes based on commit messages").description("Check commits in the local repository for the types of semantic commit messages made and return the results.").action(async r=>{await it(r.debug,r.setLabel,r.releaseNotes)});import at from"node:fs";import{get as ct}from"node:https";import E from"chalk";import{program as Pr}from"commander";import Tr from"ora";var Ds=Pr.command("pr-release").option("-n, --namespace <package-namespace>","Set namespace of the package release","@aurodesignsystem-dev").option("-p, --pr-number <number>","Set pull request number for the release","0").description("Generate the package version based off of PR number then update the package.json file. Note: this does not publish the package.").action(async r=>{await xr(r)}),xr=async r=>{let{namespace:t,prNumber:e}=r,n=Tr("Updating package.json").start();try{let o="package.json",i=JSON.parse(at.readFileSync(o,"utf8"));n.text="Checking npm registry for version information...";let s=`0.0.0-pr${e}`,a=i.name.split("/")[1],l=`${t}/${a}`,m=await Rr(s,l,n),u=`${s}.${m}`;i.name=l,i.version=u,n.text="Writing updated package.json...",at.writeFileSync(o,`${JSON.stringify(i,null,2)}
44
- `,"utf8"),n.succeed(`Package.json updated to use ${E.green(u)} and ${E.green(l)}`),process.exit(0)}catch(o){n.fail(`Failed to update package.json: ${o}`),process.exit(1)}},Rr=(r,t,e)=>new Promise(n=>{try{let a=function(l){if(l.statusCode!==200){e.info(`Package not found. Status code: ${E.red(l.statusCode)}, defaulting to version 0`),n(0);return}e.text="Processing version information...";let m="";l.on("data",u=>{m+=u}),l.on("end",()=>{try{let u=JSON.parse(m),c=u.versions?Object.keys(u.versions):[];e.text="Calculating next version number...";let d=-1,h=new RegExp(`^${r}\\.(\\d+)$`);for(let p of c){let w=p.match(h);if(w){let C=Number.parseInt(w[1],10);d=Math.max(d,C)}}d>=0?e.info(`Found existing version ${E.green(`${r}.${d}`)}. Incrementing to ${E.green(`${r}.${d+1}`)}`):e.info(`No existing version found for ${E.green(r)}. Starting with ${E.green(`${r}.0`)}`),n(d+1)}catch(u){e.warn(`Failed to parse NPM registry response: ${u instanceof Error?u.message:"Unknown error"}, defaulting to version 0`),n(0)}})};var o=a;let i=`https://registry.npmjs.org/${t}`,s=ct(i,{headers:{Accept:"application/json"}},l=>{if((l.statusCode===301||l.statusCode===302)&&l.headers.location){e.info(`Following redirect to ${l.headers.location}...`);try{ct(l.headers.location,{headers:{Accept:"application/json"}},a).on("error",m=>{e.warn(`Error following redirect: ${m.message}, defaulting to version 0`),n(0)}).end()}catch(m){e.warn(`Redirect request failed: ${m instanceof Error?m.message:"Unknown error"}, defaulting to version 0`),n(0)}return}a(l)});s.on("error",l=>{e.warn(`Request error: ${l.message}, defaulting to version 0`),n(0)}),s.end()}catch{e.warn("Error checking version in npm registry, defaulting to version 0"),n(0)}});import ge from"node:path";import{fileURLToPath as Dr}from"node:url";import{program as Sr}from"commander";import Or from"open";var jr=Dr(import.meta.url),Nr=ge.resolve(ge.dirname(jr),".."),Bs=Sr.command("test").option("-w, --watch","Set watch number for the test").option("-c, --coverage-report","Generate coverage report").option("-o, --open","Open the coverage report in the browser").option("-f, --files <String|String[]>","Test files glob pattern").description("Run the web test runner to test the component library").action(async r=>{let e=`npx wtr --config "${ge.join(Nr,"dist","configs","web-test-runner.config.mjs")}"`,n=`${process.cwd()}/coverage/index.html`;if(r.coverageReport&&(e+=" --coverage"),r.watch&&(e+=" --watch"),r.files){let o=Array.isArray(r.files)?r.files.join(" "):r.files;e+=` --files "${o}"`}b(e),r.open&&await Or(n)});import be from"node:fs/promises";import dt from"node:path";import{program as Lr}from"commander";import Ur from"inquirer";import ke from"ora";import _r from"node:fs/promises";import Fr from"node:path";import Br from"ora";var Gr={"auth-type":"workspace-token","author-email":null,"author-name":null,"base-branch":"main","base-url":null,"clone-dir":".gitter-temp","code-search":null,concurrent:4,"conflict-strategy":"replace",draft:!1,"dry-run":!0,"fetch-depth":1,fork:!1,"fork-owner":null,"git-type":"go",group:null,"include-subgroups":!1,insecure:!1,interactive:!1,labels:null,"log-file":"'-'","log-format":"'text'","log-level":"'error'","max-reviewers":0,"max-team-reviewers":0,org:null,output:"'-'","plain-output":!1,platform:"github",project:null,"push-only":!1,repo:["AlaskaAirlines/auro-accordion","AlaskaAirlines/auro-alert","AlaskaAirlines/auro-avatar","AlaskaAirlines/auro-background","AlaskaAirlines/auro-backtotop","AlaskaAirlines/auro-button","AlaskaAirlines/auro-badge","AlaskaAirlines/auro-banner","AlaskaAirlines/auro-card","AlaskaAirlines/auro-carousel","AlaskaAirlines/auro-datetime","AlaskaAirlines/auro-dialog","AlaskaAirlines/auro-drawer","AlaskaAirlines/auro-flight","AlaskaAirlines/auro-flightline","AlaskaAirlines/auro-header","AlaskaAirlines/auro-hyperlink","AlaskaAirlines/auro-icon","AlaskaAirlines/auro-loader","AlaskaAirlines/auro-lockup","AlaskaAirlines/auro-nav","AlaskaAirlines/auro-pane","AlaskaAirlines/auro-popover","AlaskaAirlines/auro-sidenav","AlaskaAirlines/auro-skeleton","AlaskaAirlines/auro-slideshow","AlaskaAirlines/auro-table","AlaskaAirlines/auro-tabs","AlaskaAirlines/auro-toast"],"repo-exclude":null,"repo-include":null,"repo-search":null,reviewers:null,"skip-forks":!1,"skip-pr":!1,"skip-repo":null,"ssh-auth":!1,"team-reviewers":null};function Mr(r){return Object.entries(r).map(([t,e])=>Array.isArray(e)?`${t}:
43
+ make sure to commit some changes before running this command.`);t&&await Cr(o,n)}catch(o){n.fail("Error getting commit messages"),console.error(o)}}async function Cr(r,t){let e=["breaking","feat","fix","perf","docs","style","refactor","test","build","ci","chore"],n=r.map(s=>s.type).filter(s=>e.includes(s)),o=null,i=Number.POSITIVE_INFINITY;for(let s of n){let a=e.indexOf(s);a<i&&(i=a,o=s)}if(o){let s=it("Checking existing labels on pull request...").start();try{if((await de()).includes(`semantic-status: ${o}`)){s.info(`Label "semantic-status: ${D(o)}" already exists on the pull request.`);return}s.text="Applying label to pull request...",await ot(o),s.succeed(`Label "semantic-status: ${D(o)}" applied to the pull request.`)}catch(a){let l=a instanceof Error?a.message:String(a);s.fail(l)}}else t.warn(st.yellow("No semantic commit type found to apply as label."))}var Es=Ir.command("check-commits").alias("cc").option("-l, --set-label","Set label on the pull request based on the commit message type").option("-d, --debug","Display detailed commit information for debugging").option("-r, --release-notes","Generate release notes based on commit messages").description("Check commits in the local repository for the types of semantic commit messages made and return the results.").action(async r=>{await at(r.debug,r.setLabel,r.releaseNotes)});import ct from"node:fs";import{get as lt}from"node:https";import E from"chalk";import{program as Pr}from"commander";import Tr from"ora";var Ds=Pr.command("pr-release").option("-n, --namespace <package-namespace>","Set namespace of the package release","@aurodesignsystem-dev").option("-p, --pr-number <number>","Set pull request number for the release","0").description("Generate the package version based off of PR number then update the package.json file. Note: this does not publish the package.").action(async r=>{await xr(r)}),xr=async r=>{let{namespace:t,prNumber:e}=r,n=Tr("Updating package.json").start();try{let o="package.json",i=JSON.parse(ct.readFileSync(o,"utf8"));n.text="Checking npm registry for version information...";let s=`0.0.0-pr${e}`,a=i.name.split("/")[1],l=`${t}/${a}`,m=await Rr(s,l,n),u=`${s}.${m}`;i.name=l,i.version=u,n.text="Writing updated package.json...",ct.writeFileSync(o,`${JSON.stringify(i,null,2)}
44
+ `,"utf8"),n.succeed(`Package.json updated to use ${E.green(u)} and ${E.green(l)}`),process.exit(0)}catch(o){n.fail(`Failed to update package.json: ${o}`),process.exit(1)}},Rr=(r,t,e)=>new Promise(n=>{try{let a=function(l){if(l.statusCode!==200){e.info(`Package not found. Status code: ${E.red(l.statusCode)}, defaulting to version 0`),n(0);return}e.text="Processing version information...";let m="";l.on("data",u=>{m+=u}),l.on("end",()=>{try{let u=JSON.parse(m),c=u.versions?Object.keys(u.versions):[];e.text="Calculating next version number...";let d=-1,h=new RegExp(`^${r}\\.(\\d+)$`);for(let p of c){let w=p.match(h);if(w){let C=Number.parseInt(w[1],10);d=Math.max(d,C)}}d>=0?e.info(`Found existing version ${E.green(`${r}.${d}`)}. Incrementing to ${E.green(`${r}.${d+1}`)}`):e.info(`No existing version found for ${E.green(r)}. Starting with ${E.green(`${r}.0`)}`),n(d+1)}catch(u){e.warn(`Failed to parse NPM registry response: ${u instanceof Error?u.message:"Unknown error"}, defaulting to version 0`),n(0)}})};var o=a;let i=`https://registry.npmjs.org/${t}`,s=lt(i,{headers:{Accept:"application/json"}},l=>{if((l.statusCode===301||l.statusCode===302)&&l.headers.location){e.info(`Following redirect to ${l.headers.location}...`);try{lt(l.headers.location,{headers:{Accept:"application/json"}},a).on("error",m=>{e.warn(`Error following redirect: ${m.message}, defaulting to version 0`),n(0)}).end()}catch(m){e.warn(`Redirect request failed: ${m instanceof Error?m.message:"Unknown error"}, defaulting to version 0`),n(0)}return}a(l)});s.on("error",l=>{e.warn(`Request error: ${l.message}, defaulting to version 0`),n(0)}),s.end()}catch{e.warn("Error checking version in npm registry, defaulting to version 0"),n(0)}});import ge from"node:path";import{fileURLToPath as Dr}from"node:url";import{program as Sr}from"commander";import Or from"open";var jr=Dr(import.meta.url),Nr=ge.resolve(ge.dirname(jr),".."),Bs=Sr.command("test").option("-w, --watch","Set watch number for the test").option("-c, --coverage-report","Generate coverage report").option("-o, --open","Open the coverage report in the browser").option("-f, --files <String|String[]>","Test files glob pattern").description("Run the web test runner to test the component library").action(async r=>{let e=`npx wtr --config "${ge.join(Nr,"dist","configs","web-test-runner.config.mjs")}"`,n=`${process.cwd()}/coverage/index.html`;if(r.coverageReport&&(e+=" --coverage"),r.watch&&(e+=" --watch"),r.files){let o=Array.isArray(r.files)?r.files.join(" "):r.files;e+=` --files "${o}"`}b(e),r.open&&await Or(n)});import be from"node:fs/promises";import pt from"node:path";import{program as Lr}from"commander";import Ur from"inquirer";import ke from"ora";import Fr from"node:fs/promises";import _r from"node:path";import Br from"ora";var Gr={"auth-type":"workspace-token","author-email":null,"author-name":null,"base-branch":"main","base-url":null,"clone-dir":".gitter-temp","code-search":null,concurrent:4,"conflict-strategy":"replace",draft:!1,"dry-run":!0,"fetch-depth":1,fork:!1,"fork-owner":null,"git-type":"go",group:null,"include-subgroups":!1,insecure:!1,interactive:!1,labels:null,"log-file":"'-'","log-format":"'text'","log-level":"'error'","max-reviewers":0,"max-team-reviewers":0,org:null,output:"'-'","plain-output":!1,platform:"github",project:null,"push-only":!1,repo:["AlaskaAirlines/auro-accordion","AlaskaAirlines/auro-alert","AlaskaAirlines/auro-avatar","AlaskaAirlines/auro-background","AlaskaAirlines/auro-backtotop","AlaskaAirlines/auro-button","AlaskaAirlines/auro-badge","AlaskaAirlines/auro-banner","AlaskaAirlines/auro-card","AlaskaAirlines/auro-carousel","AlaskaAirlines/auro-datetime","AlaskaAirlines/auro-dialog","AlaskaAirlines/auro-drawer","AlaskaAirlines/auro-flight","AlaskaAirlines/auro-flightline","AlaskaAirlines/auro-header","AlaskaAirlines/auro-hyperlink","AlaskaAirlines/auro-icon","AlaskaAirlines/auro-loader","AlaskaAirlines/auro-lockup","AlaskaAirlines/auro-nav","AlaskaAirlines/auro-pane","AlaskaAirlines/auro-popover","AlaskaAirlines/auro-sidenav","AlaskaAirlines/auro-skeleton","AlaskaAirlines/auro-slideshow","AlaskaAirlines/auro-table","AlaskaAirlines/auro-tabs","AlaskaAirlines/auro-toast"],"repo-exclude":null,"repo-include":null,"repo-search":null,reviewers:null,"skip-forks":!1,"skip-pr":!1,"skip-repo":null,"ssh-auth":!1,"team-reviewers":null};function Mr(r){return Object.entries(r).map(([t,e])=>Array.isArray(e)?`${t}:
45
45
  - ${e.join(`
46
46
  - `)}`:typeof e=="object"&&e!==null?`${t}:
47
47
  ${Object.entries(e).map(([n,o])=>` ${n}: ${o}`).join(`
48
48
  `)}`:`${t}: ${e}`).join(`
49
- `)}async function lt(r){let t=Br("Writing multi-gitter configuration...").start(),e=Mr(Gr),n=Fr.join(r,"multi-gitter_DEPENDENCY_TREE.yml");try{await _r.writeFile(n,e,"utf8"),t.succeed(`Multi-gitter configuration written to ${n}`)}catch(o){t.fail("Error writing multi-gitter configuration:"),console.error(o)}}import he from"node:fs";function ut(r){let t={},e=[],n=[],o=[];for(let i in r)t[i]=r[i].dependsOn.length;for(let i in t)t[i]===0&&o.push(i);for(;o.length>0;){n=[];let i=o.length;for(let s=0;s<i;s++){let a=o.shift();n.push(a);for(let l of r[a].dependentPackages)t[l]--,t[l]===0&&o.push(l)}e.push(n)}if(e.flat().length!==Object.keys(r).length)throw new Error("Circular dependency detected!");return e}function Hr(r){return he.readdirSync(r).filter(t=>t.endsWith(".json"))}async function mt(r,t=[]){console.log(t);let e={},n=Hr(r);for(let o of n){if(o==="dependencyTree.json")continue;let i=he.readFileSync(`${r}/${o}`,"utf-8"),s=JSON.parse(i),a=s.name,l=Object.keys(s.peerDependencies),m=Object.keys(s.devDependencies),u=Object.keys(s.dependencies);e[a]||(e[a]={dependsOn:[],dependentPackages:[]});let c=[...l,...m,...u];e[a].dependsOn=[...new Set(c)];for(let d of c)e[d]||(e[d]={dependsOn:[],dependentPackages:[]}),e[d].dependentPackages.includes(a)||e[d].dependentPackages.push(a)}if(t.length){let o=new Set;for(let[s,a]of Object.entries(e))a.dependsOn.some(l=>t.includes(l))&&o.add(s);for(let s of t)e[s]&&o.add(s);let i={};for(let s of o)i[s]={dependsOn:e[s].dependsOn.filter(a=>o.has(a)),dependentPackages:e[s].dependentPackages.filter(a=>o.has(a))};e=i}else console.log("No target dependencies provided - using all packages.");return he.writeFileSync(`${r}/dependencyTree.json`,JSON.stringify(e,null,2)),e}var ye=ee("run-migrations","config"),we=ee("run-migrations","outputs");var pt=["@aurodesignsystem/auro-accordion","@aurodesignsystem/auro-alert","@aurodesignsystem/auro-avatar","@aurodesignsystem/auro-background","@aurodesignsystem/auro-backtotop","@aurodesignsystem/auro-button","@aurodesignsystem/auro-badge","@aurodesignsystem/auro-banner","@aurodesignsystem/auro-card","@aurodesignsystem/auro-carousel","@aurodesignsystem/auro-datetime","@aurodesignsystem/auro-dialog","@aurodesignsystem/auro-drawer","@aurodesignsystem/auro-formkit","@aurodesignsystem/auro-flight","@aurodesignsystem/auro-flightline","@aurodesignsystem/auro-header","@aurodesignsystem/auro-hyperlink","@aurodesignsystem/auro-icon","@aurodesignsystem/auro-loader","@aurodesignsystem/auro-lockup","@aurodesignsystem/auro-nav","@aurodesignsystem/auro-pane","@aurodesignsystem/auro-popover","@aurodesignsystem/auro-sidenav","@aurodesignsystem/auro-skeleton","@aurodesignsystem/auro-slideshow","@aurodesignsystem/auro-table","@aurodesignsystem/auro-tabs","@aurodesignsystem/auro-toast"],ti=[...pt,"@aurodesignsystem/auro-library","@aurodesignsystem/WebCoreStyleSheets","@aurodesignsystem/AuroDesignTokens","@aurodesignsystem/auro-cli","@alaskaairux/icons"];async function Wr(r){try{await be.mkdir(we,{recursive:!0}),await be.mkdir(ye,{recursive:!0})}catch(o){console.error("Failed to create output or config directories:",o),process.exit(1)}let t=ke("Creating dependency tree...").start();t.text="Creating multi-gitter dependency tree configuration...",await lt(ye),t.text="Scraping dependencies from Auro packages...";let n=`multi-gitter run "node ${te("static","getAuroDeps.js")}" --config ${dt.join(ye,"multi-gitter_DEPENDENCY_TREE.yml")}`;try{await b(n)}catch(o){t.fail("Failed to generate dependency tree:"),console.error(o),process.exit(1)}return t.text="Generating dependency tree JSON file using packages...",await mt(we,r),t.succeed("Dependency tree generated successfully."),dt.join(we,"dependencyTree.json")}var zr=async r=>{let t=ke("Loading dependency tree...").start(),e=JSON.parse(await be.readFile(r,"utf-8"));t.text="Processing dependency tree...";let n=ut(e);return t.succeed("Dependency batches created successfully."),n},ri=Lr.command("agent").action(async r=>{let t=await Ur.prompt([{type:"select",name:"agentAction",message:"What agent action do you want to perform?",choices:[{name:"Run a migration on auro components",value:"run-migration"}],default:["run-migration"]},{type:"input",name:"migrationId",message:"What migration id do you want to run?",when:e=>e.agentAction==="run-migration",validate:e=>e.trim()!==""||"Migration id cannot be empty."},{type:"confirm",name:"useExisting",message:"Would you like to specify starting packages?",default:!0,transformer:e=>e?"Yes = Packages related to selections":"No = All packages",when:e=>e.agentAction==="run-migration"},{type:"checkbox",name:"startWithComponents",message:"Enter the components to start with (comma-separated, blank for all):",choices:pt.map(e=>({name:e.replace("@aurodesignsystem/",""),value:e})),when:e=>e.agentAction==="run-migration"&&e.useExisting}]);switch(t.agentAction){case"run-migration":{let e=ke("Running migration...").start(),n=await Wr(t.startWithComponents);e.text="Getting dependency batches from tree...";let i=(await zr(n)).map((s,a)=>`Batch ${a+1}
49
+ `)}async function ut(r){let t=Br("Writing multi-gitter configuration...").start(),e=Mr(Gr),n=_r.join(r,"multi-gitter_DEPENDENCY_TREE.yml");try{await Fr.writeFile(n,e,"utf8"),t.succeed(`Multi-gitter configuration written to ${n}`)}catch(o){t.fail("Error writing multi-gitter configuration:"),console.error(o)}}import he from"node:fs";function mt(r){let t={},e=[],n=[],o=[];for(let i in r)t[i]=r[i].dependsOn.length;for(let i in t)t[i]===0&&o.push(i);for(;o.length>0;){n=[];let i=o.length;for(let s=0;s<i;s++){let a=o.shift();n.push(a);for(let l of r[a].dependentPackages)t[l]--,t[l]===0&&o.push(l)}e.push(n)}if(e.flat().length!==Object.keys(r).length)throw new Error("Circular dependency detected!");return e}function Hr(r){return he.readdirSync(r).filter(t=>t.endsWith(".json"))}async function dt(r,t=[]){console.log(t);let e={},n=Hr(r);for(let o of n){if(o==="dependencyTree.json")continue;let i=he.readFileSync(`${r}/${o}`,"utf-8"),s=JSON.parse(i),a=s.name,l=Object.keys(s.peerDependencies),m=Object.keys(s.devDependencies),u=Object.keys(s.dependencies);e[a]||(e[a]={dependsOn:[],dependentPackages:[]});let c=[...l,...m,...u];e[a].dependsOn=[...new Set(c)];for(let d of c)e[d]||(e[d]={dependsOn:[],dependentPackages:[]}),e[d].dependentPackages.includes(a)||e[d].dependentPackages.push(a)}if(t.length){let o=new Set;for(let[s,a]of Object.entries(e))a.dependsOn.some(l=>t.includes(l))&&o.add(s);for(let s of t)e[s]&&o.add(s);let i={};for(let s of o)i[s]={dependsOn:e[s].dependsOn.filter(a=>o.has(a)),dependentPackages:e[s].dependentPackages.filter(a=>o.has(a))};e=i}else console.log("No target dependencies provided - using all packages.");return he.writeFileSync(`${r}/dependencyTree.json`,JSON.stringify(e,null,2)),e}var ye=te("run-migrations","config"),we=te("run-migrations","outputs");var ft=["@aurodesignsystem/auro-accordion","@aurodesignsystem/auro-alert","@aurodesignsystem/auro-avatar","@aurodesignsystem/auro-background","@aurodesignsystem/auro-backtotop","@aurodesignsystem/auro-button","@aurodesignsystem/auro-badge","@aurodesignsystem/auro-banner","@aurodesignsystem/auro-card","@aurodesignsystem/auro-carousel","@aurodesignsystem/auro-datetime","@aurodesignsystem/auro-dialog","@aurodesignsystem/auro-drawer","@aurodesignsystem/auro-formkit","@aurodesignsystem/auro-flight","@aurodesignsystem/auro-flightline","@aurodesignsystem/auro-header","@aurodesignsystem/auro-hyperlink","@aurodesignsystem/auro-icon","@aurodesignsystem/auro-loader","@aurodesignsystem/auro-lockup","@aurodesignsystem/auro-nav","@aurodesignsystem/auro-pane","@aurodesignsystem/auro-popover","@aurodesignsystem/auro-sidenav","@aurodesignsystem/auro-skeleton","@aurodesignsystem/auro-slideshow","@aurodesignsystem/auro-table","@aurodesignsystem/auro-tabs","@aurodesignsystem/auro-toast"],ti=[...ft,"@aurodesignsystem/auro-library","@aurodesignsystem/WebCoreStyleSheets","@aurodesignsystem/AuroDesignTokens","@aurodesignsystem/auro-cli","@alaskaairux/icons"];async function Wr(r){try{await be.mkdir(we,{recursive:!0}),await be.mkdir(ye,{recursive:!0})}catch(o){console.error("Failed to create output or config directories:",o),process.exit(1)}let t=ke("Creating dependency tree...").start();t.text="Creating multi-gitter dependency tree configuration...",await ut(ye),t.text="Scraping dependencies from Auro packages...";let n=`multi-gitter run "node ${re("static","getAuroDeps.js")}" --config ${pt.join(ye,"multi-gitter_DEPENDENCY_TREE.yml")}`;try{await b(n)}catch(o){t.fail("Failed to generate dependency tree:"),console.error(o),process.exit(1)}return t.text="Generating dependency tree JSON file using packages...",await dt(we,r),t.succeed("Dependency tree generated successfully."),pt.join(we,"dependencyTree.json")}var zr=async r=>{let t=ke("Loading dependency tree...").start(),e=JSON.parse(await be.readFile(r,"utf-8"));t.text="Processing dependency tree...";let n=mt(e);return t.succeed("Dependency batches created successfully."),n},ri=Lr.command("agent").action(async r=>{let t=await Ur.prompt([{type:"select",name:"agentAction",message:"What agent action do you want to perform?",choices:[{name:"Run a migration on auro components",value:"run-migration"}],default:["run-migration"]},{type:"input",name:"migrationId",message:"What migration id do you want to run?",when:e=>e.agentAction==="run-migration",validate:e=>e.trim()!==""||"Migration id cannot be empty."},{type:"confirm",name:"useExisting",message:"Would you like to specify starting packages?",default:!0,transformer:e=>e?"Yes = Packages related to selections":"No = All packages",when:e=>e.agentAction==="run-migration"},{type:"checkbox",name:"startWithComponents",message:"Enter the components to start with (comma-separated, blank for all):",choices:ft.map(e=>({name:e.replace("@aurodesignsystem/",""),value:e})),when:e=>e.agentAction==="run-migration"&&e.useExisting}]);switch(t.agentAction){case"run-migration":{let e=ke("Running migration...").start(),n=await Wr(t.startWithComponents);e.text="Getting dependency batches from tree...";let i=(await zr(n)).map((s,a)=>`Batch ${a+1}
50
50
  ${s.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alaskaairux/icons","AlaskaAirlines/Icons")}`).join(`
51
51
  `)}`).join(`
52
52
 
53
- `);console.log(i),e.text="Running migrations on dependency batches...",new Promise(s=>setTimeout(s,2e3)),e.succeed("Migration process completed successfully.");break}default:console.error("Unknown action selected.")}});import{program as qr}from"commander";var $e=qr.command("docs").description("Generate API documentation").option("-c, --cem","Generate Custom Elements Manifest (CEM) file",!1).option("-a, --api","Creates api md file from CEM",!1).option("--skip-readme","Skip README.md processing",!1);$e=j($e);var ai=$e.action(async r=>{r.cem&&await G(),r.api&&await M(),await Se(r),r.serve&&await Oe(r)});import{program as Xr}from"commander";import{Octokit as Ae}from"@octokit/rest";import*as Y from"azure-devops-node-api";import J from"ora";var Vr=async r=>{let t=process.env.GH_TOKEN;if(!t)throw new Error("GH_TOKEN environment variable is required");let e=new Ae({auth:t}),n,o,i;if(r.includes("github.com")){let a=r.match(/github\.com\/([^\/]+)\/([^\/]+)\/issues\/(\d+)/);if(!a)throw new Error("Invalid GitHub issue URL format");[,n,o,i]=a}else if(r.includes("#")){let a=r.match(/([^\/]+)\/([^#]+)#(\d+)/);if(!a)throw new Error("Invalid GitHub issue reference format");[,n,o,i]=a}else throw new Error("Issue must be provided as full URL or in format 'owner/repo#number'");let s=Number.parseInt(i,10);try{let{data:a}=await e.rest.issues.get({owner:n,repo:o,issue_number:s});return{title:a.title,body:a.body??null,html_url:a.html_url,number:a.number,repository:{owner:{login:n},name:o}}}catch(a){throw new Error(`Failed to fetch GitHub issue: ${a}`)}},Kr=async r=>{let t=process.env.GH_TOKEN;if(!t)return null;let e=new Ae({auth:t});try{let n=`
53
+ `);console.log(i),e.text="Running migrations on dependency batches...",new Promise(s=>setTimeout(s,2e3)),e.succeed("Migration process completed successfully.");break}default:console.error("Unknown action selected.")}});import{program as qr}from"commander";var $e=qr.command("docs").description("Generate API documentation").option("-c, --cem","Generate Custom Elements Manifest (CEM) file",!1).option("-a, --api","Creates api md file from CEM",!1).option("--skip-readme","Skip README.md processing",!1);$e=N($e);var ai=$e.action(async r=>{r.cem&&await M(),r.api&&await H(),await Se(r),r.serve&&await Oe(r)});import{program as Xr}from"commander";import{Octokit as Ae}from"@octokit/rest";import*as X from"azure-devops-node-api";import Y from"ora";var Vr=async r=>{let t=process.env.GH_TOKEN;if(!t)throw new Error("GH_TOKEN environment variable is required");let e=new Ae({auth:t}),n,o,i;if(r.includes("github.com")){let a=r.match(/github\.com\/([^\/]+)\/([^\/]+)\/issues\/(\d+)/);if(!a)throw new Error("Invalid GitHub issue URL format");[,n,o,i]=a}else if(r.includes("#")){let a=r.match(/([^\/]+)\/([^#]+)#(\d+)/);if(!a)throw new Error("Invalid GitHub issue reference format");[,n,o,i]=a}else throw new Error("Issue must be provided as full URL or in format 'owner/repo#number'");let s=Number.parseInt(i,10);try{let{data:a}=await e.rest.issues.get({owner:n,repo:o,issue_number:s});return{title:a.title,body:a.body??null,html_url:a.html_url,number:a.number,repository:{owner:{login:n},name:o}}}catch(a){throw new Error(`Failed to fetch GitHub issue: ${a}`)}},Kr=async r=>{let t=process.env.GH_TOKEN;if(!t)return null;let e=new Ae({auth:t});try{let n=`
54
54
  query($owner: String!, $repo: String!, $issueNumber: Int!) {
55
55
  repository(owner: $owner, name: $repo) {
56
56
  issue(number: $issueNumber) {
@@ -143,4 +143,4 @@ ${s.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
143
143
  }
144
144
  }
145
145
  }
146
- `,{projectId:l,itemId:c,fieldId:u.id,value:t});else if(!u)throw new Error("No 'ado' field found in GitHub project")}catch(i){console.error(`Failed to update GitHub project: ${i}`)}},Yr=async r=>{let t=process.env.ADO_TOKEN;if(!t)throw new Error("ADO_TOKEN environment variable is required");let e="https://dev.azure.com/itsals",n="E_Retain_Content",o="E_Retain_Content\\Auro Design System",i=Y.getPersonalAccessTokenHandler(t),a=await new Y.WebApi(e,i).getWorkItemTrackingApi();try{let l=[{op:"add",path:"/fields/System.Title",value:r.title},{op:"add",path:"/fields/System.Description",value:`GitHub Issue: <a href="${r.html_url}">${r.html_url}</a>`},{op:"add",path:"/fields/System.AreaPath",value:o}];return await a.createWorkItem(null,l,n,"User Story")}catch(l){throw new Error(`Failed to create ADO work item: ${l}`)}},ft=async r=>{let t=J(`Processing GitHub issue: ${r}`).start();try{if(!process.env.GH_TOKEN)throw new Error("GH_TOKEN environment variable is required");if(!process.env.ADO_TOKEN)throw new Error("ADO_TOKEN environment variable is required");t.text="Fetching GitHub issue details...";let e=await Vr(r);t.succeed(`Found issue: "${e.title}"`);let n=J("Checking for existing ADO work item...").start(),o=await Kr(e);if(o){n.succeed("ADO work item already exists for this issue!"),console.log(`${o}`);return}n.succeed("No existing ADO work item found");let i=J("Creating new ADO work item...").start(),s=await Yr(e);if(i.succeed(`Successfully created ADO work item #${s.id}`),console.log(`Work item: ${s._links?.html?.href||"N/A"}`),s._links?.html?.href){let a=J("Adding to GitHub project and updating ADO field...").start();await Jr(e,s._links.html.href),a.succeed("Updated GitHub project with ADO link")}}catch(e){t.fail(`Error: ${e instanceof Error?e.message:e}`),process.exit(1)}};var fi=Xr.command("ado").description("Generate ADO item from GitHub issue").option("-g, --gh-issue <issue>","What GitHub issue to use").action(async r=>{r.ghIssue&&await ft(r.ghIssue)});import{program as tn}from"commander";import{Octokit as Zr}from"@octokit/rest";import{simpleGit as Qr}from"simple-git";var gt="Release Candidate",D="dev",en="main",X=class r{constructor(t,e,n){this.filteredCommits=null;this.repoInfo={owner:t,repo:e},this.octokit=n}static async create(){let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN is required to run RC workflow.");let e=await v.getRepoOwnerAndName(),n=new Zr({auth:t});if(!e)throw new Error("Failed to retrieve repository information. Ensure you're in a valid git repository.");let o=await r.getTriggerBranchName();return o&&o!==D&&(console.log(`Switching from ${o} to ${D} branch...`),await Qr().checkout(D)),new r(e.owner,e.repo,n)}get owner(){return this.repoInfo.owner}get repo(){return this.repoInfo.repo}get repoData(){return{...this.repoInfo}}async createReleaseCandidate(){await this.hasCommitsReadyInDev()||console.log("No filtered commits found. Continuing to update RC issue/branch/PR.");let e=await this.getLatestOpenRcIssue(),n=e?await this.getLinkedPrByHead(e.number):null;if(n?.multipleOpen)throw new Error("Multiple open RC PRs found for the same rc/<issueNumber> branch.");if(n?.state==="closed"&&(console.log("Linked RC PR is closed. Creating a new RC issue and PR."),e=await this.createRcIssue(),n=null),e?await this.updateRcIssue(e.number):e=await this.createRcIssue(),!e)throw new Error("Failed to resolve RC issue.");await this.createOrUpdateRcBranch(e.number),n?await this.updateRcPr(e.number,n.number):n=await this.createRcPr(e.number)}async getFilteredCommits(){if(this.filteredCommits===null){let t=await v.getCommitMessages(D);this.filteredCommits=fe(t)}return this.filteredCommits}async hasCommitsReadyInDev(){return(await this.getFilteredCommits()).length>0}async getLatestOpenRcIssue(){let{data:t}=await this.octokit.rest.issues.listForRepo({owner:this.repoInfo.owner,repo:this.repoInfo.repo,labels:gt,state:"open",sort:"updated",direction:"desc",per_page:30}),e=t.filter(o=>!o.pull_request);if(e.length===0)return console.log(`No open Release Candidate issues found in ${this.repoInfo.repo}`),null;let n=e[0];return console.log(`Using latest open Release Candidate issue: #${n.number}`),{number:n.number,title:n.title||""}}async updateRcIssue(t){let e=await this.getReleaseNotes(),n=`RC ${this.getCurrentDate()}`;await this.octokit.rest.issues.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,issue_number:t,title:n,body:e})}async createRcIssue(){let t=await this.getReleaseNotes(),{data:e}=await this.octokit.rest.issues.create({owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC ${this.getCurrentDate()}`,labels:[gt],body:t});return console.log(`Created Release Candidate issue: #${e.number} (${e.html_url})`),{number:e.number,html_url:e.html_url}}async createOrUpdateRcBranch(t){let e=`heads/rc/${t}`,n=`rc/${t}`,{data:o}=await this.octokit.rest.repos.getBranch({owner:this.repoInfo.owner,repo:this.repoInfo.repo,branch:D}),{data:i}=await this.octokit.rest.git.listMatchingRefs({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e}),s=i.length>0;try{s?(console.log(`Updating existing RC branch: ${n}`),await this.octokit.rest.git.updateRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e,sha:o.commit.sha,force:!0})):(console.log(`Creating new RC branch: ${n}`),await this.octokit.rest.git.createRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:`refs/${e}`,sha:o.commit.sha}))}catch(a){throw new Error(`Failed to create or update ${n} branch: ${a}`)}}async getLinkedPrByHead(t){let e=`${this.repoInfo.owner}:rc/${t}`,{data:n}=await this.octokit.rest.pulls.list({owner:this.repoInfo.owner,repo:this.repoInfo.repo,state:"all",head:e,per_page:30}),o=n.filter(s=>s.state==="open");if(o.length>1)return{state:"open",multipleOpen:!0};if(o.length===1)return{state:"open",html_url:o[0].html_url,number:o[0].number};let i=n.filter(s=>s.state==="closed");return i.length>0?{state:"closed",html_url:i[0].html_url,number:i[0].number}:null}async fetchPrTemplate(t){try{let{data:e}=await this.octokit.rest.repos.getContent({owner:this.repoInfo.owner,repo:this.repoInfo.repo,path:".github/PULL_REQUEST_TEMPLATE.md"});if("content"in e&&e.type==="file"){let n=Buffer.from(e.content,"base64").toString("utf-8");return n=n.replace("Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.",`Release candidate pull request. See issue #${t} for details.`),n=n.replace(/<details>/g,"<details open>"),n}}catch(e){e&&typeof e=="object"&&"status"in e&&e.status===404?console.log("No PR template found in repo, using default message."):console.warn("Failed to fetch PR template:",e)}return`Release candidate pull request. See issue #${t} for details.`}async createRcPr(t){try{let e=await this.fetchPrTemplate(t),{data:n}=await this.octokit.request(`POST /repos/${this.repoInfo.owner}/${this.repoInfo.repo}/pulls`,{owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC #${t}`,body:e,head:`rc/${t}`,base:en,headers:{"X-GitHub-Api-Version":"2022-11-28"}});return console.log(`Created Release Candidate pull request: #${n.number} (${n.html_url})`),{state:"open",html_url:n.html_url,number:n.number}}catch(e){throw console.error("Failed to create RC PR:",e),e}}async updateRcPr(t,e){try{let n=await this.fetchPrTemplate(t);await this.octokit.rest.pulls.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,pull_number:e,body:n}),console.log(`Updated Release Candidate pull request: #${e}`)}catch(n){throw console.error("Failed to update RC PR:",n),n}}async getReleaseNotes(){let t=await this.getFilteredCommits();return pe(t,!1)}static async getTriggerBranchName(){return process.env.GITHUB_REF_NAME?process.env.GITHUB_REF_NAME:process.env.GITHUB_REF?.startsWith("refs/heads/")?process.env.GITHUB_REF.replace("refs/heads/",""):v.getCurrentBranchName()}getCurrentDate(){return new Date().toISOString().split("T")[0]}};var vi=tn.command("rc-workflow").description("Generate RC issue and pull request").action(async()=>{await(await X.create()).createReleaseCandidate()});ve.name("auro").version(Z()).description("A cli tool to support the Auro Design System");ve.addHelpText("beforeAll",Ee());ve.parse();
146
+ `,{projectId:l,itemId:c,fieldId:u.id,value:t});else if(!u)throw new Error("No 'ado' field found in GitHub project")}catch(i){console.error(`Failed to update GitHub project: ${i}`)}},Yr=async r=>{let t=process.env.ADO_TOKEN;if(!t)throw new Error("ADO_TOKEN environment variable is required");let e="https://dev.azure.com/itsals",n="E_Retain_Content",o="E_Retain_Content\\Auro Design System",i=X.getPersonalAccessTokenHandler(t),a=await new X.WebApi(e,i).getWorkItemTrackingApi();try{let l=[{op:"add",path:"/fields/System.Title",value:r.title},{op:"add",path:"/fields/System.Description",value:`GitHub Issue: <a href="${r.html_url}">${r.html_url}</a>`},{op:"add",path:"/fields/System.AreaPath",value:o}];return await a.createWorkItem(null,l,n,"User Story")}catch(l){throw new Error(`Failed to create ADO work item: ${l}`)}},gt=async r=>{let t=Y(`Processing GitHub issue: ${r}`).start();try{if(!process.env.GH_TOKEN)throw new Error("GH_TOKEN environment variable is required");if(!process.env.ADO_TOKEN)throw new Error("ADO_TOKEN environment variable is required");t.text="Fetching GitHub issue details...";let e=await Vr(r);t.succeed(`Found issue: "${e.title}"`);let n=Y("Checking for existing ADO work item...").start(),o=await Kr(e);if(o){n.succeed("ADO work item already exists for this issue!"),console.log(`${o}`);return}n.succeed("No existing ADO work item found");let i=Y("Creating new ADO work item...").start(),s=await Yr(e);if(i.succeed(`Successfully created ADO work item #${s.id}`),console.log(`Work item: ${s._links?.html?.href||"N/A"}`),s._links?.html?.href){let a=Y("Adding to GitHub project and updating ADO field...").start();await Jr(e,s._links.html.href),a.succeed("Updated GitHub project with ADO link")}}catch(e){t.fail(`Error: ${e instanceof Error?e.message:e}`),process.exit(1)}};var fi=Xr.command("ado").description("Generate ADO item from GitHub issue").option("-g, --gh-issue <issue>","What GitHub issue to use").action(async r=>{r.ghIssue&&await gt(r.ghIssue)});import{program as tn}from"commander";import{Octokit as Zr}from"@octokit/rest";import{simpleGit as Qr}from"simple-git";var ht="Release Candidate",S="dev",en="main",Z=class r{constructor(t,e,n){this.filteredCommits=null;this.repoInfo={owner:t,repo:e},this.octokit=n}static async create(){let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN is required to run RC workflow.");let e=await v.getRepoOwnerAndName(),n=new Zr({auth:t});if(!e)throw new Error("Failed to retrieve repository information. Ensure you're in a valid git repository.");let o=await r.getTriggerBranchName();return o&&o!==S&&(console.log(`Switching from ${o} to ${S} branch...`),await Qr().checkout(S)),new r(e.owner,e.repo,n)}get owner(){return this.repoInfo.owner}get repo(){return this.repoInfo.repo}get repoData(){return{...this.repoInfo}}async createReleaseCandidate(){await this.hasCommitsReadyInDev()||console.log("No filtered commits found. Continuing to update RC issue/branch/PR.");let e=await this.getLatestOpenRcIssue(),n=e?await this.getLinkedPrByHead(e.number):null;if(n?.multipleOpen)throw new Error("Multiple open RC PRs found for the same rc/<issueNumber> branch.");if(n?.state==="closed"&&(console.log("Linked RC PR is closed. Creating a new RC issue and PR."),e=await this.createRcIssue(),n=null),e?await this.updateRcIssue(e.number):e=await this.createRcIssue(),!e)throw new Error("Failed to resolve RC issue.");await this.createOrUpdateRcBranch(e.number),n?await this.updateRcPr(e.number,n.number):n=await this.createRcPr(e.number)}async getFilteredCommits(){if(this.filteredCommits===null){let t=await v.getCommitMessages(S);this.filteredCommits=fe(t)}return this.filteredCommits}async hasCommitsReadyInDev(){return(await this.getFilteredCommits()).length>0}async getLatestOpenRcIssue(){let{data:t}=await this.octokit.rest.issues.listForRepo({owner:this.repoInfo.owner,repo:this.repoInfo.repo,labels:ht,state:"open",sort:"updated",direction:"desc",per_page:30}),e=t.filter(o=>!o.pull_request);if(e.length===0)return console.log(`No open Release Candidate issues found in ${this.repoInfo.repo}`),null;let n=e[0];return console.log(`Using latest open Release Candidate issue: #${n.number}`),{number:n.number,title:n.title||""}}async updateRcIssue(t){let e=await this.getReleaseNotes(),n=`RC ${this.getCurrentDate()}`;await this.octokit.rest.issues.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,issue_number:t,title:n,body:e})}async createRcIssue(){let t=await this.getReleaseNotes(),{data:e}=await this.octokit.rest.issues.create({owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC ${this.getCurrentDate()}`,labels:[ht],body:t});return console.log(`Created Release Candidate issue: #${e.number} (${e.html_url})`),{number:e.number,html_url:e.html_url}}async createOrUpdateRcBranch(t){let e=`heads/rc/${t}`,n=`rc/${t}`,{data:o}=await this.octokit.rest.repos.getBranch({owner:this.repoInfo.owner,repo:this.repoInfo.repo,branch:S}),{data:i}=await this.octokit.rest.git.listMatchingRefs({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e}),s=i.length>0;try{s?(console.log(`Updating existing RC branch: ${n}`),await this.octokit.rest.git.updateRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e,sha:o.commit.sha,force:!0})):(console.log(`Creating new RC branch: ${n}`),await this.octokit.rest.git.createRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:`refs/${e}`,sha:o.commit.sha}))}catch(a){throw new Error(`Failed to create or update ${n} branch: ${a}`)}}async getLinkedPrByHead(t){let e=`${this.repoInfo.owner}:rc/${t}`,{data:n}=await this.octokit.rest.pulls.list({owner:this.repoInfo.owner,repo:this.repoInfo.repo,state:"all",head:e,per_page:30}),o=n.filter(s=>s.state==="open");if(o.length>1)return{state:"open",multipleOpen:!0};if(o.length===1)return{state:"open",html_url:o[0].html_url,number:o[0].number};let i=n.filter(s=>s.state==="closed");return i.length>0?{state:"closed",html_url:i[0].html_url,number:i[0].number}:null}async fetchPrTemplate(t){try{let{data:e}=await this.octokit.rest.repos.getContent({owner:this.repoInfo.owner,repo:this.repoInfo.repo,path:".github/PULL_REQUEST_TEMPLATE.md"});if("content"in e&&e.type==="file"){let n=Buffer.from(e.content,"base64").toString("utf-8");return n=n.replace("Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.",`Release candidate pull request. See issue #${t} for details.`),n=n.replace(/<details>/g,"<details open>"),n}}catch(e){e&&typeof e=="object"&&"status"in e&&e.status===404?console.log("No PR template found in repo, using default message."):console.warn("Failed to fetch PR template:",e)}return`Release candidate pull request. See issue #${t} for details.`}async createRcPr(t){try{let e=await this.fetchPrTemplate(t),{data:n}=await this.octokit.request(`POST /repos/${this.repoInfo.owner}/${this.repoInfo.repo}/pulls`,{owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC #${t}`,body:e,head:`rc/${t}`,base:en,headers:{"X-GitHub-Api-Version":"2022-11-28"}});return console.log(`Created Release Candidate pull request: #${n.number} (${n.html_url})`),{state:"open",html_url:n.html_url,number:n.number}}catch(e){throw console.error("Failed to create RC PR:",e),e}}async updateRcPr(t,e){try{let n=await this.fetchPrTemplate(t);await this.octokit.rest.pulls.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,pull_number:e,body:n}),console.log(`Updated Release Candidate pull request: #${e}`)}catch(n){throw console.error("Failed to update RC PR:",n),n}}async getReleaseNotes(){let t=await this.getFilteredCommits();return pe(t,!1)}static async getTriggerBranchName(){return process.env.GITHUB_REF_NAME?process.env.GITHUB_REF_NAME:process.env.GITHUB_REF?.startsWith("refs/heads/")?process.env.GITHUB_REF.replace("refs/heads/",""):v.getCurrentBranchName()}getCurrentDate(){return new Date().toISOString().split("T")[0]}};var vi=tn.command("rc-workflow").description("Generate RC issue and pull request").action(async()=>{await(await Z.create()).createReleaseCandidate()});ve.name("auro").version(Q()).description("A cli tool to support the Auro Design System");ve.addHelpText("beforeAll",Ee());ve.parse();