@aurodesignsystem-dev/auro-cli 0.0.0-pr264.0 → 0.0.0-pr266.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/auro-cli.js
CHANGED
|
@@ -1,59 +1,56 @@
|
|
|
1
1
|
#!/usr/bin/env node
|
|
2
|
-
import{program as
|
|
2
|
+
import{program as ve}from"commander";import yt from"figlet";import{mind as wt}from"gradient-string";var Ce=()=>wt(yt.textSync("Auro CLI"));import Ee from"node:fs";import Ie from"node:path";import{fileURLToPath as bt}from"node:url";function D(r){process.env.DEBUG&&console.log(`[DEBUG] ${r}`)}function Z(){try{let r=bt(import.meta.url),t=Ie.dirname(r);D(`Current module path: ${t}`);let e=Ie.resolve(t,"..","package.json");return D(`Checking package.json at: ${e}`),Ee.existsSync(e)?(D(`Found package.json at: ${e}`),JSON.parse(Ee.readFileSync(e,"utf8")).version):(D("Could not find package.json in the standard installed module location, using default version"),"0.0.0")}catch(r){return console.error("Error retrieving package version:",r),"0.0.0"}}import{program as Qt}from"commander";import qe from"ora";function O(r){return r.option("-m, --module-paths [paths...]","Path(s) to node_modules folder").option("-w, --watch","Watches for changes").option("--skip-docs","Skip documentation generation",!1).option("--wca-input [files...]","Source file(s) to analyze for API documentation").option("--wca-output [files...]","Output file(s) for API documentation")}function j(r){return r.option("-s, --serve","Starts a server").option("-p, --port <number>","Port for the server").option("-o, --open","Open the browser after starting the server")}import Jt from"@rollup/plugin-terser";import{watch as Yt}from"rollup";import{rmSync as _t}from"node:fs";import{join as Ft}from"node:path";import oe from"ora";import{rollup as je}from"rollup";import ne from"ora";import{spawn as kt}from"node:child_process";import $t from"ora";var b=(r,t)=>{let e=`${r} ${t?t.join(" "):""}`,n=$t(),o=r,i=t||[];if(!t&&typeof r=="string"){let m=r.split(" ");o=m[0],i=m.slice(1)}let s=e.includes("--watch")||e.includes(" -w"),l=kt(o,i,{stdio:s?"inherit":["inherit","pipe","pipe"],shell:!0});if(!s){let m=[];l.stdout?.on("data",u=>{let c=u.toString();m.push(c),process.stdout.write(c)}),l.stderr?.on("data",u=>{let c=u.toString();m.push(c),process.stderr.write(c)})}return new Promise((m,u)=>{l.on("close",c=>{c!==0?s?(n.info(`Watch mode terminated with code ${c}`),m()):(n.fail(`${e} failed (code ${c})`),u(new Error(`Command failed with exit code ${c}`))):(n.succeed(`${e} completed successfully`),m())})})};import N from"node:fs";import At from"node:path";import{markdownTable as Pe}from"markdown-table";var I=class r{static{this.manifest={schemaVersion:"1.0.0",readme:"",modules:[]}}static generate(t={}){let{outDir:e="./docs",outFile:n="api.md",manifestPath:o="./custom-elements.json"}=t,{getElements:i,renderAllElements:s}=r;if(o)try{let c=N.readFileSync(o,"utf8");r.manifest=JSON.parse(c)}catch(c){throw console.error(`Error reading manifest file at ${o}:`,c),c}let a=i(),l=e;N.existsSync(l)||N.mkdirSync(l,{recursive:!0});let m=s(a),u=At.join(l,n);N.writeFileSync(u,m),console.log(`Generated combined API documentation at ${u}`)}static getElements(){let t=r.manifest.modules.filter(r.isWcaModule);return r.manifest.modules.reduce((e,n)=>e.concat(n.declarations?.filter(o=>"customElement"in o&&o.customElement===!0&&"tagName"in o&&(t.length>0?r.isWcaModule(n):!0))??[]),[])}static isWcaModule(t){let{path:e}=t;return e?e.startsWith("scripts/wca/auro-")&&e.endsWith(".js"):!1}static renderAllElements(t){return`${t.sort((e,n)=>(e.tagName||"").localeCompare(n.tagName||"")).map(e=>r.renderElement(e,!0)).join(`
|
|
3
3
|
|
|
4
|
-
`)}`}static renderElement(t,
|
|
4
|
+
`)}`}static renderElement(t,e=!0){let n=[],{renderTable:o,renderPropertiesAttributesTable:i,renderParameters:s,getType:a}=r;n.push(e?`# ${t.tagName}`:""),t.description&&n.push(t.description);let l=i(t);l&&n.push(l.trim());let m=o("Methods",["name","parameters","return","description"],(t.members||[]).filter(p=>p.kind==="method"&&("privacy"in p?p.privacy!=="private":!0)&&p.name[0]!=="_").map(p=>({...p,parameters:s("parameters"in p?p.parameters:void 0),return:"return"in p&&p.return?a(p.return):""})));m&&n.push(m.trim());let u=o("Events",["name","description"],t.events);u&&n.push(u.trim());let c=o("Slots",[["name","(default)"],"description"],t.slots);c&&n.push(c.trim());let d=o("CSS Shadow Parts",["name","description"],t.cssParts);d&&n.push(d.trim());let h=o("CSS Custom Properties",["name","description"],t.cssProperties);return h&&n.push(h.trim()),n.join(`
|
|
5
5
|
|
|
6
|
-
`)}static renderPropertiesAttributesTable(t){let{getType:
|
|
6
|
+
`)}static renderPropertiesAttributesTable(t){let{getType:e,escapeMarkdown:n}=r,o=t.members?.filter(c=>c.kind==="field"&&("privacy"in c?c.privacy!=="private":!0)&&c.name[0]!=="_")||[],i=t.attributes||[],s=[],a=new Set;if(o.forEach(c=>{if(c.description?.trim()){let d=e(c)||"",p=("return"in c&&c.return?e(c.return):"")||d;s.push({name:c.name,properties:c.name,attributes:("attribute"in c?c.attribute:"")||"",modifiers:"readonly"in c&&c.readonly?"readonly":"",type:p,default:("default"in c?c.default:"")||"",description:c.description||""})}a.add(c.name),"attribute"in c&&c.attribute&&a.add(c.attribute)}),i.forEach(c=>{!a.has(c.name)&&c.description?.trim()&&s.push({name:c.name,properties:"",attributes:c.name,modifiers:"",type:e(c)||"",default:c.default||"",description:c.description||""})}),s.length===0)return"";let l=["Properties","Attributes","Modifiers","Type","Default","Description"],m=s.map(c=>{let w=(c.default||"").trim().replace(/^'([^']+)'$/,"$1").replace(/^"([^"]+)"$/,"$1"),E=w?w.startsWith("`")&&w.endsWith("`")?w:`\`${w}\``:"";return[n(c.properties),n(c.attributes),n(c.modifiers),n(c.type),n(E),n(c.description)]});return`### Properties & Attributes
|
|
7
7
|
|
|
8
|
-
${
|
|
9
|
-
`}static renderParameters(t){let{escapeMarkdown:
|
|
8
|
+
${Pe([l,...m])}
|
|
9
|
+
`}static renderParameters(t){let{escapeMarkdown:e,getType:n}=r;return!t||t.length===0?"None":t.map(o=>{let i=n(o)||"any",s=o.description?` - ${o.description}`:"";return`\`${o.name}\` (${e(i)})${e(s)}`}).join("<br>")}static renderTable(t,e,n){let{escapeMarkdown:o,get:i,capitalize:s}=r;if(n===void 0||n.length===0)return"";let a=n.filter(c=>{let{description:d}=c;return typeof d=="string"&&d.trim()});if(a.length===0)return"";let l=e.map(c=>s((Array.isArray(c)?c[0]:c).split(".")[0])),m=a.map(c=>e.map(d=>{let h=i(c,d);return o(String(h||""))})),u=Pe([l,...m]);return`### ${t}
|
|
10
10
|
|
|
11
|
-
${
|
|
12
|
-
`}static escapeMarkdown(t){return t.replace(/\\/g,"\\\\").replace(/\n/g,"<br>").replace(/\|/g,"\\|")}static getType(t){if(!t||!t.type)return"";let{type:
|
|
13
|
-
`).start();try{let r={port:Number(e.port)||void 0,open:e.open?"/":void 0,watch:e.watch??T.watch,nodeResolve:e.nodeResolve??T.nodeResolve,basePath:e.basePath??T.basePath,rootDir:e.rootDir??T.rootDir,middleware:[function(i,s){return!i.url.endsWith("/")&&!i.url.includes(".")&&(i.url+=".html"),s()}],plugins:[Et({include:e.hmrInclude??T.hmrInclude})]},n=await vt({config:r,readCliArgs:!1,readFileConfig:!1});return t.stop(),n}catch(r){throw t.fail("Server snag! Couldn't start dev server."),console.error("Error starting development server:",r),new Error(`Development server failed to start: ${r.message}`)}}async function G(){let e=ee("Generating Custom Elements Manifest...").start();try{await b(`npx --package=@custom-elements-manifest/analyzer -y -- cem analyze --config '${Ee("custom-elements-manifest.config.mjs")}'`),e.succeed("Custom Elements Manifest generated successfully!")}catch(t){let r=t instanceof Error?t.message:String(t);e.warn("CEM analyzer completed with warnings: "+r)}}async function M(){let e=ee("Generating API md file...").start();try{await x.generate(),e.succeed("API md file generated successfully!")}catch(t){let r=t instanceof Error?t.message:String(t);throw e.fail("Failed to generate API md file: "+r),t}}async function De(){let e=ee("Compiling documentation...").start();try{await N(),e.succeed("Documentation compiled successfully!")}catch(t){let r=t instanceof Error?t.message:String(t);throw e.fail("Failed to compile MD documentation: "+r),t}}async function Pe(e={}){await F(e)}async function _(){await G(),await M()}function Se(){let e=Dt("./dist"),t=te("Cleaning dist folder...").start();try{return Tt(e,{recursive:!0,force:!0}),t.succeed("All clean! Dist folder wiped."),!0}catch(r){return t.fail(`Oops! Couldn't clean dist/ folder: ${r.message}`),console.error(r),!1}}async function Ie(e,t,r,n){let o=te(e).start();try{let i=await t();return o.succeed(r),i}catch(i){throw o.fail(n),console.error(`Error: ${i.message}`),i}}async function je(e,t){return Ie(`Bundling ${e.name||"main"} and ${t.name||"demo"}...`,async()=>{let r=await Ce(e);await r.write(e.output),await r.close();let n=await Ce(t);await n.write(t.output),await n.close()},`Bundles ready! ${e.name||"Main"} and ${t.name||"demo"} built.`,"Bundle hiccup! Build failed.")}async function B(e){let{wcaInput:t,wcaOutput:r,skipDocs:n}=e;if(n){let o=te("Skipping docs generation...").start();setTimeout(()=>{o.succeed("Docs generation skipped.")},0);return}return Ie("Analyzing components and making docs...",async()=>{await _(t,r),await N()},"Docs ready! Looking good.","Doc troubles!")}import{basename as St,join as Re}from"node:path";import{nodeResolve as It}from"@rollup/plugin-node-resolve";import{glob as jt}from"glob";import{litScss as Ot}from"rollup-plugin-scss-lit";import Pt from"node:path";import{glob as Ct}from"glob";function Oe(e){return{name:"watch-globs",buildStart(){let t=Array.isArray(e)?e:[e];for(let r of t)try{for(let n of Ct.sync(Pt.resolve(r)))this.addWatchFile(n)}catch(n){this.error(`Error watching glob pattern "${r}": ${n.message}`)}}}}var re={moduleDirectories:["node_modules"],modulePaths:["../../node_modules","../node_modules","node_modules"],watchPatterns:["./apiExamples/**/*","./docs/**/*"]};function Ne(e=[],t={}){let{watchPatterns:r=re.watchPatterns,dedupe:n=["lit","lit-element","lit-html"],dev:o=!1}=t,i=[...re.modulePaths,...e];return[It({dedupe:n,preferBuiltins:!1,moduleDirectories:re.moduleDirectories}),Ot({minify:o?!1:{fast:!0},options:{loadPaths:[...i,Re(process.cwd(),"src","styles"),Re(process.cwd(),"src")]}}),Oe(r)]}function ne(e={}){let{modulePaths:t=[],watch:r=!1,input:n=["./src/index.js","./src/registered.js"],outputDir:o="./dist",format:i="esm",dev:s=!1}=e;return{name:"Main",config:{input:n,output:{format:i,dir:o,entryFileNames:c=>s||["index","registered"].includes(c.name)?"[name].js":"[name]-[hash].js",chunkFileNames:s?"[name].js":"[name]-[hash].js",assetFileNames:s?"[name][extname]":"[name]-[hash][extname]"},external:Rt(),plugins:Ne(t,{dev:s}),watch:Fe(r)}}}function oe(e={}){let{modulePaths:t=[],watch:r=!1,globPattern:n="./demo/*.js",ignorePattern:o=["./demo/*.min.js"],outputDir:i="./demo",dev:s=!1}=e;return{name:"Demo",config:{input:Object.fromEntries(jt.sync(n,{ignore:o}).map(c=>[St(c,".js"),c])),output:{format:"esm",dir:i,entryFileNames:"[name].min.js",chunkFileNames:"[name].min.js",assetFileNames:s?"[name][extname]":"[name]-[hash][extname]"},plugins:Ne(t,{dev:s}),watch:Fe(r)}}}function Fe(e){if(!e)return!1;let t=typeof e=="object"?e:{};return{clearScreen:t.clearScreen??!0,buildDelay:t.buildDelay??500,chokidar:{ignoreInitial:!0,ignored:t.ignored??["**/dist/**/*.d.ts","**/custom-elements.json","**/demo/*.md","**/demo/**/*.min.js","**/docs/api.md","**/node_modules/**","**/.git/**"],awaitWriteFinish:t.awaitWriteFinish??{stabilityThreshold:1e3,pollInterval:100}},include:t.include??["./src/**/*.scss","./src/**/*.js","./src/**/*.ts","./demo/**/*.js","./apiExamples/**/*","./docs/**/*.md"],exclude:t.exclude??["./dist/**/*","./node_modules/**/*"]}}function Rt(e=[]){return[...[/node_modules\/lit/,/node_modules\/lit-element/,/node_modules\/lit-html/,/node_modules\/@lit/],...e]}import Me from"node:path";import A from"ora";import"rollup";var H=!1,Nt={analyze:{active:!1,lastTime:0},docs:{active:!1,lastTime:0}},Ft=5e3,W=new Set,Gt=["/dist/index.d.ts","/custom-elements.json","/demo/api.md","/docs/api.md","/demo/index.min.js"];function Mt(e){if(!e||typeof e!="string")return!1;try{let t=Me.normalize(e);return Gt.some(r=>t.endsWith(r))||t.includes("/dist/")||t.endsWith(".min.js")||t.endsWith(".d.ts")}catch(t){return console.error(`Error checking path (${typeof e}):`,t.message),!1}}async function Ge(e,t){let r=Nt[e];if(r.active||Date.now()-r.lastTime<Ft)return!1;try{return r.active=!0,r.lastTime=Date.now(),await t()}catch(n){return console.error(`Error in ${e} task:`,n),!1}finally{r.active=!1}}async function _e(e,t,r){let n=!0,o={analyze:!1,docs:!1},i=null,s,c=A("Activating watch mode...").start(),l={analyze:async()=>{let{wcaInput:a,wcaOutput:d,skipDocs:g}=t;if(g){let y=A("Skipping component analysis...").start();return setTimeout(()=>{y.succeed("Component analysis skipped.")},0),!0}let f=A("Detective work: analyzing components...").start();try{return await _(a,d),f.succeed("Component analysis complete! API generated."),!0}catch(y){return f.fail("Analysis hiccup! Something went wrong."),console.error("Component analysis error:",y),!1}},docs:async()=>{if(H)return!1;if(t.skipDocs){let d=A("Skipping docs generation...").start();return setTimeout(()=>{d.succeed("Docs generation skipped.")},0),!0}let a=A("Refreshing docs...").start();try{return await B(t),a.succeed("Documentation refreshed!"),!0}catch(d){a.fail("Docs stumble! Couldn't refresh."),console.error("Documentation rebuild error:",d)}}},u=()=>{n&&o.analyze&&o.docs&&typeof r=="function"&&(n=!1,r())};function m(a=1e3){i&&clearTimeout(i),i=setTimeout(async()=>{setTimeout(async()=>{o.analyze=await Ge("analyze",l.analyze),setTimeout(async()=>{o.docs=await Ge("docs",l.docs),u()},1e3)},1e3)},a)}e.on("event",async a=>{switch(a.code){case"START":c.succeed("Watch mode active! Eyes peeled.");break;case"BUNDLE_START":if(W.clear(),a.input)try{let d=Array.isArray(a.input)?a.input:typeof a.input=="string"?[a.input]:typeof a.input=="object"&&a.input!==null?Object.values(a.input):[];for(let g of d)typeof g=="string"&&!Mt(g)&&W.add(Me.normalize(g))}catch(d){console.error("Error processing input paths:",d)}s=A("Weaving bundles...").start(),H=!0;break;case"BUNDLE_END":s&&s.succeed(`Bundle ${Array.isArray(a.input)?`of ${a.input.join("& ")} `:""}done in ${a.duration}ms! \u{1F680}`),H=!1,W.size>0&&m();break;case"END":break;case"ERROR":H=!1,s?s.fail(`Oops! Bundle hit a snag: ${a.error.message}`):A().fail(`Watch mode hiccup: ${a.error.message}`),W.clear();break}})}function Be(e){return process.on("SIGINT",()=>{let t=A("Wrapping up...").start();e.close(),t.succeed("All done! See you next time. \u2728"),process.exit(0)}),e}async function Ht(e){let t=ne(e),r=oe(e);e.dev||t.config.plugins.push(_t()),await B(e),await je(t.config,r.config)}async function Wt(e){let{dev:t}=e,r=ne({...e,watch:!0}),n=oe({...e,watch:!0}),o=Bt([r.config,n.config]);return _e(o,e,t?async()=>F(e):void 0),Be(o),o}async function U(e={}){try{let{watch:t}=e;return Se(),t?await Wt(e):await Ht(e)}catch(t){throw new Error(`Build failed: ${t.message}`)}}var D=Ut.command("dev").description("Runs development server for auro components");D=I(D);D=j(D);var so=D.action(async e=>{try{let t=He("Initializing...");e.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=e.docs===!1?"Building component (docs disabled)":"Building component",t.start(),e.watch||t.succeed("Build completed!"),await U({...e,dev:!0,watch:e.watch})}catch(t){He().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{program as Lt}from"commander";import We from"ora";var ie=Lt.command("build").description("Builds auro components");ie=I(ie);var po=ie.action(async e=>{try{let t=We("Initializing...");e.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=e.docs===!1?"Building component (docs disabled)":"Building component",t.start(),await U(e),e.watch||t.succeed("Build completed!")}catch(t){We().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{exec as zt}from"node:child_process";import Ue from"node:path";import qt from"node:process";import{fileURLToPath as Vt}from"node:url";import Jt from"node:util";import{program as Kt}from"commander";import Yt from"inquirer";var vo=Kt.command("migrate").description("Script runner to perform repetitive code change tasks").requiredOption("-i, --id <string>","Select the migration you would like to run by id").option("-m, --multi-gitter","Run the migration on all repositories in the multi-gitter config").action(async e=>{let t=Vt(import.meta.url),r=Ue.dirname(t),n=Ue.resolve(r,"migrations",e.id);if(e.multiGitter){let o=Jt.promisify(zt);try{await o("command -v multi-gitter")}catch{console.error("multi-gitter is not installed."),qt.exit(1)}(await Yt.prompt([{type:"confirm",name:"dryRun",message:"Run migration in dry-run mode? (no changes will be committed)",default:!0}])).dryRun?b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml" --dry-run`):b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml"`)}else b(`${n}/script.sh`)});import tr from"node:process";import{program as rr}from"commander";import{readFile as nr,writeFile as or}from"node:fs/promises";import{Logger as ce}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import se from"node:fs/promises";import ae from"node:path";import{Octokit as Xt}from"@octokit/rest";import k from"ora";import{processContentForFile as Zt,templateFiller as Qt}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";async function Le(e,t){let o=(await new Xt({auth:process.env.GITHUB_TOKEN||""}).request("GET /repos/{owner}/{repo}/contents/{path}",{ref:t,owner:"AlaskaAirlines",repo:"auro-templates",path:e,headers:{"X-GitHub-Api-Version":"2022-11-28"}})).data;if(typeof o!="object"||!Array.isArray(o)){let i=`Unexpected response format: ${JSON.stringify(o)}`;throw k().start().fail(i),new Error("Failed to retrieve folder items")}return o}async function ze({folderItems:e,templatePathToReplace:t,rootDir:r,ref:n}){let o=[];for(let i of e){if(i.type=="dir"){let u=k(`Processing directory: ${i.path}`).start(),m=await Le(i.path,n);u.succeed(`Found ${m.length} additional items in ${i.path}`);let a=await ze({folderItems:m,templatePathToReplace:t,rootDir:r,ref:n});o.push(...a);continue}let s=i.path.replace(`${t}/`,""),c=`${r}/.github/${s}`,l={identifier:i.name,input:{remoteUrl:i.download_url||"",fileName:c,overwrite:!0},output:c};o.push(l)}return o}async function er(e){try{await se.rm(e,{recursive:!0,force:!0}),k().start().succeed(`Successfully removed directory: ${e}`)}catch(t){throw k().start().fail(`Error removing directory ${e}: ${t.message}`),t}}async function qe(e,t="",r=!0){try{let n=await se.stat(e),o=ae.basename(e);if(!n.isDirectory())return`${t}${r?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}
|
|
14
|
-
`;let i=`${t}${
|
|
15
|
-
`;try{let
|
|
16
|
-
`}return i}catch(n){return`${t}${
|
|
17
|
-
`}}async function
|
|
11
|
+
${u}
|
|
12
|
+
`}static escapeMarkdown(t){return t.replace(/\\/g,"\\\\").replace(/\n/g,"<br>").replace(/\|/g,"\\|")}static getType(t){if(!t||!t.type)return"";let{type:e}=t,n=i=>i.replace(/\s*\|\s*/g," | ").replace(/'([^']+)'/g,"`$1`");if(typeof e=="string")return n(e);if(e.text)return n(e.text);if(Array.isArray(e))return e.map(i=>typeof i=="string"?i:i.text?i.text:i.name?i.name:String(i)).join(" \\| ");if(e.name)return n(e.name);if(e.references&&Array.isArray(e.references))return e.references.map(i=>i.name||String(i)).join(" \\| ");let o=String(e);return n(o)}static get(t,e){let n="",o=e;Array.isArray(e)&&([o,n]=e);let i=o.split("."),s=t;for(;s&&i.length;)s=s[i.shift()];return s==null||s===""?n:String(s)}static capitalize(t){return t.replace(/([A-Z])/g," $1").replace(/^./,e=>e.toUpperCase()).trim()}};import vt from"node:fs";import Ct from"node:os";import _ from"node:path";import Q from"node:process";function Et(){let r=Ct.homedir()||Q.env.HOME||Q.env.USERPROFILE;if(!r)throw new Error("Unable to determine user home directory");return _.join(r,".auro")}function ee(...r){return _.join(Et(),...r)}function te(...r){let t=vt.realpathSync(Q.argv[1]),e=_.dirname(t);return _.resolve(e,...r)}var Te=r=>te("configs",r);import{Logger as It}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{generateReadmeUrl as Pt,processContentForFile as Tt,templateFiller as Re}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";import re from"node:fs";import xt from"node:path";var Rt="/docs/pages",Se={overwriteLocalCopies:!0,remoteReadmeVersion:"master",remoteReadmeVariant:"_updated_paths"};function k(r){return`${process.cwd()}/${r}`}async function St(r,t=!1){let e=[];t||e.push({identifier:"README.md",input:{remoteUrl:r.remoteReadmeUrl||Pt(r.remoteReadmeVersion,r.remoteReadmeVariant),fileName:k("/docTemplates/README.md"),overwrite:r.overwriteLocalCopies},output:k("/README.md")}),xe("/docs/partials/index.md")&&e.push({identifier:"index.md",input:k("/docs/partials/index.md"),output:k("/demo/index.md"),mdMagicConfig:{output:{directory:k("/demo")}}}),xe("/docs/partials/api.md")&&e.push({identifier:"api.md",input:k("/docs/partials/api.md"),output:k("/demo/api.md"),preProcessors:[Re.formatApiTable]});let n=k(Rt);if(re.existsSync(n)){let i=(await re.promises.readdir(n)).map(s=>({identifier:s,input:xt.join(n,s),output:k(`/demo/${s}`)}));e.push(...i)}return e}async function Dt(r=Se,t=!1){await Re.extractNames();let e=await St(r,t);for(let n of e)try{await Tt(n)}catch(o){It.error(`Error processing ${n.identifier}: ${o.message}`)}}async function F(r={}){await Dt({...Se,remoteReadmeUrl:"https://raw.githubusercontent.com/AlaskaAirlines/auro-templates/main/templates/default/README.md"},r.skipReadme)}function xe(r){return re.existsSync(k(r))}import{startDevServer as Ot}from"@web/dev-server";import{hmrPlugin as jt}from"@web/dev-server-hmr";import Nt from"ora";var P={watch:!0,nodeResolve:!0,basePath:"/",rootDir:"./demo",hmrInclude:["src/**/*","demo/**/*","apiExamples/**/*","docs/**/*"]};async function B(r={}){if(!r.serve)return;let t=Nt(`Firing up dev server...
|
|
13
|
+
`).start();try{let e={port:Number(r.port)||void 0,open:r.open?"/":void 0,watch:r.watch??P.watch,nodeResolve:r.nodeResolve??P.nodeResolve,basePath:r.basePath??P.basePath,rootDir:r.rootDir??P.rootDir,middleware:[function(i,s){return!i.url.endsWith("/")&&!i.url.includes(".")&&(i.url+=".html"),s()}],plugins:[jt({include:r.hmrInclude??P.hmrInclude})]},n=await Ot({config:e,readCliArgs:!1,readFileConfig:!1});return t.stop(),n}catch(e){throw t.fail("Server snag! Couldn't start dev server."),console.error("Error starting development server:",e),new Error(`Development server failed to start: ${e.message}`)}}async function G(){let r=ne("Generating Custom Elements Manifest...").start();try{await b(`npx --package=@custom-elements-manifest/analyzer -y -- cem analyze --config '${Te("custom-elements-manifest.config.mjs")}'`),r.succeed("Custom Elements Manifest generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);r.warn("CEM analyzer completed with warnings: "+e)}}async function M(){let r=ne("Generating API md file...").start();try{await I.generate(),r.succeed("API md file generated successfully!")}catch(t){let e=t instanceof Error?t.message:String(t);throw r.fail("Failed to generate API md file: "+e),t}}async function De(r={}){let t=ne("Compiling documentation...").start();try{await F(r),t.succeed("Documentation compiled successfully!")}catch(e){let n=e instanceof Error?e.message:String(e);throw t.fail("Failed to compile MD documentation: "+n),e}}async function Oe(r={}){await B(r)}async function H(){await G(),await M()}function Ne(){let r=Ft("./dist"),t=oe("Cleaning dist folder...").start();try{return _t(r,{recursive:!0,force:!0}),t.succeed("All clean! Dist folder wiped."),!0}catch(e){return t.fail(`Oops! Couldn't clean dist/ folder: ${e.message}`),console.error(e),!1}}async function _e(r,t,e,n){let o=oe(r).start();try{let i=await t();return o.succeed(e),i}catch(i){throw o.fail(n),console.error(`Error: ${i.message}`),i}}async function Fe(r,t){return _e(`Bundling ${r.name||"main"} and ${t.name||"demo"}...`,async()=>{let e=await je(r);await e.write(r.output),await e.close();let n=await je(t);await n.write(t.output),await n.close()},`Bundles ready! ${r.name||"Main"} and ${t.name||"demo"} built.`,"Bundle hiccup! Build failed.")}async function L(r){let{wcaInput:t,wcaOutput:e,skipDocs:n}=r;if(n){let o=oe("Skipping docs generation...").start();setTimeout(()=>{o.succeed("Docs generation skipped.")},0);return}return _e("Analyzing components and making docs...",async()=>{await H(t,e),await F()},"Docs ready! Looking good.","Doc troubles!")}import{basename as Mt,join as Ge}from"node:path";import{nodeResolve as Ht}from"@rollup/plugin-node-resolve";import{glob as Lt}from"glob";import{litScss as Ut}from"rollup-plugin-scss-lit";import Bt from"node:path";import{glob as Gt}from"glob";function Be(r){return{name:"watch-globs",buildStart(){let t=Array.isArray(r)?r:[r];for(let e of t)try{for(let n of Gt.sync(Bt.resolve(e)))this.addWatchFile(n)}catch(n){this.error(`Error watching glob pattern "${e}": ${n.message}`)}}}}var se={moduleDirectories:["node_modules"],modulePaths:["../../node_modules","../node_modules","node_modules"],watchPatterns:["./apiExamples/**/*","./docs/**/*"]};function Me(r=[],t={}){let{watchPatterns:e=se.watchPatterns,dedupe:n=["lit","lit-element","lit-html"],dev:o=!1}=t,i=[...se.modulePaths,...r];return[Ht({dedupe:n,preferBuiltins:!1,moduleDirectories:se.moduleDirectories}),Ut({minify:o?!1:{fast:!0},options:{loadPaths:[...i,Ge(process.cwd(),"src","styles"),Ge(process.cwd(),"src")]}}),Be(e)]}function ie(r={}){let{modulePaths:t=[],watch:e=!1,input:n=["./src/index.js","./src/registered.js"],outputDir:o="./dist",format:i="esm",dev:s=!1}=r;return{name:"Main",config:{input:n,output:{format:i,dir:o,entryFileNames:a=>s||["index","registered"].includes(a.name)?"[name].js":"[name]-[hash].js",chunkFileNames:s?"[name].js":"[name]-[hash].js",assetFileNames:s?"[name][extname]":"[name]-[hash][extname]"},external:Wt(),plugins:Me(t,{dev:s}),watch:He(e)}}}function ae(r={}){let{modulePaths:t=[],watch:e=!1,globPattern:n="./demo/*.js",ignorePattern:o=["./demo/*.min.js"],outputDir:i="./demo",dev:s=!1}=r;return{name:"Demo",config:{input:Object.fromEntries(Lt.sync(n,{ignore:o}).map(a=>[Mt(a,".js"),a])),output:{format:"esm",dir:i,entryFileNames:"[name].min.js",chunkFileNames:"[name].min.js",assetFileNames:s?"[name][extname]":"[name]-[hash][extname]"},plugins:Me(t,{dev:s}),watch:He(e)}}}function He(r){if(!r)return!1;let t=typeof r=="object"?r:{};return{clearScreen:t.clearScreen??!0,buildDelay:t.buildDelay??500,chokidar:{ignoreInitial:!0,ignored:t.ignored??["**/dist/**/*.d.ts","**/custom-elements.json","**/demo/*.md","**/demo/**/*.min.js","**/docs/api.md","**/node_modules/**","**/.git/**"],awaitWriteFinish:t.awaitWriteFinish??{stabilityThreshold:1e3,pollInterval:100}},include:t.include??["./src/**/*.scss","./src/**/*.js","./src/**/*.ts","./demo/**/*.js","./apiExamples/**/*","./docs/**/*.md"],exclude:t.exclude??["./dist/**/*","./node_modules/**/*"]}}function Wt(r=[]){return[...[/node_modules\/lit/,/node_modules\/lit-element/,/node_modules\/lit-html/,/node_modules\/@lit/],...r]}import Ue from"node:path";import A from"ora";import"rollup";var U=!1,zt={analyze:{active:!1,lastTime:0},docs:{active:!1,lastTime:0}},qt=5e3,W=new Set,Vt=["/dist/index.d.ts","/custom-elements.json","/demo/api.md","/docs/api.md","/demo/index.min.js"];function Kt(r){if(!r||typeof r!="string")return!1;try{let t=Ue.normalize(r);return Vt.some(e=>t.endsWith(e))||t.includes("/dist/")||t.endsWith(".min.js")||t.endsWith(".d.ts")}catch(t){return console.error(`Error checking path (${typeof r}):`,t.message),!1}}async function Le(r,t){let e=zt[r];if(e.active||Date.now()-e.lastTime<qt)return!1;try{return e.active=!0,e.lastTime=Date.now(),await t()}catch(n){return console.error(`Error in ${r} task:`,n),!1}finally{e.active=!1}}async function We(r,t,e){let n=!0,o={analyze:!1,docs:!1},i=null,s,a=A("Activating watch mode...").start(),l={analyze:async()=>{let{wcaInput:c,wcaOutput:d,skipDocs:h}=t;if(h){let w=A("Skipping component analysis...").start();return setTimeout(()=>{w.succeed("Component analysis skipped.")},0),!0}let p=A("Detective work: analyzing components...").start();try{return await H(c,d),p.succeed("Component analysis complete! API generated."),!0}catch(w){return p.fail("Analysis hiccup! Something went wrong."),console.error("Component analysis error:",w),!1}},docs:async()=>{if(U)return!1;if(t.skipDocs){let d=A("Skipping docs generation...").start();return setTimeout(()=>{d.succeed("Docs generation skipped.")},0),!0}let c=A("Refreshing docs...").start();try{return await L(t),c.succeed("Documentation refreshed!"),!0}catch(d){c.fail("Docs stumble! Couldn't refresh."),console.error("Documentation rebuild error:",d)}}},m=()=>{n&&o.analyze&&o.docs&&typeof e=="function"&&(n=!1,e())};function u(c=1e3){i&&clearTimeout(i),i=setTimeout(async()=>{setTimeout(async()=>{o.analyze=await Le("analyze",l.analyze),setTimeout(async()=>{o.docs=await Le("docs",l.docs),m()},1e3)},1e3)},c)}r.on("event",async c=>{switch(c.code){case"START":a.succeed("Watch mode active! Eyes peeled.");break;case"BUNDLE_START":if(W.clear(),c.input)try{let d=Array.isArray(c.input)?c.input:typeof c.input=="string"?[c.input]:typeof c.input=="object"&&c.input!==null?Object.values(c.input):[];for(let h of d)typeof h=="string"&&!Kt(h)&&W.add(Ue.normalize(h))}catch(d){console.error("Error processing input paths:",d)}s=A("Weaving bundles...").start(),U=!0;break;case"BUNDLE_END":s&&s.succeed(`Bundle ${Array.isArray(c.input)?`of ${c.input.join("& ")} `:""}done in ${c.duration}ms! \u{1F680}`),U=!1,W.size>0&&u();break;case"END":break;case"ERROR":U=!1,s?s.fail(`Oops! Bundle hit a snag: ${c.error.message}`):A().fail(`Watch mode hiccup: ${c.error.message}`),W.clear();break}})}function ze(r){return process.on("SIGINT",()=>{let t=A("Wrapping up...").start();r.close(),t.succeed("All done! See you next time. \u2728"),process.exit(0)}),r}async function Xt(r){let t=ie(r),e=ae(r);r.dev||t.config.plugins.push(Jt()),await L(r),await Fe(t.config,e.config)}async function Zt(r){let{dev:t}=r,e=ie({...r,watch:!0}),n=ae({...r,watch:!0}),o=Yt([e.config,n.config]);return We(o,r,t?async()=>B(r):void 0),ze(o),o}async function z(r={}){try{let{watch:t}=r;return Ne(),t?await Zt(r):await Xt(r)}catch(t){throw new Error(`Build failed: ${t.message}`)}}var T=Qt.command("dev").description("Runs development server for auro components");T=O(T);T=j(T);var Ao=T.action(async r=>{try{let t=qe("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),r.watch||t.succeed("Build completed!"),await z({...r,dev:!0,watch:r.watch})}catch(t){qe().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{program as er}from"commander";import Ve from"ora";var ce=er.command("build").description("Builds auro components");ce=O(ce);var To=ce.action(async r=>{try{let t=Ve("Initializing...");r.watch?(t.text="Waiting for changes...",t.spinner="bouncingBar",t.color="green"):t.text=r.docs===!1?"Building component (docs disabled)":"Building component",t.start(),await z(r),r.watch||t.succeed("Build completed!")}catch(t){Ve().fail(`Build failed: ${t.message}`),console.error(t),process.exit(1)}});import{exec as tr}from"node:child_process";import Ke from"node:path";import rr from"node:process";import{fileURLToPath as nr}from"node:url";import or from"node:util";import{program as sr}from"commander";import ir from"inquirer";var Bo=sr.command("migrate").description("Script runner to perform repetitive code change tasks").requiredOption("-i, --id <string>","Select the migration you would like to run by id").option("-m, --multi-gitter","Run the migration on all repositories in the multi-gitter config").action(async r=>{let t=nr(import.meta.url),e=Ke.dirname(t),n=Ke.resolve(e,"migrations",r.id);if(r.multiGitter){let o=or.promisify(tr);try{await o("command -v multi-gitter")}catch{console.error("multi-gitter is not installed."),rr.exit(1)}(await ir.prompt([{type:"confirm",name:"dryRun",message:"Run migration in dry-run mode? (no changes will be committed)",default:!0}])).dryRun?b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml" --dry-run`):b(`multi-gitter run ${n}/script.sh --config "${n}/multi-gitter.yml"`)}else b(`${n}/script.sh`)});import mr from"node:process";import{program as dr}from"commander";import{readFile as pr,writeFile as fr}from"node:fs/promises";import{Logger as me}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import le from"node:fs/promises";import ue from"node:path";import{Octokit as ar}from"@octokit/rest";import $ from"ora";import{processContentForFile as cr,templateFiller as lr}from"@aurodesignsystem/auro-library/scripts/utils/sharedFileProcessorUtils.mjs";async function Je(r,t){let o=(await new ar({auth:process.env.GITHUB_TOKEN||""}).request("GET /repos/{owner}/{repo}/contents/{path}",{ref:t,owner:"AlaskaAirlines",repo:"auro-templates",path:r,headers:{"X-GitHub-Api-Version":"2022-11-28"}})).data;if(typeof o!="object"||!Array.isArray(o)){let i=`Unexpected response format: ${JSON.stringify(o)}`;throw $().start().fail(i),new Error("Failed to retrieve folder items")}return o}async function Ye({folderItems:r,templatePathToReplace:t,rootDir:e,ref:n}){let o=[];for(let i of r){if(i.type=="dir"){let m=$(`Processing directory: ${i.path}`).start(),u=await Je(i.path,n);m.succeed(`Found ${u.length} additional items in ${i.path}`);let c=await Ye({folderItems:u,templatePathToReplace:t,rootDir:e,ref:n});o.push(...c);continue}let s=i.path.replace(`${t}/`,""),a=`${e}/.github/${s}`,l={identifier:i.name,input:{remoteUrl:i.download_url||"",fileName:a,overwrite:!0},output:a};o.push(l)}return o}async function ur(r){try{await le.rm(r,{recursive:!0,force:!0}),$().start().succeed(`Successfully removed directory: ${r}`)}catch(t){throw $().start().fail(`Error removing directory ${r}: ${t.message}`),t}}async function Xe(r,t="",e=!0){try{let n=await le.stat(r),o=ue.basename(r);if(!n.isDirectory())return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}
|
|
14
|
+
`;let i=`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}${o}/
|
|
15
|
+
`;try{let a=(await le.readdir(r)).sort();for(let l=0;l<a.length;l++){let m=a[l],u=ue.join(r,m),c=l===a.length-1,d=t+(e?" ":"\u2502 ");i+=await Xe(u,d,c)}}catch{i+=`${t}${e?" ":"\u2502 "}\u2514\u2500\u2500 [Permission denied or error reading directory]
|
|
16
|
+
`}return i}catch(n){return`${t}${e?"\u2514\u2500\u2500 ":"\u251C\u2500\u2500 "}[Error: ${n}]
|
|
17
|
+
`}}async function Ze(r,t="main"){r||($().start().fail("Root directory must be specified"),process.exit(1));let e=".github",n=$("Removing existing .github directory...").start();try{await ur(e),n.succeed(".github directory removed successfully")}catch(l){n.fail(`Error removing .github directory: ${l.message}`),process.exit(1)}await lr.extractNames(),process.env.GITHUB_TOKEN||($().start().fail("GITHUB_TOKEN environment variable is not set."),process.exit(1));let o="templates/default/.github",i=await Je(o,t),s=await Ye({folderItems:i,templatePathToReplace:o,rootDir:r,ref:t}),a=$("Processing all files...").start();try{await Promise.all(s.map(m=>cr(m))),a.succeed("All files processed.");let l=$("Generating directory tree...").start();try{let m=ue.join(r,".github"),u=await Xe(m);l.succeed("Synced .github directory structure:"),console.log(u)}catch(m){l.fail(`Error generating directory tree: ${m.message}`)}}catch(l){a.fail(`Error processing files: ${l.message}`),process.exit(1)}}var Xo=dr.command("sync").option("-r, --ref <branch/tag/commit>","Git reference (branch/tag/commit) to use","main").description("Script runner to synchronize local repository configuration files").action(async r=>{me.info("Synchronizing repository configuration files..."),me.warn("Note: sync does not create a new git branch. Changes are added to the current branch.");let t=mr.cwd();await Ze(t,r.ref);let e=`${t}/.github/CODEOWNERS`,o=(await pr(e,{encoding:"utf-8"})).replace(/\r\n/gu,`
|
|
18
18
|
`).replace(/\n\n/gu,`
|
|
19
|
-
`);await
|
|
19
|
+
`);await fr(e,o,{encoding:"utf-8"}),(o.includes("\r")||o.includes(`
|
|
20
20
|
|
|
21
|
-
`))&&
|
|
21
|
+
`))&&me.error("CODEOWNERS file still has Windows line endings.")});import q from"node:fs";import x from"node:path";import{Logger as et}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{program as gr}from"commander";import{glob as hr}from"glob";var Qe=(r,t)=>{let e=(r.match(/static register\(name \= (.+)\)/)||r.match(/customElements.get\((.+?)\)/))[1],n=r.match(/export class (.+) extends/)?.[1],o=r.match(/\/\*\*((.|\n)*?)(\*\n|\*\/|[@])/)?.[1]||"";return!e||!n?r:`
|
|
22
22
|
import { ${n} } from '${t}';
|
|
23
23
|
|
|
24
24
|
/**${o}*/
|
|
25
25
|
class ${n}WCA extends ${n} {}
|
|
26
26
|
|
|
27
|
-
if (!customElements.get(${
|
|
28
|
-
customElements.define(${
|
|
27
|
+
if (!customElements.get(${e})) {
|
|
28
|
+
customElements.define(${e}, ${n}WCA);
|
|
29
29
|
}
|
|
30
|
-
`};var
|
|
30
|
+
`};var V=x.resolve(process.cwd(),"./scripts/wca");async function yr(r){try{return(await Promise.all(r.map(e=>hr(e)))).flat()}catch(t){throw console.error("Error processing glob patterns:",t),t}}async function wr(r){q.existsSync(V)||await q.promises.mkdir(V,{recursive:!0});for(let t of r){let e=x.resolve(process.cwd(),t),n=await q.promises.readFile(e,"utf-8"),o=x.resolve(V,`${x.basename(t)}`),i=Qe(n,x.relative(V,t));await q.promises.writeFile(o,i)}}async function br(){let r=await yr(["./src/auro-*.js"]);await wr(r)}var is=gr.command("wca-setup").description("Set up WCA (Web Component Analyzer) for the project").action(()=>{br().then(()=>{et.success("WCA setup completed successfully.")}).catch(r=>{et.error(`WCA setup failed: ${r.message}`)})});import{program as Ir}from"commander";import ot from"chalk";import st from"ora";import{appendFile as kr,readFile as $r}from"node:fs/promises";import{Logger as f}from"@aurodesignsystem/auro-library/scripts/utils/logger.mjs";import{simpleGit as Ar}from"simple-git";var g;try{g=Ar({baseDir:process.cwd(),binary:"git",maxConcurrentProcesses:1})}catch(r){f.error(`Failed to initialize git: ${r}`),g={}}var v=class r{static async checkGitignore(t){if(t==="")return!1;try{return(await $r(".gitignore","utf-8")).includes(t)}catch(e){return f.error(`Error reading file: ${e}`),!1}}static async getCommitMessages(t=""){try{let e=t;e||(e=(await g.branchLocal()).current);let n="main",o="";if(!!process.env.GITHUB_ACTIONS){n=process.env.GITHUB_BASE_REF||"main";try{if(await g.fetch("origin",n),e!=="HEAD")try{await g.raw(["rev-parse","--verify",e])}catch{await g.fetch("origin",e)}o=`${(await g.raw(["merge-base",`origin/${n}`,e])).trim()}..${e}`}catch(s){f.warn(`Error setting up commit range in CI: ${s}`),o=`origin/${n}..${e}`}}else try{try{await g.raw(["rev-parse","--verify",`origin/${n}`])}catch{f.info(`Fetching ${n} from origin`),await g.fetch("origin",n)}if(e!=="HEAD")try{await g.raw(["rev-parse","--verify",e])}catch{await g.fetch("origin",e)}o=`${(await g.raw(["merge-base",`origin/${n}`,e])).trim()}..${e}`}catch(s){f.warn(`Error determining commits locally: ${s}`),o=`${e}~10..${e}`}return await r.getFormattedCommits(o)}catch(e){return f.error(`Error getting commit messages: ${e}`),[]}}static async getRepoOwnerAndName(){try{let t=await g.getRemotes(!0);if(t.length===0)return f.warn("No remotes found"),null;let e=t.find(o=>o.name==="origin")||t[0],n=e.refs.fetch||e.refs.push;return r.parseGitUrl(n)}catch(t){return f.error(`Error getting repo owner and name: ${t}`),null}}static async getCurrentBranchName(){try{return(await g.branchLocal()).current||null}catch(t){return f.error(`Error getting current branch name: ${t}`),null}}static parseGitUrl(t){let e;return t.includes("@")&&t.includes(":")&&(e=t.match(/@([^:]+):([^/]+)\/(.+?)(?:\.git)?$/),e)?{owner:e[2],repo:e[3]}:(e=t.match(/https?:\/\/(?:[^@]+@)?[^/]+\/([^/]+)\/(.+?)(?:\.git)?$/),e?{owner:e[1],repo:e[2]}:(f.warn(`Could not parse git URL: ${t}`),null))}static async getFormattedCommits(t){let n=(await g.raw(["log","--pretty=format:COMMIT_START%n%H%n%ad%n%an%n%s%n%b%nCOMMIT_END","--date=short",t])).split(`COMMIT_START
|
|
31
31
|
`).filter(i=>i.trim()!==""),o=[];for(let i of n){let s=i.split(`
|
|
32
|
-
`);if(s.length>=4){let
|
|
32
|
+
`);if(s.length>=4){let a=s[0],l=s[1],m=s[2],u=s[3],c=s.slice(4).filter(E=>E!=="COMMIT_END"),d=c.length>0?c.join(""):"",h=a.substring(0,7),p=u.match(/^(feat|fix|docs|style|refactor|perf|test|build|ci|chore)(\(.+\))?:/),w=p?p[1]:"unknown";d.includes("BREAKING CHANGE")&&(w="breaking"),o.push({type:w,hash:h,date:l,subject:u,body:d,message:`${u}${d?`
|
|
33
33
|
|
|
34
|
-
${d}`:""}`,author_name:
|
|
35
|
-
${t}`),
|
|
36
|
-
`,o=""),o=`${o}${i} `;return o.length>0&&(n+=o.trim()),n}function
|
|
37
|
-
`)}import
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
`),
|
|
41
|
-
`);
|
|
42
|
-
------
|
|
43
|
-
`),r.length>0?console.log(me.green(`\u2713 Showing ${r.length} commits of types: ${t.join(", ")}`)):console.log(me.yellow(`\u26A0 No feat/fix/breaking commits found. Showing all ${e.length} commits for your selection.`))}async function et(e=!1,t=!1,r=!1){let n=Qe(`Checking commits...
|
|
44
|
-
`).start();try{let o=await q.getCommitMessages();if(r){n.succeed(`Total commits analyzed: ${o.length}`),gr(o);return}if(e&&Xe(o),n.succeed(`Total commits analyzed: ${o.length}`),o.length!==0){let i=o.map(l=>l.type),c=Array.from(new Set(i)).map(l=>C(l)).join(", ");n.succeed(`Found commit types: ${c}`)}else n.info(`The list of commits is created by comparing the current branch
|
|
34
|
+
${d}`:""}`,author_name:m})}}return o}static async addToGitignore(t,e=!0){await r.checkGitignore(t).then(async n=>{if(n)f.warn(`${t} already exists`);else try{await kr(".gitignore",`
|
|
35
|
+
${t}`),e&&f.success(`${t} added to .gitignore`)}catch(o){f.error(o)}})}static async removeFromGitCache(t){try{await g.rmKeepLocal(t),f.success(`${t.join(", ")} are removed from git cache`)}catch(e){f.error(e)}}static async createBranch(t){try{await g.checkoutLocalBranch(t),f.success(`Created and switched to ${t} branch`)}catch(e){f.error(e)}}static async commitStagedFiles(t){try{await g.add("."),await g.commit(t),f.success(`Committed with message: ${t}`)}catch(e){f.error(e)}}};import y from"chalk";var vr=60,Cr=100;function R(r){switch(r){case"breaking":return y.bold.red(r);case"feat":return y.bold.green(r);case"fix":return y.bold.green(r);case"perf":return y.bold.green(r);case"docs":return y.bold.cyan(r);case"style":return y.bold.cyan(r);case"refactor":return y.bold.cyan(r);case"test":return y.bold.cyan(r);case"build":return y.bold.cyan(r);case"ci":return y.bold.cyan(r);case"chore":return y.bold.cyan(r);default:return y.bold.white(r)}}function tt(r,t){if(!r)return"";if(r.length<=t)return r;let e=r.split(" "),n="",o="";for(let i of e)(o+i).length>t&&o.length>0&&(n+=`${o.trim()}
|
|
36
|
+
`,o=""),o=`${o}${i} `;return o.length>0&&(n+=o.trim()),n}function rt(r){for(let t of r){console.log("\u2500".repeat(60));let e=tt(t.subject,vr),n=tt(t.body,Cr);console.log(y.bold(`${R(t.type)}`)),console.log(y.dim(`${t.hash} | ${t.date} | ${t.author_name}`)),console.log(y.bold(`${y.white(e)}`)),t.body&&console.log(y.dim(n))}console.log("\u2500".repeat(60)),console.log(`
|
|
37
|
+
`)}import K from"@actions/github";async function de(){try{let r=process.env.GITHUB_TOKEN;if(!r)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let t=K.getOctokit(r),{context:e}=K;if(!e.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[n,o]=process.env.GITHUB_REPOSITORY.split("/"),i=e.payload.pull_request.number,{data:s}=await t.rest.issues.listLabelsOnIssue({owner:n,repo:o,issue_number:i});return s.map(a=>a.name)}catch(r){throw r instanceof Error?new Error(`Failed to get existing labels: ${r.message}`):r}}async function nt(r){try{let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN environment variable is not set");if(!process.env.GITHUB_REPOSITORY||!process.env.GITHUB_EVENT_PATH)throw new Error("This function can only be used in a GitHub Actions environment");let e=K.getOctokit(t),{context:n}=K;if(!n.payload.pull_request)throw new Error("No pull request found in the GitHub context");let[o,i]=process.env.GITHUB_REPOSITORY.split("/"),s=n.payload.pull_request.number,a=`semantic-status: ${r}`,l=await de();if(l.includes(a))return;let m=l.filter(u=>u.startsWith("semantic-status:")&&u!==a);for(let u of m)await e.rest.issues.removeLabel({owner:o,repo:i,issue_number:s,name:u});await e.rest.issues.addLabels({owner:o,repo:i,issue_number:s,labels:[a]});return}catch(t){throw t instanceof Error?new Error(`Failed to apply label: ${t.message}`):t}}var it=["feat","fix","breaking","perf"];function pe(r,t=!0){let e=`### In this release
|
|
38
|
+
`;for(let n of r)if(e+=`- ${n.hash} ${n.subject}
|
|
39
|
+
`,n.body?.trim()){let i=n.body.trim().split(/\n+/).map(s=>s.trim()).filter(s=>s.length>0);for(let s of i){let a=s;a=a.replace(/([^\s])(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)/g,"$1 $2"),a=a.replace(/(AlaskaAirlines\/[a-zA-Z0-9-]+#\d+)([^\s])/g,"$1 $2"),e+=` - ${a}`}}return r.length===0?"":(t&&console.log(ot.green(`\u2713 Showing ${r.length} commits of types: ${it.join(", ")}`)),e)}function fe(r,t=!0){let e=r.filter(o=>it.includes(o.type)),n;return t?n=e.length>0?e:r:n=e,n.length===0&&console.log(`No commits found to include in release notes.
|
|
40
|
+
`),n}async function at(r=!1,t=!1,e=!1){let n=st(`Checking commits...
|
|
41
|
+
`).start();try{let o=await v.getCommitMessages();if(e){n.succeed(`Total commits analyzed: ${o.length}`);let i=fe(o);console.log(pe(i));return}if(r&&rt(o),n.succeed(`Total commits analyzed: ${o.length}`),o.length!==0){let i=o.map(l=>l.type),a=Array.from(new Set(i)).map(l=>R(l)).join(", ");n.succeed(`Found commit types: ${a}`)}else n.info(`The list of commits is created by comparing the current branch
|
|
45
42
|
with the main branch. If you are on a new branch, please
|
|
46
|
-
make sure to commit some changes before running this command.`);t&&await
|
|
47
|
-
`,"utf8"),n.succeed(`Package.json updated to use ${
|
|
48
|
-
- ${
|
|
49
|
-
- `)}`:typeof
|
|
50
|
-
${Object.entries(
|
|
51
|
-
`)}`:`${t}: ${
|
|
52
|
-
`)}async function
|
|
43
|
+
make sure to commit some changes before running this command.`);t&&await Er(o,n)}catch(o){n.fail("Error getting commit messages"),console.error(o)}}async function Er(r,t){let e=["breaking","feat","fix","perf","docs","style","refactor","test","build","ci","chore"],n=r.map(s=>s.type).filter(s=>e.includes(s)),o=null,i=Number.POSITIVE_INFINITY;for(let s of n){let a=e.indexOf(s);a<i&&(i=a,o=s)}if(o){let s=st("Checking existing labels on pull request...").start();try{if((await de()).includes(`semantic-status: ${o}`)){s.info(`Label "semantic-status: ${R(o)}" already exists on the pull request.`);return}s.text="Applying label to pull request...",await nt(o),s.succeed(`Label "semantic-status: ${R(o)}" applied to the pull request.`)}catch(a){let l=a instanceof Error?a.message:String(a);s.fail(l)}}else t.warn(ot.yellow("No semantic commit type found to apply as label."))}var Cs=Ir.command("check-commits").alias("cc").option("-l, --set-label","Set label on the pull request based on the commit message type").option("-d, --debug","Display detailed commit information for debugging").option("-r, --release-notes","Generate release notes based on commit messages").description("Check commits in the local repository for the types of semantic commit messages made and return the results.").action(async r=>{await at(r.debug,r.setLabel,r.releaseNotes)});import ct from"node:fs";import{get as lt}from"node:https";import C from"chalk";import{program as Pr}from"commander";import Tr from"ora";var Ss=Pr.command("pr-release").option("-n, --namespace <package-namespace>","Set namespace of the package release","@aurodesignsystem-dev").option("-p, --pr-number <number>","Set pull request number for the release","0").description("Generate the package version based off of PR number then update the package.json file. Note: this does not publish the package.").action(async r=>{await xr(r)}),xr=async r=>{let{namespace:t,prNumber:e}=r,n=Tr("Updating package.json").start();try{let o="package.json",i=JSON.parse(ct.readFileSync(o,"utf8"));n.text="Checking npm registry for version information...";let s=`0.0.0-pr${e}`,a=i.name.split("/")[1],l=`${t}/${a}`,m=await Rr(s,l,n),u=`${s}.${m}`;i.name=l,i.version=u,n.text="Writing updated package.json...",ct.writeFileSync(o,`${JSON.stringify(i,null,2)}
|
|
44
|
+
`,"utf8"),n.succeed(`Package.json updated to use ${C.green(u)} and ${C.green(l)}`),process.exit(0)}catch(o){n.fail(`Failed to update package.json: ${o}`),process.exit(1)}},Rr=(r,t,e)=>new Promise(n=>{try{let a=function(l){if(l.statusCode!==200){e.info(`Package not found. Status code: ${C.red(l.statusCode)}, defaulting to version 0`),n(0);return}e.text="Processing version information...";let m="";l.on("data",u=>{m+=u}),l.on("end",()=>{try{let u=JSON.parse(m),c=u.versions?Object.keys(u.versions):[];e.text="Calculating next version number...";let d=-1,h=new RegExp(`^${r}\\.(\\d+)$`);for(let p of c){let w=p.match(h);if(w){let E=Number.parseInt(w[1],10);d=Math.max(d,E)}}d>=0?e.info(`Found existing version ${C.green(`${r}.${d}`)}. Incrementing to ${C.green(`${r}.${d+1}`)}`):e.info(`No existing version found for ${C.green(r)}. Starting with ${C.green(`${r}.0`)}`),n(d+1)}catch(u){e.warn(`Failed to parse NPM registry response: ${u instanceof Error?u.message:"Unknown error"}, defaulting to version 0`),n(0)}})};var o=a;let i=`https://registry.npmjs.org/${t}`,s=lt(i,{headers:{Accept:"application/json"}},l=>{if((l.statusCode===301||l.statusCode===302)&&l.headers.location){e.info(`Following redirect to ${l.headers.location}...`);try{lt(l.headers.location,{headers:{Accept:"application/json"}},a).on("error",m=>{e.warn(`Error following redirect: ${m.message}, defaulting to version 0`),n(0)}).end()}catch(m){e.warn(`Redirect request failed: ${m instanceof Error?m.message:"Unknown error"}, defaulting to version 0`),n(0)}return}a(l)});s.on("error",l=>{e.warn(`Request error: ${l.message}, defaulting to version 0`),n(0)}),s.end()}catch{e.warn("Error checking version in npm registry, defaulting to version 0"),n(0)}});import ge from"node:path";import{fileURLToPath as Sr}from"node:url";import{program as Dr}from"commander";import Or from"open";var jr=Sr(import.meta.url),Nr=ge.resolve(ge.dirname(jr),".."),Bs=Dr.command("test").option("-w, --watch","Set watch number for the test").option("-c, --coverage-report","Generate coverage report").option("-o, --open","Open the coverage report in the browser").option("-f, --files <String|String[]>","Test files glob pattern").description("Run the web test runner to test the component library").action(async r=>{let e=`npx wtr --config "${ge.join(Nr,"dist","configs","web-test-runner.config.mjs")}"`,n=`${process.cwd()}/coverage/index.html`;if(r.coverageReport&&(e+=" --coverage"),r.watch&&(e+=" --watch"),r.files){let o=Array.isArray(r.files)?r.files.join(" "):r.files;e+=` --files "${o}"`}b(e),r.open&&await Or(n)});import be from"node:fs/promises";import pt from"node:path";import{program as Lr}from"commander";import Ur from"inquirer";import ke from"ora";import _r from"node:fs/promises";import Fr from"node:path";import Br from"ora";var Gr={"auth-type":"workspace-token","author-email":null,"author-name":null,"base-branch":"main","base-url":null,"clone-dir":".gitter-temp","code-search":null,concurrent:4,"conflict-strategy":"replace",draft:!1,"dry-run":!0,"fetch-depth":1,fork:!1,"fork-owner":null,"git-type":"go",group:null,"include-subgroups":!1,insecure:!1,interactive:!1,labels:null,"log-file":"'-'","log-format":"'text'","log-level":"'error'","max-reviewers":0,"max-team-reviewers":0,org:null,output:"'-'","plain-output":!1,platform:"github",project:null,"push-only":!1,repo:["AlaskaAirlines/auro-accordion","AlaskaAirlines/auro-alert","AlaskaAirlines/auro-avatar","AlaskaAirlines/auro-background","AlaskaAirlines/auro-backtotop","AlaskaAirlines/auro-button","AlaskaAirlines/auro-badge","AlaskaAirlines/auro-banner","AlaskaAirlines/auro-card","AlaskaAirlines/auro-carousel","AlaskaAirlines/auro-datetime","AlaskaAirlines/auro-dialog","AlaskaAirlines/auro-drawer","AlaskaAirlines/auro-flight","AlaskaAirlines/auro-flightline","AlaskaAirlines/auro-header","AlaskaAirlines/auro-hyperlink","AlaskaAirlines/auro-icon","AlaskaAirlines/auro-loader","AlaskaAirlines/auro-lockup","AlaskaAirlines/auro-nav","AlaskaAirlines/auro-pane","AlaskaAirlines/auro-popover","AlaskaAirlines/auro-sidenav","AlaskaAirlines/auro-skeleton","AlaskaAirlines/auro-slideshow","AlaskaAirlines/auro-table","AlaskaAirlines/auro-tabs","AlaskaAirlines/auro-toast"],"repo-exclude":null,"repo-include":null,"repo-search":null,reviewers:null,"skip-forks":!1,"skip-pr":!1,"skip-repo":null,"ssh-auth":!1,"team-reviewers":null};function Mr(r){return Object.entries(r).map(([t,e])=>Array.isArray(e)?`${t}:
|
|
45
|
+
- ${e.join(`
|
|
46
|
+
- `)}`:typeof e=="object"&&e!==null?`${t}:
|
|
47
|
+
${Object.entries(e).map(([n,o])=>` ${n}: ${o}`).join(`
|
|
48
|
+
`)}`:`${t}: ${e}`).join(`
|
|
49
|
+
`)}async function ut(r){let t=Br("Writing multi-gitter configuration...").start(),e=Mr(Gr),n=Fr.join(r,"multi-gitter_DEPENDENCY_TREE.yml");try{await _r.writeFile(n,e,"utf8"),t.succeed(`Multi-gitter configuration written to ${n}`)}catch(o){t.fail("Error writing multi-gitter configuration:"),console.error(o)}}import he from"node:fs";function mt(r){let t={},e=[],n=[],o=[];for(let i in r)t[i]=r[i].dependsOn.length;for(let i in t)t[i]===0&&o.push(i);for(;o.length>0;){n=[];let i=o.length;for(let s=0;s<i;s++){let a=o.shift();n.push(a);for(let l of r[a].dependentPackages)t[l]--,t[l]===0&&o.push(l)}e.push(n)}if(e.flat().length!==Object.keys(r).length)throw new Error("Circular dependency detected!");return e}function Hr(r){return he.readdirSync(r).filter(t=>t.endsWith(".json"))}async function dt(r,t=[]){console.log(t);let e={},n=Hr(r);for(let o of n){if(o==="dependencyTree.json")continue;let i=he.readFileSync(`${r}/${o}`,"utf-8"),s=JSON.parse(i),a=s.name,l=Object.keys(s.peerDependencies),m=Object.keys(s.devDependencies),u=Object.keys(s.dependencies);e[a]||(e[a]={dependsOn:[],dependentPackages:[]});let c=[...l,...m,...u];e[a].dependsOn=[...new Set(c)];for(let d of c)e[d]||(e[d]={dependsOn:[],dependentPackages:[]}),e[d].dependentPackages.includes(a)||e[d].dependentPackages.push(a)}if(t.length){let o=new Set;for(let[s,a]of Object.entries(e))a.dependsOn.some(l=>t.includes(l))&&o.add(s);for(let s of t)e[s]&&o.add(s);let i={};for(let s of o)i[s]={dependsOn:e[s].dependsOn.filter(a=>o.has(a)),dependentPackages:e[s].dependentPackages.filter(a=>o.has(a))};e=i}else console.log("No target dependencies provided - using all packages.");return he.writeFileSync(`${r}/dependencyTree.json`,JSON.stringify(e,null,2)),e}var ye=ee("run-migrations","config"),we=ee("run-migrations","outputs");var ft=["@aurodesignsystem/auro-accordion","@aurodesignsystem/auro-alert","@aurodesignsystem/auro-avatar","@aurodesignsystem/auro-background","@aurodesignsystem/auro-backtotop","@aurodesignsystem/auro-button","@aurodesignsystem/auro-badge","@aurodesignsystem/auro-banner","@aurodesignsystem/auro-card","@aurodesignsystem/auro-carousel","@aurodesignsystem/auro-datetime","@aurodesignsystem/auro-dialog","@aurodesignsystem/auro-drawer","@aurodesignsystem/auro-formkit","@aurodesignsystem/auro-flight","@aurodesignsystem/auro-flightline","@aurodesignsystem/auro-header","@aurodesignsystem/auro-hyperlink","@aurodesignsystem/auro-icon","@aurodesignsystem/auro-loader","@aurodesignsystem/auro-lockup","@aurodesignsystem/auro-nav","@aurodesignsystem/auro-pane","@aurodesignsystem/auro-popover","@aurodesignsystem/auro-sidenav","@aurodesignsystem/auro-skeleton","@aurodesignsystem/auro-slideshow","@aurodesignsystem/auro-table","@aurodesignsystem/auro-tabs","@aurodesignsystem/auro-toast"],ti=[...ft,"@aurodesignsystem/auro-library","@aurodesignsystem/WebCoreStyleSheets","@aurodesignsystem/AuroDesignTokens","@aurodesignsystem/auro-cli","@alaskaairux/icons"];async function Wr(r){try{await be.mkdir(we,{recursive:!0}),await be.mkdir(ye,{recursive:!0})}catch(o){console.error("Failed to create output or config directories:",o),process.exit(1)}let t=ke("Creating dependency tree...").start();t.text="Creating multi-gitter dependency tree configuration...",await ut(ye),t.text="Scraping dependencies from Auro packages...";let n=`multi-gitter run "node ${te("static","getAuroDeps.js")}" --config ${pt.join(ye,"multi-gitter_DEPENDENCY_TREE.yml")}`;try{await b(n)}catch(o){t.fail("Failed to generate dependency tree:"),console.error(o),process.exit(1)}return t.text="Generating dependency tree JSON file using packages...",await dt(we,r),t.succeed("Dependency tree generated successfully."),pt.join(we,"dependencyTree.json")}var zr=async r=>{let t=ke("Loading dependency tree...").start(),e=JSON.parse(await be.readFile(r,"utf-8"));t.text="Processing dependency tree...";let n=mt(e);return t.succeed("Dependency batches created successfully."),n},ri=Lr.command("agent").action(async r=>{let t=await Ur.prompt([{type:"select",name:"agentAction",message:"What agent action do you want to perform?",choices:[{name:"Run a migration on auro components",value:"run-migration"}],default:["run-migration"]},{type:"input",name:"migrationId",message:"What migration id do you want to run?",when:e=>e.agentAction==="run-migration",validate:e=>e.trim()!==""||"Migration id cannot be empty."},{type:"confirm",name:"useExisting",message:"Would you like to specify starting packages?",default:!0,transformer:e=>e?"Yes = Packages related to selections":"No = All packages",when:e=>e.agentAction==="run-migration"},{type:"checkbox",name:"startWithComponents",message:"Enter the components to start with (comma-separated, blank for all):",choices:ft.map(e=>({name:e.replace("@aurodesignsystem/",""),value:e})),when:e=>e.agentAction==="run-migration"&&e.useExisting}]);switch(t.agentAction){case"run-migration":{let e=ke("Running migration...").start(),n=await Wr(t.startWithComponents);e.text="Getting dependency batches from tree...";let i=(await zr(n)).map((s,a)=>`Batch ${a+1}
|
|
53
50
|
${s.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alaskaairux/icons","AlaskaAirlines/Icons")}`).join(`
|
|
54
51
|
`)}`).join(`
|
|
55
52
|
|
|
56
|
-
`);console.log(i),
|
|
53
|
+
`);console.log(i),e.text="Running migrations on dependency batches...",new Promise(s=>setTimeout(s,2e3)),e.succeed("Migration process completed successfully.");break}default:console.error("Unknown action selected.")}});import{program as qr}from"commander";var $e=qr.command("docs").description("Generate API documentation").option("-c, --cem","Generate Custom Elements Manifest (CEM) file",!1).option("-a, --api","Creates api md file from CEM",!1).option("--skip-readme","Skip README.md processing",!1);$e=j($e);var ai=$e.action(async r=>{r.cem&&await G(),r.api&&await M(),await De(r),r.serve&&await Oe(r)});import{program as Xr}from"commander";import{Octokit as Ae}from"@octokit/rest";import*as Y from"azure-devops-node-api";import J from"ora";var Vr=async r=>{let t=process.env.GH_TOKEN;if(!t)throw new Error("GH_TOKEN environment variable is required");let e=new Ae({auth:t}),n,o,i;if(r.includes("github.com")){let a=r.match(/github\.com\/([^\/]+)\/([^\/]+)\/issues\/(\d+)/);if(!a)throw new Error("Invalid GitHub issue URL format");[,n,o,i]=a}else if(r.includes("#")){let a=r.match(/([^\/]+)\/([^#]+)#(\d+)/);if(!a)throw new Error("Invalid GitHub issue reference format");[,n,o,i]=a}else throw new Error("Issue must be provided as full URL or in format 'owner/repo#number'");let s=Number.parseInt(i,10);try{let{data:a}=await e.rest.issues.get({owner:n,repo:o,issue_number:s});return{title:a.title,body:a.body??null,html_url:a.html_url,number:a.number,repository:{owner:{login:n},name:o}}}catch(a){throw new Error(`Failed to fetch GitHub issue: ${a}`)}},Kr=async r=>{let t=process.env.GH_TOKEN;if(!t)return null;let e=new Ae({auth:t});try{let n=`
|
|
57
54
|
query($owner: String!, $repo: String!, $issueNumber: Int!) {
|
|
58
55
|
repository(owner: $owner, name: $repo) {
|
|
59
56
|
issue(number: $issueNumber) {
|
|
@@ -79,7 +76,7 @@ ${s.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
|
|
|
79
76
|
}
|
|
80
77
|
}
|
|
81
78
|
}
|
|
82
|
-
`,o={owner:
|
|
79
|
+
`,o={owner:r.repository.owner.login,repo:r.repository.name,issueNumber:r.number},s=(await e.graphql(n,o)).repository.issue.projectItems.nodes.find(a=>a.project.number===19);if(s){let a=s.fieldValues.nodes.find(l=>l.field?.name?.toLowerCase()==="ado"&&l.text?.trim());if(a?.text?.trim())return a.text.trim()}return null}catch(n){return console.error(`Failed to check existing ADO link: ${n}`),null}},Jr=async(r,t)=>{let e=process.env.GH_TOKEN;if(!e)throw new Error("GH_TOKEN environment variable is required");let n=new Ae({auth:e}),o=19;try{let i=`
|
|
83
80
|
query($org: String!, $projectNumber: Int!, $owner: String!, $repo: String!, $issueNumber: Int!) {
|
|
84
81
|
organization(login: $org) {
|
|
85
82
|
projectV2(number: $projectNumber) {
|
|
@@ -116,7 +113,7 @@ ${s.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
|
|
|
116
113
|
}
|
|
117
114
|
}
|
|
118
115
|
}
|
|
119
|
-
`,s={org:"AlaskaAirlines",projectNumber:o,owner:
|
|
116
|
+
`,s={org:"AlaskaAirlines",projectNumber:o,owner:r.repository.owner.login,repo:r.repository.name,issueNumber:r.number},a=await n.graphql(i,s),l=a.organization.projectV2.id,m=a.repository.issue.id,u=a.organization.projectV2.fields.nodes.find(d=>d.name?.toLowerCase()==="ado"),c=a.repository.issue.projectItems.nodes.find(d=>d.project.number===o)?.id;if(c||(c=(await n.graphql(`
|
|
120
117
|
mutation($projectId: ID!, $contentId: ID!) {
|
|
121
118
|
addProjectV2ItemById(
|
|
122
119
|
input: {
|
|
@@ -129,7 +126,7 @@ ${s.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
|
|
|
129
126
|
}
|
|
130
127
|
}
|
|
131
128
|
}
|
|
132
|
-
`,{projectId:l,contentId:
|
|
129
|
+
`,{projectId:l,contentId:m})).addProjectV2ItemById.item.id),u&&c)await n.graphql(`
|
|
133
130
|
mutation($projectId: ID!, $itemId: ID!, $fieldId: ID!, $value: String!) {
|
|
134
131
|
updateProjectV2ItemFieldValue(
|
|
135
132
|
input: {
|
|
@@ -146,4 +143,4 @@ ${s.map(l=>` - ${l.replace("@aurodesignsystem","AlaskaAirlines").replace("@alas
|
|
|
146
143
|
}
|
|
147
144
|
}
|
|
148
145
|
}
|
|
149
|
-
`,{projectId:l,itemId:
|
|
146
|
+
`,{projectId:l,itemId:c,fieldId:u.id,value:t});else if(!u)throw new Error("No 'ado' field found in GitHub project")}catch(i){console.error(`Failed to update GitHub project: ${i}`)}},Yr=async r=>{let t=process.env.ADO_TOKEN;if(!t)throw new Error("ADO_TOKEN environment variable is required");let e="https://dev.azure.com/itsals",n="E_Retain_Content",o="E_Retain_Content\\Auro Design System",i=Y.getPersonalAccessTokenHandler(t),a=await new Y.WebApi(e,i).getWorkItemTrackingApi();try{let l=[{op:"add",path:"/fields/System.Title",value:r.title},{op:"add",path:"/fields/System.Description",value:`GitHub Issue: <a href="${r.html_url}">${r.html_url}</a>`},{op:"add",path:"/fields/System.AreaPath",value:o}];return await a.createWorkItem(null,l,n,"User Story")}catch(l){throw new Error(`Failed to create ADO work item: ${l}`)}},gt=async r=>{let t=J(`Processing GitHub issue: ${r}`).start();try{if(!process.env.GH_TOKEN)throw new Error("GH_TOKEN environment variable is required");if(!process.env.ADO_TOKEN)throw new Error("ADO_TOKEN environment variable is required");t.text="Fetching GitHub issue details...";let e=await Vr(r);t.succeed(`Found issue: "${e.title}"`);let n=J("Checking for existing ADO work item...").start(),o=await Kr(e);if(o){n.succeed("ADO work item already exists for this issue!"),console.log(`${o}`);return}n.succeed("No existing ADO work item found");let i=J("Creating new ADO work item...").start(),s=await Yr(e);if(i.succeed(`Successfully created ADO work item #${s.id}`),console.log(`Work item: ${s._links?.html?.href||"N/A"}`),s._links?.html?.href){let a=J("Adding to GitHub project and updating ADO field...").start();await Jr(e,s._links.html.href),a.succeed("Updated GitHub project with ADO link")}}catch(e){t.fail(`Error: ${e instanceof Error?e.message:e}`),process.exit(1)}};var fi=Xr.command("ado").description("Generate ADO item from GitHub issue").option("-g, --gh-issue <issue>","What GitHub issue to use").action(async r=>{r.ghIssue&&await gt(r.ghIssue)});import{program as tn}from"commander";import{Octokit as Zr}from"@octokit/rest";import{simpleGit as Qr}from"simple-git";var ht="Release Candidate",S="dev",en="main",X=class r{constructor(t,e,n){this.filteredCommits=null;this.repoInfo={owner:t,repo:e},this.octokit=n}static async create(){let t=process.env.GITHUB_TOKEN;if(!t)throw new Error("GITHUB_TOKEN is required to run RC workflow.");let e=await v.getRepoOwnerAndName(),n=new Zr({auth:t});if(!e)throw new Error("Failed to retrieve repository information. Ensure you're in a valid git repository.");let o=await r.getTriggerBranchName();return o&&o!==S&&(console.log(`Switching from ${o} to ${S} branch...`),await Qr().checkout(S)),new r(e.owner,e.repo,n)}get owner(){return this.repoInfo.owner}get repo(){return this.repoInfo.repo}get repoData(){return{...this.repoInfo}}async createReleaseCandidate(){await this.hasCommitsReadyInDev()||console.log("No filtered commits found. Continuing to update RC issue/branch/PR.");let e=await this.getLatestOpenRcIssue(),n=e?await this.getLinkedPrByHead(e.number):null;if(n?.multipleOpen)throw new Error("Multiple open RC PRs found for the same rc/<issueNumber> branch.");if(n?.state==="closed"&&(console.log("Linked RC PR is closed. Creating a new RC issue and PR."),e=await this.createRcIssue(),n=null),e?await this.updateRcIssue(e.number):e=await this.createRcIssue(),!e)throw new Error("Failed to resolve RC issue.");await this.createOrUpdateRcBranch(e.number),n?await this.updateRcPr(e.number,n.number):n=await this.createRcPr(e.number)}async getFilteredCommits(){if(this.filteredCommits===null){let t=await v.getCommitMessages(S);this.filteredCommits=fe(t)}return this.filteredCommits}async hasCommitsReadyInDev(){return(await this.getFilteredCommits()).length>0}async getLatestOpenRcIssue(){let{data:t}=await this.octokit.rest.issues.listForRepo({owner:this.repoInfo.owner,repo:this.repoInfo.repo,labels:ht,state:"open",sort:"updated",direction:"desc",per_page:30}),e=t.filter(o=>!o.pull_request);if(e.length===0)return console.log(`No open Release Candidate issues found in ${this.repoInfo.repo}`),null;let n=e[0];return console.log(`Using latest open Release Candidate issue: #${n.number}`),{number:n.number,title:n.title||""}}async updateRcIssue(t){let e=await this.getReleaseNotes(),n=`RC ${this.getCurrentDate()}`;await this.octokit.rest.issues.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,issue_number:t,title:n,body:e})}async createRcIssue(){let t=await this.getReleaseNotes(),{data:e}=await this.octokit.rest.issues.create({owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC ${this.getCurrentDate()}`,labels:[ht],body:t});return console.log(`Created Release Candidate issue: #${e.number} (${e.html_url})`),{number:e.number,html_url:e.html_url}}async createOrUpdateRcBranch(t){let e=`heads/rc/${t}`,n=`rc/${t}`,{data:o}=await this.octokit.rest.repos.getBranch({owner:this.repoInfo.owner,repo:this.repoInfo.repo,branch:S}),{data:i}=await this.octokit.rest.git.listMatchingRefs({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e}),s=i.length>0;try{s?(console.log(`Updating existing RC branch: ${n}`),await this.octokit.rest.git.updateRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:e,sha:o.commit.sha,force:!0})):(console.log(`Creating new RC branch: ${n}`),await this.octokit.rest.git.createRef({owner:this.repoInfo.owner,repo:this.repoInfo.repo,ref:`refs/${e}`,sha:o.commit.sha}))}catch(a){throw new Error(`Failed to create or update ${n} branch: ${a}`)}}async getLinkedPrByHead(t){let e=`${this.repoInfo.owner}:rc/${t}`,{data:n}=await this.octokit.rest.pulls.list({owner:this.repoInfo.owner,repo:this.repoInfo.repo,state:"all",head:e,per_page:30}),o=n.filter(s=>s.state==="open");if(o.length>1)return{state:"open",multipleOpen:!0};if(o.length===1)return{state:"open",html_url:o[0].html_url,number:o[0].number};let i=n.filter(s=>s.state==="closed");return i.length>0?{state:"closed",html_url:i[0].html_url,number:i[0].number}:null}async fetchPrTemplate(t){try{let{data:e}=await this.octokit.rest.repos.getContent({owner:this.repoInfo.owner,repo:this.repoInfo.repo,path:".github/PULL_REQUEST_TEMPLATE.md"});if("content"in e&&e.type==="file"){let n=Buffer.from(e.content,"base64").toString("utf-8");return n=n.replace("Please include a summary of the change and which issue is fixed. Please also include relevant motivation and context. List any dependencies that are required for this change.",`Release candidate pull request. See issue #${t} for details.`),n=n.replace(/<details>/g,"<details open>"),n}}catch(e){e&&typeof e=="object"&&"status"in e&&e.status===404?console.log("No PR template found in repo, using default message."):console.warn("Failed to fetch PR template:",e)}return`Release candidate pull request. See issue #${t} for details.`}async createRcPr(t){try{let e=await this.fetchPrTemplate(t),{data:n}=await this.octokit.request(`POST /repos/${this.repoInfo.owner}/${this.repoInfo.repo}/pulls`,{owner:this.repoInfo.owner,repo:this.repoInfo.repo,title:`RC #${t}`,body:e,head:`rc/${t}`,base:en,headers:{"X-GitHub-Api-Version":"2022-11-28"}});return console.log(`Created Release Candidate pull request: #${n.number} (${n.html_url})`),{state:"open",html_url:n.html_url,number:n.number}}catch(e){throw console.error("Failed to create RC PR:",e),e}}async updateRcPr(t,e){try{let n=await this.fetchPrTemplate(t);await this.octokit.rest.pulls.update({owner:this.repoInfo.owner,repo:this.repoInfo.repo,pull_number:e,body:n}),console.log(`Updated Release Candidate pull request: #${e}`)}catch(n){throw console.error("Failed to update RC PR:",n),n}}async getReleaseNotes(){let t=await this.getFilteredCommits();return pe(t,!1)}static async getTriggerBranchName(){return process.env.GITHUB_REF_NAME?process.env.GITHUB_REF_NAME:process.env.GITHUB_REF?.startsWith("refs/heads/")?process.env.GITHUB_REF.replace("refs/heads/",""):v.getCurrentBranchName()}getCurrentDate(){return new Date().toISOString().split("T")[0]}};var vi=tn.command("rc-workflow").description("Generate RC issue and pull request").action(async()=>{await(await X.create()).createReleaseCandidate()});ve.name("auro").version(Z()).description("A cli tool to support the Auro Design System");ve.addHelpText("beforeAll",Ce());ve.parse();
|