@olegkuibar/plunk 0.7.6 → 0.8.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (55) hide show
  1. package/dist/add-S5F3UPAP.mjs +14 -0
  2. package/dist/chunk-2GDRDQA5.mjs +3 -0
  3. package/dist/{chunk-HHV6VEYA.mjs → chunk-5EOIBIQO.mjs} +3 -3
  4. package/dist/chunk-7ZWOJDCA.mjs +4 -0
  5. package/dist/chunk-ETKOGSRN.mjs +3 -0
  6. package/dist/chunk-GAS7XU4A.mjs +3 -0
  7. package/dist/chunk-HN3SZGD3.mjs +4 -0
  8. package/dist/chunk-PH7M73X2.mjs +13 -0
  9. package/dist/chunk-S3IAOIXI.mjs +3 -0
  10. package/dist/chunk-SG6M7SAC.mjs +4 -0
  11. package/dist/chunk-ULPTRGDP.mjs +3 -0
  12. package/dist/chunk-W4LHCWGD.mjs +3 -0
  13. package/dist/chunk-WSJHZ2ND.mjs +19 -0
  14. package/dist/clean-STIN6OS6.mjs +3 -0
  15. package/dist/cli.mjs +1 -1
  16. package/dist/{dev-DBGYZOIV.mjs → dev-YR4ZYYT4.mjs} +2 -2
  17. package/dist/doctor-DAEAGACF.mjs +4 -0
  18. package/dist/fs-6ZUVEONC.mjs +2 -0
  19. package/dist/index.d.ts +31 -5
  20. package/dist/index.mjs +250 -148
  21. package/dist/{init-C6XCSFCU.mjs → init-G6X5O3A3.mjs} +2 -2
  22. package/dist/list-N2B4TO6R.mjs +5 -0
  23. package/dist/migrate-YZRZUG54.mjs +8 -0
  24. package/dist/{publish-6A7PX5IH.mjs → publish-DGEQG3L5.mjs} +2 -2
  25. package/dist/{push-AUAGCBYK.mjs → push-VS3VZKU4.mjs} +1 -1
  26. package/dist/remove-GPLRNTDM.mjs +2 -0
  27. package/dist/reset-E5P4ZKQ7.mjs +3 -0
  28. package/dist/{restore-JVH6INAG.mjs → restore-WIYNZSQL.mjs} +1 -1
  29. package/dist/{status-22YV26A3.mjs → status-ADDZJ66C.mjs} +2 -2
  30. package/dist/tailwind-source-C2ANIANJ.mjs +5 -0
  31. package/dist/update-2Q46L6NR.mjs +3 -0
  32. package/dist/{vite-config-LXMLHR7V.mjs → vite-config-JEUUATXJ.mjs} +4 -4
  33. package/dist/watcher-KAPGRKYP.mjs +3 -0
  34. package/dist/workspace-4U2WM4MX.mjs +3 -0
  35. package/package.json +1 -1
  36. package/dist/add-GYBX4VAZ.mjs +0 -14
  37. package/dist/chunk-22YCXJTS.mjs +0 -4
  38. package/dist/chunk-4ZGIZZSF.mjs +0 -12
  39. package/dist/chunk-AC5FETT7.mjs +0 -4
  40. package/dist/chunk-CIY4E6PA.mjs +0 -3
  41. package/dist/chunk-EMRPZYLU.mjs +0 -3
  42. package/dist/chunk-LLVBXPQN.mjs +0 -3
  43. package/dist/chunk-PVMVWPLG.mjs +0 -4
  44. package/dist/chunk-SMIGYQFG.mjs +0 -15
  45. package/dist/chunk-TMH7HIJ2.mjs +0 -3
  46. package/dist/clean-F2IWAVRP.mjs +0 -3
  47. package/dist/doctor-GJGAAT6J.mjs +0 -4
  48. package/dist/fs-BUNURH4P.mjs +0 -2
  49. package/dist/list-QSPN7FE5.mjs +0 -5
  50. package/dist/migrate-4TFDXO4G.mjs +0 -8
  51. package/dist/remove-5DAQD627.mjs +0 -3
  52. package/dist/tailwind-source-YCRZUHUP.mjs +0 -5
  53. package/dist/update-33ICRFYZ.mjs +0 -3
  54. package/dist/watcher-SY45L3SS.mjs +0 -3
  55. package/dist/workspace-SKJJQMMK.mjs +0 -3
@@ -0,0 +1,14 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {a as a$6}from'./chunk-5EOIBIQO.mjs';import {b as b$5,a as a$5,e}from'./chunk-WSJHZ2ND.mjs';import {d}from'./chunk-PH7M73X2.mjs';import {a as a$4,b as b$4,c as c$1,d as d$4}from'./chunk-7ZWOJDCA.mjs';import {f,d as d$1,h}from'./chunk-W4LHCWGD.mjs';import {a as a$3}from'./chunk-HN3SZGD3.mjs';import {a as a$2}from'./chunk-HPF6K6WO.mjs';import {b as b$3,c}from'./chunk-ETKOGSRN.mjs';import {a as a$1}from'./chunk-MBKCCWSD.mjs';import {d as d$2}from'./chunk-GAS7XU4A.mjs';import {b as b$1,a as a$7}from'./chunk-OJJZ7CLB.mjs';import'./chunk-OL7SNXMX.mjs';import'./chunk-HW7AEGI3.mjs';import {w,j}from'./chunk-SG6M7SAC.mjs';import {c as c$2,d as d$3}from'./chunk-R3RSOZXN.mjs';import {b as b$2}from'./chunk-ICCM7US5.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import'fs/promises';import {spawn}from'child_process';import {platform}from'os';import {resolve,basename}from'path';globalThis.require=createRequire(import.meta.url);
3
+ var In={meta:{name:"add",description:"Link a package from the plunk store into this project"},args:{package:{type:"positional",description:"Package name to add",required:true},from:{type:"string",description:"Path to package source (will publish first)"},yes:{type:"boolean",alias:"y",description:"Auto-accept prompts (install missing deps, etc.)",default:false}},async run({args:t}){b$1();let o=new a$1,e$1=resolve("."),{name:n,version:p}=Z(t.package);if((!n||n==="@"||n.startsWith("@")&&!n.includes("/"))&&(a$2(`Invalid package name "${t.package}". Use format: package-name or @scope/package-name`),process.exit(1)),t.from){let s=resolve(t.from);b$2.info(`Publishing from ${s}...`),await a$3(s);}let a=p?await b$3(n,p):await c(n);if(!a){let s=p?`@${p} `:"";a$2(`Package "${n}"${s?" "+s:""} not found in store. Run 'plunk publish' in the package directory first, or use --from <path>.`),process.exit(1);}let f$1=!await w(j(e$1)),r=await a$4(e$1);if(f$1&&(await d(e$1,r),b$2.success("Auto-initialized plunk (consumer mode)")),b$2.info(`Detected package manager: ${r}`),r==="yarn"){let s=await b$4(e$1);(s==="pnp"||s===null&&await c$1(e$1))&&(b$2.error(`Yarn PnP mode is not compatible with plunk.
4
+
5
+ plunk works by copying files into node_modules/, but PnP eliminates
6
+ node_modules/ entirely. To use plunk with Yarn Berry, add this to
7
+ .yarnrc.yml:
8
+
9
+ nodeLinker: node-modules
10
+
11
+ Then run: yarn install`),process.exit(1));}let k=await f(e$1,n);k&&(k.version===a.version?b$2.info(`Updating ${n}@${a.version} (already linked)`):b$2.info(`Updating ${n}: ${k.version} \u2192 ${a.version}`));let v=await b$5(e$1,n,r);v&&b$2.info(`Backed up existing ${n} installation`);let d$5=await a$5(a,e$1,r);b$2.success(`Linked ${n}@${a.version} \u2192 node_modules/${n} (${d$5.copied} files copied, ${d$5.skipped} unchanged)`),d$5.binLinks>0&&b$2.info(`Created ${d$5.binLinks} bin link(s)`);let z={version:a.version,contentHash:a.meta.contentHash,linkedAt:new Date().toISOString(),sourcePath:a.meta.sourcePath,backupExists:v,packageManager:r,buildId:a.meta.buildId??""};await d$1(e$1,n,z),await h(n,e$1),await d$2(e$1,n,a.version);let l=await e(a,e$1);if(l.length>0)if(c$2())d$3(`[add] Missing transitive deps (json mode): ${l.join(", ")}`);else if(t.yes){let s=y(r,l);b$2.info(`Installing missing dependencies: ${l.join(", ")}`),await b(s,e$1)?b$2.success("Installed missing dependencies"):b$2.warn(`Install failed. Run manually: ${s}`);}else if(await b$2.prompt(`Install ${l.length} missing dependencies? (${l.join(", ")})`,{type:"confirm",initial:true})){let m=y(r,l);await b(m,e$1)?b$2.success("Installed missing dependencies"):b$2.warn(`Install failed. Run manually: ${m}`);}else b$2.warn(`Missing transitive dependencies: ${l.join(", ")}
12
+ Run: ${y(r,l)}`);let c$3=await d$4(e$1);if(c$3.type==="next"&&c$3.configFile){let s=await a$6(c$3.configFile,n);s.modified?b$2.success(`Added ${n} to transpilePackages in ${basename(c$3.configFile)}`):s.error&&b$2.info(`Add to next.config manually: transpilePackages: ['${n}']`);}else if(c$3.type==="vite"&&c$3.configFile){let{addPlunkVitePlugin:s}=await import('./vite-config-JEUUATXJ.mjs'),m=await s(c$3.configFile);if(m.modified){b$2.success(`Added plunk plugin to ${basename(c$3.configFile)}`);let w=X(r,"@olegkuibar/plunk");b$2.info("Installing @olegkuibar/plunk as devDependency..."),await b(w,e$1)?b$2.success("Installed @olegkuibar/plunk"):b$2.warn(`Install failed. Run manually: ${w}`);}else m.error&&b$2.info(`Add manually:
13
+ import plunk from "@olegkuibar/plunk/vite"
14
+ plugins: [plunk()]`);}let{findTailwindCss:J,addTailwindSource:G}=await import('./tailwind-source-C2ANIANJ.mjs'),g=await J(e$1);if(g){let s=await G(g,n,e$1);s.modified?b$2.success(`Added @source for ${n} to ${basename(g)}`):s.error&&b$2.info(`Add to your CSS manually: @source "../node_modules/${n}";`);}b$2.info(`Done in ${o.elapsed()}`),a$7({package:n,version:a.version,copied:d$5.copied,skipped:d$5.skipped,binLinks:d$5.binLinks,elapsed:o.elapsedMs()});}};function y(t,o){let e=o.join(" ");switch(t){case "pnpm":return `pnpm add ${e}`;case "yarn":return `yarn add ${e}`;case "bun":return `bun add ${e}`;default:return `npm install ${e}`}}a(y,"buildInstallCommand");function X(t,o){switch(t){case "pnpm":return `pnpm add -D ${o}`;case "yarn":return `yarn add -D ${o}`;case "bun":return `bun add -d ${o}`;default:return `npm install -D ${o}`}}a(X,"buildDevInstallCommand");function b(t,o){return new Promise(e=>{let n=platform()==="win32",f=spawn(n?"cmd":"sh",[n?"/c":"-c",t],{cwd:o,stdio:"inherit"});f.on("close",r=>e(r===0)),f.on("error",()=>e(false));})}a(b,"runInstallCommand");function Z(t){if(t.startsWith("@")){let e=t.indexOf("/");if(e>0){let n=t.indexOf("@",e);if(n>e)return {name:t.slice(0,n),version:t.slice(n+1)}}return {name:t,version:null}}let o=t.lastIndexOf("@");return o>0?{name:t.slice(0,o),version:t.slice(o+1)}:{name:t,version:null}}a(Z,"parsePackageArg");export{In as default};
@@ -0,0 +1,3 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {a}from'./chunk-2VCW5RWI.mjs';globalThis.require=createRequire(import.meta.url);
3
+ function B(t){if(t===0)return "0 B";let o=["B","KB","MB","GB"],n=Math.min(Math.floor(Math.log(t)/Math.log(1024)),o.length-1),i=t/1024**n;return n===0?`${t} B`:`${i.toFixed(1)} ${o[n]}`}a(B,"formatBytes");export{B as a};
@@ -1,7 +1,7 @@
1
1
  #!/usr/bin/env node
2
- import {createRequire}from'node:module';import {l}from'./chunk-TMH7HIJ2.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {readFile}from'fs/promises';globalThis.require=createRequire(import.meta.url);
2
+ import {createRequire}from'node:module';import {x as x$1}from'./chunk-SG6M7SAC.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {readFile}from'fs/promises';globalThis.require=createRequire(import.meta.url);
3
3
  function m(e){return e.match(/^(\s+)\S/m)?.[1]||" "}a(m,"detectIndent");function p(e){let n=[],t=/['"]([^'"]*)['"]/g,r;for(;(r=t.exec(e))!==null;)n.push(r[1]);return n}a(p,"parseArrayItems");function x(e,n){return e.length===0?"[]":e.length<=3?`[${e.map(r=>`'${r}'`).join(", ")}]`:`[
4
4
  ${e.map(r=>`${n}${n}'${r}',`).join(`
5
5
  `)}
6
- ${n}]`}a(x,"formatArray");async function w(e,n){let t;try{t=await readFile(e,"utf-8");}catch{return {modified:false,error:"could not read config file"}}let r=m(t),a=/transpilePackages\s*:\s*\[([^\]]*)\]/s,c=a.exec(t);if(c){let i=p(c[1]);if(i.includes(n))return {modified:false};i.push(n);let l$1=x(i,r),d=t.replace(a,`transpilePackages: ${l$1}`);return await l(e,d),{modified:true}}let s=/(?:module\.exports\s*=\s*\{|export\s+default\s+\{|nextConfig\s*=\s*\{)/.exec(t);if(s){let i=s.index+s[0].length,l$1=`
7
- ${r}transpilePackages: ['${n}'],`,d=t.slice(0,i)+l$1+t.slice(i);return await l(e,d),{modified:true}}return {modified:false,error:"unrecognized config pattern"}}a(w,"addToTranspilePackages");async function y(e,n){let t;try{t=await readFile(e,"utf-8");}catch{return {modified:false}}let r=m(t),a=/transpilePackages\s*:\s*\[([^\]]*)\]/s,c=a.exec(t);if(!c)return {modified:false};let g=p(c[1]),s=g.filter(d=>d!==n);if(s.length===g.length)return {modified:false};let i=x(s,r),l$1=t.replace(a,`transpilePackages: ${i}`);return await l(e,l$1),{modified:true}}a(y,"removeFromTranspilePackages");export{w as a,y as b};
6
+ ${n}]`}a(x,"formatArray");async function w(e,n){let t;try{t=await readFile(e,"utf-8");}catch{return {modified:false,error:"could not read config file"}}let r=m(t),a=/transpilePackages\s*:\s*\[([^\]]*)\]/s,c=a.exec(t);if(c){let i=p(c[1]);if(i.includes(n))return {modified:false};i.push(n);let l=x(i,r),d=t.replace(a,`transpilePackages: ${l}`);return await x$1(e,d),{modified:true}}let s=/(?:module\.exports\s*=\s*\{|export\s+default\s+\{|nextConfig\s*=\s*\{)/.exec(t);if(s){let i=s.index+s[0].length,l=`
7
+ ${r}transpilePackages: ['${n}'],`,d=t.slice(0,i)+l+t.slice(i);return await x$1(e,d),{modified:true}}return {modified:false,error:"unrecognized config pattern"}}a(w,"addToTranspilePackages");async function y(e,n){let t;try{t=await readFile(e,"utf-8");}catch{return {modified:false}}let r=m(t),a=/transpilePackages\s*:\s*\[([^\]]*)\]/s,c=a.exec(t);if(!c)return {modified:false};let g=p(c[1]),s=g.filter(d=>d!==n);if(s.length===g.length)return {modified:false};let i=x(s,r),l=t.replace(a,`transpilePackages: ${i}`);return await x$1(e,l),{modified:true}}a(y,"removeFromTranspilePackages");export{w as a,y as b};
@@ -0,0 +1,4 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {w}from'./chunk-SG6M7SAC.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {readFile,stat}from'fs/promises';import {join,dirname}from'path';globalThis.require=createRequire(import.meta.url);
3
+ var y=new Set(["npm","pnpm","yarn","bun"]),k=[["pnpm-lock.yaml","pnpm"],["bun.lockb","bun"],["bun.lock","bun"],["yarn.lock","yarn"],["package-lock.json","npm"]];async function P(r){try{let n=await readFile(join(r,"package.json"),"utf-8"),e=JSON.parse(n);if(typeof e.packageManager!="string")return null;let t=e.packageManager.split("@")[0];return y.has(t)?t:null}catch{return null}}a(P,"readPackageManagerField");async function x(r){let n=r;for(;;){let e=await P(n);if(e)return e;let o=(await Promise.all(k.map(async([a,s])=>{try{return await stat(join(n,a)),s}catch{return null}}))).find(a=>a!==null);if(o)return o;let i=dirname(n);if(i===n)return "npm";n=i;}}a(x,"detectPackageManager");async function h(r){let n=r;for(;;){let e;try{e=await readFile(join(n,".yarnrc.yml"),"utf-8");}catch{let t=dirname(n);if(t===n)return null;n=t;continue}for(let t of e.split(`
4
+ `)){let o=t.trim();if(o.startsWith("#")||!o.includes("nodeLinker"))continue;let i=o.match(/^nodeLinker:\s*(.+)$/);if(i){let a=i[1].trim().replace(/^["']|["']$/g,"");if(a==="node-modules"||a==="pnpm"||a==="pnp")return a}}return null}}a(h,"detectYarnNodeLinker");async function j(r){let n=r;for(;;)try{return await stat(join(n,".yarnrc.yml")),!0}catch{let e=dirname(n);if(e===n)return false;n=e;}}a(j,"hasYarnrcYml");var d=[["vite",["vite.config.ts","vite.config.js","vite.config.mts","vite.config.mjs"]],["next",["next.config.js","next.config.ts","next.config.mjs"]],["webpack",["webpack.config.js","webpack.config.ts"]],["turbo",["turbo.json"]],["rollup",["rollup.config.js","rollup.config.ts","rollup.config.mjs"]]];async function v(r){for(let[n,e]of d)for(let t of e){let o=join(r,t);if(await w(o))return {type:n,configFile:o}}return {type:null,configFile:null}}a(v,"detectBundler");async function I(r){let n=d.flatMap(([i,a])=>a.map(async s=>({type:i,configFile:join(r,s),found:await w(join(r,s))}))),e=await Promise.all(n),t=new Set,o=[];for(let{type:i,configFile:a,found:s}of e)s&&!t.has(i)&&(t.add(i),o.push({type:i,configFile:a}));return o}a(I,"detectAllBundlers");export{x as a,h as b,j as c,v as d,I as e};
@@ -0,0 +1,3 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {b}from'./chunk-GAS7XU4A.mjs';import {h,o,g,w,b as b$2,d,e,t,f}from'./chunk-SG6M7SAC.mjs';import {b as b$1}from'./chunk-ICCM7US5.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {readFile,readdir}from'fs/promises';import'path';globalThis.require=createRequire(import.meta.url);
3
+ async function y(t,r){let s=h(t,r);try{let a=await readFile(s,"utf-8"),e=JSON.parse(a);return b(e)?e:(b$1.warn(`Invalid metadata for ${t}@${r}, ignoring`),null)}catch(a){return o(a)&&a.code!=="ENOENT"&&b$1.warn(`Failed to read metadata for ${t}@${r}: ${a instanceof Error?a.message:String(a)}`),null}}a(y,"readMeta");async function A(t,r){let s=g(t,r),a=await y(t,r);return !a||!await w(s)?null:{name:t,version:r,packageDir:s,meta:a}}a(A,"getStoreEntry");async function J(t){let r=b$2();if(!await w(r))return null;let s=d(t)+"@",e=(await readdir(r,{withFileTypes:true})).filter(n=>n.isDirectory()&&n.name.startsWith(s)),u=(await Promise.all(e.map(async n=>{let i=n.name.slice(s.length),c=await y(t,i);return c?{name:t,version:i,packageDir:g(t,i),meta:c}:null}))).filter(n=>n!==null);return u.length===0?null:u.reduce((n,i)=>new Date(i.meta.publishedAt).getTime()>new Date(n.meta.publishedAt).getTime()?i:n)}a(J,"findStoreEntry");async function W(){let t=b$2();if(!await w(t))return [];let s=(await readdir(t,{withFileTypes:true})).filter(e=>e.isDirectory()?e.name.lastIndexOf("@")>0:false);return (await Promise.all(s.map(async e$1=>{let l=e$1.name.lastIndexOf("@"),u=e$1.name.slice(0,l),n=e$1.name.slice(l+1),i=e(u),c=await y(i,n);return c?{name:i,version:n,packageDir:g(i,n),meta:c}:null}))).filter(e=>e!==null)}a(W,"listStoreEntries");async function q(t$1,r){await t(f(t$1,r));}a(q,"removeStoreEntry");export{y as a,A as b,J as c,W as d,q as e};
@@ -0,0 +1,3 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {o}from'./chunk-SG6M7SAC.mjs';import {b}from'./chunk-ICCM7US5.mjs';import {a as a$1}from'./chunk-2VCW5RWI.mjs';import {readFile,mkdir,stat,rm}from'fs/promises';import {join,dirname}from'path';import {setTimeout}from'timers/promises';globalThis.require=createRequire(import.meta.url);
3
+ function R(e){if(typeof e!="object"||e===null)return false;let t=e;return typeof t.contentHash=="string"&&typeof t.publishedAt=="string"&&typeof t.sourcePath=="string"&&(t.buildId===void 0||typeof t.buildId=="string")&&(t.schemaVersion===void 0||typeof t.schemaVersion=="number")}a$1(R,"isPlunkMeta");function k(e){if(typeof e!="object"||e===null)return false;let t=e;return typeof t.version=="string"&&typeof t.contentHash=="string"&&typeof t.linkedAt=="string"&&typeof t.sourcePath=="string"&&typeof t.backupExists=="boolean"&&typeof t.packageManager=="string"&&["npm","pnpm","yarn","bun"].includes(t.packageManager)&&(t.buildId===void 0||typeof t.buildId=="string")}a$1(k,"isLinkEntry");function T(e){if(typeof e!="object"||e===null)return false;let t=e;if(t.version!=="1"||typeof t.links!="object"||t.links===null)return false;let n=t.links;for(let r of Object.values(n))if(!k(r))return false;return true}a$1(T,"isConsumerState");async function x(e,t,n){try{let r=await readFile(join(e,"package.json"),"utf-8"),u=JSON.parse(r),o=u.dependencies?.[t]??u.devDependencies?.[t]??u.peerDependencies?.[t];if(!o||/^(workspace:|catalog:|\*)/.test(o))return;let i=o.match(/(\d+)\.\d+\.\d+/);if(!i)return;let c=parseInt(i[1],10),f=parseInt(n.split(".")[0],10);c!==f&&b.warn(`Version mismatch: store has ${t}@${n} but your package.json declares "${o}". Consider updating your dependency range.`);}catch{}}a$1(x,"warnVersionMismatch");function E(e){if(typeof e!="object"||e===null)return false;let t=e;for(let n of Object.values(t)){if(!Array.isArray(n))return false;for(let r of n)if(typeof r!="string")return false}return true}a$1(E,"isConsumersRegistry");var a={retries:5,minTimeout:100,maxTimeout:1e3,factor:2,stale:3e4};async function L(e,t,n){await mkdir(dirname(e),{recursive:true});let r=e+".lk",u=n?.stale??a.stale,o$1=false;for(let i=0;i<=a.retries;i++)try{await mkdir(r),o$1=!0;break}catch(c){if(o(c)&&c.code==="EEXIST"){try{let f=await stat(r);if(Date.now()-f.mtimeMs>u){await rm(r,{recursive:!0,force:!0});continue}}catch{continue}if(i<a.retries){let f=Math.min(a.minTimeout*a.factor**i,a.maxTimeout);await setTimeout(f);}}else throw c}if(!o$1)throw new Error(`Failed to acquire lock after ${a.retries} attempts. Another plunk process may be running. If this persists, delete ${r} and retry.`);try{return await t()}finally{await rm(r,{recursive:true,force:true});}}a$1(L,"withFileLock");export{L as a,R as b,T as c,x as d,E as e};
@@ -0,0 +1,4 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {a as a$3}from'./chunk-ETKOGSRN.mjs';import {a as a$4}from'./chunk-GAS7XU4A.mjs';import {a}from'./chunk-HW7AEGI3.mjs';import {a as a$2,l,o,n,f,v,u,p,w,s,t}from'./chunk-SG6M7SAC.mjs';import {d}from'./chunk-R3RSOZXN.mjs';import {b}from'./chunk-ICCM7US5.mjs';import {c,a as a$1}from'./chunk-2VCW5RWI.mjs';import {stat,readFile,readdir,writeFile,rename}from'fs/promises';import {resolve,join,relative,sep,dirname}from'path';import {spawn}from'child_process';import {availableParallelism,platform}from'os';globalThis.require=createRequire(import.meta.url);
3
+ var P=c(a(),1);async function Q(s,n){let o$1=[],t=resolve(s);o$1.push(join(t,"package.json"));let e=await X(t,t),r=e.map(a=>l(relative(t,a)));if(n.files&&n.files.length>0)for(let a of n.files){let i=join(t,a),u=resolve(i);if(!u.startsWith(t+sep)&&u!==t){b.warn(`files pattern "${a}" escapes package directory, skipping`);continue}let l$1=false;try{if((await stat(i)).isDirectory()){let w=l(relative(t,i))+"/";for(let f=0;f<r.length;f++)r[f].startsWith(w)&&o$1.push(e[f]);l$1=!0;}else o$1.push(i),l$1=!0;}catch(d){if(o(d)&&d.code!=="ENOENT")throw d}if(!l$1){let d=(0, P.default)(a,{dot:true}),w=0;for(let f=0;f<r.length;f++)if(d(r[f])){let E=resolve(e[f]);if(!E.startsWith(t+sep)&&E!==t)continue;o$1.push(e[f]),w++;}w===0&&b.warn(`files pattern "${a}" matched no files`);}}else {let a=await fe(t);for(let i=0;i<r.length;i++)ce(r[i],a)||o$1.push(e[i]);}let p=new Set(o$1),c=new Set(e);for(let a of ["README.md","README","LICENSE","LICENCE","CHANGELOG.md"]){let i=join(t,a);p.has(i)||c.has(i)&&(o$1.push(i),p.add(i));}return [...p]}a$1(Q,"resolvePackFiles");var le=new Set(["node_modules",".git",".svn",".hg",".DS_Store",".npmrc",".plunk","test","tests","__tests__",".github",".vscode",".idea","coverage",".nyc_output","tsconfig.json","tsconfig.build.json",".eslintrc",".eslintrc.js",".eslintrc.json",".prettierrc",".prettierrc.js","jest.config.js","jest.config.ts","vitest.config.ts","vitest.config.js"]);function ce(s,n){for(let t of n.negations)if(t(s))return false;let o=s.split(/[\\/]/);for(let t of o)if(le.has(t)||n.literals.has(t))return true;if(n.literals.has(s))return true;for(let t of n.patterns)if(t(s))return true;return false}a$1(ce,"shouldIgnore");async function fe(s){let n={literals:new Set,patterns:[],negations:[]};try{let o=await readFile(join(s,".npmignore"),"utf-8");for(let t of o.split(`
4
+ `)){let e=t.trim();if(!(!e||e.startsWith("#")))if(e.startsWith("!")){let r=e.slice(1);B(r)?n.negations.push((0,P.default)(r,{dot:!0})):n.negations.push((0,P.default)(r,{dot:!0}));}else B(e)?n.patterns.push((0,P.default)(e,{dot:!0})):n.literals.add(e.replace(/\/$/,""));}}catch(o$1){if(o(o$1)&&o$1.code!=="ENOENT")throw o$1}return n}a$1(fe,"loadNpmIgnore");function B(s){return /[*?[\]{}()]/.test(s)}a$1(B,"hasGlobChars");async function X(s,n){let o$1=[];try{let t=await readdir(s,{withFileTypes:!0});for(let e of t){let r=join(s,e.name);if(e.isDirectory()){if(e.name===".git"||s===n&&e.name==="node_modules"||e.isSymbolicLink())continue;o$1.push(...await X(r,n));}else e.isSymbolicLink()||o$1.push(r);}}catch(t){if(o(t)&&t.code==="ENOENT")return [];throw t}return o$1}a$1(X,"collectAllFiles");var we=a$2(Math.max(availableParallelism(),8));async function ze(s$1,n$1={}){let o=join(s$1,"package.json"),t$1;try{t$1=await readFile(o,"utf-8");}catch{throw new Error(`No package.json found in ${s$1}`)}let e=JSON.parse(t$1);if(!e.name)throw new Error("package.json missing 'name' field");if(!e.version)throw new Error("package.json missing 'version' field");if(e.private&&!n$1.allowPrivate)throw new Error(`Package "${e.name}" is private. Use --private flag to publish private packages.`);await S(s$1,e,"preplunk"),n$1.runScripts!==false&&await S(s$1,e,"prepack");let r=s$1;if(e.publishConfig?.directory){r=resolve(s$1,e.publishConfig.directory);try{if(!(await stat(r)).isDirectory())throw new Error(`publishConfig.directory "${e.publishConfig.directory}" is not a directory`)}catch(l){throw l instanceof Error&&"code"in l&&l.code==="ENOENT"?new Error(`publishConfig.directory "${e.publishConfig.directory}" does not exist`):l}d(`[publish] Using publishConfig.directory: ${r}`);}let p$1=r!==s$1?JSON.parse(await readFile(join(r,"package.json"),"utf-8").catch(()=>JSON.stringify(e))):e,c=await Q(r,p$1);if(c.length===0)throw new Error("No publishable files found");d(`[publish] Resolved ${c.length} files for ${e.name}@${e.version}`);let a=await n(c,r);if(await ke(e,s$1),await Ce(e,s$1),!n$1.force){let l=await a$3(e.name,e.version);if(l&&l.contentHash===a)return b.info(`${e.name}@${e.version} already up to date (no changes since last publish)`),{name:e.name,version:e.version,fileCount:c.length,skipped:true,contentHash:a,buildId:l.buildId??""}}let i=f(e.name,e.version),u$1=await a$4(i+".lock",async()=>{if(!n$1.force){let f=await a$3(e.name,e.version);if(f&&f.contentHash===a)return b.info(`${e.name}@${e.version} already up to date (no changes since last publish)`),{name:e.name,version:e.version,fileCount:c.length,skipped:true,contentHash:a,buildId:f.buildId??""}}let l=i+`.tmp-${process.pid}-${Date.now()}`,d$1=join(l,"package"),w$1=a.slice(9,17);try{await v(d$1);let f=ye(e);f=ve(f),d("[publish] Copying files to temp store...");let E=new Set(c.map(v=>dirname(join(d$1,relative(r,v)))));await Promise.all([...E].map(v=>u(v))),await Promise.all(c.map(v=>we(async()=>{let D=relative(r,v),F=join(d$1,D);D==="package.json"&&f!==e?await writeFile(F,JSON.stringify(f,null,2)):await p(v,F,{ensureParent:!1});}))),r!==s$1&&await writeFile(join(d$1,"package.json"),JSON.stringify(f,null,2));let se={schemaVersion:1,contentHash:a,publishedAt:new Date().toISOString(),sourcePath:s$1,buildId:w$1};await writeFile(join(l,".plunk-meta.json"),JSON.stringify(se,null,2));let R=await w(i),_=i+".old-"+Date.now();R&&await rename(i,_),await s(l,i),R&&await t(_),d(`[publish] Stored at ${i}`);}catch(f){throw await t(l),f}return {name:e.name,version:e.version,fileCount:c.length,skipped:false,contentHash:a,buildId:w$1}},{stale:6e4});return u$1.skipped||(n$1.runScripts!==false&&await S(s$1,e,"postpack"),await S(s$1,e,"postplunk"),b.success(`Published ${e.name}@${e.version} (${c.length} files) [${u$1.buildId}]`)),u$1}a$1(ze,"publish");var Z=parseInt(process.env.PLUNK_HOOK_TIMEOUT??"30000",10);async function S(s,n,o){let t=n.scripts?.[o];if(t)return d(`[lifecycle] Running ${o}: ${t}`),new Promise((e,r)=>{let p=platform()==="win32",i=spawn(p?"cmd":"sh",[p?"/c":"-c",t],{cwd:s,stdio:"inherit"}),u=setTimeout(()=>{i.kill("SIGTERM"),r(new Error(`${o} script timed out after ${Z/1e3}s. Increase PLUNK_HOOK_TIMEOUT env var if the script needs more time.`));},Z);i.on("close",l=>{clearTimeout(u),l===0?e():r(new Error(`${o} script failed with exit code ${l}`));}),i.on("error",l=>{clearTimeout(u),r(new Error(`${o} script error: ${l.message}`));});})}a$1(S,"runLifecycleHook");var be=["main","module","exports","types","typings","browser","bin"];function ve(s){if(!s.publishConfig)return s;let n={...s};for(let o of be)o in s.publishConfig&&(n[o]=s.publishConfig[o]);return delete n.publishConfig,n}a$1(ve,"applyPublishConfig");function ye(s){let n=false,o={...s},t=null,e=false;for(let r of ["dependencies","devDependencies","peerDependencies","optionalDependencies"]){let p=s[r];if(!p)continue;let c=false,a={...p};for(let[i,u]of Object.entries(p))if(u.startsWith("workspace:")){let l=u.slice(10);if(l==="*"||l==="^"||l==="~"){let d=$?.versions.get(i)??s.version;a[i]=l==="*"?d:l+d;}else a[i]=l;c=true,n=true;}else if(u.startsWith("catalog:"))if(e||(t=Ee(),e=true),t){let l=Pe(u,i,t);l?(a[i]=l,c=true,n=true):b.warn(`catalog: specifier for "${i}" could not be resolved \u2014 published package.json will contain "${u}" which may cause install failures`);}else d(`[publish] No pnpm-workspace.yaml found, cannot resolve catalog: for "${i}"`);c&&(o[r]=a);}return n?o:s}a$1(ye,"rewriteProtocolVersions");function Pe(s,n,o){let t=s.slice(8);return t===""||t==="default"?o.default[n]??null:o.named[t]?.[n]??null}a$1(Pe,"resolveCatalogVersion");var J=null;async function te(s){if(J?.dir===s)return J.root;let{findWorkspaceRoot:n}=await import('./workspace-4U2WM4MX.mjs'),o=await n(s);return J={dir:s,root:o},o}a$1(te,"getWorkspaceRoot");var $=null;async function ke(s,n){if(!["dependencies","devDependencies","peerDependencies","optionalDependencies"].some(c=>{let a=s[c];return a&&Object.values(a).some(i=>i.startsWith("workspace:"))}))return;let t=await te(n);if(!t){$=null;return}if($?.root===t)return;let{findWorkspacePackages:e}=await import('./workspace-4U2WM4MX.mjs'),r=await e(t),p=new Map;await Promise.all(r.map(async c=>{try{let a=JSON.parse(await readFile(join(c,"package.json"),"utf-8"));a.name&&a.version&&p.set(a.name,a.version);}catch{}})),$={root:t,versions:p};}a$1(ke,"preloadWorkspaceVersions");var k=null;function Ee(){return k?.catalogs??null}a$1(Ee,"loadCatalogsFromCache");async function Ce(s,n){if(!["dependencies","devDependencies","peerDependencies","optionalDependencies"].some(a=>{let i=s[a];return i&&Object.values(i).some(u=>u.startsWith("catalog:"))}))return;let t=await te(n);if(!t){k=null;return}let e=join(t,"pnpm-workspace.yaml"),r=(await stat(e).catch(()=>null))?.mtimeMs??0;if(k?.root===t&&k.mtimeMs===r)return;let{parseCatalogs:p}=await import('./workspace-4U2WM4MX.mjs'),c=await p(t);k={root:t,mtimeMs:r,catalogs:c};}a$1(Ce,"preloadCatalogs");export{ze as a};
@@ -0,0 +1,13 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {c as c$1,a as a$2}from'./chunk-W4LHCWGD.mjs';import {i,j,w,u}from'./chunk-SG6M7SAC.mjs';import {a,b}from'./chunk-ICCM7US5.mjs';import {c,a as a$1}from'./chunk-2VCW5RWI.mjs';import {readFile,writeFile}from'fs/promises';import {join}from'path';globalThis.require=createRequire(import.meta.url);
3
+ var y=c(a(),1);async function x(t){let n="";try{n=await readFile(t,"utf-8");}catch{}if(n.split(`
4
+ `).some(s=>s.trim()===".plunk/"||s.trim()===".plunk"||s.trim()==="/.plunk/"||s.trim()==="/.plunk"))return false;let o=n.length>0&&!n.endsWith(`
5
+ `)?`
6
+ `:"",i=n.length>0?`
7
+ # plunk local links
8
+ .plunk/
9
+ `:`# plunk local links
10
+ .plunk/
11
+ `;return await writeFile(t,n+o+i),true}a$1(x,"ensureGitignore");async function S(t){let n=await readFile(t,"utf-8"),e=JSON.parse(n);if(e.scripts?.postinstall)return e.scripts.postinstall.includes("plunk")||b.warn(`Existing postinstall script found. Add ${y.default.cyan("npx @olegkuibar/plunk restore")} manually if needed.`),false;e.scripts||(e.scripts={}),e.scripts.postinstall="npx @olegkuibar/plunk restore || true";let a=n.match(/^(\s+)"/m)?.[1]||" ";return await writeFile(t,JSON.stringify(e,null,a)+`
12
+ `),true}a$1(S,"addPostinstall");async function v(t){let n;try{n=await readFile(t,"utf-8");}catch{return false}let e=JSON.parse(n);if(!e.scripts?.postinstall?.includes("plunk"))return false;delete e.scripts.postinstall,Object.keys(e.scripts).length===0&&delete e.scripts;let a=n.match(/^(\s+)"/m)?.[1]||" ";return await writeFile(t,JSON.stringify(e,null,a)+`
13
+ `),true}a$1(v,"removePostinstall");async function I(t,n){let e=i(t),a=j(t);if(!await w(a))await u(e),await c$1(t,{version:"1",packageManager:n,role:"consumer",links:{}});else {let s=await a$2(t);s.packageManager||(s.packageManager=n,s.role=s.role??"consumer",await c$1(t,s));}let o=join(t,".gitignore");await x(o);let i$1=join(t,"package.json");await w(i$1)&&await S(i$1);}a$1(I,"ensureConsumerInit");export{x as a,S as b,v as c,I as d};
@@ -0,0 +1,3 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {b as b$1}from'./chunk-5EOIBIQO.mjs';import {d as d$1,c}from'./chunk-WSJHZ2ND.mjs';import {d as d$2}from'./chunk-7ZWOJDCA.mjs';import {e,i,a as a$1,f}from'./chunk-W4LHCWGD.mjs';import {a as a$4}from'./chunk-HPF6K6WO.mjs';import {a as a$2}from'./chunk-MBKCCWSD.mjs';import {b as b$2,a as a$3}from'./chunk-OJJZ7CLB.mjs';import {d}from'./chunk-R3RSOZXN.mjs';import {b}from'./chunk-ICCM7US5.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {basename,resolve}from'path';globalThis.require=createRequire(import.meta.url);
3
+ var W={meta:{name:"remove",description:"Remove a plunk-linked package and restore the original"},args:{package:{type:"positional",description:"Package name to remove",required:false},all:{type:"boolean",description:"Remove all linked packages",default:false},force:{type:"boolean",description:"Skip error checking",default:false},yes:{type:"boolean",alias:"y",description:"Skip confirmation prompts",default:false}},async run({args:i}){b$2();let e=new a$2,n=resolve(".");if(i.all){let u=await a$1(n),a=Object.entries(u.links);if(a.length===0){b.info("No linked packages to remove"),a$3({removed:0});return}if(!i.yes){let l=a.map(([m])=>m).join(", "),d=await b.prompt(`Remove ${a.length} linked package(s)? (${l})`,{type:"confirm"});if(!d||typeof d=="symbol"){b.info("Cancelled");return}}let r=0,s=0;for(let[l,d]of a)try{await M(n,l,d),r++;}catch(m){s++,b.warn(`Failed to remove ${l}: ${m instanceof Error?m.message:String(m)}`);}b.success(`Removed ${r} plunk link(s)${s>0?`, ${s} failed`:""} in ${e.elapsed()}`),a$3({removed:r,failed:s,elapsed:e.elapsedMs()});return}let o=i.package;o||(a$4("Package name required. Use --all to remove all linked packages."),process.exit(1));let p=await f(n,o);if(!p){if(i.force){b.warn(`Package "${o}" is not linked, skipping`),a$3({removed:0});return}a$4(`Package "${o}" is not linked in this project`),process.exit(1);}await M(n,o,p),b.success(`Removed plunk link for ${o} in ${e.elapsed()}`),a$3({removed:1,package:o,elapsed:e.elapsedMs()});}};async function M(i$1,e$1,n){if(d(`[remove] Removing ${e$1}`),await d$1(i$1,e$1,n.packageManager),n.backupExists)try{await c(i$1,e$1,n.packageManager)&&b.success(`Restored original ${e$1} from backup`);}catch(r){b.warn(`Failed to restore backup for ${e$1}: ${r instanceof Error?r.message:String(r)}. Run your package manager's install command to restore it.`);}let o=await d$2(i$1);o.type==="next"&&o.configFile&&(await b$1(o.configFile,e$1)).modified&&d(`[remove] Removed ${e$1} from ${basename(o.configFile)}`);let{findTailwindCss:p,removeTailwindSource:u}=await import('./tailwind-source-C2ANIANJ.mjs'),a=await p(i$1);if(a&&(await u(a,e$1)).modified&&d(`[remove] Removed @source for ${e$1} from ${basename(a)}`),await e(i$1,e$1),await i(e$1,i$1),o.type==="vite"&&o.configFile){let r=await a$1(i$1);if(Object.keys(r.links).length===0){let{removeFromViteConfig:s}=await import('./vite-config-JEUUATXJ.mjs');(await s(o.configFile)).modified&&d(`[remove] Removed plunk plugin from ${basename(o.configFile)}`);}}}a(M,"removeSinglePackage");export{W as a,M as b};
@@ -0,0 +1,4 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {d,b}from'./chunk-R3RSOZXN.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {stat,readFile,mkdir,copyFile,constants,readdir,utimes,rm,rename,cp,writeFile}from'fs/promises';import {join,relative,parse,dirname}from'path';import {availableParallelism,homedir}from'os';import {createHash}from'crypto';globalThis.require=createRequire(import.meta.url);
3
+ function v(t){let e=0,n=[],i=0,a$1=a(()=>{i<n.length&&e<t&&(e++,n[i++]()),i>0&&i===n.length&&(n.length=0,i=0);},"next");return h=>new Promise((u,l)=>{let d=a(()=>{h().then(s=>{e--,u(s),a$1();},s=>{e--,l(s),a$1();});},"run");e<t?(e++,d()):n.push(d);})}a(v,"pLimit");function X(){return process.env.PLUNK_HOME||join(homedir(),".plunk")}a(X,"getPlunkHome");function G(){return join(X(),"store")}a(G,"getStorePath");function dt(){return join(X(),"consumers.json")}a(dt,"getConsumersPath");function _(t){return t.replace(/\//g,"+")}a(_,"encodePackageName");function wt(t){if(t.startsWith("@")){let e=t.indexOf("+");if(e!==-1)return t.slice(0,e)+"/"+t.slice(e+1)}return t}a(wt,"decodePackageName");function A(t,e){return join(G(),`${_(t)}@${e}`)}a(A,"getStoreEntryPath");function Pt(t,e){return join(A(t,e),"package")}a(Pt,"getStorePackagePath");function yt(t,e){return join(A(t,e),".plunk-meta.json")}a(yt,"getStoreMetaPath");function xt(t){return join(t,".plunk")}a(xt,"getConsumerPlunkDir");function vt(t){return join(t,".plunk","state.json")}a(vt,"getConsumerStatePath");function kt(t,e){return join(t,".plunk","backups",_(e))}a(kt,"getConsumerBackupPath");function E(t){return t.replace(/\\/g,"/")}a(E,"normalizePath");function $t(t,e){return join(t,"node_modules",e)}a($t,"getNodeModulesPackagePath");var z=null;function tt(){return z||(z=import('./xxhash-wasm-DTW44IIQ.mjs').then(t=>t.default()).catch(t=>{throw z=null,t})),z}a(tt,"getXXHash");var et=1024*1024,rt=v(Math.max(availableParallelism(),8)),S=new Map;async function Nt(t,e){let n=[...t].sort((s,f)=>{let w=E(relative(e,s)),g=E(relative(e,f));return w.localeCompare(g)}),i=new Set(n),a=0,h=await Promise.all(n.map(s=>rt(async()=>{let f=E(relative(e,s)),w=await stat(s),g=S.get(s);if(g&&g.mtimeMs===w.mtimeMs&&g.size===w.size)return a++,{rel:f,content:g.content};let o=await readFile(s);return S.set(s,{mtimeMs:w.mtimeMs,size:w.size,content:o}),{rel:f,content:o}})));for(let s of S.keys())i.has(s)||S.delete(s);d(`[hash] Computing content hash for ${t.length} files (${a} cached)`);let u=createHash("sha256"),l=Buffer.alloc(4);for(let{rel:s,content:f}of h)u.update(s),u.update("\0"),l.writeUInt32LE(f.length),u.update(l),u.update(f);let d$1="sha256v2:"+u.digest("hex");return d(`[hash] Result: ${d$1.slice(0,20)}...`),d$1}a(Nt,"computeContentHash");async function I(t,e){let n=e??(await stat(t)).size,i=await tt();if(n>et)return nt(t,i);let a=await readFile(t);return i.h64Raw(a).toString(16)}a(I,"hashFile");async function nt(t,e){let{createReadStream:n}=await import('fs'),i=e.create64();return new Promise((a,h)=>{let u=n(t);u.on("data",l=>i.update(l)),u.on("end",()=>a(i.digest().toString(16))),u.on("error",h);})}a(nt,"hashFileStream");var W=v(Math.max(availableParallelism(),8));function O(t){return t instanceof Error&&"code"in t}a(O,"isNodeError");var N=new Map;function mt(t){let{root:e}=parse(t);return e||"/"}a(mt,"volumeRoot");async function pt(t,e,n){if(b()){d(`[dry-run] would copy ${t} \u2192 ${e}`);return}n?.ensureParent!==false&&await mkdir(dirname(e),{recursive:true});let i=mt(e),a=N.get(i);if(a===false){await copyFile(t,e);return}if(a===true){await copyFile(t,e,constants.COPYFILE_FICLONE);return}try{await copyFile(t,e,constants.COPYFILE_FICLONE_FORCE),N.set(i,!0);}catch{N.set(i,false),d(`[copy] reflink not supported on ${i}, using plain copy`),await copyFile(t,e);}}a(pt,"copyWithCoW");async function R(t){return (await readdir(t,{recursive:true,withFileTypes:true})).filter(n=>n.isFile()).map(n=>join(n.parentPath,n.name))}a(R,"collectFiles");async function Dt(t,e,n={}){let i=R(t),a=R(e).catch(o=>{if(O(o)&&o.code==="ENOENT")return [];throw o}),[h,u]=await Promise.all([i,a]),l=0,d$1=0,s=0,f=await Promise.all(h.map(o=>W(async()=>{let y=relative(t,o),$=join(e,y),M=true,x=null;if(n.force)d(`[copy] ${y} (forced)`);else try{let[m,C]=await Promise.all([stat(o),stat($)]);if(x={atime:m.atime,mtime:m.mtime},m.size!==C.size)d(`[copy] ${y} (size differs: ${m.size} vs ${C.size})`);else if(m.mtimeMs===C.mtimeMs)M=!1,d(`[skip] ${y} (size+mtime match)`);else {let[J,K]=await Promise.all([I(o,m.size),I($,C.size)]);J===K?(M=!1,d(`[skip] ${y} (unchanged)`)):d(`[copy] ${y} (hash differs)`);}}catch(m){if(O(m)&&m.code==="ENOENT")d(`[copy] ${y} (new file)`);else throw m}if(M){if(await pt(o,$),!x){let m=await stat(o);x={atime:m.atime,mtime:m.mtime};}return await utimes($,x.atime,x.mtime),"copied"}return "skipped"})));for(let o of f)o==="copied"?l++:s++;let w=new Set(h.map(o=>relative(t,o))),g=u.filter(o=>!w.has(relative(e,o)));return await Promise.all(g.map(o=>W(async()=>{d(`[remove] ${relative(e,o)} (no longer in source)`),b()||await rm(o);}))),d$1=g.length,{copied:l,removed:d$1,skipped:s}}a(Dt,"incrementalCopy");async function Wt(t,e){if(b()){d(`[dry-run] would move ${t} \u2192 ${e}`);return}try{await rename(t,e);}catch(n){if(O(n)&&n.code==="EXDEV")await cp(t,e,{recursive:true}),await rm(t,{recursive:true,force:true});else throw n}}a(Wt,"moveDir");async function Ut(t){if(b()){d(`[dry-run] would remove ${t}`);return}await rm(t,{recursive:true,force:true});}a(Ut,"removeDir");async function Yt(t){if(b()){d(`[dry-run] would ensure dir ${t}`);return}await mkdir(t,{recursive:true});}a(Yt,"ensureDir");async function qt(t){if(b()){d(`[dry-run] would ensure private dir ${t}`);return}await mkdir(t,{recursive:true,mode:448});}a(qt,"ensurePrivateDir");async function Jt(t){try{return await stat(t),!0}catch{return false}}a(Jt,"exists");async function Kt(t,e){if(b()){d(`[dry-run] would write ${t}`);return}let n=t+`.tmp-${process.pid}-${Date.now()}`;await writeFile(n,e),await rename(n,t);}a(Kt,"atomicWriteFile");async function Vt(t){try{let e=await R(t);return (await Promise.all(e.map(i=>stat(i).then(a=>a.size).catch(()=>0)))).reduce((i,a)=>i+a,0)}catch{return 0}}a(Vt,"dirSize");async function Gt(t,e){await cp(t,e,{recursive:true});}a(Gt,"copyDir");
4
+ export{v as a,G as b,dt as c,_ as d,wt as e,A as f,Pt as g,yt as h,xt as i,vt as j,kt as k,E as l,$t as m,Nt as n,O as o,pt as p,R as q,Dt as r,Wt as s,Ut as t,Yt as u,qt as v,Jt as w,Kt as x,Vt as y,Gt as z};
@@ -0,0 +1,3 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {a as a$6}from'./chunk-WSJHZ2ND.mjs';import {a as a$8}from'./chunk-7JG555TZ.mjs';import {a as a$7}from'./chunk-7ZWOJDCA.mjs';import {j,f,d as d$1}from'./chunk-W4LHCWGD.mjs';import {a as a$3}from'./chunk-HN3SZGD3.mjs';import {a as a$4}from'./chunk-HPF6K6WO.mjs';import {b as b$1}from'./chunk-ETKOGSRN.mjs';import {a as a$2}from'./chunk-MBKCCWSD.mjs';import {a as a$5}from'./chunk-OJJZ7CLB.mjs';import {a as a$1}from'./chunk-SG6M7SAC.mjs';import {d}from'./chunk-R3RSOZXN.mjs';import {b}from'./chunk-ICCM7US5.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {readFile}from'fs/promises';import {join}from'path';globalThis.require=createRequire(import.meta.url);
3
+ async function x(i){try{let t=await readFile(join(i,"package.json"),"utf-8"),r=JSON.parse(t);if(!r.plunk||typeof r.plunk!="object")return {};let e={},n=r.plunk;return typeof n.buildCmd=="string"&&(e.buildCmd=n.buildCmd),Array.isArray(n.watchPatterns)&&(e.watchPatterns=n.watchPatterns.filter(o=>typeof o=="string")),typeof n.debounce=="number"&&Number.isFinite(n.debounce)&&(e.debounce=n.debounce),typeof n.cooldown=="number"&&Number.isFinite(n.cooldown)&&(e.cooldown=n.cooldown),d(`[config] Loaded plunk config from package.json: ${JSON.stringify(e)}`),e}catch{return {}}}a(x,"loadPlunkConfig");var L=a$1(4);async function ie(i,t={}){let r=new a$2,e=await a$3(i,{runScripts:t.runScripts,force:t.force});if(e.skipped){b.info("No changes to push");return}let n=await b$1(e.name,e.version);if(!n){a$4(`Failed to read store entry for ${e.name}@${e.version} after publish`);return}let o=await j(e.name);if(o.length===0){b.success(`Published ${e.name}@${e.version} to store`),b.info("No consumers registered yet. Run 'plunk add "+e.name+"' in a consumer project to start receiving pushes."),a$5({name:e.name,version:e.version,buildId:e.buildId,consumers:0,failedConsumers:0,copied:0,skipped:0,elapsed:r.elapsedMs()});return}let c=0,u=0,s=0,f$1=0,A=await Promise.all(o.map(d$2=>L(async()=>{let g=await f(d$2,e.name);if(!g)return d(`[push] No link found for ${e.name} in ${d$2}, skipping`),null;try{let p=await a$6(n,d$2,g.packageManager,{force:t.force});return await d$1(d$2,e.name,{...g,contentHash:n.meta.contentHash,linkedAt:new Date().toISOString(),buildId:n.meta.buildId??""}),p}catch(p){return b.warn(`Failed to push to ${d$2}: ${p instanceof Error?p.message:String(p)}`),null}})));for(let d of A)d?(c+=d.copied,u+=d.skipped,s++):f$1++;let F=e.buildId?` [${e.buildId}]`:"";b.success(`Pushed ${e.name}@${e.version}${F} to ${s} consumer(s) in ${r.elapsed()} (${c} files changed, ${u} unchanged)`),a$5({name:e.name,version:e.version,buildId:e.buildId,consumers:s,failedConsumers:f$1,copied:c,skipped:u,elapsed:r.elapsedMs()});}a(ie,"doPush");function W(i){if(!i)return;let t=parseInt(i,10);return Number.isFinite(t)?t:void 0}a(W,"parseMs");async function re(i,t,r){let{startWatcher:e}=await import('./watcher-KAPGRKYP.mjs'),n=await x(i),{buildCmd:o,patterns:c}=await E(i,t,n),u=await e(i,{patterns:c,buildCmd:o,debounce:W(t.debounce)??n.debounce,cooldown:W(t.cooldown)??n.cooldown},r);await new Promise(s=>{let f=a(async()=>{b.info("Stopping watcher..."),await u.close(),s();},"cleanup");process.once("SIGINT",f),process.once("SIGTERM",f);});}a(re,"startWatchMode");async function E(i,t,r){let e=t.build,n=r?.watchPatterns;if(!t.build){if(!t["skip-build"])if(r?.buildCmd)e=r.buildCmd,b.info(`Using build command from config: ${e}`);else {let o=await a$7(i),c=await a$8(i,o);c&&(e=c,b.info(`Auto-detected build command: ${c}`));}}if(e){let{exists:o}=await import('./fs-6ZUVEONC.mjs'),c=["src","lib","source","app","pages","components"],u=(await Promise.all(c.map(async s=>({dir:s,exists:await o(join(i,s))})))).filter(s=>s.exists).map(s=>s.dir);n=u.length>0?u:["src","lib"],d(`[watch] Using source patterns with build command: ${n.join(", ")}`);}else {b.info("No build command detected \u2014 watching output directories directly");try{let o=JSON.parse(await readFile(join(i,"package.json"),"utf-8"));o.files&&o.files.length>0?(n=o.files,b.info(`Watching from package.json "files": ${n.join(", ")}`)):b.warn('No "files" field in package.json \u2014 falling back to watching src/ and lib/. Add a "files" field or use --build to specify a build command.');}catch(o){d(`[watch] Could not read package.json: ${o instanceof Error?o.message:String(o)}`);}}return {buildCmd:e,patterns:n}}a(E,"resolveWatchConfig");export{ie as a,re as b};
@@ -0,0 +1,3 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {c,a as a$1,e}from'./chunk-GAS7XU4A.mjs';import {j,o,u,i,x,c as c$1,v,l,w}from'./chunk-SG6M7SAC.mjs';import {b}from'./chunk-ICCM7US5.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {readFile}from'fs/promises';import {dirname}from'path';globalThis.require=createRequire(import.meta.url);
3
+ async function R(t){let{state:e}=await p(t);return e}a(R,"readConsumerState");async function p(t){let e=j(t);try{let r=await readFile(e,"utf-8"),s=JSON.parse(r);return c(s)?{state:s,reliable:!0}:(b.warn(`Invalid consumer state in ${e}, using defaults`),{state:{version:"1",links:{}},reliable:!1})}catch(r){return o(r)&&r.code==="ENOENT"?{state:{version:"1",links:{}},reliable:true}:(b.warn(`Failed to read consumer state: ${r instanceof Error?r.message:String(r)}`),{state:{version:"1",links:{}},reliable:false})}}a(p,"readConsumerStateSafe");async function L(t,e){await u(i(t));let r=j(t);await x(r,JSON.stringify(e,null,2));}a(L,"writeConsumerState");async function T(t,e,r){let s=j(t);await a$1(s,async()=>{let{state:i,reliable:c}=await p(t);if(!c)throw new Error(`Consumer state in ${s} is corrupt \u2014 refusing to write to avoid destroying existing links. Delete .plunk/state.json and re-run 'plunk add' for each package.`);i.links[e]=r,await L(t,i);});}a(T,"addLink");async function M(t,e){let r=j(t);await a$1(r,async()=>{let{state:s,reliable:i}=await p(t);if(!i)throw new Error(`Consumer state in ${r} is corrupt \u2014 refusing to write to avoid destroying existing links. Delete .plunk/state.json and re-run 'plunk add' for each package.`);delete s.links[e],await L(t,s);});}a(M,"removeLink");async function W(t,e){return (await R(t)).links[e]??null}a(W,"getLink");async function g(){let t=c$1();try{let e$1=await readFile(t,"utf-8"),r=JSON.parse(e$1);return e(r)?r:(b.warn("Invalid consumers registry, using empty registry"),{})}catch(e){return o(e)&&e.code!=="ENOENT"&&b.warn(`Failed to read consumers registry: ${e instanceof Error?e.message:String(e)}`),{}}}a(g,"readConsumersRegistry");async function C(t){let e=c$1();await v(dirname(c$1())),await x(e,JSON.stringify(t,null,2));}a(C,"writeConsumersRegistry");async function q(t,e){let r=c$1();await a$1(r,async()=>{let s=await g();s[t]||(s[t]=[]);let i=l(e);s[t].includes(i)||s[t].push(i),await C(s);});}a(q,"registerConsumer");async function A(t,e){let r=c$1();await a$1(r,async()=>{let s=await g();if(!s[t])return;let i=l(e);s[t]=s[t].filter(c=>c!==i),s[t].length===0&&delete s[t],await C(s);});}a(A,"unregisterConsumer");async function B(t){return (await g())[t]??[]}a(B,"getConsumers");async function G(){let t=c$1(),e=0,r=0;return await a$1(t,async()=>{let s=await g(),i={};for(let[c,P]of Object.entries(s)){let d=(await Promise.all(P.map(async a=>({consumerPath:a,valid:await w(a)})))).filter(a=>a.valid).map(a=>a.consumerPath);e+=P.length-d.length,d.length>0?i[c]=d:r++;}await C(i);}),{removedConsumers:e,removedPackages:r}}a(G,"cleanStaleConsumers");export{R as a,p as b,L as c,T as d,M as e,W as f,g,q as h,A as i,B as j,G as k};
@@ -0,0 +1,19 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {e,b as b$2}from'./chunk-7ZWOJDCA.mjs';import {l,o,w,t,u,r,k,z,m}from'./chunk-SG6M7SAC.mjs';import {b,d}from'./chunk-R3RSOZXN.mjs';import {b as b$1}from'./chunk-ICCM7US5.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {mkdir,writeFile,rm,symlink,chmod,readdir,stat,realpath,readFile}from'fs/promises';import {join,resolve,sep,relative}from'path';import {platform}from'os';globalThis.require=createRequire(import.meta.url);
3
+ function A(e){return e.bin?typeof e.bin=="string"?{[e.name.startsWith("@")?e.name.split("/")[1]:e.name]:e.bin}:e.bin:{}}a(A,"resolveBinEntries");async function W(e,n,i){let r=A(i);if(Object.keys(r).length===0)return 0;if(b())return d(`[dry-run] would create ${Object.keys(r).length} bin link(s) for ${n}`),Object.keys(r).length;let t=join(e,"node_modules",".bin");await mkdir(t,{recursive:true});let d$1=platform()==="win32",a=0;for(let[o$1,g]of Object.entries(r)){let p=join(e,"node_modules",n),l$1=join(p,g),R=resolve(l$1);if(!R.startsWith(resolve(p)+sep)&&R!==resolve(p)){b$1.warn(`bin "${o$1}" points outside package directory, skipping`);continue}let $=l(relative(t,l$1));if(d$1){let w=join(t,`${o$1}.cmd`),x=`@ECHO off\r
4
+ GOTO start\r
5
+ :find_dp0\r
6
+ SET dp0=%~dp0\r
7
+ EXIT /b\r
8
+ :start\r
9
+ CALL :find_dp0\r
10
+ "%dp0%\\${$.replace(/\//g,"\\")}" %*\r
11
+ `;await writeFile(w,x);let G=join(t,`${o$1}.ps1`),Y=`#!/usr/bin/env pwsh
12
+ $basedir=Split-Path $MyInvocation.MyCommand.Definition -Parent
13
+ & node "$basedir/${$}" $args
14
+ exit $LASTEXITCODE
15
+ `;await writeFile(G,Y);let q=join(t,o$1),K=`#!/bin/sh
16
+ exec node "${$}" "$@"
17
+ `;await writeFile(q,K);}else {let w=join(t,o$1);try{await rm(w,{force:!0});}catch{}try{await symlink($,w),await chmod(l$1,493);}catch(b){if(o(b)&&(b.code==="EPERM"||b.code==="EACCES")){d(`[bin-linker] Symlink failed (${b.code}), using shell wrapper for ${o$1}`);let x=`#!/bin/sh
18
+ exec node "${$}" "$@"
19
+ `;await writeFile(w,x),await chmod(w,493);}else throw b}}a++;}return a}a(W,"createBinLinks");async function F(e,n){let i=A(n);if(b()){d(`[dry-run] would remove ${Object.keys(i).length} bin link(s)`);return}let r=join(e,"node_modules",".bin"),t=platform()==="win32";for(let d of Object.keys(i))try{await rm(join(r,d),{force:!0}),t&&(await rm(join(r,`${d}.cmd`),{force:!0}),await rm(join(r,`${d}.ps1`),{force:!0}));}catch{}}a(F,"removeBinLinks");var ne={next:[".next/cache"],webpack:["node_modules/.cache"]},H=new Map;async function X(e$1){if(b()){d(`[dry-run] would invalidate bundler caches for ${e$1}`);return}let n=H.get(e$1);n||(n=await e(e$1),H.set(e$1,n));for(let i of n){if(!i.type)continue;let r=ne[i.type];if(r)for(let t$1 of r){let d$1=join(e$1,t$1);if(await w(d$1))try{await t(d$1),d(`[inject] Invalidated ${i.type} cache: ${t$1}`);}catch{d(`[inject] Could not clear ${i.type} cache: ${t$1} (locked?)`);}}}}a(X,"invalidateBundlerCache");async function Se(e,n,i,r$1={}){let t=await P(n,e.name,i,e.version);d(`[inject] ${e.name}@${e.version} \u2192 ${t}`),await u(t);let{copied:d$1,removed:a,skipped:o}=await r(e.packageDir,t,{force:r$1.force});d(`[inject] ${d$1} copied, ${a} removed, ${o} skipped`),(d$1>0||a>0)&&await X(n);let g=await B(e.packageDir),p=g?await W(n,e.name,g):0;return p>0&&d(`[inject] Created ${p} bin link(s)`),{copied:d$1,removed:a,skipped:o,binLinks:p}}a(Se,"inject");async function Ie(e,n,i){let r=await P(e,n,i);if(!await w(r))return false;let t$1=k(e,n);return await t(t$1),await z(r,t$1),true}a(Ie,"backupExisting");async function Me(e,n,i){let r=k(e,n);if(!await w(r))return false;let t$1=await P(e,n,i);return await t(t$1),await z(r,t$1),await t(r),true}a(Me,"restoreBackup");async function Te(e,n,i){let r=await P(e,n,i),t$1=await B(r);t$1&&await F(e,t$1),await t(r);}a(Te,"removeInjected");async function _e(e,n){let i=await B(e.packageDir);if(!i)return [];let r={...i.dependencies,...Object.fromEntries(Object.entries(i.peerDependencies??{}).filter(([a])=>!i.peerDependenciesMeta?.[a]?.optional))};if(Object.keys(r).length===0)return [];let t=Object.keys(r);return (await Promise.all(t.map(async a=>({dep:a,installed:await w(join(n,"node_modules",a))})))).filter(a=>!a.installed).map(a=>a.dep)}a(_e,"checkMissingDeps");async function P(e,n,i,r){let t=m(e,n);if(!(i==="pnpm"||i==="yarn"&&await b$2(e)==="pnpm"))return t;try{let o=await ae(t);if(o!==resolve(t))return d(`[inject] pnpm: resolved symlink \u2192 ${o}`),o}catch(o$1){o(o$1)&&o$1.code!=="ENOENT"&&b$1.debug(`pnpm symlink resolution error: ${o$1 instanceof Error?o$1.message:String(o$1)}`);}let a=join(e,"node_modules",".pnpm");if(await w(a)){d(`[inject] pnpm: scanning .pnpm/ for ${n}`);let o=n.replaceAll("/","+");if(r){let p=`${o}@${r}`,l=join(a,p,"node_modules",n);if(await w(l))return d(`[inject] pnpm: exact version match in .pnpm/ \u2192 ${l}`),l}let g=await readdir(a);for(let p of g)if(p.startsWith(o+"@")){let l=join(a,p,"node_modules",n);if(await w(l))return d(`[inject] pnpm: found in .pnpm/ \u2192 ${l}`),l}}return b$1.warn(`pnpm: Could not find ${n} in .pnpm/ virtual store, using direct node_modules path. If this causes issues, run 'pnpm install' to rebuild the virtual store, then 'plunk add' again.`),t}a(P,"resolveTargetDir");async function ae(e){try{return await stat(e),await realpath(e)}catch(n){if(o(n)&&n.code==="ENOENT")return resolve(e);throw n}}a(ae,"resolveRealPath");async function B(e){try{let n=await readFile(join(e,"package.json"),"utf-8");return JSON.parse(n)}catch(n){return o(n)&&n.code!=="ENOENT"&&b$1.warn(`Failed to read package.json in ${e}: ${n instanceof Error?n.message:String(n)}`),null}}a(B,"readPackageJson");export{Se as a,Ie as b,Me as c,Te as d,_e as e};
@@ -0,0 +1,3 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {a as a$1}from'./chunk-2GDRDQA5.mjs';import {k,g,b as b$2}from'./chunk-W4LHCWGD.mjs';import {d as d$1,e}from'./chunk-ETKOGSRN.mjs';import {a}from'./chunk-MBKCCWSD.mjs';import'./chunk-GAS7XU4A.mjs';import {b,a as a$2}from'./chunk-OJJZ7CLB.mjs';import'./chunk-OL7SNXMX.mjs';import {l,b as b$3,w,y,t}from'./chunk-SG6M7SAC.mjs';import {d}from'./chunk-R3RSOZXN.mjs';import {b as b$1}from'./chunk-ICCM7US5.mjs';import'./chunk-2VCW5RWI.mjs';import {readdir}from'fs/promises';import {join}from'path';globalThis.require=createRequire(import.meta.url);
3
+ var ee={meta:{name:"clean",description:"Remove unreferenced store entries and stale consumer registrations"},args:{yes:{type:"boolean",alias:"y",description:"Skip confirmation prompts",default:false}},async run({args:M}){b();let y$1=new a;b$1.start("Checking consumer registrations...");let{removedConsumers:l$1,removedPackages:v}=await k();l$1>0&&b$1.success(`Removed ${l$1} stale consumer registration(s) across ${v} package(s)`);let $=await g(),f=new Set,N=[...new Set(Object.values($).flat())],w$1=await Promise.all(N.map(async e=>{let{state:t,reliable:o}=await b$2(e);return {path:e,state:t,reliable:o}}));for(let{state:e,reliable:t}of w$1)if(t)for(let[o,g]of Object.entries(e.links))f.add(`${o}@${g.version}`);let d$2=new Set,u=w$1.filter(e=>!e.reliable);if(u.length>0){b$1.warn(`${u.length} consumer(s) have corrupt state \u2014 their store entries will be preserved`);let e=new Set(u.map(t=>l(t.path)));for(let[t,o]of Object.entries($))o.some(g=>e.has(g))&&d$2.add(t);d(`[clean] Protected packages (corrupt state): ${[...d$2].join(", ")}`);}d(`[clean] Referenced entries: ${[...f].join(", ")||"(none)"}`);let A=await d$1(),s=[];for(let e of A){let t=`${e.name}@${e.version}`;if(!f.has(t)){if(d$2.has(e.name)){d(`[clean] Preserving ${t} (consumer state unreadable)`);continue}let o=Date.now()-new Date(e.meta.publishedAt).getTime();if(o<300*1e3){d(`[clean] Skipping recently published entry: ${t} (${Math.round(o/1e3)}s old)`);continue}s.push(e);}}let n=[],p=b$3();if(await w(p)){let e=await readdir(p,{withFileTypes:true});for(let t of e)t.isDirectory()&&(t.name.includes(".tmp-")||t.name.includes(".old-"))&&n.push(t.name);}if(s.length+n.length>0&&!M.yes){let e=[];s.length>0&&e.push(`${s.length} unreferenced store entry(ies)`),n.length>0&&e.push(`${n.length} orphaned temp directory(ies)`);let t=await b$1.prompt(`Remove ${e.join(" and ")}?`,{type:"confirm"});if(!t||typeof t=="symbol"){b$1.info("Cancelled");return}}let a$3=0,F=await Promise.all(s.map(e=>y(e.packageDir))),H=await Promise.all(n.map(e=>y(join(p,e)))),c=0;for(let e$1=0;e$1<s.length;e$1++){let t=s[e$1];d(`[clean] Removing unreferenced store entry: ${t.name}@${t.version}`),await e(t.name,t.version),a$3+=F[e$1],c++;}c>0&&b$1.success(`Removed ${c} unreferenced store entry(ies)`);let m=0;for(let e=0;e<n.length;e++)d(`[clean] Removing orphaned directory: ${n[e]}`),await t(join(p,n[e])),a$3+=H[e],m++;m>0&&b$1.success(`Removed ${m} orphaned temp directory(ies)`),l$1===0&&c===0&&m===0&&b$1.info("Store is clean \u2014 no stale entries or registrations found");let q=a$3>0?` (reclaimed ${a$1(a$3)})`:"";b$1.info(`Clean complete in ${y$1.elapsed()}${q}`),a$2({removedConsumers:l$1,removedPackages:v,removedEntries:c,removedOrphans:m,reclaimedBytes:a$3,elapsed:y$1.elapsedMs()});}};export{ee as default};
package/dist/cli.mjs CHANGED
@@ -8,4 +8,4 @@ var o=c(a$1(),1);var r=`
8
8
  \u2588\u2588\u2551 \u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2557\u255A\u2588\u2588\u2588\u2588\u2588\u2588\u2554\u255D\u2588\u2588\u2551 \u255A\u2588\u2588\u2588\u2588\u2551\u2588\u2588\u2551 \u2588\u2588\u2557
9
9
  \u255A\u2550\u255D \u255A\u2550\u2550\u2550\u2550\u2550\u2550\u255D \u255A\u2550\u2550\u2550\u2550\u2550\u255D \u255A\u2550\u255D \u255A\u2550\u2550\u2550\u255D\u255A\u2550\u255D \u255A\u2550\u255D
10
10
  `;function a(){console.log(o.default.yellow(r)),console.log(o.default.cyan(" \u{1F4E6} Local npm package development without symlinks")),console.log(o.default.dim(` Copies built files into consumer node_modules with incremental sync
11
- `));}a$2(a,"showBanner");process.env.UV_THREADPOOL_SIZE??=String(Math.max(availableParallelism(),8));a$3();var d=process.argv.slice(2),p=d.some(e=>!e.startsWith("-")&&["init","publish","add","remove","push","dev","restore","list","status","update","clean","gc","doctor","migrate"].includes(e));p||a();var u={meta:{name:"plunk",version:"0.7.6",description:"Local npm package development \u2014 copies built files into consumer node_modules"},args:{verbose:{type:"boolean",alias:"v",description:"Enable verbose debug logging",default:false},"dry-run":{type:"boolean",description:"Preview changes without writing files",default:false},json:{type:"boolean",description:"Output machine-readable JSON",default:false}},subCommands:{init:a$2(()=>import('./init-C6XCSFCU.mjs').then(e=>e.default),"init"),publish:a$2(()=>import('./publish-6A7PX5IH.mjs').then(e=>e.default),"publish"),add:a$2(()=>import('./add-GYBX4VAZ.mjs').then(e=>e.default),"add"),remove:a$2(()=>import('./remove-5DAQD627.mjs').then(e=>e.default),"remove"),push:a$2(()=>import('./push-AUAGCBYK.mjs').then(e=>e.default),"push"),dev:a$2(()=>import('./dev-DBGYZOIV.mjs').then(e=>e.default),"dev"),restore:a$2(()=>import('./restore-JVH6INAG.mjs').then(e=>e.default),"restore"),list:a$2(()=>import('./list-QSPN7FE5.mjs').then(e=>e.default),"list"),status:a$2(()=>import('./status-22YV26A3.mjs').then(e=>e.default),"status"),update:a$2(()=>import('./update-33ICRFYZ.mjs').then(e=>e.default),"update"),clean:a$2(()=>import('./clean-F2IWAVRP.mjs').then(e=>e.default),"clean"),gc:a$2(()=>import('./clean-F2IWAVRP.mjs').then(e=>e.default),"gc"),doctor:a$2(()=>import('./doctor-GJGAAT6J.mjs').then(e=>e.default),"doctor"),migrate:a$2(()=>import('./migrate-4TFDXO4G.mjs').then(e=>e.default),"migrate")}};a$4(u);
11
+ `));}a$2(a,"showBanner");process.env.UV_THREADPOOL_SIZE??=String(Math.max(availableParallelism(),8));a$3();var d=process.argv.slice(2),p=d.some(e=>!e.startsWith("-")&&["init","publish","add","remove","push","dev","restore","list","status","update","clean","gc","doctor","migrate","reset"].includes(e));p||a();var u={meta:{name:"plunk",version:"0.8.0",description:"Local npm package development \u2014 copies built files into consumer node_modules"},args:{verbose:{type:"boolean",alias:"v",description:"Enable verbose debug logging",default:false},"dry-run":{type:"boolean",description:"Preview changes without writing files",default:false},json:{type:"boolean",description:"Output machine-readable JSON",default:false}},subCommands:{init:a$2(()=>import('./init-G6X5O3A3.mjs').then(e=>e.default),"init"),publish:a$2(()=>import('./publish-DGEQG3L5.mjs').then(e=>e.default),"publish"),add:a$2(()=>import('./add-S5F3UPAP.mjs').then(e=>e.default),"add"),remove:a$2(()=>import('./remove-GPLRNTDM.mjs').then(e=>e.default),"remove"),push:a$2(()=>import('./push-VS3VZKU4.mjs').then(e=>e.default),"push"),dev:a$2(()=>import('./dev-YR4ZYYT4.mjs').then(e=>e.default),"dev"),restore:a$2(()=>import('./restore-WIYNZSQL.mjs').then(e=>e.default),"restore"),list:a$2(()=>import('./list-N2B4TO6R.mjs').then(e=>e.default),"list"),status:a$2(()=>import('./status-ADDZJ66C.mjs').then(e=>e.default),"status"),update:a$2(()=>import('./update-2Q46L6NR.mjs').then(e=>e.default),"update"),clean:a$2(()=>import('./clean-STIN6OS6.mjs').then(e=>e.default),"clean"),gc:a$2(()=>import('./clean-STIN6OS6.mjs').then(e=>e.default),"gc"),doctor:a$2(()=>import('./doctor-DAEAGACF.mjs').then(e=>e.default),"doctor"),migrate:a$2(()=>import('./migrate-YZRZUG54.mjs').then(e=>e.default),"migrate"),reset:a$2(()=>import('./reset-E5P4ZKQ7.mjs').then(e=>e.default),"reset")}};a$4(u);
@@ -1,3 +1,3 @@
1
1
  #!/usr/bin/env node
2
- import {createRequire}from'node:module';import {a as a$1,b as b$1}from'./chunk-CIY4E6PA.mjs';import'./chunk-SMIGYQFG.mjs';import'./chunk-7JG555TZ.mjs';import'./chunk-AC5FETT7.mjs';import'./chunk-LLVBXPQN.mjs';import'./chunk-PVMVWPLG.mjs';import'./chunk-HPF6K6WO.mjs';import'./chunk-EMRPZYLU.mjs';import'./chunk-MBKCCWSD.mjs';import'./chunk-22YCXJTS.mjs';import {b}from'./chunk-OJJZ7CLB.mjs';import'./chunk-OL7SNXMX.mjs';import'./chunk-HW7AEGI3.mjs';import'./chunk-TMH7HIJ2.mjs';import'./chunk-R3RSOZXN.mjs';import'./chunk-ICCM7US5.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {resolve}from'path';globalThis.require=createRequire(import.meta.url);
3
- var u={meta:{name:"dev",description:"Watch, rebuild, and push to all consumers. Auto-detects build command."},args:{build:{type:"string",description:"Override build command (default: auto-detect from package.json)"},"skip-build":{type:"boolean",description:"Watch output dirs directly, skip build command detection",default:false},debounce:{type:"string",description:"Debounce delay in ms (default: 500)"},cooldown:{type:"string",description:"Minimum time between builds in ms (default: 500)"},"no-scripts":{type:"boolean",description:"Skip prepack/postpack lifecycle hooks",default:false}},async run({args:e}){b();let t=resolve("."),i=a(()=>a$1(t,{runScripts:!e["no-scripts"]}),"push");await i(),await b$1(t,e,i);}};export{u as default};
2
+ import {createRequire}from'node:module';import {a as a$1,b as b$1}from'./chunk-ULPTRGDP.mjs';import'./chunk-WSJHZ2ND.mjs';import'./chunk-7JG555TZ.mjs';import'./chunk-7ZWOJDCA.mjs';import'./chunk-W4LHCWGD.mjs';import'./chunk-HN3SZGD3.mjs';import'./chunk-HPF6K6WO.mjs';import'./chunk-ETKOGSRN.mjs';import'./chunk-MBKCCWSD.mjs';import'./chunk-GAS7XU4A.mjs';import {b}from'./chunk-OJJZ7CLB.mjs';import'./chunk-OL7SNXMX.mjs';import'./chunk-HW7AEGI3.mjs';import'./chunk-SG6M7SAC.mjs';import'./chunk-R3RSOZXN.mjs';import'./chunk-ICCM7US5.mjs';import {a}from'./chunk-2VCW5RWI.mjs';import {resolve}from'path';globalThis.require=createRequire(import.meta.url);
3
+ var l={meta:{name:"dev",description:"Watch, rebuild, and push to all consumers. Auto-detects build command."},args:{build:{type:"string",description:"Override build command (default: auto-detect from package.json)"},"skip-build":{type:"boolean",description:"Watch output dirs directly, skip build command detection",default:false},debounce:{type:"string",description:"Debounce delay in ms (default: 500)"},cooldown:{type:"string",description:"Minimum time between builds in ms (default: 500)"},"no-scripts":{type:"boolean",description:"Skip prepack/postpack lifecycle hooks",default:false},force:{type:"boolean",alias:"f",description:"Force copy all files, bypassing hash comparison",default:false}},async run({args:e}){b();let t=resolve("."),o=a(()=>a$1(t,{runScripts:!e["no-scripts"],force:e.force}),"push");await o(),await b$1(t,e,o);}};export{l as default};
@@ -0,0 +1,4 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';import {a as a$1,b as b$5,c as c$2,d as d$1}from'./chunk-7ZWOJDCA.mjs';import {g,b as b$3}from'./chunk-W4LHCWGD.mjs';import {d,b as b$4}from'./chunk-ETKOGSRN.mjs';import'./chunk-GAS7XU4A.mjs';import {b,a as a$2}from'./chunk-OJJZ7CLB.mjs';import'./chunk-OL7SNXMX.mjs';import {b as b$2,w,c as c$1}from'./chunk-SG6M7SAC.mjs';import'./chunk-R3RSOZXN.mjs';import {a,b as b$1}from'./chunk-ICCM7US5.mjs';import {c}from'./chunk-2VCW5RWI.mjs';import {resolve,join}from'path';globalThis.require=createRequire(import.meta.url);
3
+ var o=c(a(),1);var K={meta:{name:"doctor",description:"Run diagnostic checks on your plunk setup"},async run(){b();let t=resolve("."),s=[];b$1.info(`Running plunk diagnostics...
4
+ `);let l=b$2();if(await w(l)){let e=await d();s.push({name:"Store directory",status:"pass",message:`${e.length} entries in ${l}`});}else s.push({name:"Store directory",status:"warn",message:`Store not found at ${l}. Run 'plunk publish' to create it.`});let _=c$1();if(await w(_)){let e=await g(),n=Object.values(e).flat().length;s.push({name:"Global registry",status:"pass",message:`${Object.keys(e).length} packages, ${n} consumer registrations`});}else s.push({name:"Global registry",status:"warn",message:"No consumers registered yet. Use 'plunk add' to link packages."});let{state:x,reliable:A}=await b$3(t);A||s.push({name:"Consumer state",status:"fail",message:"state.json is corrupt or unreadable. Delete .plunk/state.json and re-run 'plunk add' for each package."});let m=Object.entries(x.links);if(m.length>0){s.push({name:"Consumer state",status:"pass",message:`${m.length} linked package(s)`});for(let[e,n]of m){let u=await b$4(e,n.version);u?u.meta.contentHash!==n.contentHash?s.push({name:`Store: ${e}`,status:"warn",message:"Store has newer content. Run 'plunk update' to sync."}):s.push({name:`Store: ${e}`,status:"pass",message:`${e}@${n.version} in sync`}):s.push({name:`Store: ${e}`,status:"fail",message:`Store entry missing for ${e}@${n.version}. Re-publish it.`});let k=join(t,"node_modules",e);if(!await w(k))s.push({name:`node_modules: ${e}`,status:"fail",message:"Missing from node_modules. Run 'plunk restore'."});else try{let{readFile:F}=await import('fs/promises'),d=JSON.parse(await F(join(k,"package.json"),"utf-8"));d.version&&d.version!==n.version&&s.push({name:`node_modules: ${e}`,status:"warn",message:`node_modules has v${d.version} but plunk linked v${n.version}. Run 'plunk restore'.`});}catch{}}}else s.push({name:"Consumer state",status:"warn",message:"No packages linked. Use 'plunk add' to link a package."});let f=await a$1(t);if(s.push({name:"Package manager",status:"pass",message:f}),f==="yarn"){let e=await b$5(t);await c$2(t)?e==="node-modules"?s.push({name:"Yarn linker",status:"pass",message:"Yarn Berry with node-modules linker"}):e==="pnpm"?s.push({name:"Yarn linker",status:"pass",message:"Yarn pnpm linker mode (plunk handles this)"}):e==="pnp"?s.push({name:"Yarn linker",status:"fail",message:"Yarn PnP is not compatible. Set `nodeLinker: node-modules` in .yarnrc.yml"}):s.push({name:"Yarn linker",status:"warn",message:"Yarn Berry defaults to PnP. Add `nodeLinker: node-modules` to .yarnrc.yml"}):s.push({name:"Yarn linker",status:"pass",message:"Yarn Classic, node_modules mode"});}let i=await d$1(t);i.type?s.push({name:"Bundler",status:"pass",message:`${i.type}${i.configFile?` (${i.configFile})`:""}`}):s.push({name:"Bundler",status:"warn",message:"No bundler config detected"});let h=join(t,".gitignore");if(await w(h)){let{readFile:e}=await import('fs/promises');(await e(h,"utf-8")).includes(".plunk")?s.push({name:".gitignore",status:"pass",message:".plunk/ is ignored"}):s.push({name:".gitignore",status:"warn",message:".plunk/ not in .gitignore. Run 'plunk init' to fix."});}parseInt(process.versions.node.split(".")[0],10)>=22?s.push({name:"Node.js version",status:"pass",message:`v${process.versions.node}`}):s.push({name:"Node.js version",status:"fail",message:`v${process.versions.node} \u2014 plunk requires Node.js >= 22`});let B={pass:o.default.green("PASS"),fail:o.default.red("FAIL"),warn:o.default.yellow("WARN")};for(let e of s)b$1.log(` ${B[e.status]} ${e.name}: ${o.default.dim(e.message)}`);let p=s.filter(e=>e.status==="fail").length,c=s.filter(e=>e.status==="warn").length;b$1.log(""),p>0?b$1.error(`${p} issue(s) found that need attention`):c>0?b$1.warn(`${c} warning(s), but no critical issues`):b$1.success("All checks passed!"),a$2({results:s.map(({name:e,status:n,message:u})=>({name:e,status:n,message:u})),failures:p,warnings:c});}};export{K as default};
@@ -0,0 +1,2 @@
1
+ #!/usr/bin/env node
2
+ import {createRequire}from'node:module';export{x as atomicWriteFile,q as collectFiles,z as copyDir,p as copyWithCoW,y as dirSize,u as ensureDir,v as ensurePrivateDir,w as exists,r as incrementalCopy,o as isNodeError,s as moveDir,t as removeDir}from'./chunk-SG6M7SAC.mjs';import'./chunk-R3RSOZXN.mjs';import'./chunk-ICCM7US5.mjs';import'./chunk-2VCW5RWI.mjs';globalThis.require=createRequire(import.meta.url);
package/dist/index.d.ts CHANGED
@@ -105,6 +105,8 @@ interface PackageJson {
105
105
  types?: string;
106
106
  typings?: string;
107
107
  browser?: string | Record<string, string>;
108
+ /** Corepack packageManager field, e.g. "pnpm@9.0.0" */
109
+ packageManager?: string;
108
110
  publishConfig?: {
109
111
  main?: string;
110
112
  module?: string;
@@ -193,6 +195,26 @@ declare function cleanStaleConsumers(): Promise<{
193
195
  removedPackages: number;
194
196
  }>;
195
197
 
198
+ /**
199
+ * plunk configuration stored in package.json under the "plunk" key.
200
+ * Provides persistent defaults for watch/build/dev behavior.
201
+ */
202
+ interface PlunkConfig {
203
+ /** Build command override (same as --build CLI flag) */
204
+ buildCmd?: string;
205
+ /** Glob patterns to watch (same as watch patterns) */
206
+ watchPatterns?: string[];
207
+ /** Debounce delay in ms */
208
+ debounce?: number;
209
+ /** Minimum time between builds in ms */
210
+ cooldown?: number;
211
+ }
212
+ /**
213
+ * Load plunk configuration from package.json#plunk.
214
+ * Returns an empty config if the field is missing or invalid.
215
+ */
216
+ declare function loadPlunkConfig(projectDir: string): Promise<PlunkConfig>;
217
+
196
218
  interface PushOptions {
197
219
  runScripts?: boolean;
198
220
  /** Force copy all files, bypassing hash comparison */
@@ -220,10 +242,11 @@ declare function startWatcher(watchDir: string, options: WatchOptions, onChange:
220
242
  }>;
221
243
 
222
244
  /**
223
- * Detect the package manager used in a project directory
224
- * by checking for lockfile presence, walking up to the filesystem root.
225
- * Closest lockfile wins. Within the same directory, priority order is maintained.
226
- * Falls back to "npm" if no lockfile is found.
245
+ * Detect the package manager used in a project directory.
246
+ * Checks `packageManager` field in package.json first (Corepack convention),
247
+ * then falls back to lockfile presence, walking up to the filesystem root.
248
+ * Closest match wins. Within the same directory, priority order is maintained.
249
+ * Falls back to "npm" if nothing is found.
227
250
  */
228
251
  declare function detectPackageManager(projectDir: string): Promise<PackageManager>;
229
252
 
@@ -241,4 +264,7 @@ declare class Timer {
241
264
  /** Type guard for Node.js system errors with an error code */
242
265
  declare function isNodeError(err: unknown): err is NodeJS.ErrnoException;
243
266
 
244
- export { type ConsumerState, type ConsumersRegistry, type InjectOptions, type InjectResult, type LinkEntry, type PackageJson, type PackageManager, type PlunkMeta, type PublishOptions, type PublishResult, type PushOptions, type StoreEntry, Timer, type WatchOptions, addLink, backupExisting, checkMissingDeps, cleanStaleConsumers, detectPackageManager, doPush, findStoreEntry, getConsumers, getLink, getStoreEntry, inject, isNodeError, killActiveBuild, listStoreEntries, publish, readConsumerState, readConsumerStateSafe, registerConsumer, removeInjected, removeLink, restoreBackup, startWatcher, unregisterConsumer };
267
+ /** Normalize a file path to use forward slashes (for cross-platform consistency). */
268
+ declare function normalizePath(p: string): string;
269
+
270
+ export { type ConsumerState, type ConsumersRegistry, type InjectOptions, type InjectResult, type LinkEntry, type PackageJson, type PackageManager, type PlunkConfig, type PlunkMeta, type PublishOptions, type PublishResult, type PushOptions, type StoreEntry, Timer, type WatchOptions, addLink, backupExisting, checkMissingDeps, cleanStaleConsumers, detectPackageManager, doPush, findStoreEntry, getConsumers, getLink, getStoreEntry, inject, isNodeError, killActiveBuild, listStoreEntries, loadPlunkConfig, normalizePath, publish, readConsumerState, readConsumerStateSafe, registerConsumer, removeInjected, removeLink, restoreBackup, startWatcher, unregisterConsumer };