@better-openclaw/core 1.0.24 → 1.0.26

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (226) hide show
  1. package/dist/addon-stack.cjs +725 -0
  2. package/dist/addon-stack.cjs.map +1 -0
  3. package/dist/addon-stack.d.cts +23 -0
  4. package/dist/addon-stack.d.cts.map +1 -0
  5. package/dist/addon-stack.d.mts +23 -0
  6. package/dist/addon-stack.d.mts.map +1 -0
  7. package/dist/addon-stack.mjs +723 -0
  8. package/dist/addon-stack.mjs.map +1 -0
  9. package/dist/addon-stack.test.cjs +461 -0
  10. package/dist/addon-stack.test.cjs.map +1 -0
  11. package/dist/addon-stack.test.d.cts +1 -0
  12. package/dist/addon-stack.test.d.mts +1 -0
  13. package/dist/addon-stack.test.mjs +461 -0
  14. package/dist/addon-stack.test.mjs.map +1 -0
  15. package/dist/bare-metal-partition.test.cjs +20 -20
  16. package/dist/bare-metal-partition.test.cjs.map +1 -1
  17. package/dist/bare-metal-partition.test.mjs +2 -2
  18. package/dist/compose-validation.test.cjs +1 -1
  19. package/dist/composer.cjs +5 -1
  20. package/dist/composer.cjs.map +1 -1
  21. package/dist/composer.d.cts +24 -1
  22. package/dist/composer.d.cts.map +1 -1
  23. package/dist/composer.d.mts +24 -1
  24. package/dist/composer.d.mts.map +1 -1
  25. package/dist/composer.mjs +1 -1
  26. package/dist/composer.mjs.map +1 -1
  27. package/dist/composer.snapshot.test.cjs +20 -20
  28. package/dist/composer.snapshot.test.cjs.map +1 -1
  29. package/dist/composer.snapshot.test.mjs +2 -2
  30. package/dist/composer.test.cjs +54 -54
  31. package/dist/composer.test.cjs.map +1 -1
  32. package/dist/composer.test.mjs +2 -2
  33. package/dist/deployers/strip-host-ports.cjs +1 -1
  34. package/dist/deployers/strip-host-ports.test.cjs +26 -26
  35. package/dist/deployers/strip-host-ports.test.cjs.map +1 -1
  36. package/dist/deployers/strip-host-ports.test.mjs +1 -1
  37. package/dist/generate.cjs +3 -3
  38. package/dist/generate.mjs +3 -3
  39. package/dist/generate.test.cjs +56 -56
  40. package/dist/generate.test.cjs.map +1 -1
  41. package/dist/generate.test.mjs +1 -1
  42. package/dist/generators/bare-metal-install.test.cjs +18 -18
  43. package/dist/generators/bare-metal-install.test.cjs.map +1 -1
  44. package/dist/generators/bare-metal-install.test.mjs +1 -1
  45. package/dist/generators/caddy.test.cjs +13 -13
  46. package/dist/generators/caddy.test.cjs.map +1 -1
  47. package/dist/generators/caddy.test.mjs +1 -1
  48. package/dist/generators/clone-repos.test.cjs +27 -27
  49. package/dist/generators/clone-repos.test.cjs.map +1 -1
  50. package/dist/generators/clone-repos.test.mjs +1 -1
  51. package/dist/generators/env.cjs +1 -1
  52. package/dist/generators/env.test.cjs +17 -17
  53. package/dist/generators/env.test.cjs.map +1 -1
  54. package/dist/generators/env.test.mjs +1 -1
  55. package/dist/generators/health-check.test.cjs +39 -39
  56. package/dist/generators/health-check.test.cjs.map +1 -1
  57. package/dist/generators/health-check.test.mjs +1 -1
  58. package/dist/generators/postgres-init.cjs +5 -0
  59. package/dist/generators/postgres-init.cjs.map +1 -1
  60. package/dist/generators/postgres-init.d.cts.map +1 -1
  61. package/dist/generators/postgres-init.d.mts.map +1 -1
  62. package/dist/generators/postgres-init.mjs +5 -0
  63. package/dist/generators/postgres-init.mjs.map +1 -1
  64. package/dist/generators/scripts.test.cjs +39 -39
  65. package/dist/generators/scripts.test.cjs.map +1 -1
  66. package/dist/generators/scripts.test.mjs +1 -1
  67. package/dist/generators/skills.cjs +1 -1
  68. package/dist/generators/skills.d.cts.map +1 -1
  69. package/dist/generators/skills.d.mts.map +1 -1
  70. package/dist/generators/skills.mjs +141 -0
  71. package/dist/generators/skills.mjs.map +1 -1
  72. package/dist/generators/traefik.test.cjs +32 -32
  73. package/dist/generators/traefik.test.cjs.map +1 -1
  74. package/dist/generators/traefik.test.mjs +1 -1
  75. package/dist/index.cjs +21 -5
  76. package/dist/index.d.cts +5 -4
  77. package/dist/index.d.mts +5 -4
  78. package/dist/index.mjs +7 -6
  79. package/dist/migrations.test.cjs +16 -16
  80. package/dist/migrations.test.cjs.map +1 -1
  81. package/dist/migrations.test.mjs +1 -1
  82. package/dist/presets/presets.test.cjs +1 -1
  83. package/dist/presets/registry.test.cjs +14 -14
  84. package/dist/presets/registry.test.cjs.map +1 -1
  85. package/dist/presets/registry.test.mjs +1 -1
  86. package/dist/resolver.test.cjs +95 -95
  87. package/dist/resolver.test.cjs.map +1 -1
  88. package/dist/resolver.test.mjs +1 -1
  89. package/dist/{schema-eX44HhRp.d.mts → schema-BQnZrcw8.d.cts} +300 -2
  90. package/dist/schema-BQnZrcw8.d.cts.map +1 -0
  91. package/dist/{schema-tn5RK8CM.d.cts → schema-SBpL0bdI.d.mts} +300 -2
  92. package/dist/schema-SBpL0bdI.d.mts.map +1 -0
  93. package/dist/schema.cjs +148 -2
  94. package/dist/schema.cjs.map +1 -1
  95. package/dist/schema.d.cts +2 -2
  96. package/dist/schema.d.mts +2 -2
  97. package/dist/schema.mjs +139 -2
  98. package/dist/schema.mjs.map +1 -1
  99. package/dist/schema.test.cjs +86 -86
  100. package/dist/schema.test.cjs.map +1 -1
  101. package/dist/schema.test.mjs +1 -1
  102. package/dist/services/definitions/browserless.cjs +4 -1
  103. package/dist/services/definitions/browserless.cjs.map +1 -1
  104. package/dist/services/definitions/browserless.mjs +4 -1
  105. package/dist/services/definitions/browserless.mjs.map +1 -1
  106. package/dist/services/definitions/burnlink.cjs +142 -0
  107. package/dist/services/definitions/burnlink.cjs.map +1 -0
  108. package/dist/services/definitions/burnlink.d.cts +7 -0
  109. package/dist/services/definitions/burnlink.d.cts.map +1 -0
  110. package/dist/services/definitions/burnlink.d.mts +7 -0
  111. package/dist/services/definitions/burnlink.d.mts.map +1 -0
  112. package/dist/services/definitions/burnlink.mjs +141 -0
  113. package/dist/services/definitions/burnlink.mjs.map +1 -0
  114. package/dist/services/definitions/convex.cjs +43 -1
  115. package/dist/services/definitions/convex.cjs.map +1 -1
  116. package/dist/services/definitions/convex.mjs +43 -1
  117. package/dist/services/definitions/convex.mjs.map +1 -1
  118. package/dist/services/definitions/grafana.cjs +11 -1
  119. package/dist/services/definitions/grafana.cjs.map +1 -1
  120. package/dist/services/definitions/grafana.mjs +11 -1
  121. package/dist/services/definitions/grafana.mjs.map +1 -1
  122. package/dist/services/definitions/hindsight.cjs +130 -0
  123. package/dist/services/definitions/hindsight.cjs.map +1 -0
  124. package/dist/services/definitions/hindsight.d.cts +7 -0
  125. package/dist/services/definitions/hindsight.d.cts.map +1 -0
  126. package/dist/services/definitions/hindsight.d.mts +7 -0
  127. package/dist/services/definitions/hindsight.d.mts.map +1 -0
  128. package/dist/services/definitions/hindsight.mjs +129 -0
  129. package/dist/services/definitions/hindsight.mjs.map +1 -0
  130. package/dist/services/definitions/index.cjs +9 -0
  131. package/dist/services/definitions/index.cjs.map +1 -1
  132. package/dist/services/definitions/index.d.cts +4 -1
  133. package/dist/services/definitions/index.d.cts.map +1 -1
  134. package/dist/services/definitions/index.d.mts +4 -1
  135. package/dist/services/definitions/index.d.mts.map +1 -1
  136. package/dist/services/definitions/index.mjs +7 -1
  137. package/dist/services/definitions/index.mjs.map +1 -1
  138. package/dist/services/definitions/meilisearch.cjs +11 -1
  139. package/dist/services/definitions/meilisearch.cjs.map +1 -1
  140. package/dist/services/definitions/meilisearch.mjs +11 -1
  141. package/dist/services/definitions/meilisearch.mjs.map +1 -1
  142. package/dist/services/definitions/minio.cjs +3 -1
  143. package/dist/services/definitions/minio.cjs.map +1 -1
  144. package/dist/services/definitions/minio.mjs +3 -1
  145. package/dist/services/definitions/minio.mjs.map +1 -1
  146. package/dist/services/definitions/n8n.cjs +11 -1
  147. package/dist/services/definitions/n8n.cjs.map +1 -1
  148. package/dist/services/definitions/n8n.mjs +11 -1
  149. package/dist/services/definitions/n8n.mjs.map +1 -1
  150. package/dist/services/definitions/ollama.cjs +3 -1
  151. package/dist/services/definitions/ollama.cjs.map +1 -1
  152. package/dist/services/definitions/ollama.mjs +3 -1
  153. package/dist/services/definitions/ollama.mjs.map +1 -1
  154. package/dist/services/definitions/opensandbox.cjs +149 -0
  155. package/dist/services/definitions/opensandbox.cjs.map +1 -0
  156. package/dist/services/definitions/opensandbox.d.cts +7 -0
  157. package/dist/services/definitions/opensandbox.d.cts.map +1 -0
  158. package/dist/services/definitions/opensandbox.d.mts +7 -0
  159. package/dist/services/definitions/opensandbox.d.mts.map +1 -0
  160. package/dist/services/definitions/opensandbox.mjs +148 -0
  161. package/dist/services/definitions/opensandbox.mjs.map +1 -0
  162. package/dist/services/definitions/qdrant.cjs +3 -1
  163. package/dist/services/definitions/qdrant.cjs.map +1 -1
  164. package/dist/services/definitions/qdrant.mjs +3 -1
  165. package/dist/services/definitions/qdrant.mjs.map +1 -1
  166. package/dist/services/definitions/searxng.cjs +8 -1
  167. package/dist/services/definitions/searxng.cjs.map +1 -1
  168. package/dist/services/definitions/searxng.mjs +8 -1
  169. package/dist/services/definitions/searxng.mjs.map +1 -1
  170. package/dist/services/definitions/uptime-kuma.cjs +8 -1
  171. package/dist/services/definitions/uptime-kuma.cjs.map +1 -1
  172. package/dist/services/definitions/uptime-kuma.mjs +8 -1
  173. package/dist/services/definitions/uptime-kuma.mjs.map +1 -1
  174. package/dist/services/registry.test.cjs +36 -36
  175. package/dist/services/registry.test.cjs.map +1 -1
  176. package/dist/services/registry.test.mjs +1 -1
  177. package/dist/{skills-BlzpHmpH.cjs → skills-BSF7iNa4.cjs} +142 -1
  178. package/dist/{skills-BlzpHmpH.cjs.map → skills-BSF7iNa4.cjs.map} +1 -1
  179. package/dist/{vi.2VT5v0um-C_jmO7m2.mjs → test.CTcmp4Su-ClCHJ3FA.mjs} +6793 -6403
  180. package/dist/test.CTcmp4Su-ClCHJ3FA.mjs.map +1 -0
  181. package/dist/{vi.2VT5v0um-iVBt6Fyq.cjs → test.CTcmp4Su-DlzTarwH.cjs} +6793 -6403
  182. package/dist/test.CTcmp4Su-DlzTarwH.cjs.map +1 -0
  183. package/dist/track-analytics.test.cjs +28 -28
  184. package/dist/track-analytics.test.cjs.map +1 -1
  185. package/dist/track-analytics.test.mjs +1 -1
  186. package/dist/types.cjs.map +1 -1
  187. package/dist/types.d.cts +10 -2
  188. package/dist/types.d.cts.map +1 -1
  189. package/dist/types.d.mts +10 -2
  190. package/dist/types.d.mts.map +1 -1
  191. package/dist/types.mjs.map +1 -1
  192. package/dist/validator.cjs +1 -1
  193. package/dist/validator.test.cjs +15 -15
  194. package/dist/validator.test.cjs.map +1 -1
  195. package/dist/validator.test.mjs +2 -2
  196. package/dist/version-manager.test.cjs +37 -37
  197. package/dist/version-manager.test.cjs.map +1 -1
  198. package/dist/version-manager.test.mjs +1 -1
  199. package/package.json +4 -4
  200. package/src/__snapshots__/composer.snapshot.test.ts.snap +5 -0
  201. package/src/addon-stack.test.ts +648 -0
  202. package/src/addon-stack.ts +1046 -0
  203. package/src/composer.ts +4 -4
  204. package/src/generators/postgres-init.ts +2 -0
  205. package/src/generators/skills.ts +142 -0
  206. package/src/index.ts +20 -2
  207. package/src/schema.ts +190 -0
  208. package/src/services/definitions/browserless.ts +3 -0
  209. package/src/services/definitions/burnlink.ts +142 -0
  210. package/src/services/definitions/convex.ts +31 -0
  211. package/src/services/definitions/grafana.ts +9 -0
  212. package/src/services/definitions/hindsight.ts +131 -0
  213. package/src/services/definitions/index.ts +10 -0
  214. package/src/services/definitions/meilisearch.ts +9 -0
  215. package/src/services/definitions/minio.ts +2 -0
  216. package/src/services/definitions/n8n.ts +9 -0
  217. package/src/services/definitions/ollama.ts +2 -0
  218. package/src/services/definitions/opensandbox.ts +156 -0
  219. package/src/services/definitions/qdrant.ts +2 -0
  220. package/src/services/definitions/searxng.ts +3 -0
  221. package/src/services/definitions/uptime-kuma.ts +3 -0
  222. package/src/types.ts +18 -0
  223. package/dist/schema-eX44HhRp.d.mts.map +0 -1
  224. package/dist/schema-tn5RK8CM.d.cts.map +0 -1
  225. package/dist/vi.2VT5v0um-C_jmO7m2.mjs.map +0 -1
  226. package/dist/vi.2VT5v0um-iVBt6Fyq.cjs.map +0 -1
@@ -1 +1 @@
1
- {"version":3,"file":"postgres-init.mjs","names":[],"sources":["../../src/generators/postgres-init.ts"],"sourcesContent":["import type { ResolverOutput } from \"../types.js\";\n\n/**\n * Database requirement descriptor for a service.\n */\ninterface DbRequirement {\n\tserviceId: string;\n\tserviceName: string;\n\tdbName: string;\n\tdbUser: string;\n\tpasswordEnvVar: string;\n}\n\n/**\n * Known database requirements for services that need their own PostgreSQL database.\n * Maps service ID to the DB name, user, and password env var they expect.\n */\nconst DB_REQUIREMENTS: Record<string, Omit<DbRequirement, \"serviceId\" | \"serviceName\">> = {\n\tn8n: { dbName: \"n8n\", dbUser: \"n8n\", passwordEnvVar: \"N8N_DB_PASSWORD\" },\n\tpostiz: { dbName: \"postiz\", dbUser: \"postiz\", passwordEnvVar: \"POSTIZ_DB_PASSWORD\" },\n\toutline: { dbName: \"outline\", dbUser: \"outline\", passwordEnvVar: \"OUTLINE_DB_PASSWORD\" },\n\tdify: { dbName: \"dify\", dbUser: \"dify\", passwordEnvVar: \"DIFY_DB_PASSWORD\" },\n\ttemporal: { dbName: \"temporal\", dbUser: \"temporal\", passwordEnvVar: \"TEMPORAL_DB_PASSWORD\" },\n\tmattermost: {\n\t\tdbName: \"mattermost\",\n\t\tdbUser: \"mattermost\",\n\t\tpasswordEnvVar: \"MATTERMOST_DB_PASSWORD\",\n\t},\n\t\"matrix-synapse\": { dbName: \"synapse\", dbUser: \"synapse\", passwordEnvVar: \"SYNAPSE_DB_PASSWORD\" },\n\t\"paperless-ngx\": {\n\t\tdbName: \"paperless\",\n\t\tdbUser: \"paperless\",\n\t\tpasswordEnvVar: \"PAPERLESS_DB_PASSWORD\",\n\t},\n\tumami: { dbName: \"umami\", dbUser: \"umami\", passwordEnvVar: \"UMAMI_DB_PASSWORD\" },\n\tmatomo: { dbName: \"matomo\", dbUser: \"matomo\", passwordEnvVar: \"MATOMO_DB_PASSWORD\" },\n\tmixpost: { dbName: \"mixpost\", dbUser: \"mixpost\", passwordEnvVar: \"MIXPOST_DB_PASSWORD\" },\n\t\"cal-com\": { dbName: \"calcom\", dbUser: \"calcom\", passwordEnvVar: \"CALCOM_DB_PASSWORD\" },\n\timmich: { dbName: \"immich\", dbUser: \"immich\", passwordEnvVar: \"IMMICH_DB_PASSWORD\" },\n\tauthentik: { dbName: \"authentik\", dbUser: \"authentik\", passwordEnvVar: \"AUTHENTIK_DB_PASSWORD\" },\n\tchatwoot: { dbName: \"chatwoot\", dbUser: \"chatwoot\", passwordEnvVar: \"CHATWOOT_DB_PASSWORD\" },\n\tfirecrawl: { dbName: \"firecrawl\", dbUser: \"firecrawl\", passwordEnvVar: \"FIRECRAWL_DB_PASSWORD\" },\n\tflagsmith: { dbName: \"flagsmith\", dbUser: \"flagsmith\", passwordEnvVar: \"FLAGSMITH_DB_PASSWORD\" },\n\tinfisical: { dbName: \"infisical\", dbUser: \"infisical\", passwordEnvVar: \"INFISICAL_DB_PASSWORD\" },\n\tkeycloak: { dbName: \"keycloak\", dbUser: \"keycloak\", passwordEnvVar: \"KEYCLOAK_DB_PASSWORD\" },\n\tlistmonk: { dbName: \"listmonk\", dbUser: \"listmonk\", passwordEnvVar: \"LISTMONK_DB_PASSWORD\" },\n\t\"lasuite-meet-backend\": {\n\t\tdbName: \"meet\",\n\t\tdbUser: \"meet\",\n\t\tpasswordEnvVar: \"LASUITE_MEET_DB_PASSWORD\",\n\t},\n\topenpanel: { dbName: \"openpanel\", dbUser: \"openpanel\", passwordEnvVar: \"OPENPANEL_DB_PASSWORD\" },\n\tusesend: { dbName: \"usesend\", dbUser: \"usesend\", passwordEnvVar: \"USESEND_DB_PASSWORD\" },\n\tnextcloud: { dbName: \"nextcloud\", dbUser: \"nextcloud\", passwordEnvVar: \"NEXTCLOUD_DB_PASSWORD\" },\n\t// ── SaaS Boilerplates ────────────────────────────────────────────────────\n\t\"open-saas\": { dbName: \"opensaas\", dbUser: \"opensaas\", passwordEnvVar: \"OPENSAAS_DB_PASSWORD\" },\n\t\"apptension-saas\": {\n\t\tdbName: \"apptensionsaas\",\n\t\tdbUser: \"apptensionsaas\",\n\t\tpasswordEnvVar: \"APPTENSION_SAAS_DB_PASSWORD\",\n\t},\n\t\"boxyhq-saas\": {\n\t\tdbName: \"boxyhqsaas\",\n\t\tdbUser: \"boxyhqsaas\",\n\t\tpasswordEnvVar: \"BOXYHQ_SAAS_DB_PASSWORD\",\n\t},\n\t\"ixartz-saas\": {\n\t\tdbName: \"ixartzsaas\",\n\t\tdbUser: \"ixartzsaas\",\n\t\tpasswordEnvVar: \"IXARTZ_SAAS_DB_PASSWORD\",\n\t},\n};\n\n/**\n * Get the list of database requirements for all resolved services.\n * Exported so other generators (env, composer) can access the same data.\n */\nexport function getDbRequirements(resolved: ResolverOutput): DbRequirement[] {\n\tconst reqs: DbRequirement[] = [];\n\tfor (const { definition } of resolved.services) {\n\t\tconst req = DB_REQUIREMENTS[definition.id];\n\t\tif (req) {\n\t\t\treqs.push({\n\t\t\t\tserviceId: definition.id,\n\t\t\t\tserviceName: definition.name,\n\t\t\t\t...req,\n\t\t\t});\n\t\t}\n\t}\n\treturn reqs;\n}\n\n/**\n * Generates a PostgreSQL initialization script that creates databases and users\n * for all companion services that need their own DB.\n *\n * Returns null if PostgreSQL is not in the resolved stack or no services need extra DBs.\n * The script is designed to be mounted at /docker-entrypoint-initdb.d/init-databases.sh\n * and runs automatically during PostgreSQL's first initialization.\n */\nexport function generatePostgresInit(resolved: ResolverOutput): string | null {\n\tconst hasPostgres = resolved.services.some((s) => s.definition.id === \"postgresql\");\n\tif (!hasPostgres) return null;\n\n\tconst reqs = getDbRequirements(resolved);\n\tif (reqs.length === 0) return null;\n\n\tconst createCalls = reqs\n\t\t.map(\n\t\t\t(r) =>\n\t\t\t\t`create_db_and_user \"${r.dbName}\" \"${r.dbUser}\" \"\\${${r.passwordEnvVar}:-$POSTGRES_PASSWORD}\"`,\n\t\t)\n\t\t.join(\"\\n\");\n\n\treturn `#!/bin/bash\nset -e\n\n# ═══════════════════════════════════════════════════════════════════════════════\n# PostgreSQL Initialization Script\n# Generated by better-openclaw\n#\n# This script runs automatically during PostgreSQL's first initialization.\n# It creates databases and users for all companion services in your stack.\n# ═══════════════════════════════════════════════════════════════════════════════\n\ncreate_db_and_user() {\n local db=\"$1\"\n local user=\"$2\"\n local password=\"$3\"\n\n echo \">>> Creating database '$db' with user '$user'...\"\n\n psql -v ON_ERROR_STOP=1 --username \"$POSTGRES_USER\" --dbname \"$POSTGRES_DB\" <<-EOSQL\n -- Create user if not exists\n DO \\\\$\\\\$\n BEGIN\n IF NOT EXISTS (SELECT FROM pg_catalog.pg_roles WHERE rolname = '$user') THEN\n CREATE ROLE \"$user\" WITH LOGIN PASSWORD '$password';\n RAISE NOTICE 'Created user: $user';\n ELSE\n -- Update password in case it changed\n ALTER ROLE \"$user\" WITH PASSWORD '$password';\n RAISE NOTICE 'User already exists, updated password: $user';\n END IF;\n END\n \\\\$\\\\$;\n\n -- Create database if not exists\n SELECT 'CREATE DATABASE \"$db\" OWNER \"$user\"'\n WHERE NOT EXISTS (SELECT FROM pg_database WHERE datname = '$db')\\\\gexec\n\n -- Grant privileges\n GRANT ALL PRIVILEGES ON DATABASE \"$db\" TO \"$user\";\nEOSQL\n\n echo \">>> Done: database '$db' with user '$user'\"\n}\n\necho \"═══════════════════════════════════════════════════════════\"\necho \" Initializing databases for companion services...\"\necho \"═══════════════════════════════════════════════════════════\"\n\n${createCalls}\n\necho \"\"\necho \"═══════════════════════════════════════════════════════════\"\necho \" All databases initialized successfully!\"\necho \" Services: ${reqs.map((r) => r.serviceName).join(\", \")}\"\necho \"═══════════════════════════════════════════════════════════\"\n`;\n}\n"],"mappings":";;;;;AAiBA,MAAM,kBAAoF;CACzF,KAAK;EAAE,QAAQ;EAAO,QAAQ;EAAO,gBAAgB;EAAmB;CACxE,QAAQ;EAAE,QAAQ;EAAU,QAAQ;EAAU,gBAAgB;EAAsB;CACpF,SAAS;EAAE,QAAQ;EAAW,QAAQ;EAAW,gBAAgB;EAAuB;CACxF,MAAM;EAAE,QAAQ;EAAQ,QAAQ;EAAQ,gBAAgB;EAAoB;CAC5E,UAAU;EAAE,QAAQ;EAAY,QAAQ;EAAY,gBAAgB;EAAwB;CAC5F,YAAY;EACX,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD,kBAAkB;EAAE,QAAQ;EAAW,QAAQ;EAAW,gBAAgB;EAAuB;CACjG,iBAAiB;EAChB,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD,OAAO;EAAE,QAAQ;EAAS,QAAQ;EAAS,gBAAgB;EAAqB;CAChF,QAAQ;EAAE,QAAQ;EAAU,QAAQ;EAAU,gBAAgB;EAAsB;CACpF,SAAS;EAAE,QAAQ;EAAW,QAAQ;EAAW,gBAAgB;EAAuB;CACxF,WAAW;EAAE,QAAQ;EAAU,QAAQ;EAAU,gBAAgB;EAAsB;CACvF,QAAQ;EAAE,QAAQ;EAAU,QAAQ;EAAU,gBAAgB;EAAsB;CACpF,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAChG,UAAU;EAAE,QAAQ;EAAY,QAAQ;EAAY,gBAAgB;EAAwB;CAC5F,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAChG,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAChG,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAChG,UAAU;EAAE,QAAQ;EAAY,QAAQ;EAAY,gBAAgB;EAAwB;CAC5F,UAAU;EAAE,QAAQ;EAAY,QAAQ;EAAY,gBAAgB;EAAwB;CAC5F,wBAAwB;EACvB,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAChG,SAAS;EAAE,QAAQ;EAAW,QAAQ;EAAW,gBAAgB;EAAuB;CACxF,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAEhG,aAAa;EAAE,QAAQ;EAAY,QAAQ;EAAY,gBAAgB;EAAwB;CAC/F,mBAAmB;EAClB,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD,eAAe;EACd,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD,eAAe;EACd,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD;;;;;AAMD,SAAgB,kBAAkB,UAA2C;CAC5E,MAAM,OAAwB,EAAE;AAChC,MAAK,MAAM,EAAE,gBAAgB,SAAS,UAAU;EAC/C,MAAM,MAAM,gBAAgB,WAAW;AACvC,MAAI,IACH,MAAK,KAAK;GACT,WAAW,WAAW;GACtB,aAAa,WAAW;GACxB,GAAG;GACH,CAAC;;AAGJ,QAAO;;;;;;;;;;AAWR,SAAgB,qBAAqB,UAAyC;AAE7E,KAAI,CADgB,SAAS,SAAS,MAAM,MAAM,EAAE,WAAW,OAAO,aAAa,CACjE,QAAO;CAEzB,MAAM,OAAO,kBAAkB,SAAS;AACxC,KAAI,KAAK,WAAW,EAAG,QAAO;AAS9B,QAAO;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAPa,KAClB,KACC,MACA,uBAAuB,EAAE,OAAO,KAAK,EAAE,OAAO,QAAQ,EAAE,eAAe,wBACxE,CACA,KAAK,KAAK,CAkDC;;;;;mBAKK,KAAK,KAAK,MAAM,EAAE,YAAY,CAAC,KAAK,KAAK,CAAC"}
1
+ {"version":3,"file":"postgres-init.mjs","names":[],"sources":["../../src/generators/postgres-init.ts"],"sourcesContent":["import type { ResolverOutput } from \"../types.js\";\n\n/**\n * Database requirement descriptor for a service.\n */\ninterface DbRequirement {\n\tserviceId: string;\n\tserviceName: string;\n\tdbName: string;\n\tdbUser: string;\n\tpasswordEnvVar: string;\n}\n\n/**\n * Known database requirements for services that need their own PostgreSQL database.\n * Maps service ID to the DB name, user, and password env var they expect.\n */\nconst DB_REQUIREMENTS: Record<string, Omit<DbRequirement, \"serviceId\" | \"serviceName\">> = {\n\tn8n: { dbName: \"n8n\", dbUser: \"n8n\", passwordEnvVar: \"N8N_DB_PASSWORD\" },\n\tpostiz: { dbName: \"postiz\", dbUser: \"postiz\", passwordEnvVar: \"POSTIZ_DB_PASSWORD\" },\n\toutline: { dbName: \"outline\", dbUser: \"outline\", passwordEnvVar: \"OUTLINE_DB_PASSWORD\" },\n\tdify: { dbName: \"dify\", dbUser: \"dify\", passwordEnvVar: \"DIFY_DB_PASSWORD\" },\n\ttemporal: { dbName: \"temporal\", dbUser: \"temporal\", passwordEnvVar: \"TEMPORAL_DB_PASSWORD\" },\n\tmattermost: {\n\t\tdbName: \"mattermost\",\n\t\tdbUser: \"mattermost\",\n\t\tpasswordEnvVar: \"MATTERMOST_DB_PASSWORD\",\n\t},\n\t\"matrix-synapse\": { dbName: \"synapse\", dbUser: \"synapse\", passwordEnvVar: \"SYNAPSE_DB_PASSWORD\" },\n\t\"paperless-ngx\": {\n\t\tdbName: \"paperless\",\n\t\tdbUser: \"paperless\",\n\t\tpasswordEnvVar: \"PAPERLESS_DB_PASSWORD\",\n\t},\n\tumami: { dbName: \"umami\", dbUser: \"umami\", passwordEnvVar: \"UMAMI_DB_PASSWORD\" },\n\tmatomo: { dbName: \"matomo\", dbUser: \"matomo\", passwordEnvVar: \"MATOMO_DB_PASSWORD\" },\n\tmixpost: { dbName: \"mixpost\", dbUser: \"mixpost\", passwordEnvVar: \"MIXPOST_DB_PASSWORD\" },\n\t\"cal-com\": { dbName: \"calcom\", dbUser: \"calcom\", passwordEnvVar: \"CALCOM_DB_PASSWORD\" },\n\timmich: { dbName: \"immich\", dbUser: \"immich\", passwordEnvVar: \"IMMICH_DB_PASSWORD\" },\n\tauthentik: { dbName: \"authentik\", dbUser: \"authentik\", passwordEnvVar: \"AUTHENTIK_DB_PASSWORD\" },\n\tchatwoot: { dbName: \"chatwoot\", dbUser: \"chatwoot\", passwordEnvVar: \"CHATWOOT_DB_PASSWORD\" },\n\tfirecrawl: { dbName: \"firecrawl\", dbUser: \"firecrawl\", passwordEnvVar: \"FIRECRAWL_DB_PASSWORD\" },\n\tflagsmith: { dbName: \"flagsmith\", dbUser: \"flagsmith\", passwordEnvVar: \"FLAGSMITH_DB_PASSWORD\" },\n\tinfisical: { dbName: \"infisical\", dbUser: \"infisical\", passwordEnvVar: \"INFISICAL_DB_PASSWORD\" },\n\tkeycloak: { dbName: \"keycloak\", dbUser: \"keycloak\", passwordEnvVar: \"KEYCLOAK_DB_PASSWORD\" },\n\tlistmonk: { dbName: \"listmonk\", dbUser: \"listmonk\", passwordEnvVar: \"LISTMONK_DB_PASSWORD\" },\n\t\"lasuite-meet-backend\": {\n\t\tdbName: \"meet\",\n\t\tdbUser: \"meet\",\n\t\tpasswordEnvVar: \"LASUITE_MEET_DB_PASSWORD\",\n\t},\n\topenpanel: { dbName: \"openpanel\", dbUser: \"openpanel\", passwordEnvVar: \"OPENPANEL_DB_PASSWORD\" },\n\tusesend: { dbName: \"usesend\", dbUser: \"usesend\", passwordEnvVar: \"USESEND_DB_PASSWORD\" },\n\tnextcloud: { dbName: \"nextcloud\", dbUser: \"nextcloud\", passwordEnvVar: \"NEXTCLOUD_DB_PASSWORD\" },\n\t// ── Agent Memory ────────────────────────────────────────────────────────\n\thindsight: { dbName: \"hindsight\", dbUser: \"hindsight\", passwordEnvVar: \"HINDSIGHT_DB_PASSWORD\" },\n\t// ── SaaS Boilerplates ────────────────────────────────────────────────────\n\t\"open-saas\": { dbName: \"opensaas\", dbUser: \"opensaas\", passwordEnvVar: \"OPENSAAS_DB_PASSWORD\" },\n\t\"apptension-saas\": {\n\t\tdbName: \"apptensionsaas\",\n\t\tdbUser: \"apptensionsaas\",\n\t\tpasswordEnvVar: \"APPTENSION_SAAS_DB_PASSWORD\",\n\t},\n\t\"boxyhq-saas\": {\n\t\tdbName: \"boxyhqsaas\",\n\t\tdbUser: \"boxyhqsaas\",\n\t\tpasswordEnvVar: \"BOXYHQ_SAAS_DB_PASSWORD\",\n\t},\n\t\"ixartz-saas\": {\n\t\tdbName: \"ixartzsaas\",\n\t\tdbUser: \"ixartzsaas\",\n\t\tpasswordEnvVar: \"IXARTZ_SAAS_DB_PASSWORD\",\n\t},\n};\n\n/**\n * Get the list of database requirements for all resolved services.\n * Exported so other generators (env, composer) can access the same data.\n */\nexport function getDbRequirements(resolved: ResolverOutput): DbRequirement[] {\n\tconst reqs: DbRequirement[] = [];\n\tfor (const { definition } of resolved.services) {\n\t\tconst req = DB_REQUIREMENTS[definition.id];\n\t\tif (req) {\n\t\t\treqs.push({\n\t\t\t\tserviceId: definition.id,\n\t\t\t\tserviceName: definition.name,\n\t\t\t\t...req,\n\t\t\t});\n\t\t}\n\t}\n\treturn reqs;\n}\n\n/**\n * Generates a PostgreSQL initialization script that creates databases and users\n * for all companion services that need their own DB.\n *\n * Returns null if PostgreSQL is not in the resolved stack or no services need extra DBs.\n * The script is designed to be mounted at /docker-entrypoint-initdb.d/init-databases.sh\n * and runs automatically during PostgreSQL's first initialization.\n */\nexport function generatePostgresInit(resolved: ResolverOutput): string | null {\n\tconst hasPostgres = resolved.services.some((s) => s.definition.id === \"postgresql\");\n\tif (!hasPostgres) return null;\n\n\tconst reqs = getDbRequirements(resolved);\n\tif (reqs.length === 0) return null;\n\n\tconst createCalls = reqs\n\t\t.map(\n\t\t\t(r) =>\n\t\t\t\t`create_db_and_user \"${r.dbName}\" \"${r.dbUser}\" \"\\${${r.passwordEnvVar}:-$POSTGRES_PASSWORD}\"`,\n\t\t)\n\t\t.join(\"\\n\");\n\n\treturn `#!/bin/bash\nset -e\n\n# ═══════════════════════════════════════════════════════════════════════════════\n# PostgreSQL Initialization Script\n# Generated by better-openclaw\n#\n# This script runs automatically during PostgreSQL's first initialization.\n# It creates databases and users for all companion services in your stack.\n# ═══════════════════════════════════════════════════════════════════════════════\n\ncreate_db_and_user() {\n local db=\"$1\"\n local user=\"$2\"\n local password=\"$3\"\n\n echo \">>> Creating database '$db' with user '$user'...\"\n\n psql -v ON_ERROR_STOP=1 --username \"$POSTGRES_USER\" --dbname \"$POSTGRES_DB\" <<-EOSQL\n -- Create user if not exists\n DO \\\\$\\\\$\n BEGIN\n IF NOT EXISTS (SELECT FROM pg_catalog.pg_roles WHERE rolname = '$user') THEN\n CREATE ROLE \"$user\" WITH LOGIN PASSWORD '$password';\n RAISE NOTICE 'Created user: $user';\n ELSE\n -- Update password in case it changed\n ALTER ROLE \"$user\" WITH PASSWORD '$password';\n RAISE NOTICE 'User already exists, updated password: $user';\n END IF;\n END\n \\\\$\\\\$;\n\n -- Create database if not exists\n SELECT 'CREATE DATABASE \"$db\" OWNER \"$user\"'\n WHERE NOT EXISTS (SELECT FROM pg_database WHERE datname = '$db')\\\\gexec\n\n -- Grant privileges\n GRANT ALL PRIVILEGES ON DATABASE \"$db\" TO \"$user\";\nEOSQL\n\n echo \">>> Done: database '$db' with user '$user'\"\n}\n\necho \"═══════════════════════════════════════════════════════════\"\necho \" Initializing databases for companion services...\"\necho \"═══════════════════════════════════════════════════════════\"\n\n${createCalls}\n\necho \"\"\necho \"═══════════════════════════════════════════════════════════\"\necho \" All databases initialized successfully!\"\necho \" Services: ${reqs.map((r) => r.serviceName).join(\", \")}\"\necho \"═══════════════════════════════════════════════════════════\"\n`;\n}\n"],"mappings":";;;;;AAiBA,MAAM,kBAAoF;CACzF,KAAK;EAAE,QAAQ;EAAO,QAAQ;EAAO,gBAAgB;EAAmB;CACxE,QAAQ;EAAE,QAAQ;EAAU,QAAQ;EAAU,gBAAgB;EAAsB;CACpF,SAAS;EAAE,QAAQ;EAAW,QAAQ;EAAW,gBAAgB;EAAuB;CACxF,MAAM;EAAE,QAAQ;EAAQ,QAAQ;EAAQ,gBAAgB;EAAoB;CAC5E,UAAU;EAAE,QAAQ;EAAY,QAAQ;EAAY,gBAAgB;EAAwB;CAC5F,YAAY;EACX,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD,kBAAkB;EAAE,QAAQ;EAAW,QAAQ;EAAW,gBAAgB;EAAuB;CACjG,iBAAiB;EAChB,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD,OAAO;EAAE,QAAQ;EAAS,QAAQ;EAAS,gBAAgB;EAAqB;CAChF,QAAQ;EAAE,QAAQ;EAAU,QAAQ;EAAU,gBAAgB;EAAsB;CACpF,SAAS;EAAE,QAAQ;EAAW,QAAQ;EAAW,gBAAgB;EAAuB;CACxF,WAAW;EAAE,QAAQ;EAAU,QAAQ;EAAU,gBAAgB;EAAsB;CACvF,QAAQ;EAAE,QAAQ;EAAU,QAAQ;EAAU,gBAAgB;EAAsB;CACpF,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAChG,UAAU;EAAE,QAAQ;EAAY,QAAQ;EAAY,gBAAgB;EAAwB;CAC5F,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAChG,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAChG,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAChG,UAAU;EAAE,QAAQ;EAAY,QAAQ;EAAY,gBAAgB;EAAwB;CAC5F,UAAU;EAAE,QAAQ;EAAY,QAAQ;EAAY,gBAAgB;EAAwB;CAC5F,wBAAwB;EACvB,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAChG,SAAS;EAAE,QAAQ;EAAW,QAAQ;EAAW,gBAAgB;EAAuB;CACxF,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAEhG,WAAW;EAAE,QAAQ;EAAa,QAAQ;EAAa,gBAAgB;EAAyB;CAEhG,aAAa;EAAE,QAAQ;EAAY,QAAQ;EAAY,gBAAgB;EAAwB;CAC/F,mBAAmB;EAClB,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD,eAAe;EACd,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD,eAAe;EACd,QAAQ;EACR,QAAQ;EACR,gBAAgB;EAChB;CACD;;;;;AAMD,SAAgB,kBAAkB,UAA2C;CAC5E,MAAM,OAAwB,EAAE;AAChC,MAAK,MAAM,EAAE,gBAAgB,SAAS,UAAU;EAC/C,MAAM,MAAM,gBAAgB,WAAW;AACvC,MAAI,IACH,MAAK,KAAK;GACT,WAAW,WAAW;GACtB,aAAa,WAAW;GACxB,GAAG;GACH,CAAC;;AAGJ,QAAO;;;;;;;;;;AAWR,SAAgB,qBAAqB,UAAyC;AAE7E,KAAI,CADgB,SAAS,SAAS,MAAM,MAAM,EAAE,WAAW,OAAO,aAAa,CACjE,QAAO;CAEzB,MAAM,OAAO,kBAAkB,SAAS;AACxC,KAAI,KAAK,WAAW,EAAG,QAAO;AAS9B,QAAO;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAPa,KAClB,KACC,MACA,uBAAuB,EAAE,OAAO,KAAK,EAAE,OAAO,QAAQ,EAAE,eAAe,wBACxE,CACA,KAAK,KAAK,CAkDC;;;;;mBAKK,KAAK,KAAK,MAAM,EAAE,YAAY,CAAC,KAAK,KAAK,CAAC"}
@@ -1,8 +1,8 @@
1
- const require_vi_2VT5v0um = require("../vi.2VT5v0um-iVBt6Fyq.cjs");
1
+ const require_test_CTcmp4Su = require("../test.CTcmp4Su-DlzTarwH.cjs");
2
2
  const require_generators_scripts = require("./scripts.cjs");
3
3
  //#region src/generators/scripts.test.ts
4
- require_vi_2VT5v0um.describe("generateScripts", () => {
5
- require_vi_2VT5v0um.it("generates all 10 expected scripts (5 bash + 5 PowerShell)", () => {
4
+ require_test_CTcmp4Su.describe("generateScripts", () => {
5
+ require_test_CTcmp4Su.it("generates all 10 expected scripts (5 bash + 5 PowerShell)", () => {
6
6
  const result = require_generators_scripts.generateScripts();
7
7
  for (const script of [
8
8
  "scripts/start.sh",
@@ -16,67 +16,67 @@ require_vi_2VT5v0um.describe("generateScripts", () => {
16
16
  "scripts/backup.ps1",
17
17
  "scripts/status.ps1"
18
18
  ]) {
19
- require_vi_2VT5v0um.globalExpect(result).toHaveProperty(script);
20
- require_vi_2VT5v0um.globalExpect(result[script].length).toBeGreaterThan(0);
19
+ require_test_CTcmp4Su.globalExpect(result).toHaveProperty(script);
20
+ require_test_CTcmp4Su.globalExpect(result[script].length).toBeGreaterThan(0);
21
21
  }
22
22
  });
23
- require_vi_2VT5v0um.it("start.sh calls docker compose up", () => {
23
+ require_test_CTcmp4Su.it("start.sh calls docker compose up", () => {
24
24
  const result = require_generators_scripts.generateScripts();
25
- require_vi_2VT5v0um.globalExpect(result["scripts/start.sh"]).toContain("docker compose");
26
- require_vi_2VT5v0um.globalExpect(result["scripts/start.sh"]).toContain("up");
25
+ require_test_CTcmp4Su.globalExpect(result["scripts/start.sh"]).toContain("docker compose");
26
+ require_test_CTcmp4Su.globalExpect(result["scripts/start.sh"]).toContain("up");
27
27
  });
28
- require_vi_2VT5v0um.it("stop.sh calls docker compose down", () => {
28
+ require_test_CTcmp4Su.it("stop.sh calls docker compose down", () => {
29
29
  const result = require_generators_scripts.generateScripts();
30
- require_vi_2VT5v0um.globalExpect(result["scripts/stop.sh"]).toContain("docker compose");
31
- require_vi_2VT5v0um.globalExpect(result["scripts/stop.sh"]).toContain("down");
30
+ require_test_CTcmp4Su.globalExpect(result["scripts/stop.sh"]).toContain("docker compose");
31
+ require_test_CTcmp4Su.globalExpect(result["scripts/stop.sh"]).toContain("down");
32
32
  });
33
- require_vi_2VT5v0um.it("update.sh calls docker compose pull", () => {
33
+ require_test_CTcmp4Su.it("update.sh calls docker compose pull", () => {
34
34
  const result = require_generators_scripts.generateScripts();
35
- require_vi_2VT5v0um.globalExpect(result["scripts/update.sh"]).toContain("docker compose");
36
- require_vi_2VT5v0um.globalExpect(result["scripts/update.sh"]).toContain("pull");
35
+ require_test_CTcmp4Su.globalExpect(result["scripts/update.sh"]).toContain("docker compose");
36
+ require_test_CTcmp4Su.globalExpect(result["scripts/update.sh"]).toContain("pull");
37
37
  });
38
- require_vi_2VT5v0um.it("backup.sh references volumes or backup", () => {
38
+ require_test_CTcmp4Su.it("backup.sh references volumes or backup", () => {
39
39
  const backup = require_generators_scripts.generateScripts()["scripts/backup.sh"];
40
- require_vi_2VT5v0um.globalExpect(backup).toBeDefined();
41
- require_vi_2VT5v0um.globalExpect(backup.length).toBeGreaterThan(50);
40
+ require_test_CTcmp4Su.globalExpect(backup).toBeDefined();
41
+ require_test_CTcmp4Su.globalExpect(backup.length).toBeGreaterThan(50);
42
42
  });
43
- require_vi_2VT5v0um.it("status.sh calls docker compose ps", () => {
43
+ require_test_CTcmp4Su.it("status.sh calls docker compose ps", () => {
44
44
  const result = require_generators_scripts.generateScripts();
45
- require_vi_2VT5v0um.globalExpect(result["scripts/status.sh"]).toContain("docker compose");
46
- require_vi_2VT5v0um.globalExpect(result["scripts/status.sh"]).toContain("ps");
45
+ require_test_CTcmp4Su.globalExpect(result["scripts/status.sh"]).toContain("docker compose");
46
+ require_test_CTcmp4Su.globalExpect(result["scripts/status.sh"]).toContain("ps");
47
47
  });
48
- require_vi_2VT5v0um.it("all bash scripts start with shebang", () => {
48
+ require_test_CTcmp4Su.it("all bash scripts start with shebang", () => {
49
49
  const result = require_generators_scripts.generateScripts();
50
- for (const [path, content] of Object.entries(result)) if (path.endsWith(".sh")) require_vi_2VT5v0um.globalExpect(content.startsWith("#!/")).toBe(true);
50
+ for (const [path, content] of Object.entries(result)) if (path.endsWith(".sh")) require_test_CTcmp4Su.globalExpect(content.startsWith("#!/")).toBe(true);
51
51
  });
52
- require_vi_2VT5v0um.it("all PowerShell scripts start with #Requires", () => {
52
+ require_test_CTcmp4Su.it("all PowerShell scripts start with #Requires", () => {
53
53
  const result = require_generators_scripts.generateScripts();
54
- for (const [path, content] of Object.entries(result)) if (path.endsWith(".ps1")) require_vi_2VT5v0um.globalExpect(content.startsWith("#Requires")).toBe(true);
54
+ for (const [path, content] of Object.entries(result)) if (path.endsWith(".ps1")) require_test_CTcmp4Su.globalExpect(content.startsWith("#Requires")).toBe(true);
55
55
  });
56
- require_vi_2VT5v0um.it("start.ps1 calls docker compose up", () => {
56
+ require_test_CTcmp4Su.it("start.ps1 calls docker compose up", () => {
57
57
  const result = require_generators_scripts.generateScripts();
58
- require_vi_2VT5v0um.globalExpect(result["scripts/start.ps1"]).toContain("docker compose");
59
- require_vi_2VT5v0um.globalExpect(result["scripts/start.ps1"]).toContain("up");
58
+ require_test_CTcmp4Su.globalExpect(result["scripts/start.ps1"]).toContain("docker compose");
59
+ require_test_CTcmp4Su.globalExpect(result["scripts/start.ps1"]).toContain("up");
60
60
  });
61
- require_vi_2VT5v0um.it("stop.ps1 calls docker compose down", () => {
61
+ require_test_CTcmp4Su.it("stop.ps1 calls docker compose down", () => {
62
62
  const result = require_generators_scripts.generateScripts();
63
- require_vi_2VT5v0um.globalExpect(result["scripts/stop.ps1"]).toContain("docker compose");
64
- require_vi_2VT5v0um.globalExpect(result["scripts/stop.ps1"]).toContain("down");
63
+ require_test_CTcmp4Su.globalExpect(result["scripts/stop.ps1"]).toContain("docker compose");
64
+ require_test_CTcmp4Su.globalExpect(result["scripts/stop.ps1"]).toContain("down");
65
65
  });
66
- require_vi_2VT5v0um.it("update.ps1 calls docker compose pull", () => {
66
+ require_test_CTcmp4Su.it("update.ps1 calls docker compose pull", () => {
67
67
  const result = require_generators_scripts.generateScripts();
68
- require_vi_2VT5v0um.globalExpect(result["scripts/update.ps1"]).toContain("docker compose");
69
- require_vi_2VT5v0um.globalExpect(result["scripts/update.ps1"]).toContain("pull");
68
+ require_test_CTcmp4Su.globalExpect(result["scripts/update.ps1"]).toContain("docker compose");
69
+ require_test_CTcmp4Su.globalExpect(result["scripts/update.ps1"]).toContain("pull");
70
70
  });
71
- require_vi_2VT5v0um.it("backup.ps1 references volumes or backup", () => {
71
+ require_test_CTcmp4Su.it("backup.ps1 references volumes or backup", () => {
72
72
  const backup = require_generators_scripts.generateScripts()["scripts/backup.ps1"];
73
- require_vi_2VT5v0um.globalExpect(backup).toBeDefined();
74
- require_vi_2VT5v0um.globalExpect(backup.length).toBeGreaterThan(50);
73
+ require_test_CTcmp4Su.globalExpect(backup).toBeDefined();
74
+ require_test_CTcmp4Su.globalExpect(backup.length).toBeGreaterThan(50);
75
75
  });
76
- require_vi_2VT5v0um.it("status.ps1 calls docker compose ps", () => {
76
+ require_test_CTcmp4Su.it("status.ps1 calls docker compose ps", () => {
77
77
  const result = require_generators_scripts.generateScripts();
78
- require_vi_2VT5v0um.globalExpect(result["scripts/status.ps1"]).toContain("docker compose");
79
- require_vi_2VT5v0um.globalExpect(result["scripts/status.ps1"]).toContain("ps");
78
+ require_test_CTcmp4Su.globalExpect(result["scripts/status.ps1"]).toContain("docker compose");
79
+ require_test_CTcmp4Su.globalExpect(result["scripts/status.ps1"]).toContain("ps");
80
80
  });
81
81
  });
82
82
  //#endregion
@@ -1 +1 @@
1
- {"version":3,"file":"scripts.test.cjs","names":["describe","generateScripts"],"sources":["../../src/generators/scripts.test.ts"],"sourcesContent":["import { describe, expect, it } from \"vitest\";\nimport { generateScripts } from \"./scripts.js\";\n\ndescribe(\"generateScripts\", () => {\n\tit(\"generates all 10 expected scripts (5 bash + 5 PowerShell)\", () => {\n\t\tconst result = generateScripts();\n\n\t\tconst expectedScripts = [\n\t\t\t\"scripts/start.sh\",\n\t\t\t\"scripts/stop.sh\",\n\t\t\t\"scripts/update.sh\",\n\t\t\t\"scripts/backup.sh\",\n\t\t\t\"scripts/status.sh\",\n\t\t\t\"scripts/start.ps1\",\n\t\t\t\"scripts/stop.ps1\",\n\t\t\t\"scripts/update.ps1\",\n\t\t\t\"scripts/backup.ps1\",\n\t\t\t\"scripts/status.ps1\",\n\t\t];\n\n\t\tfor (const script of expectedScripts) {\n\t\t\texpect(result).toHaveProperty(script);\n\t\t\texpect(result[script]!.length).toBeGreaterThan(0);\n\t\t}\n\t});\n\n\tit(\"start.sh calls docker compose up\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/start.sh\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/start.sh\"]).toContain(\"up\");\n\t});\n\n\tit(\"stop.sh calls docker compose down\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/stop.sh\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/stop.sh\"]).toContain(\"down\");\n\t});\n\n\tit(\"update.sh calls docker compose pull\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/update.sh\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/update.sh\"]).toContain(\"pull\");\n\t});\n\n\tit(\"backup.sh references volumes or backup\", () => {\n\t\tconst result = generateScripts();\n\t\tconst backup = result[\"scripts/backup.sh\"]!;\n\t\texpect(backup).toBeDefined();\n\t\texpect(backup.length).toBeGreaterThan(50);\n\t});\n\n\tit(\"status.sh calls docker compose ps\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/status.sh\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/status.sh\"]).toContain(\"ps\");\n\t});\n\n\tit(\"all bash scripts start with shebang\", () => {\n\t\tconst result = generateScripts();\n\n\t\tfor (const [path, content] of Object.entries(result)) {\n\t\t\tif (path.endsWith(\".sh\")) {\n\t\t\t\texpect(content.startsWith(\"#!/\")).toBe(true);\n\t\t\t}\n\t\t}\n\t});\n\n\tit(\"all PowerShell scripts start with #Requires\", () => {\n\t\tconst result = generateScripts();\n\n\t\tfor (const [path, content] of Object.entries(result)) {\n\t\t\tif (path.endsWith(\".ps1\")) {\n\t\t\t\texpect(content.startsWith(\"#Requires\")).toBe(true);\n\t\t\t}\n\t\t}\n\t});\n\n\tit(\"start.ps1 calls docker compose up\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/start.ps1\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/start.ps1\"]).toContain(\"up\");\n\t});\n\n\tit(\"stop.ps1 calls docker compose down\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/stop.ps1\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/stop.ps1\"]).toContain(\"down\");\n\t});\n\n\tit(\"update.ps1 calls docker compose pull\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/update.ps1\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/update.ps1\"]).toContain(\"pull\");\n\t});\n\n\tit(\"backup.ps1 references volumes or backup\", () => {\n\t\tconst result = generateScripts();\n\t\tconst backup = result[\"scripts/backup.ps1\"]!;\n\t\texpect(backup).toBeDefined();\n\t\texpect(backup.length).toBeGreaterThan(50);\n\t});\n\n\tit(\"status.ps1 calls docker compose ps\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/status.ps1\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/status.ps1\"]).toContain(\"ps\");\n\t});\n});\n"],"mappings":";;;AAGAA,oBAAAA,SAAS,yBAAyB;AACjC,qBAAA,GAAG,mEAAmE;EACrE,MAAM,SAASC,2BAAAA,iBAAiB;AAehC,OAAK,MAAM,UAba;GACvB;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA,EAEqC;AACrC,uBAAA,aAAO,OAAO,CAAC,eAAe,OAAO;AACrC,uBAAA,aAAO,OAAO,QAAS,OAAO,CAAC,gBAAgB,EAAE;;GAEjD;AAEF,qBAAA,GAAG,0CAA0C;EAC5C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,sBAAA,aAAO,OAAO,oBAAoB,CAAC,UAAU,iBAAiB;AAC9D,sBAAA,aAAO,OAAO,oBAAoB,CAAC,UAAU,KAAK;GACjD;AAEF,qBAAA,GAAG,2CAA2C;EAC7C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,sBAAA,aAAO,OAAO,mBAAmB,CAAC,UAAU,iBAAiB;AAC7D,sBAAA,aAAO,OAAO,mBAAmB,CAAC,UAAU,OAAO;GAClD;AAEF,qBAAA,GAAG,6CAA6C;EAC/C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,sBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,iBAAiB;AAC/D,sBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,OAAO;GACpD;AAEF,qBAAA,GAAG,gDAAgD;EAElD,MAAM,SADSA,2BAAAA,iBAAiB,CACV;AACtB,sBAAA,aAAO,OAAO,CAAC,aAAa;AAC5B,sBAAA,aAAO,OAAO,OAAO,CAAC,gBAAgB,GAAG;GACxC;AAEF,qBAAA,GAAG,2CAA2C;EAC7C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,sBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,iBAAiB;AAC/D,sBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,KAAK;GAClD;AAEF,qBAAA,GAAG,6CAA6C;EAC/C,MAAM,SAASA,2BAAAA,iBAAiB;AAEhC,OAAK,MAAM,CAAC,MAAM,YAAY,OAAO,QAAQ,OAAO,CACnD,KAAI,KAAK,SAAS,MAAM,CACvB,qBAAA,aAAO,QAAQ,WAAW,MAAM,CAAC,CAAC,KAAK,KAAK;GAG7C;AAEF,qBAAA,GAAG,qDAAqD;EACvD,MAAM,SAASA,2BAAAA,iBAAiB;AAEhC,OAAK,MAAM,CAAC,MAAM,YAAY,OAAO,QAAQ,OAAO,CACnD,KAAI,KAAK,SAAS,OAAO,CACxB,qBAAA,aAAO,QAAQ,WAAW,YAAY,CAAC,CAAC,KAAK,KAAK;GAGnD;AAEF,qBAAA,GAAG,2CAA2C;EAC7C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,sBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,iBAAiB;AAC/D,sBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,KAAK;GAClD;AAEF,qBAAA,GAAG,4CAA4C;EAC9C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,sBAAA,aAAO,OAAO,oBAAoB,CAAC,UAAU,iBAAiB;AAC9D,sBAAA,aAAO,OAAO,oBAAoB,CAAC,UAAU,OAAO;GACnD;AAEF,qBAAA,GAAG,8CAA8C;EAChD,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,sBAAA,aAAO,OAAO,sBAAsB,CAAC,UAAU,iBAAiB;AAChE,sBAAA,aAAO,OAAO,sBAAsB,CAAC,UAAU,OAAO;GACrD;AAEF,qBAAA,GAAG,iDAAiD;EAEnD,MAAM,SADSA,2BAAAA,iBAAiB,CACV;AACtB,sBAAA,aAAO,OAAO,CAAC,aAAa;AAC5B,sBAAA,aAAO,OAAO,OAAO,CAAC,gBAAgB,GAAG;GACxC;AAEF,qBAAA,GAAG,4CAA4C;EAC9C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,sBAAA,aAAO,OAAO,sBAAsB,CAAC,UAAU,iBAAiB;AAChE,sBAAA,aAAO,OAAO,sBAAsB,CAAC,UAAU,KAAK;GACnD;EACD"}
1
+ {"version":3,"file":"scripts.test.cjs","names":["describe","generateScripts"],"sources":["../../src/generators/scripts.test.ts"],"sourcesContent":["import { describe, expect, it } from \"vitest\";\nimport { generateScripts } from \"./scripts.js\";\n\ndescribe(\"generateScripts\", () => {\n\tit(\"generates all 10 expected scripts (5 bash + 5 PowerShell)\", () => {\n\t\tconst result = generateScripts();\n\n\t\tconst expectedScripts = [\n\t\t\t\"scripts/start.sh\",\n\t\t\t\"scripts/stop.sh\",\n\t\t\t\"scripts/update.sh\",\n\t\t\t\"scripts/backup.sh\",\n\t\t\t\"scripts/status.sh\",\n\t\t\t\"scripts/start.ps1\",\n\t\t\t\"scripts/stop.ps1\",\n\t\t\t\"scripts/update.ps1\",\n\t\t\t\"scripts/backup.ps1\",\n\t\t\t\"scripts/status.ps1\",\n\t\t];\n\n\t\tfor (const script of expectedScripts) {\n\t\t\texpect(result).toHaveProperty(script);\n\t\t\texpect(result[script]!.length).toBeGreaterThan(0);\n\t\t}\n\t});\n\n\tit(\"start.sh calls docker compose up\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/start.sh\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/start.sh\"]).toContain(\"up\");\n\t});\n\n\tit(\"stop.sh calls docker compose down\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/stop.sh\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/stop.sh\"]).toContain(\"down\");\n\t});\n\n\tit(\"update.sh calls docker compose pull\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/update.sh\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/update.sh\"]).toContain(\"pull\");\n\t});\n\n\tit(\"backup.sh references volumes or backup\", () => {\n\t\tconst result = generateScripts();\n\t\tconst backup = result[\"scripts/backup.sh\"]!;\n\t\texpect(backup).toBeDefined();\n\t\texpect(backup.length).toBeGreaterThan(50);\n\t});\n\n\tit(\"status.sh calls docker compose ps\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/status.sh\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/status.sh\"]).toContain(\"ps\");\n\t});\n\n\tit(\"all bash scripts start with shebang\", () => {\n\t\tconst result = generateScripts();\n\n\t\tfor (const [path, content] of Object.entries(result)) {\n\t\t\tif (path.endsWith(\".sh\")) {\n\t\t\t\texpect(content.startsWith(\"#!/\")).toBe(true);\n\t\t\t}\n\t\t}\n\t});\n\n\tit(\"all PowerShell scripts start with #Requires\", () => {\n\t\tconst result = generateScripts();\n\n\t\tfor (const [path, content] of Object.entries(result)) {\n\t\t\tif (path.endsWith(\".ps1\")) {\n\t\t\t\texpect(content.startsWith(\"#Requires\")).toBe(true);\n\t\t\t}\n\t\t}\n\t});\n\n\tit(\"start.ps1 calls docker compose up\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/start.ps1\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/start.ps1\"]).toContain(\"up\");\n\t});\n\n\tit(\"stop.ps1 calls docker compose down\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/stop.ps1\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/stop.ps1\"]).toContain(\"down\");\n\t});\n\n\tit(\"update.ps1 calls docker compose pull\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/update.ps1\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/update.ps1\"]).toContain(\"pull\");\n\t});\n\n\tit(\"backup.ps1 references volumes or backup\", () => {\n\t\tconst result = generateScripts();\n\t\tconst backup = result[\"scripts/backup.ps1\"]!;\n\t\texpect(backup).toBeDefined();\n\t\texpect(backup.length).toBeGreaterThan(50);\n\t});\n\n\tit(\"status.ps1 calls docker compose ps\", () => {\n\t\tconst result = generateScripts();\n\t\texpect(result[\"scripts/status.ps1\"]).toContain(\"docker compose\");\n\t\texpect(result[\"scripts/status.ps1\"]).toContain(\"ps\");\n\t});\n});\n"],"mappings":";;;AAGAA,sBAAAA,SAAS,yBAAyB;AACjC,uBAAA,GAAG,mEAAmE;EACrE,MAAM,SAASC,2BAAAA,iBAAiB;AAehC,OAAK,MAAM,UAba;GACvB;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA;GACA,EAEqC;AACrC,yBAAA,aAAO,OAAO,CAAC,eAAe,OAAO;AACrC,yBAAA,aAAO,OAAO,QAAS,OAAO,CAAC,gBAAgB,EAAE;;GAEjD;AAEF,uBAAA,GAAG,0CAA0C;EAC5C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,wBAAA,aAAO,OAAO,oBAAoB,CAAC,UAAU,iBAAiB;AAC9D,wBAAA,aAAO,OAAO,oBAAoB,CAAC,UAAU,KAAK;GACjD;AAEF,uBAAA,GAAG,2CAA2C;EAC7C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,wBAAA,aAAO,OAAO,mBAAmB,CAAC,UAAU,iBAAiB;AAC7D,wBAAA,aAAO,OAAO,mBAAmB,CAAC,UAAU,OAAO;GAClD;AAEF,uBAAA,GAAG,6CAA6C;EAC/C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,wBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,iBAAiB;AAC/D,wBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,OAAO;GACpD;AAEF,uBAAA,GAAG,gDAAgD;EAElD,MAAM,SADSA,2BAAAA,iBAAiB,CACV;AACtB,wBAAA,aAAO,OAAO,CAAC,aAAa;AAC5B,wBAAA,aAAO,OAAO,OAAO,CAAC,gBAAgB,GAAG;GACxC;AAEF,uBAAA,GAAG,2CAA2C;EAC7C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,wBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,iBAAiB;AAC/D,wBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,KAAK;GAClD;AAEF,uBAAA,GAAG,6CAA6C;EAC/C,MAAM,SAASA,2BAAAA,iBAAiB;AAEhC,OAAK,MAAM,CAAC,MAAM,YAAY,OAAO,QAAQ,OAAO,CACnD,KAAI,KAAK,SAAS,MAAM,CACvB,uBAAA,aAAO,QAAQ,WAAW,MAAM,CAAC,CAAC,KAAK,KAAK;GAG7C;AAEF,uBAAA,GAAG,qDAAqD;EACvD,MAAM,SAASA,2BAAAA,iBAAiB;AAEhC,OAAK,MAAM,CAAC,MAAM,YAAY,OAAO,QAAQ,OAAO,CACnD,KAAI,KAAK,SAAS,OAAO,CACxB,uBAAA,aAAO,QAAQ,WAAW,YAAY,CAAC,CAAC,KAAK,KAAK;GAGnD;AAEF,uBAAA,GAAG,2CAA2C;EAC7C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,wBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,iBAAiB;AAC/D,wBAAA,aAAO,OAAO,qBAAqB,CAAC,UAAU,KAAK;GAClD;AAEF,uBAAA,GAAG,4CAA4C;EAC9C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,wBAAA,aAAO,OAAO,oBAAoB,CAAC,UAAU,iBAAiB;AAC9D,wBAAA,aAAO,OAAO,oBAAoB,CAAC,UAAU,OAAO;GACnD;AAEF,uBAAA,GAAG,8CAA8C;EAChD,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,wBAAA,aAAO,OAAO,sBAAsB,CAAC,UAAU,iBAAiB;AAChE,wBAAA,aAAO,OAAO,sBAAsB,CAAC,UAAU,OAAO;GACrD;AAEF,uBAAA,GAAG,iDAAiD;EAEnD,MAAM,SADSA,2BAAAA,iBAAiB,CACV;AACtB,wBAAA,aAAO,OAAO,CAAC,aAAa;AAC5B,wBAAA,aAAO,OAAO,OAAO,CAAC,gBAAgB,GAAG;GACxC;AAEF,uBAAA,GAAG,4CAA4C;EAC9C,MAAM,SAASA,2BAAAA,iBAAiB;AAChC,wBAAA,aAAO,OAAO,sBAAsB,CAAC,UAAU,iBAAiB;AAChE,wBAAA,aAAO,OAAO,sBAAsB,CAAC,UAAU,KAAK;GACnD;EACD"}
@@ -1,4 +1,4 @@
1
- import { n as describe, r as it, t as globalExpect } from "../vi.2VT5v0um-C_jmO7m2.mjs";
1
+ import { n as describe, r as it, t as globalExpect } from "../test.CTcmp4Su-ClCHJ3FA.mjs";
2
2
  import { generateScripts } from "./scripts.mjs";
3
3
  //#region src/generators/scripts.test.ts
4
4
  describe("generateScripts", () => {
@@ -1,3 +1,3 @@
1
1
  Object.defineProperty(exports, Symbol.toStringTag, { value: "Module" });
2
- const require_skills = require("../skills-BlzpHmpH.cjs");
2
+ const require_skills = require("../skills-BSF7iNa4.cjs");
3
3
  exports.generateSkillFiles = require_skills.generateSkillFiles;
@@ -1 +1 @@
1
- {"version":3,"file":"skills.d.cts","names":[],"sources":["../../src/generators/skills.ts"],"mappings":";;;;;AAmiBA;;;;;iBAAgB,kBAAA,CAAmB,QAAA,EAAU,cAAA,GAAiB,MAAA"}
1
+ {"version":3,"file":"skills.d.cts","names":[],"sources":["../../src/generators/skills.ts"],"mappings":";;;;;AAirBA;;;;;iBAAgB,kBAAA,CAAmB,QAAA,EAAU,cAAA,GAAiB,MAAA"}
@@ -1 +1 @@
1
- {"version":3,"file":"skills.d.mts","names":[],"sources":["../../src/generators/skills.ts"],"mappings":";;;;;;AAmiBA;;;;iBAAgB,kBAAA,CAAmB,QAAA,EAAU,cAAA,GAAiB,MAAA"}
1
+ {"version":3,"file":"skills.d.mts","names":[],"sources":["../../src/generators/skills.ts"],"mappings":";;;;;;AAirBA;;;;iBAAgB,kBAAA,CAAmB,QAAA,EAAU,cAAA,GAAiB,MAAA"}
@@ -523,6 +523,147 @@ curl -X POST http://{{STEEL_HOST}}:{{STEEL_PORT}}/v1/scrape \\
523
523
  - Proxy support and IP rotation
524
524
  - Auto CAPTCHA solving
525
525
  - Puppeteer/Playwright/Selenium compatible
526
+ `,
527
+ "code-sandbox": `---
528
+ name: code-sandbox
529
+ description: "Execute code safely in an isolated OpenSandbox container"
530
+ metadata:
531
+ openclaw:
532
+ emoji: "📦"
533
+ ---
534
+
535
+ # Code Sandbox
536
+
537
+ Execute code safely in an isolated OpenSandbox container.
538
+
539
+ ## Description
540
+
541
+ This skill provides secure, containerized code execution for AI agents. Code runs in ephemeral Docker containers with resource limits, network isolation, and automatic cleanup.
542
+
543
+ ## Connection Details
544
+
545
+ - **Host:** \`{{OPENSANDBOX_HOST}}\`
546
+ - **Port:** \`{{OPENSANDBOX_PORT}}\`
547
+ - **Auth:** API key (auto-configured)
548
+
549
+ ## Supported Languages
550
+
551
+ - Python 3.12
552
+ - JavaScript / TypeScript (Node.js 22)
553
+ - Java 21
554
+ - Go 1.24
555
+ - Bash
556
+
557
+ ## Available Actions
558
+
559
+ ### execute_code
560
+
561
+ Run a code snippet in a fresh sandbox.
562
+
563
+ **Parameters:**
564
+ - \`language\` (required): Programming language ("python", "javascript", "typescript", "java", "go", "bash")
565
+ - \`code\` (required): The code to execute
566
+ - \`timeout_seconds\` (optional): Max execution time (default: 60, max: 300)
567
+
568
+ **Returns:** stdout, stderr, exit_code, execution_time_ms
569
+
570
+ ### execute_shell
571
+
572
+ Run a shell command in an existing or new sandbox.
573
+
574
+ **Parameters:**
575
+ - \`command\` (required): Shell command to execute
576
+ - \`sandbox_id\` (optional): Reuse an existing sandbox (for multi-step workflows)
577
+ - \`background\` (optional): Run in background (default: false)
578
+
579
+ **Returns:** stdout, stderr, exit_code
580
+
581
+ ### upload_file
582
+
583
+ Upload a file to a sandbox for processing.
584
+
585
+ **Parameters:**
586
+ - \`sandbox_id\` (required): Target sandbox
587
+ - \`path\` (required): Destination path inside sandbox
588
+ - \`content\` (required): File content (text or base64 for binary)
589
+
590
+ ### download_file
591
+
592
+ Download a file from a sandbox.
593
+
594
+ **Parameters:**
595
+ - \`sandbox_id\` (required): Source sandbox
596
+ - \`path\` (required): File path inside sandbox
597
+
598
+ **Returns:** File content
599
+
600
+ ### list_sandboxes
601
+
602
+ List active sandboxes on this instance.
603
+
604
+ **Returns:** Array of { id, status, image, created_at, expires_at }
605
+
606
+ ### terminate_sandbox
607
+
608
+ Terminate a running sandbox immediately.
609
+
610
+ **Parameters:**
611
+ - \`sandbox_id\` (required): Sandbox to terminate
612
+
613
+ ### create_desktop
614
+
615
+ Create a GUI desktop sandbox with VNC access (for Homespace live preview).
616
+
617
+ **Parameters:**
618
+ - \`image\` (optional): Desktop image (default: "opensandbox/desktop:latest", also: "opensandbox/chrome:latest", "opensandbox/vscode:latest")
619
+ - \`resolution\` (optional): Screen resolution (default: "1280x800x24")
620
+
621
+ **Returns:** sandbox_id, vnc_endpoint (port 5900), novnc_url (port 6080 WebSocket), devtools_url (port 9222, chrome only)
622
+
623
+ ### get_preview_url
624
+
625
+ Get the browser-accessible noVNC URL for an existing desktop sandbox.
626
+
627
+ **Parameters:**
628
+ - \`sandbox_id\` (required): Desktop sandbox ID
629
+
630
+ **Returns:** novnc_url (embeddable in iframe), vnc_endpoint, status
631
+
632
+ ## Examples
633
+
634
+ ### Run Python code
635
+
636
+ \`\`\`bash
637
+ curl -X POST http://{{OPENSANDBOX_HOST}}:{{OPENSANDBOX_PORT}}/v1/sandboxes \\
638
+ -H "Authorization: Bearer $OPENSANDBOX_API_KEY" \\
639
+ -H "Content-Type: application/json" \\
640
+ -d '{"image": "opensandbox/code-interpreter:python"}'
641
+ \`\`\`
642
+
643
+ ### Execute code in a sandbox
644
+
645
+ \`\`\`bash
646
+ curl -X POST http://{{OPENSANDBOX_HOST}}:{{OPENSANDBOX_PORT}}/v1/sandboxes/{id}/code \\
647
+ -H "Authorization: Bearer $OPENSANDBOX_API_KEY" \\
648
+ -H "Content-Type: application/json" \\
649
+ -d '{"language": "python", "code": "print(42 * 42)"}'
650
+ \`\`\`
651
+
652
+ ## Configuration
653
+
654
+ - **Default timeout:** 60 seconds
655
+ - **Max concurrent sandboxes:** Determined by VPS RAM
656
+ - **Idle cleanup:** Sandboxes with no activity for 30 minutes are auto-terminated
657
+ - **Network:** Bridge mode (isolated from host services)
658
+ - **Security:** gVisor runtime, capability dropping, PID limits
659
+
660
+ ## Limitations
661
+
662
+ - No persistent storage between sandbox sessions (ephemeral by design)
663
+ - No GPU access (CPU-only execution)
664
+ - No outbound network access by default (egress blocked)
665
+ - Max 512 PIDs per sandbox (fork bomb protection)
666
+ - Memory capped per sandbox (default 512MB)
526
667
  `
527
668
  };
528
669
  /**
@@ -1 +1 @@
1
- {"version":3,"file":"skills.mjs","names":[],"sources":["../../src/generators/skills.ts"],"sourcesContent":["import Handlebars from \"handlebars\";\nimport type { ResolverOutput } from \"../types.js\";\n\n/**\n * Basic skill templates keyed by skillId.\n *\n * Templates use Handlebars `{{VAR}}` syntax. Variables are resolved from the\n * service's `openclawEnvVars` array at generation time.\n */\nconst skillTemplates: Record<string, string> = {\n\t\"redis-cache\": `---\nname: redis-cache\ndescription: \"Cache data and manage key-value state using Redis\"\nmetadata:\n openclaw:\n emoji: \"🔴\"\n---\n\n# Redis Cache\n\nUse Redis as a high-performance in-memory cache and key-value store.\n\n## Connection Details\n\n- **Host:** \\`{{REDIS_HOST}}\\`\n- **Port:** \\`{{REDIS_PORT}}\\`\n\n## Example Commands\n\n### Set a value\n\\`\\`\\`bash\nredis-cli -h {{REDIS_HOST}} -p {{REDIS_PORT}} -a $REDIS_PASSWORD SET mykey \"myvalue\"\n\\`\\`\\`\n\n### Get a value\n\\`\\`\\`bash\nredis-cli -h {{REDIS_HOST}} -p {{REDIS_PORT}} -a $REDIS_PASSWORD GET mykey\n\\`\\`\\`\n\n### List all keys\n\\`\\`\\`bash\nredis-cli -h {{REDIS_HOST}} -p {{REDIS_PORT}} -a $REDIS_PASSWORD KEYS \"*\"\n\\`\\`\\`\n\n## Usage Notes\n\n- Use Redis for caching frequently accessed data, session storage, and pub/sub messaging.\n- Data is persisted to disk via RDB snapshots in the mounted volume.\n- Password authentication is required.\n`,\n\n\t\"qdrant-memory\": `---\nname: qdrant-memory\ndescription: \"Store and retrieve vector embeddings for semantic search and RAG\"\nmetadata:\n openclaw:\n emoji: \"🧠\"\n---\n\n# Qdrant Memory\n\nUse Qdrant as a vector database for storing embeddings, enabling semantic search and retrieval-augmented generation (RAG).\n\n## Connection Details\n\n- **Host:** \\`{{QDRANT_HOST}}\\`\n- **Port:** \\`{{QDRANT_PORT}}\\` (REST API)\n\n## Example API Calls\n\n### Create a collection\n\\`\\`\\`bash\ncurl -X PUT \"http://{{QDRANT_HOST}}:{{QDRANT_PORT}}/collections/my_collection\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"vectors\": {\"size\": 384, \"distance\": \"Cosine\"}}'\n\\`\\`\\`\n\n### Upsert points\n\\`\\`\\`bash\ncurl -X PUT \"http://{{QDRANT_HOST}}:{{QDRANT_PORT}}/collections/my_collection/points\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"points\": [{\"id\": 1, \"vector\": [0.1, 0.2, ...], \"payload\": {\"text\": \"hello\"}}]}'\n\\`\\`\\`\n\n### Search similar vectors\n\\`\\`\\`bash\ncurl -X POST \"http://{{QDRANT_HOST}}:{{QDRANT_PORT}}/collections/my_collection/points/search\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"vector\": [0.1, 0.2, ...], \"limit\": 5}'\n\\`\\`\\`\n\n## Usage Notes\n\n- Use 384-dimensional vectors for MiniLM-based embeddings or 1536 for OpenAI ada-002.\n- Qdrant supports filtering, batch operations, and named vectors.\n`,\n\n\t\"n8n-trigger\": `---\nname: n8n-trigger\ndescription: \"Trigger and manage automation workflows using n8n\"\nmetadata:\n openclaw:\n emoji: \"🔄\"\n---\n\n# n8n Workflow Trigger\n\nUse n8n to create, trigger, and manage automation workflows via its REST API.\n\n## Connection Details\n\n- **Host:** \\`{{N8N_HOST}}\\`\n- **Port:** \\`{{N8N_PORT}}\\`\n- **Webhook URL:** \\`{{N8N_WEBHOOK_URL}}\\`\n\n## Example API Calls\n\n### Trigger a webhook workflow\n\\`\\`\\`bash\ncurl -X POST \"{{N8N_WEBHOOK_URL}}webhook/<your-webhook-id>\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"message\": \"Hello from OpenClaw\"}'\n\\`\\`\\`\n\n### List workflows\n\\`\\`\\`bash\ncurl -X GET \"http://{{N8N_HOST}}:{{N8N_PORT}}/api/v1/workflows\" \\\\\n -H \"Content-Type: application/json\"\n\\`\\`\\`\n\n### Activate a workflow\n\\`\\`\\`bash\ncurl -X PATCH \"http://{{N8N_HOST}}:{{N8N_PORT}}/api/v1/workflows/<id>\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"active\": true}'\n\\`\\`\\`\n\n## Usage Notes\n\n- Create workflows in the n8n UI at http://{{N8N_HOST}}:{{N8N_PORT}}.\n- Webhook nodes allow external triggers from OpenClaw.\n- n8n stores workflow state in PostgreSQL.\n`,\n\n\t\"ffmpeg-process\": `---\nname: ffmpeg-process\ndescription: \"Process media files using FFmpeg for transcoding, conversion, and manipulation\"\nmetadata:\n openclaw:\n emoji: \"🎬\"\n---\n\n# FFmpeg Media Processing\n\nUse FFmpeg for video/audio transcoding, format conversion, and media manipulation.\n\n## Shared Directory\n\n- **Media directory:** \\`{{FFMPEG_SHARED_DIR}}\\`\n\n## Example Commands\n\n### Convert video to MP4\n\\`\\`\\`bash\ndocker exec ffmpeg ffmpeg -i /data/input.avi -c:v libx264 -c:a aac /data/output.mp4\n\\`\\`\\`\n\n### Extract audio from video\n\\`\\`\\`bash\ndocker exec ffmpeg ffmpeg -i /data/video.mp4 -vn -acodec libmp3lame /data/audio.mp3\n\\`\\`\\`\n\n### Create thumbnail from video\n\\`\\`\\`bash\ndocker exec ffmpeg ffmpeg -i /data/video.mp4 -ss 00:00:05 -vframes 1 /data/thumb.jpg\n\\`\\`\\`\n\n### Resize video\n\\`\\`\\`bash\ndocker exec ffmpeg ffmpeg -i /data/input.mp4 -vf scale=1280:720 /data/output_720p.mp4\n\\`\\`\\`\n\n## Usage Notes\n\n- Place input files in the shared media directory.\n- FFmpeg runs as a sidecar container sharing a volume with OpenClaw.\n- Output files appear in the same shared directory.\n`,\n\n\t\"minio-storage\": `---\nname: minio-storage\ndescription: \"Store and retrieve files using S3-compatible object storage\"\nmetadata:\n openclaw:\n emoji: \"💾\"\n---\n\n# MinIO Object Storage\n\nUse MinIO as S3-compatible object storage for files, assets, and backups.\n\n## Connection Details\n\n- **Host:** \\`{{MINIO_HOST}}\\`\n- **API Port:** \\`{{MINIO_PORT}}\\`\n- **Access Key:** Uses \\`MINIO_ACCESS_KEY\\` env var\n- **Secret Key:** Uses \\`MINIO_SECRET_KEY\\` env var\n\n## Example API Calls\n\n### Create a bucket\n\\`\\`\\`bash\nmc alias set local http://{{MINIO_HOST}}:{{MINIO_PORT}} $MINIO_ACCESS_KEY $MINIO_SECRET_KEY\nmc mb local/my-bucket\n\\`\\`\\`\n\n### Upload a file\n\\`\\`\\`bash\nmc cp /path/to/file.txt local/my-bucket/file.txt\n\\`\\`\\`\n\n### Download a file\n\\`\\`\\`bash\nmc cp local/my-bucket/file.txt /path/to/local/file.txt\n\\`\\`\\`\n\n### List bucket contents\n\\`\\`\\`bash\nmc ls local/my-bucket/\n\\`\\`\\`\n\n## Usage Notes\n\n- MinIO is fully S3-compatible—use any S3 SDK or CLI.\n- Access the web console at http://{{MINIO_HOST}}:9001 for a visual file browser.\n- Create separate buckets for different data types (assets, backups, uploads).\n`,\n\n\t\"browserless-browse\": `---\nname: browserless-browse\ndescription: \"Automate browser interactions, scrape web pages, and generate PDFs\"\nmetadata:\n openclaw:\n emoji: \"🌐\"\n---\n\n# Browserless Browser Automation\n\nUse Browserless for headless Chrome browser automation, web scraping, screenshots, and PDF generation.\n\n## Connection Details\n\n- **Host:** \\`{{BROWSERLESS_HOST}}\\`\n- **Port:** \\`{{BROWSERLESS_PORT}}\\`\n- **Token:** Uses \\`BROWSERLESS_TOKEN\\` env var\n\n## Example API Calls\n\n### Take a screenshot\n\\`\\`\\`bash\ncurl -X POST \"http://{{BROWSERLESS_HOST}}:{{BROWSERLESS_PORT}}/screenshot?token=$BROWSERLESS_TOKEN\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"url\": \"https://example.com\", \"options\": {\"fullPage\": true}}' \\\\\n --output screenshot.png\n\\`\\`\\`\n\n### Generate a PDF\n\\`\\`\\`bash\ncurl -X POST \"http://{{BROWSERLESS_HOST}}:{{BROWSERLESS_PORT}}/pdf?token=$BROWSERLESS_TOKEN\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"url\": \"https://example.com\"}' \\\\\n --output page.pdf\n\\`\\`\\`\n\n### Scrape page content\n\\`\\`\\`bash\ncurl -X POST \"http://{{BROWSERLESS_HOST}}:{{BROWSERLESS_PORT}}/content?token=$BROWSERLESS_TOKEN\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"url\": \"https://example.com\"}'\n\\`\\`\\`\n\n## Usage Notes\n\n- Browserless manages a pool of Chrome instances (max concurrent sessions configured via env).\n- Token authentication is required for all API calls.\n- Supports Puppeteer and Playwright WebSocket connections.\n`,\n\n\t\"searxng-search\": `---\nname: searxng-search\ndescription: \"Search the web using a privacy-respecting metasearch engine\"\nmetadata:\n openclaw:\n emoji: \"🔍\"\n---\n\n# SearXNG Web Search\n\nUse SearXNG as a privacy-respecting metasearch engine to query the web.\n\n## Connection Details\n\n- **Host:** \\`{{SEARXNG_HOST}}\\`\n- **Port:** \\`{{SEARXNG_PORT}}\\`\n\n## Example API Calls\n\n### Search the web\n\\`\\`\\`bash\ncurl \"http://{{SEARXNG_HOST}}:{{SEARXNG_PORT}}/search?q=your+search+query&format=json\"\n\\`\\`\\`\n\n### Search with category filter\n\\`\\`\\`bash\ncurl \"http://{{SEARXNG_HOST}}:{{SEARXNG_PORT}}/search?q=openai&categories=it&format=json\"\n\\`\\`\\`\n\n### Search with language filter\n\\`\\`\\`bash\ncurl \"http://{{SEARXNG_HOST}}:{{SEARXNG_PORT}}/search?q=hello&language=en&format=json\"\n\\`\\`\\`\n\n## Usage Notes\n\n- Always use \\`format=json\\` for machine-readable results.\n- Available categories: general, images, videos, news, map, music, it, science, files.\n- SearXNG aggregates results from many search engines without tracking.\n`,\n\n\t\"whisper-transcribe\": `---\nname: whisper-transcribe\ndescription: \"Transcribe audio files to text using Faster Whisper\"\nmetadata:\n openclaw:\n emoji: \"🎙️\"\n---\n\n# Whisper Transcription\n\nUse the Faster Whisper server for speech-to-text transcription.\n\n## Connection Details\n\n- **Host:** \\`{{WHISPER_HOST}}\\`\n- **Port:** \\`{{WHISPER_PORT}}\\`\n\n## Example API Calls\n\n### Transcribe an audio file\n\\`\\`\\`bash\ncurl -X POST \"http://{{WHISPER_HOST}}:{{WHISPER_PORT}}/v1/audio/transcriptions\" \\\\\n -F \"file=@/path/to/audio.mp3\" \\\\\n -F \"model=base\"\n\\`\\`\\`\n\n### Transcribe with language hint\n\\`\\`\\`bash\ncurl -X POST \"http://{{WHISPER_HOST}}:{{WHISPER_PORT}}/v1/audio/transcriptions\" \\\\\n -F \"file=@/path/to/audio.wav\" \\\\\n -F \"model=base\" \\\\\n -F \"language=en\"\n\\`\\`\\`\n\n### Get available models\n\\`\\`\\`bash\ncurl \"http://{{WHISPER_HOST}}:{{WHISPER_PORT}}/v1/models\"\n\\`\\`\\`\n\n## Usage Notes\n\n- Supports MP3, WAV, FLAC, and other common audio formats.\n- Model sizes: tiny, base, small, medium, large (larger = more accurate but slower).\n- GPU acceleration significantly improves transcription speed.\n`,\n\n\t\"ollama-local-llm\": `---\nname: ollama-local-llm\ndescription: \"Run local language models for text generation, chat, and embeddings\"\nmetadata:\n openclaw:\n emoji: \"🦙\"\n---\n\n# Ollama Local LLM\n\nUse Ollama to run large language models locally for text generation, chat, and embeddings.\n\n## Connection Details\n\n- **Host:** \\`{{OLLAMA_HOST}}\\`\n- **Port:** \\`{{OLLAMA_PORT}}\\`\n\n## Example API Calls\n\n### Generate text\n\\`\\`\\`bash\ncurl -X POST \"http://{{OLLAMA_HOST}}:{{OLLAMA_PORT}}/api/generate\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"model\": \"llama3.2\", \"prompt\": \"Explain quantum computing in simple terms\"}'\n\\`\\`\\`\n\n### Chat completion\n\\`\\`\\`bash\ncurl -X POST \"http://{{OLLAMA_HOST}}:{{OLLAMA_PORT}}/api/chat\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"model\": \"llama3.2\", \"messages\": [{\"role\": \"user\", \"content\": \"Hello!\"}]}'\n\\`\\`\\`\n\n### Pull a model\n\\`\\`\\`bash\ncurl -X POST \"http://{{OLLAMA_HOST}}:{{OLLAMA_PORT}}/api/pull\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"name\": \"llama3.2\"}'\n\\`\\`\\`\n\n### Generate embeddings\n\\`\\`\\`bash\ncurl -X POST \"http://{{OLLAMA_HOST}}:{{OLLAMA_PORT}}/api/embeddings\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"model\": \"nomic-embed-text\", \"prompt\": \"Hello world\"}'\n\\`\\`\\`\n\n## Usage Notes\n\n- Pull models before first use (they are cached in the persistent volume).\n- Recommended models: llama3.2 (general), codellama (code), nomic-embed-text (embeddings).\n- GPU passthrough dramatically improves inference speed.\n- The Ollama API is OpenAI-compatible at /v1/ endpoints.\n`,\n\n\t\"remotion-render\": `---\nname: remotion-render\ndescription: \"Create and render videos programmatically using React\"\nmetadata:\n openclaw:\n emoji: \"🎥\"\n---\n\n# Remotion Video Rendering\n\nUse Remotion Studio to create and render videos programmatically with React.\n\n## Connection Details\n\n- **Host:** \\`{{REMOTION_HOST}}\\`\n- **Port:** \\`{{REMOTION_PORT}}\\`\n\n## Example Usage\n\n### Access the Studio UI\nOpen \\`http://{{REMOTION_HOST}}:{{REMOTION_PORT}}\\` in your browser to use the Remotion Studio visual editor.\n\n### Render a video via CLI\n\\`\\`\\`bash\ndocker exec remotion npx remotion render src/index.tsx MyComposition out/video.mp4\n\\`\\`\\`\n\n### Render with custom props\n\\`\\`\\`bash\ndocker exec remotion npx remotion render src/index.tsx MyComposition out/video.mp4 \\\\\n --props='{\"title\": \"Hello World\"}'\n\\`\\`\\`\n\n## Usage Notes\n\n- Define video compositions in React components.\n- Remotion supports MP4, WebM, and GIF output formats.\n- Use the Studio UI for previewing before rendering.\n- Combine with FFmpeg for post-processing.\n`,\n\n\t\"lightpanda-browse\": `---\nname: lightpanda-browse\ndescription: \"Browse the web using the ultra-fast LightPanda headless browser via CDP\"\nmetadata:\n openclaw:\n emoji: \"🐼\"\n---\n\n# LightPanda Browse\n\nLightPanda is an ultra-fast headless browser available via CDP WebSocket at \\`ws://{{LIGHTPANDA_HOST}}:{{LIGHTPANDA_PORT}}\\`.\n\n## Connect via Puppeteer\n\n\\`\\`\\`javascript\nconst browser = await puppeteer.connect({\n browserWSEndpoint: \"ws://{{LIGHTPANDA_HOST}}:{{LIGHTPANDA_PORT}}\"\n});\nconst page = await browser.newPage();\nawait page.goto('https://example.com');\nconst content = await page.evaluate(() => document.body.innerText);\n\\`\\`\\`\n\n## Key Advantages\n\n- 9x less memory than Chrome (ideal for parallel scraping)\n- 11x faster page loading\n- Instant startup\n- Full CDP compatibility with Puppeteer and Playwright\n`,\n\n\t\"steel-browse\": `---\nname: steel-browse\ndescription: \"Browse the web using Steel Browser API with session management and anti-detection\"\nmetadata:\n openclaw:\n emoji: \"🔥\"\n---\n\n# Steel Browser\n\nSteel provides a REST API at \\`http://{{STEEL_HOST}}:{{STEEL_PORT}}\\` for AI agent web automation.\n\n## Create a Session\n\n\\`\\`\\`bash\ncurl -X POST http://{{STEEL_HOST}}:{{STEEL_PORT}}/v1/sessions \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"blockAds\": true}'\n\\`\\`\\`\n\n## Scrape a Page\n\n\\`\\`\\`bash\ncurl -X POST http://{{STEEL_HOST}}:{{STEEL_PORT}}/v1/scrape \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"url\": \"https://example.com\", \"format\": \"markdown\"}'\n\\`\\`\\`\n\n## Features\n\n- Session management with persistent cookies\n- Anti-detection and stealth plugins\n- Proxy support and IP rotation\n- Auto CAPTCHA solving\n- Puppeteer/Playwright/Selenium compatible\n`,\n};\n\n/**\n * Generates SKILL.md files for each service that has skills defined.\n *\n * Returns a map of file paths (relative to project root) to file contents.\n * Handlebars is used to replace `{{VAR}}` placeholders with actual values\n * from each service's `openclawEnvVars`.\n */\nexport function generateSkillFiles(resolved: ResolverOutput): Record<string, string> {\n\tconst files: Record<string, string> = {};\n\n\tfor (const { definition } of resolved.services) {\n\t\tif (definition.skills.length === 0) continue;\n\n\t\t// Build a variable map from openclawEnvVars for Handlebars\n\t\tconst vars: Record<string, string> = {};\n\t\tfor (const envVar of definition.openclawEnvVars) {\n\t\t\t// Resolve ${REFERENCES} to the default value (just strip the ${} wrapper)\n\t\t\tconst val = envVar.defaultValue.startsWith(\"${\")\n\t\t\t\t? envVar.defaultValue.slice(2, -1)\n\t\t\t\t: envVar.defaultValue;\n\t\t\tvars[envVar.key] = val;\n\t\t}\n\n\t\tfor (const skill of definition.skills) {\n\t\t\tconst template = skillTemplates[skill.skillId];\n\t\t\tif (!template) {\n\t\t\t\t// Generate a generic skill file if no template exists\n\t\t\t\tconst generic = generateGenericSkill(\n\t\t\t\t\tskill.skillId,\n\t\t\t\t\tdefinition.name,\n\t\t\t\t\tdefinition.icon,\n\t\t\t\t\tdefinition.description,\n\t\t\t\t\tvars,\n\t\t\t\t);\n\t\t\t\tfiles[`openclaw/workspace/skills/${skill.skillId}/SKILL.md`] = generic;\n\t\t\t\tcontinue;\n\t\t\t}\n\n\t\t\tconst compiled = Handlebars.compile(template, { noEscape: true });\n\t\t\tconst rendered = compiled(vars);\n\t\t\tfiles[`openclaw/workspace/skills/${skill.skillId}/SKILL.md`] = rendered;\n\t\t}\n\t}\n\n\treturn files;\n}\n\n/**\n * Generate a generic SKILL.md for skills that don't have a dedicated template.\n */\nfunction generateGenericSkill(\n\tskillId: string,\n\tserviceName: string,\n\ticon: string,\n\tdescription: string,\n\tvars: Record<string, string>,\n): string {\n\tconst title = skillId\n\t\t.split(\"-\")\n\t\t.map((w) => w.charAt(0).toUpperCase() + w.slice(1))\n\t\t.join(\" \");\n\n\tconst envSection = Object.entries(vars)\n\t\t.map(([key, value]) => `- **${key}:** \\`${value}\\``)\n\t\t.join(\"\\n\");\n\n\treturn `---\nname: ${skillId}\ndescription: \"${description}\"\nmetadata:\n openclaw:\n emoji: \"${icon}\"\n---\n\n# ${title}\n\nInteract with ${serviceName} through this skill.\n\n## Connection Details\n\n${envSection || \"No specific connection variables configured.\"}\n\n## Usage Notes\n\n- This skill provides OpenClaw access to ${serviceName}.\n- Refer to the service documentation for available API endpoints.\n`;\n}\n"],"mappings":";;;;;;;;AASA,MAAM,iBAAyC;CAC9C,eAAe;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAyCf,iBAAiB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA8CjB,eAAe;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA+Cf,kBAAkB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA6ClB,iBAAiB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAiDjB,sBAAsB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAkDtB,kBAAkB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAyClB,sBAAsB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA8CtB,oBAAoB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAuDpB,mBAAmB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAyCnB,qBAAqB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA+BrB,gBAAgB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAoChB;;;;;;;;AASD,SAAgB,mBAAmB,UAAkD;CACpF,MAAM,QAAgC,EAAE;AAExC,MAAK,MAAM,EAAE,gBAAgB,SAAS,UAAU;AAC/C,MAAI,WAAW,OAAO,WAAW,EAAG;EAGpC,MAAM,OAA+B,EAAE;AACvC,OAAK,MAAM,UAAU,WAAW,iBAAiB;GAEhD,MAAM,MAAM,OAAO,aAAa,WAAW,KAAK,GAC7C,OAAO,aAAa,MAAM,GAAG,GAAG,GAChC,OAAO;AACV,QAAK,OAAO,OAAO;;AAGpB,OAAK,MAAM,SAAS,WAAW,QAAQ;GACtC,MAAM,WAAW,eAAe,MAAM;AACtC,OAAI,CAAC,UAAU;IAEd,MAAM,UAAU,qBACf,MAAM,SACN,WAAW,MACX,WAAW,MACX,WAAW,aACX,KACA;AACD,UAAM,6BAA6B,MAAM,QAAQ,cAAc;AAC/D;;GAID,MAAM,WADW,WAAW,QAAQ,UAAU,EAAE,UAAU,MAAM,CAAC,CACvC,KAAK;AAC/B,SAAM,6BAA6B,MAAM,QAAQ,cAAc;;;AAIjE,QAAO;;;;;AAMR,SAAS,qBACR,SACA,aACA,MACA,aACA,MACS;AAUT,QAAO;QACA,QAAQ;gBACA,YAAY;;;cAGd,KAAK;;;IAdJ,QACZ,MAAM,IAAI,CACV,KAAK,MAAM,EAAE,OAAO,EAAE,CAAC,aAAa,GAAG,EAAE,MAAM,EAAE,CAAC,CAClD,KAAK,IAAI,CAcF;;gBAEM,YAAY;;;;EAdR,OAAO,QAAQ,KAAK,CACrC,KAAK,CAAC,KAAK,WAAW,OAAO,IAAI,QAAQ,MAAM,IAAI,CACnD,KAAK,KAAK,IAgBG,+CAA+C;;;;2CAIpB,YAAY"}
1
+ {"version":3,"file":"skills.mjs","names":[],"sources":["../../src/generators/skills.ts"],"sourcesContent":["import Handlebars from \"handlebars\";\nimport type { ResolverOutput } from \"../types.js\";\n\n/**\n * Basic skill templates keyed by skillId.\n *\n * Templates use Handlebars `{{VAR}}` syntax. Variables are resolved from the\n * service's `openclawEnvVars` array at generation time.\n */\nconst skillTemplates: Record<string, string> = {\n\t\"redis-cache\": `---\nname: redis-cache\ndescription: \"Cache data and manage key-value state using Redis\"\nmetadata:\n openclaw:\n emoji: \"🔴\"\n---\n\n# Redis Cache\n\nUse Redis as a high-performance in-memory cache and key-value store.\n\n## Connection Details\n\n- **Host:** \\`{{REDIS_HOST}}\\`\n- **Port:** \\`{{REDIS_PORT}}\\`\n\n## Example Commands\n\n### Set a value\n\\`\\`\\`bash\nredis-cli -h {{REDIS_HOST}} -p {{REDIS_PORT}} -a $REDIS_PASSWORD SET mykey \"myvalue\"\n\\`\\`\\`\n\n### Get a value\n\\`\\`\\`bash\nredis-cli -h {{REDIS_HOST}} -p {{REDIS_PORT}} -a $REDIS_PASSWORD GET mykey\n\\`\\`\\`\n\n### List all keys\n\\`\\`\\`bash\nredis-cli -h {{REDIS_HOST}} -p {{REDIS_PORT}} -a $REDIS_PASSWORD KEYS \"*\"\n\\`\\`\\`\n\n## Usage Notes\n\n- Use Redis for caching frequently accessed data, session storage, and pub/sub messaging.\n- Data is persisted to disk via RDB snapshots in the mounted volume.\n- Password authentication is required.\n`,\n\n\t\"qdrant-memory\": `---\nname: qdrant-memory\ndescription: \"Store and retrieve vector embeddings for semantic search and RAG\"\nmetadata:\n openclaw:\n emoji: \"🧠\"\n---\n\n# Qdrant Memory\n\nUse Qdrant as a vector database for storing embeddings, enabling semantic search and retrieval-augmented generation (RAG).\n\n## Connection Details\n\n- **Host:** \\`{{QDRANT_HOST}}\\`\n- **Port:** \\`{{QDRANT_PORT}}\\` (REST API)\n\n## Example API Calls\n\n### Create a collection\n\\`\\`\\`bash\ncurl -X PUT \"http://{{QDRANT_HOST}}:{{QDRANT_PORT}}/collections/my_collection\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"vectors\": {\"size\": 384, \"distance\": \"Cosine\"}}'\n\\`\\`\\`\n\n### Upsert points\n\\`\\`\\`bash\ncurl -X PUT \"http://{{QDRANT_HOST}}:{{QDRANT_PORT}}/collections/my_collection/points\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"points\": [{\"id\": 1, \"vector\": [0.1, 0.2, ...], \"payload\": {\"text\": \"hello\"}}]}'\n\\`\\`\\`\n\n### Search similar vectors\n\\`\\`\\`bash\ncurl -X POST \"http://{{QDRANT_HOST}}:{{QDRANT_PORT}}/collections/my_collection/points/search\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"vector\": [0.1, 0.2, ...], \"limit\": 5}'\n\\`\\`\\`\n\n## Usage Notes\n\n- Use 384-dimensional vectors for MiniLM-based embeddings or 1536 for OpenAI ada-002.\n- Qdrant supports filtering, batch operations, and named vectors.\n`,\n\n\t\"n8n-trigger\": `---\nname: n8n-trigger\ndescription: \"Trigger and manage automation workflows using n8n\"\nmetadata:\n openclaw:\n emoji: \"🔄\"\n---\n\n# n8n Workflow Trigger\n\nUse n8n to create, trigger, and manage automation workflows via its REST API.\n\n## Connection Details\n\n- **Host:** \\`{{N8N_HOST}}\\`\n- **Port:** \\`{{N8N_PORT}}\\`\n- **Webhook URL:** \\`{{N8N_WEBHOOK_URL}}\\`\n\n## Example API Calls\n\n### Trigger a webhook workflow\n\\`\\`\\`bash\ncurl -X POST \"{{N8N_WEBHOOK_URL}}webhook/<your-webhook-id>\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"message\": \"Hello from OpenClaw\"}'\n\\`\\`\\`\n\n### List workflows\n\\`\\`\\`bash\ncurl -X GET \"http://{{N8N_HOST}}:{{N8N_PORT}}/api/v1/workflows\" \\\\\n -H \"Content-Type: application/json\"\n\\`\\`\\`\n\n### Activate a workflow\n\\`\\`\\`bash\ncurl -X PATCH \"http://{{N8N_HOST}}:{{N8N_PORT}}/api/v1/workflows/<id>\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"active\": true}'\n\\`\\`\\`\n\n## Usage Notes\n\n- Create workflows in the n8n UI at http://{{N8N_HOST}}:{{N8N_PORT}}.\n- Webhook nodes allow external triggers from OpenClaw.\n- n8n stores workflow state in PostgreSQL.\n`,\n\n\t\"ffmpeg-process\": `---\nname: ffmpeg-process\ndescription: \"Process media files using FFmpeg for transcoding, conversion, and manipulation\"\nmetadata:\n openclaw:\n emoji: \"🎬\"\n---\n\n# FFmpeg Media Processing\n\nUse FFmpeg for video/audio transcoding, format conversion, and media manipulation.\n\n## Shared Directory\n\n- **Media directory:** \\`{{FFMPEG_SHARED_DIR}}\\`\n\n## Example Commands\n\n### Convert video to MP4\n\\`\\`\\`bash\ndocker exec ffmpeg ffmpeg -i /data/input.avi -c:v libx264 -c:a aac /data/output.mp4\n\\`\\`\\`\n\n### Extract audio from video\n\\`\\`\\`bash\ndocker exec ffmpeg ffmpeg -i /data/video.mp4 -vn -acodec libmp3lame /data/audio.mp3\n\\`\\`\\`\n\n### Create thumbnail from video\n\\`\\`\\`bash\ndocker exec ffmpeg ffmpeg -i /data/video.mp4 -ss 00:00:05 -vframes 1 /data/thumb.jpg\n\\`\\`\\`\n\n### Resize video\n\\`\\`\\`bash\ndocker exec ffmpeg ffmpeg -i /data/input.mp4 -vf scale=1280:720 /data/output_720p.mp4\n\\`\\`\\`\n\n## Usage Notes\n\n- Place input files in the shared media directory.\n- FFmpeg runs as a sidecar container sharing a volume with OpenClaw.\n- Output files appear in the same shared directory.\n`,\n\n\t\"minio-storage\": `---\nname: minio-storage\ndescription: \"Store and retrieve files using S3-compatible object storage\"\nmetadata:\n openclaw:\n emoji: \"💾\"\n---\n\n# MinIO Object Storage\n\nUse MinIO as S3-compatible object storage for files, assets, and backups.\n\n## Connection Details\n\n- **Host:** \\`{{MINIO_HOST}}\\`\n- **API Port:** \\`{{MINIO_PORT}}\\`\n- **Access Key:** Uses \\`MINIO_ACCESS_KEY\\` env var\n- **Secret Key:** Uses \\`MINIO_SECRET_KEY\\` env var\n\n## Example API Calls\n\n### Create a bucket\n\\`\\`\\`bash\nmc alias set local http://{{MINIO_HOST}}:{{MINIO_PORT}} $MINIO_ACCESS_KEY $MINIO_SECRET_KEY\nmc mb local/my-bucket\n\\`\\`\\`\n\n### Upload a file\n\\`\\`\\`bash\nmc cp /path/to/file.txt local/my-bucket/file.txt\n\\`\\`\\`\n\n### Download a file\n\\`\\`\\`bash\nmc cp local/my-bucket/file.txt /path/to/local/file.txt\n\\`\\`\\`\n\n### List bucket contents\n\\`\\`\\`bash\nmc ls local/my-bucket/\n\\`\\`\\`\n\n## Usage Notes\n\n- MinIO is fully S3-compatible—use any S3 SDK or CLI.\n- Access the web console at http://{{MINIO_HOST}}:9001 for a visual file browser.\n- Create separate buckets for different data types (assets, backups, uploads).\n`,\n\n\t\"browserless-browse\": `---\nname: browserless-browse\ndescription: \"Automate browser interactions, scrape web pages, and generate PDFs\"\nmetadata:\n openclaw:\n emoji: \"🌐\"\n---\n\n# Browserless Browser Automation\n\nUse Browserless for headless Chrome browser automation, web scraping, screenshots, and PDF generation.\n\n## Connection Details\n\n- **Host:** \\`{{BROWSERLESS_HOST}}\\`\n- **Port:** \\`{{BROWSERLESS_PORT}}\\`\n- **Token:** Uses \\`BROWSERLESS_TOKEN\\` env var\n\n## Example API Calls\n\n### Take a screenshot\n\\`\\`\\`bash\ncurl -X POST \"http://{{BROWSERLESS_HOST}}:{{BROWSERLESS_PORT}}/screenshot?token=$BROWSERLESS_TOKEN\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"url\": \"https://example.com\", \"options\": {\"fullPage\": true}}' \\\\\n --output screenshot.png\n\\`\\`\\`\n\n### Generate a PDF\n\\`\\`\\`bash\ncurl -X POST \"http://{{BROWSERLESS_HOST}}:{{BROWSERLESS_PORT}}/pdf?token=$BROWSERLESS_TOKEN\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"url\": \"https://example.com\"}' \\\\\n --output page.pdf\n\\`\\`\\`\n\n### Scrape page content\n\\`\\`\\`bash\ncurl -X POST \"http://{{BROWSERLESS_HOST}}:{{BROWSERLESS_PORT}}/content?token=$BROWSERLESS_TOKEN\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"url\": \"https://example.com\"}'\n\\`\\`\\`\n\n## Usage Notes\n\n- Browserless manages a pool of Chrome instances (max concurrent sessions configured via env).\n- Token authentication is required for all API calls.\n- Supports Puppeteer and Playwright WebSocket connections.\n`,\n\n\t\"searxng-search\": `---\nname: searxng-search\ndescription: \"Search the web using a privacy-respecting metasearch engine\"\nmetadata:\n openclaw:\n emoji: \"🔍\"\n---\n\n# SearXNG Web Search\n\nUse SearXNG as a privacy-respecting metasearch engine to query the web.\n\n## Connection Details\n\n- **Host:** \\`{{SEARXNG_HOST}}\\`\n- **Port:** \\`{{SEARXNG_PORT}}\\`\n\n## Example API Calls\n\n### Search the web\n\\`\\`\\`bash\ncurl \"http://{{SEARXNG_HOST}}:{{SEARXNG_PORT}}/search?q=your+search+query&format=json\"\n\\`\\`\\`\n\n### Search with category filter\n\\`\\`\\`bash\ncurl \"http://{{SEARXNG_HOST}}:{{SEARXNG_PORT}}/search?q=openai&categories=it&format=json\"\n\\`\\`\\`\n\n### Search with language filter\n\\`\\`\\`bash\ncurl \"http://{{SEARXNG_HOST}}:{{SEARXNG_PORT}}/search?q=hello&language=en&format=json\"\n\\`\\`\\`\n\n## Usage Notes\n\n- Always use \\`format=json\\` for machine-readable results.\n- Available categories: general, images, videos, news, map, music, it, science, files.\n- SearXNG aggregates results from many search engines without tracking.\n`,\n\n\t\"whisper-transcribe\": `---\nname: whisper-transcribe\ndescription: \"Transcribe audio files to text using Faster Whisper\"\nmetadata:\n openclaw:\n emoji: \"🎙️\"\n---\n\n# Whisper Transcription\n\nUse the Faster Whisper server for speech-to-text transcription.\n\n## Connection Details\n\n- **Host:** \\`{{WHISPER_HOST}}\\`\n- **Port:** \\`{{WHISPER_PORT}}\\`\n\n## Example API Calls\n\n### Transcribe an audio file\n\\`\\`\\`bash\ncurl -X POST \"http://{{WHISPER_HOST}}:{{WHISPER_PORT}}/v1/audio/transcriptions\" \\\\\n -F \"file=@/path/to/audio.mp3\" \\\\\n -F \"model=base\"\n\\`\\`\\`\n\n### Transcribe with language hint\n\\`\\`\\`bash\ncurl -X POST \"http://{{WHISPER_HOST}}:{{WHISPER_PORT}}/v1/audio/transcriptions\" \\\\\n -F \"file=@/path/to/audio.wav\" \\\\\n -F \"model=base\" \\\\\n -F \"language=en\"\n\\`\\`\\`\n\n### Get available models\n\\`\\`\\`bash\ncurl \"http://{{WHISPER_HOST}}:{{WHISPER_PORT}}/v1/models\"\n\\`\\`\\`\n\n## Usage Notes\n\n- Supports MP3, WAV, FLAC, and other common audio formats.\n- Model sizes: tiny, base, small, medium, large (larger = more accurate but slower).\n- GPU acceleration significantly improves transcription speed.\n`,\n\n\t\"ollama-local-llm\": `---\nname: ollama-local-llm\ndescription: \"Run local language models for text generation, chat, and embeddings\"\nmetadata:\n openclaw:\n emoji: \"🦙\"\n---\n\n# Ollama Local LLM\n\nUse Ollama to run large language models locally for text generation, chat, and embeddings.\n\n## Connection Details\n\n- **Host:** \\`{{OLLAMA_HOST}}\\`\n- **Port:** \\`{{OLLAMA_PORT}}\\`\n\n## Example API Calls\n\n### Generate text\n\\`\\`\\`bash\ncurl -X POST \"http://{{OLLAMA_HOST}}:{{OLLAMA_PORT}}/api/generate\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"model\": \"llama3.2\", \"prompt\": \"Explain quantum computing in simple terms\"}'\n\\`\\`\\`\n\n### Chat completion\n\\`\\`\\`bash\ncurl -X POST \"http://{{OLLAMA_HOST}}:{{OLLAMA_PORT}}/api/chat\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"model\": \"llama3.2\", \"messages\": [{\"role\": \"user\", \"content\": \"Hello!\"}]}'\n\\`\\`\\`\n\n### Pull a model\n\\`\\`\\`bash\ncurl -X POST \"http://{{OLLAMA_HOST}}:{{OLLAMA_PORT}}/api/pull\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"name\": \"llama3.2\"}'\n\\`\\`\\`\n\n### Generate embeddings\n\\`\\`\\`bash\ncurl -X POST \"http://{{OLLAMA_HOST}}:{{OLLAMA_PORT}}/api/embeddings\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"model\": \"nomic-embed-text\", \"prompt\": \"Hello world\"}'\n\\`\\`\\`\n\n## Usage Notes\n\n- Pull models before first use (they are cached in the persistent volume).\n- Recommended models: llama3.2 (general), codellama (code), nomic-embed-text (embeddings).\n- GPU passthrough dramatically improves inference speed.\n- The Ollama API is OpenAI-compatible at /v1/ endpoints.\n`,\n\n\t\"remotion-render\": `---\nname: remotion-render\ndescription: \"Create and render videos programmatically using React\"\nmetadata:\n openclaw:\n emoji: \"🎥\"\n---\n\n# Remotion Video Rendering\n\nUse Remotion Studio to create and render videos programmatically with React.\n\n## Connection Details\n\n- **Host:** \\`{{REMOTION_HOST}}\\`\n- **Port:** \\`{{REMOTION_PORT}}\\`\n\n## Example Usage\n\n### Access the Studio UI\nOpen \\`http://{{REMOTION_HOST}}:{{REMOTION_PORT}}\\` in your browser to use the Remotion Studio visual editor.\n\n### Render a video via CLI\n\\`\\`\\`bash\ndocker exec remotion npx remotion render src/index.tsx MyComposition out/video.mp4\n\\`\\`\\`\n\n### Render with custom props\n\\`\\`\\`bash\ndocker exec remotion npx remotion render src/index.tsx MyComposition out/video.mp4 \\\\\n --props='{\"title\": \"Hello World\"}'\n\\`\\`\\`\n\n## Usage Notes\n\n- Define video compositions in React components.\n- Remotion supports MP4, WebM, and GIF output formats.\n- Use the Studio UI for previewing before rendering.\n- Combine with FFmpeg for post-processing.\n`,\n\n\t\"lightpanda-browse\": `---\nname: lightpanda-browse\ndescription: \"Browse the web using the ultra-fast LightPanda headless browser via CDP\"\nmetadata:\n openclaw:\n emoji: \"🐼\"\n---\n\n# LightPanda Browse\n\nLightPanda is an ultra-fast headless browser available via CDP WebSocket at \\`ws://{{LIGHTPANDA_HOST}}:{{LIGHTPANDA_PORT}}\\`.\n\n## Connect via Puppeteer\n\n\\`\\`\\`javascript\nconst browser = await puppeteer.connect({\n browserWSEndpoint: \"ws://{{LIGHTPANDA_HOST}}:{{LIGHTPANDA_PORT}}\"\n});\nconst page = await browser.newPage();\nawait page.goto('https://example.com');\nconst content = await page.evaluate(() => document.body.innerText);\n\\`\\`\\`\n\n## Key Advantages\n\n- 9x less memory than Chrome (ideal for parallel scraping)\n- 11x faster page loading\n- Instant startup\n- Full CDP compatibility with Puppeteer and Playwright\n`,\n\n\t\"steel-browse\": `---\nname: steel-browse\ndescription: \"Browse the web using Steel Browser API with session management and anti-detection\"\nmetadata:\n openclaw:\n emoji: \"🔥\"\n---\n\n# Steel Browser\n\nSteel provides a REST API at \\`http://{{STEEL_HOST}}:{{STEEL_PORT}}\\` for AI agent web automation.\n\n## Create a Session\n\n\\`\\`\\`bash\ncurl -X POST http://{{STEEL_HOST}}:{{STEEL_PORT}}/v1/sessions \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"blockAds\": true}'\n\\`\\`\\`\n\n## Scrape a Page\n\n\\`\\`\\`bash\ncurl -X POST http://{{STEEL_HOST}}:{{STEEL_PORT}}/v1/scrape \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"url\": \"https://example.com\", \"format\": \"markdown\"}'\n\\`\\`\\`\n\n## Features\n\n- Session management with persistent cookies\n- Anti-detection and stealth plugins\n- Proxy support and IP rotation\n- Auto CAPTCHA solving\n- Puppeteer/Playwright/Selenium compatible\n`,\n\n\t\"code-sandbox\": `---\nname: code-sandbox\ndescription: \"Execute code safely in an isolated OpenSandbox container\"\nmetadata:\n openclaw:\n emoji: \"📦\"\n---\n\n# Code Sandbox\n\nExecute code safely in an isolated OpenSandbox container.\n\n## Description\n\nThis skill provides secure, containerized code execution for AI agents. Code runs in ephemeral Docker containers with resource limits, network isolation, and automatic cleanup.\n\n## Connection Details\n\n- **Host:** \\`{{OPENSANDBOX_HOST}}\\`\n- **Port:** \\`{{OPENSANDBOX_PORT}}\\`\n- **Auth:** API key (auto-configured)\n\n## Supported Languages\n\n- Python 3.12\n- JavaScript / TypeScript (Node.js 22)\n- Java 21\n- Go 1.24\n- Bash\n\n## Available Actions\n\n### execute_code\n\nRun a code snippet in a fresh sandbox.\n\n**Parameters:**\n- \\`language\\` (required): Programming language (\"python\", \"javascript\", \"typescript\", \"java\", \"go\", \"bash\")\n- \\`code\\` (required): The code to execute\n- \\`timeout_seconds\\` (optional): Max execution time (default: 60, max: 300)\n\n**Returns:** stdout, stderr, exit_code, execution_time_ms\n\n### execute_shell\n\nRun a shell command in an existing or new sandbox.\n\n**Parameters:**\n- \\`command\\` (required): Shell command to execute\n- \\`sandbox_id\\` (optional): Reuse an existing sandbox (for multi-step workflows)\n- \\`background\\` (optional): Run in background (default: false)\n\n**Returns:** stdout, stderr, exit_code\n\n### upload_file\n\nUpload a file to a sandbox for processing.\n\n**Parameters:**\n- \\`sandbox_id\\` (required): Target sandbox\n- \\`path\\` (required): Destination path inside sandbox\n- \\`content\\` (required): File content (text or base64 for binary)\n\n### download_file\n\nDownload a file from a sandbox.\n\n**Parameters:**\n- \\`sandbox_id\\` (required): Source sandbox\n- \\`path\\` (required): File path inside sandbox\n\n**Returns:** File content\n\n### list_sandboxes\n\nList active sandboxes on this instance.\n\n**Returns:** Array of { id, status, image, created_at, expires_at }\n\n### terminate_sandbox\n\nTerminate a running sandbox immediately.\n\n**Parameters:**\n- \\`sandbox_id\\` (required): Sandbox to terminate\n\n### create_desktop\n\nCreate a GUI desktop sandbox with VNC access (for Homespace live preview).\n\n**Parameters:**\n- \\`image\\` (optional): Desktop image (default: \"opensandbox/desktop:latest\", also: \"opensandbox/chrome:latest\", \"opensandbox/vscode:latest\")\n- \\`resolution\\` (optional): Screen resolution (default: \"1280x800x24\")\n\n**Returns:** sandbox_id, vnc_endpoint (port 5900), novnc_url (port 6080 WebSocket), devtools_url (port 9222, chrome only)\n\n### get_preview_url\n\nGet the browser-accessible noVNC URL for an existing desktop sandbox.\n\n**Parameters:**\n- \\`sandbox_id\\` (required): Desktop sandbox ID\n\n**Returns:** novnc_url (embeddable in iframe), vnc_endpoint, status\n\n## Examples\n\n### Run Python code\n\n\\`\\`\\`bash\ncurl -X POST http://{{OPENSANDBOX_HOST}}:{{OPENSANDBOX_PORT}}/v1/sandboxes \\\\\n -H \"Authorization: Bearer $OPENSANDBOX_API_KEY\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"image\": \"opensandbox/code-interpreter:python\"}'\n\\`\\`\\`\n\n### Execute code in a sandbox\n\n\\`\\`\\`bash\ncurl -X POST http://{{OPENSANDBOX_HOST}}:{{OPENSANDBOX_PORT}}/v1/sandboxes/{id}/code \\\\\n -H \"Authorization: Bearer $OPENSANDBOX_API_KEY\" \\\\\n -H \"Content-Type: application/json\" \\\\\n -d '{\"language\": \"python\", \"code\": \"print(42 * 42)\"}'\n\\`\\`\\`\n\n## Configuration\n\n- **Default timeout:** 60 seconds\n- **Max concurrent sandboxes:** Determined by VPS RAM\n- **Idle cleanup:** Sandboxes with no activity for 30 minutes are auto-terminated\n- **Network:** Bridge mode (isolated from host services)\n- **Security:** gVisor runtime, capability dropping, PID limits\n\n## Limitations\n\n- No persistent storage between sandbox sessions (ephemeral by design)\n- No GPU access (CPU-only execution)\n- No outbound network access by default (egress blocked)\n- Max 512 PIDs per sandbox (fork bomb protection)\n- Memory capped per sandbox (default 512MB)\n`,\n};\n\n/**\n * Generates SKILL.md files for each service that has skills defined.\n *\n * Returns a map of file paths (relative to project root) to file contents.\n * Handlebars is used to replace `{{VAR}}` placeholders with actual values\n * from each service's `openclawEnvVars`.\n */\nexport function generateSkillFiles(resolved: ResolverOutput): Record<string, string> {\n\tconst files: Record<string, string> = {};\n\n\tfor (const { definition } of resolved.services) {\n\t\tif (definition.skills.length === 0) continue;\n\n\t\t// Build a variable map from openclawEnvVars for Handlebars\n\t\tconst vars: Record<string, string> = {};\n\t\tfor (const envVar of definition.openclawEnvVars) {\n\t\t\t// Resolve ${REFERENCES} to the default value (just strip the ${} wrapper)\n\t\t\tconst val = envVar.defaultValue.startsWith(\"${\")\n\t\t\t\t? envVar.defaultValue.slice(2, -1)\n\t\t\t\t: envVar.defaultValue;\n\t\t\tvars[envVar.key] = val;\n\t\t}\n\n\t\tfor (const skill of definition.skills) {\n\t\t\tconst template = skillTemplates[skill.skillId];\n\t\t\tif (!template) {\n\t\t\t\t// Generate a generic skill file if no template exists\n\t\t\t\tconst generic = generateGenericSkill(\n\t\t\t\t\tskill.skillId,\n\t\t\t\t\tdefinition.name,\n\t\t\t\t\tdefinition.icon,\n\t\t\t\t\tdefinition.description,\n\t\t\t\t\tvars,\n\t\t\t\t);\n\t\t\t\tfiles[`openclaw/workspace/skills/${skill.skillId}/SKILL.md`] = generic;\n\t\t\t\tcontinue;\n\t\t\t}\n\n\t\t\tconst compiled = Handlebars.compile(template, { noEscape: true });\n\t\t\tconst rendered = compiled(vars);\n\t\t\tfiles[`openclaw/workspace/skills/${skill.skillId}/SKILL.md`] = rendered;\n\t\t}\n\t}\n\n\treturn files;\n}\n\n/**\n * Generate a generic SKILL.md for skills that don't have a dedicated template.\n */\nfunction generateGenericSkill(\n\tskillId: string,\n\tserviceName: string,\n\ticon: string,\n\tdescription: string,\n\tvars: Record<string, string>,\n): string {\n\tconst title = skillId\n\t\t.split(\"-\")\n\t\t.map((w) => w.charAt(0).toUpperCase() + w.slice(1))\n\t\t.join(\" \");\n\n\tconst envSection = Object.entries(vars)\n\t\t.map(([key, value]) => `- **${key}:** \\`${value}\\``)\n\t\t.join(\"\\n\");\n\n\treturn `---\nname: ${skillId}\ndescription: \"${description}\"\nmetadata:\n openclaw:\n emoji: \"${icon}\"\n---\n\n# ${title}\n\nInteract with ${serviceName} through this skill.\n\n## Connection Details\n\n${envSection || \"No specific connection variables configured.\"}\n\n## Usage Notes\n\n- This skill provides OpenClaw access to ${serviceName}.\n- Refer to the service documentation for available API endpoints.\n`;\n}\n"],"mappings":";;;;;;;;AASA,MAAM,iBAAyC;CAC9C,eAAe;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAyCf,iBAAiB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA8CjB,eAAe;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA+Cf,kBAAkB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA6ClB,iBAAiB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAiDjB,sBAAsB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAkDtB,kBAAkB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAyClB,sBAAsB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA8CtB,oBAAoB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAuDpB,mBAAmB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAyCnB,qBAAqB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA+BrB,gBAAgB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CAqChB,gBAAgB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA6IhB;;;;;;;;AASD,SAAgB,mBAAmB,UAAkD;CACpF,MAAM,QAAgC,EAAE;AAExC,MAAK,MAAM,EAAE,gBAAgB,SAAS,UAAU;AAC/C,MAAI,WAAW,OAAO,WAAW,EAAG;EAGpC,MAAM,OAA+B,EAAE;AACvC,OAAK,MAAM,UAAU,WAAW,iBAAiB;GAEhD,MAAM,MAAM,OAAO,aAAa,WAAW,KAAK,GAC7C,OAAO,aAAa,MAAM,GAAG,GAAG,GAChC,OAAO;AACV,QAAK,OAAO,OAAO;;AAGpB,OAAK,MAAM,SAAS,WAAW,QAAQ;GACtC,MAAM,WAAW,eAAe,MAAM;AACtC,OAAI,CAAC,UAAU;IAEd,MAAM,UAAU,qBACf,MAAM,SACN,WAAW,MACX,WAAW,MACX,WAAW,aACX,KACA;AACD,UAAM,6BAA6B,MAAM,QAAQ,cAAc;AAC/D;;GAID,MAAM,WADW,WAAW,QAAQ,UAAU,EAAE,UAAU,MAAM,CAAC,CACvC,KAAK;AAC/B,SAAM,6BAA6B,MAAM,QAAQ,cAAc;;;AAIjE,QAAO;;;;;AAMR,SAAS,qBACR,SACA,aACA,MACA,aACA,MACS;AAUT,QAAO;QACA,QAAQ;gBACA,YAAY;;;cAGd,KAAK;;;IAdJ,QACZ,MAAM,IAAI,CACV,KAAK,MAAM,EAAE,OAAO,EAAE,CAAC,aAAa,GAAG,EAAE,MAAM,EAAE,CAAC,CAClD,KAAK,IAAI,CAcF;;gBAEM,YAAY;;;;EAdR,OAAO,QAAQ,KAAK,CACrC,KAAK,CAAC,KAAK,WAAW,OAAO,IAAI,QAAQ,MAAM,IAAI,CACnD,KAAK,KAAK,IAgBG,+CAA+C;;;;2CAIpB,YAAY"}
@@ -1,4 +1,4 @@
1
- const require_vi_2VT5v0um = require("../vi.2VT5v0um-iVBt6Fyq.cjs");
1
+ const require_test_CTcmp4Su = require("../test.CTcmp4Su-DlzTarwH.cjs");
2
2
  const require_resolver = require("../resolver.cjs");
3
3
  const require_generators_traefik = require("./traefik.cjs");
4
4
  //#region src/generators/traefik.test.ts
@@ -10,56 +10,56 @@ function resolveWith(services) {
10
10
  gpu: false
11
11
  });
12
12
  }
13
- require_vi_2VT5v0um.describe("generateTraefikConfig", () => {
13
+ require_test_CTcmp4Su.describe("generateTraefikConfig", () => {
14
14
  const domain = "example.com";
15
- require_vi_2VT5v0um.it("generates static config with domain email and entrypoints", () => {
15
+ require_test_CTcmp4Su.it("generates static config with domain email and entrypoints", () => {
16
16
  const { staticConfig } = require_generators_traefik.generateTraefikConfig(resolveWith(["redis"]), domain);
17
- require_vi_2VT5v0um.globalExpect(staticConfig).toContain("admin@example.com");
18
- require_vi_2VT5v0um.globalExpect(staticConfig).toContain("address: \":80\"");
19
- require_vi_2VT5v0um.globalExpect(staticConfig).toContain("address: \":443\"");
20
- require_vi_2VT5v0um.globalExpect(staticConfig).toContain("exposedByDefault: false");
21
- require_vi_2VT5v0um.globalExpect(staticConfig).toContain("openclaw-network");
17
+ require_test_CTcmp4Su.globalExpect(staticConfig).toContain("admin@example.com");
18
+ require_test_CTcmp4Su.globalExpect(staticConfig).toContain("address: \":80\"");
19
+ require_test_CTcmp4Su.globalExpect(staticConfig).toContain("address: \":443\"");
20
+ require_test_CTcmp4Su.globalExpect(staticConfig).toContain("exposedByDefault: false");
21
+ require_test_CTcmp4Su.globalExpect(staticConfig).toContain("openclaw-network");
22
22
  });
23
- require_vi_2VT5v0um.it("generates labels for services with exposed ports", () => {
23
+ require_test_CTcmp4Su.it("generates labels for services with exposed ports", () => {
24
24
  const { serviceLabels } = require_generators_traefik.generateTraefikConfig(resolveWith(["redis"]), domain);
25
25
  const redisLabels = serviceLabels.get("redis");
26
- require_vi_2VT5v0um.globalExpect(redisLabels).toBeDefined();
27
- require_vi_2VT5v0um.globalExpect(redisLabels["traefik.enable"]).toBe("true");
28
- require_vi_2VT5v0um.globalExpect(redisLabels["traefik.http.routers.redis.rule"]).toBe("Host(`redis.example.com`)");
29
- require_vi_2VT5v0um.globalExpect(redisLabels["traefik.http.routers.redis.entrypoints"]).toBe("websecure");
30
- require_vi_2VT5v0um.globalExpect(redisLabels["traefik.http.routers.redis.tls.certresolver"]).toBe("letsencrypt");
31
- require_vi_2VT5v0um.globalExpect(redisLabels["traefik.http.services.redis.loadbalancer.server.port"]).toBe("6379");
26
+ require_test_CTcmp4Su.globalExpect(redisLabels).toBeDefined();
27
+ require_test_CTcmp4Su.globalExpect(redisLabels["traefik.enable"]).toBe("true");
28
+ require_test_CTcmp4Su.globalExpect(redisLabels["traefik.http.routers.redis.rule"]).toBe("Host(`redis.example.com`)");
29
+ require_test_CTcmp4Su.globalExpect(redisLabels["traefik.http.routers.redis.entrypoints"]).toBe("websecure");
30
+ require_test_CTcmp4Su.globalExpect(redisLabels["traefik.http.routers.redis.tls.certresolver"]).toBe("letsencrypt");
31
+ require_test_CTcmp4Su.globalExpect(redisLabels["traefik.http.services.redis.loadbalancer.server.port"]).toBe("6379");
32
32
  });
33
- require_vi_2VT5v0um.it("generates HTTP to HTTPS redirect labels", () => {
33
+ require_test_CTcmp4Su.it("generates HTTP to HTTPS redirect labels", () => {
34
34
  const { serviceLabels } = require_generators_traefik.generateTraefikConfig(resolveWith(["redis"]), domain);
35
35
  const redisLabels = serviceLabels.get("redis");
36
- require_vi_2VT5v0um.globalExpect(redisLabels["traefik.http.routers.redis-http.entrypoints"]).toBe("web");
37
- require_vi_2VT5v0um.globalExpect(redisLabels["traefik.http.routers.redis-http.middlewares"]).toBe("redirect-to-https");
36
+ require_test_CTcmp4Su.globalExpect(redisLabels["traefik.http.routers.redis-http.entrypoints"]).toBe("web");
37
+ require_test_CTcmp4Su.globalExpect(redisLabels["traefik.http.routers.redis-http.middlewares"]).toBe("redirect-to-https");
38
38
  });
39
- require_vi_2VT5v0um.it("assigns root domain to gateway", () => {
39
+ require_test_CTcmp4Su.it("assigns root domain to gateway", () => {
40
40
  const { serviceLabels } = require_generators_traefik.generateTraefikConfig(resolveWith(["redis"]), domain);
41
41
  const gwLabels = serviceLabels.get("openclaw-gateway");
42
- require_vi_2VT5v0um.globalExpect(gwLabels).toBeDefined();
43
- require_vi_2VT5v0um.globalExpect(gwLabels["traefik.http.routers.gateway.rule"]).toBe("Host(`example.com`)");
44
- require_vi_2VT5v0um.globalExpect(gwLabels["traefik.http.services.gateway.loadbalancer.server.port"]).toBe("18789");
42
+ require_test_CTcmp4Su.globalExpect(gwLabels).toBeDefined();
43
+ require_test_CTcmp4Su.globalExpect(gwLabels["traefik.http.routers.gateway.rule"]).toBe("Host(`example.com`)");
44
+ require_test_CTcmp4Su.globalExpect(gwLabels["traefik.http.services.gateway.loadbalancer.server.port"]).toBe("18789");
45
45
  });
46
- require_vi_2VT5v0um.it("adds global redirect middleware on traefik service", () => {
46
+ require_test_CTcmp4Su.it("adds global redirect middleware on traefik service", () => {
47
47
  const { serviceLabels } = require_generators_traefik.generateTraefikConfig(resolveWith(["redis"]), domain);
48
48
  const traefikLabels = serviceLabels.get("traefik");
49
- require_vi_2VT5v0um.globalExpect(traefikLabels).toBeDefined();
50
- require_vi_2VT5v0um.globalExpect(traefikLabels["traefik.http.middlewares.redirect-to-https.redirectscheme.scheme"]).toBe("https");
51
- require_vi_2VT5v0um.globalExpect(traefikLabels["traefik.http.middlewares.redirect-to-https.redirectscheme.permanent"]).toBe("true");
49
+ require_test_CTcmp4Su.globalExpect(traefikLabels).toBeDefined();
50
+ require_test_CTcmp4Su.globalExpect(traefikLabels["traefik.http.middlewares.redirect-to-https.redirectscheme.scheme"]).toBe("https");
51
+ require_test_CTcmp4Su.globalExpect(traefikLabels["traefik.http.middlewares.redirect-to-https.redirectscheme.permanent"]).toBe("true");
52
52
  });
53
- require_vi_2VT5v0um.it("skips proxy services and services without exposed ports", () => {
53
+ require_test_CTcmp4Su.it("skips proxy services and services without exposed ports", () => {
54
54
  const { serviceLabels } = require_generators_traefik.generateTraefikConfig(resolveWith(["redis", "ffmpeg"]), domain);
55
- require_vi_2VT5v0um.globalExpect(serviceLabels.has("ffmpeg")).toBe(false);
56
- require_vi_2VT5v0um.globalExpect(serviceLabels.get("traefik")["traefik.http.routers.traefik.rule"]).toBeUndefined();
55
+ require_test_CTcmp4Su.globalExpect(serviceLabels.has("ffmpeg")).toBe(false);
56
+ require_test_CTcmp4Su.globalExpect(serviceLabels.get("traefik")["traefik.http.routers.traefik.rule"]).toBeUndefined();
57
57
  });
58
- require_vi_2VT5v0um.it("sanitizes service names with hyphens in router names", () => {
58
+ require_test_CTcmp4Su.it("sanitizes service names with hyphens in router names", () => {
59
59
  const { serviceLabels } = require_generators_traefik.generateTraefikConfig(resolveWith(["open-webui"]), domain);
60
60
  const labels = serviceLabels.get("open-webui");
61
- require_vi_2VT5v0um.globalExpect(labels).toBeDefined();
62
- require_vi_2VT5v0um.globalExpect(labels["traefik.http.routers.openwebui.rule"]).toBe("Host(`open-webui.example.com`)");
61
+ require_test_CTcmp4Su.globalExpect(labels).toBeDefined();
62
+ require_test_CTcmp4Su.globalExpect(labels["traefik.http.routers.openwebui.rule"]).toBe("Host(`open-webui.example.com`)");
63
63
  });
64
64
  });
65
65
  //#endregion